mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-10-27 08:21:30 +00:00
llama: print memory breakdown on exit (#15860)
* llama: print memory breakdown on exit
This commit is contained in:
@@ -2060,6 +2060,7 @@ int main(int argc, char ** argv) {
|
||||
|
||||
LOG("\n");
|
||||
llama_perf_context_print(ctx);
|
||||
llama_memory_breakdown_print(ctx);
|
||||
|
||||
llama_backend_free();
|
||||
|
||||
|
||||
Reference in New Issue
Block a user