mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-11-10 10:27:03 +00:00
* bench : cache llama_context state at depth * cont : handle failures to restore the old state * cont : print information when the state is being reused
86 KiB
86 KiB