llama : add check for KV cache shifts (#10401)

ggml-ci
This commit is contained in:
Georgi Gerganov
2024-11-19 13:29:26 +02:00
committed by GitHub
parent a88ad007de
commit 8e752a777b
3 changed files with 14 additions and 1 deletions

View File

@@ -667,6 +667,9 @@ extern "C" {
// Apply the KV cache updates (such as K-shifts, defragmentation, etc.)
LLAMA_API void llama_kv_cache_update(struct llama_context * ctx);
// Check if the context supports KV cache shifting
LLAMA_API bool llama_kv_cache_can_shift(struct llama_context * ctx);
//
// State / sessions
//