llama : remove deprecated llama_kv_self API (#15472)

ggml-ci
This commit is contained in:
Georgi Gerganov
2025-08-21 19:13:45 +03:00
committed by GitHub
parent 3f196be84b
commit cd36b5e5c7
3 changed files with 6 additions and 297 deletions

View File

@@ -46,10 +46,8 @@ struct llama_context {
llama_memory_t get_memory() const;
// return true of the KV cache was updated
// TODO: remove
bool kv_self_update(bool optimize);
void kv_self_defrag_sched();
// return true if the memory was updated
bool memory_update(bool optimize);
enum llama_pooling_type pooling_type() const;
@@ -230,9 +228,6 @@ private:
std::unique_ptr<llama_memory_i> memory;
// TODO: temporary, until the llama_kv_self_defrag() API is removed
bool memory_force_optimize = false;
// decode output (2-dimensional array: [n_outputs][n_vocab])
size_t logits_size = 0; // capacity (of floats) for logits
float * logits = nullptr;