mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-11-02 09:12:03 +00:00
examples : replace llama_kv_cache_seq_* with llama_past_seq_*
This commit is contained in:
@@ -98,7 +98,7 @@ if llama_decode(context, batch) != 0 {
|
||||
}
|
||||
|
||||
for i in 1 ..< n_parallel {
|
||||
llama_kv_cache_seq_cp(context, 0, Int32(i), 0, batch.n_tokens)
|
||||
llama_past_seq_cp(context, 0, Int32(i), -1, -1)
|
||||
}
|
||||
|
||||
if n_parallel > 1 {
|
||||
|
||||
Reference in New Issue
Block a user