mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-11-07 09:57:00 +00:00
push more fixes
This commit is contained in:
@@ -32,7 +32,7 @@ llama_memory_hybrid::llama_memory_hybrid(
|
||||
mem_attn(new llama_kv_cache_unified(
|
||||
model,
|
||||
filter_attn == nullptr ?
|
||||
[&](int32_t il) { return !hparams.is_recurrent(il); }
|
||||
[&](int32_t il) { return hparams.is_recurrent(il); }
|
||||
: filter_attn,
|
||||
type_k,
|
||||
type_v,
|
||||
|
||||
Reference in New Issue
Block a user