mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-10-29 08:41:22 +00:00
llama: fix missing k_cache store for rwkv6qwen2 (#11445)
Signed-off-by: Molly Sophia <mollysophia379@gmail.com>
This commit is contained in:
@@ -7700,17 +7700,13 @@ struct llm_build_context {
|
|||||||
1
|
1
|
||||||
);
|
);
|
||||||
|
|
||||||
|
struct ggml_tensor * last_norm_att = ggml_view_3d(ctx0, x_norm_att, n_embd, 1, n_seqs, x_norm_att->nb[1], x_norm_att->nb[2], (n_seq_tokens-1)*n_embd*ggml_element_size(x_norm_att));
|
||||||
ggml_build_forward_expand(
|
ggml_build_forward_expand(
|
||||||
gf,
|
gf,
|
||||||
ggml_cpy(
|
ggml_cpy(
|
||||||
ctx0,
|
ctx0,
|
||||||
wkv_states,
|
ggml_view_1d(ctx0, last_norm_att, n_embd * n_seqs, 0),
|
||||||
ggml_view_1d(
|
ggml_view_1d(ctx0, kv_self.k_l[il], hparams.n_embd_k_s() * n_seqs, hparams.n_embd_k_s() * kv_head * ggml_element_size(kv_self.k_l[il]))
|
||||||
ctx0,
|
|
||||||
kv_self.v_l[il],
|
|
||||||
hparams.n_embd_v_s() * n_seqs,
|
|
||||||
hparams.n_embd_v_s() * kv_head * ggml_element_size(kv_self.v_l[il])
|
|
||||||
)
|
|
||||||
)
|
)
|
||||||
);
|
);
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user