pre-norm -> norm

This commit is contained in:
younesbelkada
2025-07-04 14:58:33 +04:00
parent 243e4d1a50
commit cce35498d5
3 changed files with 3 additions and 4 deletions

View File

@@ -4595,7 +4595,7 @@ bool llama_model::load_tensors(llama_model_loader & ml) {
// feed forward (w/ optional biases)
layer.ffn_pre_norm = create_tensor(tn(LLM_TENSOR_FFN_PRE_NORM, i), {hidden_size}, 0);
layer.ffn_norm = create_tensor(tn(LLM_TENSOR_FFN_PRE_NORM, i), {hidden_size}, 0);
layer.rope_freqs = create_tensor(tn(LLM_TENSOR_ROPE_FREQS, "weight", i), {n_rot/2}, llama_model_loader::TENSOR_NOT_REQUIRED | (i != 0 ? llama_model_loader::TENSOR_DUPLICATED : 0));
layer.ffn_gate = create_tensor(tn(LLM_TENSOR_FFN_GATE, "weight", i), {hidden_size, ffn_intermediate_size}, 0);
layer.ffn_down = create_tensor(tn(LLM_TENSOR_FFN_DOWN, "weight", i), { ffn_intermediate_size, hidden_size}, 0);
@@ -14751,7 +14751,7 @@ struct llm_build_falcon_h1 : public llm_graph_context {
// feed-forward network
cur = build_norm(ffn_inp,
model.layers[il].ffn_pre_norm, NULL,
model.layers[il].ffn_norm, NULL,
LLM_NORM_RMS, il);
cb(cur, "ffn_norm", il);