From 15138df48f2637da9696772ded9d0c013c169946 Mon Sep 17 00:00:00 2001 From: ibrahimkhadraoui Date: Fri, 4 Jul 2025 15:37:40 +0400 Subject: [PATCH] small fix ffn_norm --- gguf-py/gguf/tensor_mapping.py | 2 +- src/llama-arch.cpp | 2 +- src/llama-model.cpp | 2 +- 3 files changed, 3 insertions(+), 3 deletions(-) diff --git a/gguf-py/gguf/tensor_mapping.py b/gguf-py/gguf/tensor_mapping.py index 65311cee00..245385dfdf 100644 --- a/gguf-py/gguf/tensor_mapping.py +++ b/gguf-py/gguf/tensor_mapping.py @@ -286,7 +286,7 @@ class TensorNameMap: # Post feed-forward norm MODEL_TENSOR.FFN_PRE_NORM: ( "model.layers.{bid}.pre_feedforward_layernorm", # gemma2 - "model.layers.{bid}.pre_ff_layernorm.weight", + "model.layers.{bid}.pre_ff_layernorm.weight", ), # Post feed-forward norm diff --git a/src/llama-arch.cpp b/src/llama-arch.cpp index ebae132b7a..ae6ea76fff 100644 --- a/src/llama-arch.cpp +++ b/src/llama-arch.cpp @@ -1068,7 +1068,7 @@ static const std::map> LLM_TENSOR_N { LLM_TENSOR_SSM_D, "blk.%d.ssm_d" }, { LLM_TENSOR_SSM_NORM, "blk.%d.ssm_norm" }, { LLM_TENSOR_SSM_OUT, "blk.%d.ssm_out" }, - { LLM_TENSOR_FFN_NORM, "blk.%d.ffn_norm" }, + { LLM_TENSOR_FFN_NORM, "blk.%d.ffn_norm" }, { LLM_TENSOR_FFN_GATE, "blk.%d.ffn_gate" }, { LLM_TENSOR_FFN_DOWN, "blk.%d.ffn_down" }, { LLM_TENSOR_FFN_UP, "blk.%d.ffn_up" }, diff --git a/src/llama-model.cpp b/src/llama-model.cpp index e632fa7770..93e7dae59f 100644 --- a/src/llama-model.cpp +++ b/src/llama-model.cpp @@ -4595,7 +4595,7 @@ bool llama_model::load_tensors(llama_model_loader & ml) { // feed forward (w/ optional biases) - layer.ffn_norm = create_tensor(tn(LLM_TENSOR_FFN_PRE_NORM, i), {hidden_size}, 0); + layer.ffn_norm = create_tensor(tn(LLM_TENSOR_FFN_NORM, i), {hidden_size}, 0); layer.rope_freqs = create_tensor(tn(LLM_TENSOR_ROPE_FREQS, "weight", i), {n_rot/2}, llama_model_loader::TENSOR_NOT_REQUIRED | (i != 0 ? llama_model_loader::TENSOR_DUPLICATED : 0)); layer.ffn_gate = create_tensor(tn(LLM_TENSOR_FFN_GATE, "weight", i), {hidden_size, ffn_intermediate_size}, 0); layer.ffn_down = create_tensor(tn(LLM_TENSOR_FFN_DOWN, "weight", i), { ffn_intermediate_size, hidden_size}, 0);