diff --git a/src/llama-arch.cpp b/src/llama-arch.cpp index 4b555754ee..469c9b51e8 100644 --- a/src/llama-arch.cpp +++ b/src/llama-arch.cpp @@ -1068,7 +1068,7 @@ static const std::map> LLM_TENSOR_N { LLM_TENSOR_SSM_D, "blk.%d.ssm_d" }, { LLM_TENSOR_SSM_NORM, "blk.%d.ssm_norm" }, { LLM_TENSOR_SSM_OUT, "blk.%d.ssm_out" }, - { LLM_TENSOR_FFN_PRE_NORM, "blk.%d.ffn_pre_norm" }, + { LLM_TENSOR_FFN_NORM, "blk.%d.ffn_pre_norm" }, { LLM_TENSOR_FFN_GATE, "blk.%d.ffn_gate" }, { LLM_TENSOR_FFN_DOWN, "blk.%d.ffn_down" }, { LLM_TENSOR_FFN_UP, "blk.%d.ffn_up" }, diff --git a/src/llama-model.cpp b/src/llama-model.cpp index fc822331aa..e632fa7770 100644 --- a/src/llama-model.cpp +++ b/src/llama-model.cpp @@ -4595,7 +4595,7 @@ bool llama_model::load_tensors(llama_model_loader & ml) { // feed forward (w/ optional biases) - layer.ffn_pre_norm = create_tensor(tn(LLM_TENSOR_FFN_PRE_NORM, i), {hidden_size}, 0); + layer.ffn_norm = create_tensor(tn(LLM_TENSOR_FFN_PRE_NORM, i), {hidden_size}, 0); layer.rope_freqs = create_tensor(tn(LLM_TENSOR_ROPE_FREQS, "weight", i), {n_rot/2}, llama_model_loader::TENSOR_NOT_REQUIRED | (i != 0 ? llama_model_loader::TENSOR_DUPLICATED : 0)); layer.ffn_gate = create_tensor(tn(LLM_TENSOR_FFN_GATE, "weight", i), {hidden_size, ffn_intermediate_size}, 0); layer.ffn_down = create_tensor(tn(LLM_TENSOR_FFN_DOWN, "weight", i), { ffn_intermediate_size, hidden_size}, 0); @@ -14751,7 +14751,7 @@ struct llm_build_falcon_h1 : public llm_graph_context { // feed-forward network cur = build_norm(ffn_inp, - model.layers[il].ffn_pre_norm, NULL, + model.layers[il].ffn_norm, NULL, LLM_NORM_RMS, il); cb(cur, "ffn_norm", il); diff --git a/src/llama-model.h b/src/llama-model.h index 8e14be82b5..506fcd4789 100644 --- a/src/llama-model.h +++ b/src/llama-model.h @@ -216,7 +216,6 @@ struct llama_layer { struct ggml_tensor * layer_out_norm_b = nullptr; struct ggml_tensor * ffn_norm_exps = nullptr; struct ggml_tensor * ffn_norm_enc = nullptr; - struct ggml_tensor * ffn_pre_norm = nullptr; // ff struct ggml_tensor * ffn_gate = nullptr; // w1