mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-30 08:42:00 +00:00 
			
		
		
		
	llama : support LiquidAI LFM2-MoE hybrid model (#16464)
* llama : support LiquidAI LFM2-MoE hybrid model Add support for [LiquidAI/LFM2-8B-A1B](https://huggingface.co/LiquidAI/LFM2-8B-A1B) model. For more information about models, please read [the blog post](https://www.liquid.ai/company/news). [HF PR](https://github.com/huggingface/transformers/pull/41401) [GGUFs](https://huggingface.co/LiquidAI/LFM2-8B-A1B-GGUF) * Do not use defaultdict * Address PR feedback
This commit is contained in:
		| @@ -358,6 +358,7 @@ class TensorNameMap: | ||||
|             "model.layers.{bid}.mlp.router",                    # openai-moe | ||||
|             "model.layers.{bid}.mlp.gate.wg",                   # hunyuan | ||||
|             "model.layers.{bid}.block_sparse_moe.primary_router", # smallthinker | ||||
|             "model.layers.{bid}.feed_forward.gate",               # lfm2moe | ||||
|         ), | ||||
|  | ||||
|         MODEL_TENSOR.FFN_GATE_INP_SHEXP: ( | ||||
| @@ -367,6 +368,7 @@ class TensorNameMap: | ||||
|         MODEL_TENSOR.FFN_EXP_PROBS_B: ( | ||||
|             "model.layers.{bid}.mlp.gate.e_score_correction",               # deepseek-v3 dots1 | ||||
|             "model.layers.{bid}.mlp.moe_statics.e_score_correction",        # ernie4.5-moe | ||||
|             "model.layers.{bid}.feed_forward.expert_bias",                  # lfm2moe | ||||
|         ), | ||||
|  | ||||
|         # Feed-forward up | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Tarek Dakhran
					Tarek Dakhran