mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-30 08:42:00 +00:00 
			
		
		
		
	chat : fix hunyuan auto-detection (#15114)
Signed-off-by: stevenkuang <stevenkuang@tencent.com>
This commit is contained in:
		| @@ -193,11 +193,11 @@ llm_chat_template llm_chat_detect_template(const std::string & tmpl) { | |||||||
|         return LLM_CHAT_TEMPLATE_LLAMA4; |         return LLM_CHAT_TEMPLATE_LLAMA4; | ||||||
|     } else if (tmpl_contains("<|endofuserprompt|>")) { |     } else if (tmpl_contains("<|endofuserprompt|>")) { | ||||||
|         return LLM_CHAT_TEMPLATE_DOTS1; |         return LLM_CHAT_TEMPLATE_DOTS1; | ||||||
|     } else if (tmpl_contains("<|startoftext|>") && tmpl_contains("<|extra_4|>")) { |     } else if (tmpl_contains("<|extra_0|>") && tmpl_contains("<|extra_4|>")) { | ||||||
|         return LLM_CHAT_TEMPLATE_HUNYUAN_MOE; |         return LLM_CHAT_TEMPLATE_HUNYUAN_MOE; | ||||||
|     } else if (tmpl_contains("<|start|>") && tmpl_contains("<|channel|>")) { |     } else if (tmpl_contains("<|start|>") && tmpl_contains("<|channel|>")) { | ||||||
|         return LLM_CHAT_TEMPLATE_OPENAI_MOE; |         return LLM_CHAT_TEMPLATE_OPENAI_MOE; | ||||||
|     } else if (tmpl_contains("<|hy_place▁holder▁no▁2|>") && tmpl_contains("<|hy_place▁holder▁no▁3|>")) { |     } else if (tmpl_contains("<|hy_Assistant|>") && tmpl_contains("<|hy_place▁holder▁no▁3|>")) { | ||||||
|         return LLM_CHAT_TEMPLATE_HUNYUAN_DENSE; |         return LLM_CHAT_TEMPLATE_HUNYUAN_DENSE; | ||||||
|     } else if (tmpl_contains("<|im_assistant|>assistant<|im_middle|>")) { |     } else if (tmpl_contains("<|im_assistant|>assistant<|im_middle|>")) { | ||||||
|         return LLM_CHAT_TEMPLATE_KIMI_K2; |         return LLM_CHAT_TEMPLATE_KIMI_K2; | ||||||
|   | |||||||
		Reference in New Issue
	
	Block a user
	 stevenkuang
					stevenkuang