mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-31 08:51:55 +00:00 
			
		
		
		
	server : fallback to chatml, add AlphaMonarch chat template (#5628)
* server: fallback to chatml * add new chat template * server: add AlphaMonarch to test chat template * server: only check model template if there is no custom tmpl * remove TODO
This commit is contained in:
		| @@ -12773,6 +12773,15 @@ static int32_t llama_chat_apply_template_internal( | ||||
|         if (add_ass) { | ||||
|             ss << "<|assistant|>\n"; | ||||
|         } | ||||
|     } else if (tmpl.find("bos_token + message['role']") != std::string::npos) { | ||||
|         // mlabonne/AlphaMonarch-7B template (the <s> is included inside history) | ||||
|         for (auto message : chat) { | ||||
|             std::string bos = (message == chat.front()) ? "" : "<s>"; // skip BOS for first message | ||||
|             ss << bos << message->role << "\n" << message->content << "</s>\n"; | ||||
|         } | ||||
|         if (add_ass) { | ||||
|             ss << "<s>assistant\n"; | ||||
|         } | ||||
|     } else { | ||||
|         // template not supported | ||||
|         return -1; | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Xuan Son Nguyen
					Xuan Son Nguyen