mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-30 08:42:00 +00:00 
			
		
		
		
	server : Log original chat template parsing error (#12233)
This commit is contained in:
		| @@ -1900,6 +1900,7 @@ struct server_context { | |||||||
|         try { |         try { | ||||||
|             common_chat_format_example(chat_templates.get(), params.use_jinja); |             common_chat_format_example(chat_templates.get(), params.use_jinja); | ||||||
|         } catch (const std::exception & e) { |         } catch (const std::exception & e) { | ||||||
|  |             SRV_WRN("%s: Chat template parsing error: %s\n", __func__, e.what()); | ||||||
|             SRV_WRN("%s: The chat template that comes with this model is not yet supported, falling back to chatml. This may cause the model to output suboptimal responses\n", __func__); |             SRV_WRN("%s: The chat template that comes with this model is not yet supported, falling back to chatml. This may cause the model to output suboptimal responses\n", __func__); | ||||||
|             chat_templates = common_chat_templates_init(model, "chatml"); |             chat_templates = common_chat_templates_init(model, "chatml"); | ||||||
|         } |         } | ||||||
|   | |||||||
		Reference in New Issue
	
	Block a user
	 Sigbjørn Skjæret
					Sigbjørn Skjæret