mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-30 08:42:00 +00:00 
			
		
		
		
	Refactor common_chat_* functions to accept minja template + use_jinja option
This commit is contained in:
		| @@ -714,7 +714,7 @@ static void add_message(const char * role, const std::string & text, LlamaData & | ||||
| } | ||||
|  | ||||
| // Function to apply the chat template and resize `formatted` if needed | ||||
| static int apply_chat_template(const minja::chat_template & tmpl, LlamaData & llama_data, const bool append, bool use_jinja) { | ||||
| static int apply_chat_template(const llama_chat_template & tmpl, LlamaData & llama_data, const bool append, bool use_jinja) { | ||||
|     if (use_jinja) { | ||||
|         json messages = json::array(); | ||||
|         for (const auto & msg : llama_data.messages) { | ||||
| @@ -868,7 +868,7 @@ static int generate_response(LlamaData & llama_data, const std::string & prompt, | ||||
| } | ||||
|  | ||||
| // Helper function to apply the chat template and handle errors | ||||
| static int apply_chat_template_with_error_handling(const minja::chat_template & tmpl, LlamaData & llama_data, const bool append, int & output_length, bool use_jinja) { | ||||
| static int apply_chat_template_with_error_handling(const llama_chat_template & tmpl, LlamaData & llama_data, const bool append, int & output_length, bool use_jinja) { | ||||
|     const int new_len = apply_chat_template(tmpl, llama_data, append, use_jinja); | ||||
|     if (new_len < 0) { | ||||
|         printe("failed to apply the chat template\n"); | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 ochafik
					ochafik