mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-31 08:51:55 +00:00 
			
		
		
		
	Wire LLM_KV_TOKENIZER_CHAT_TEMPLATE_N in llama_model_chat_template
This commit is contained in:
		| @@ -503,7 +503,7 @@ extern "C" { | ||||
|     LLAMA_API uint64_t llama_model_size(const struct llama_model * model); | ||||
|  | ||||
|     // Get the default chat template. Returns nullptr if not available | ||||
|     LLAMA_API const char * llama_model_chat_template(const struct llama_model * model); | ||||
|     LLAMA_API const char * llama_model_chat_template(const struct llama_model * model, const char * name); | ||||
|  | ||||
|     // Returns the total number of parameters in the model | ||||
|     LLAMA_API uint64_t llama_model_n_params(const struct llama_model * model); | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 ochafik
					ochafik