mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-31 08:51:55 +00:00 
			
		
		
		
	 03c0946d73
			
		
	
	03c0946d73
	
	
	
		
			
			* Support converting models with multiple chat templates Adds the following metadata: * tokenizer.chat_templates * tokenizer.chat_template.<name1> * tokenizer.chat_template.<name2> * tokenizer.chat_template.<...> Where `tokenizer.chat_templates` is an array of the template names (except `default`), `default` is added to the regular `tokenizer.chat_template`. * replace filtered characters with underscore * New script to add/modify/remove metadata This scripts creates a copy of a GGUF file and allows you to add/modify/remove metadata in the process. Most importantly this allows you to update chat templates, either as a string or directly from an updated tokenizer_config.json file. * Add files via upload add new script to project/readme * flake--
		
			
				
	
	
		
			14 lines
		
	
	
		
			428 B
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			14 lines
		
	
	
		
			428 B
		
	
	
	
		
			Python
		
	
	
	
	
	
| import os
 | |
| 
 | |
| from importlib import import_module
 | |
| 
 | |
| 
 | |
| os.environ["NO_LOCAL_GGUF"] = "TRUE"
 | |
| 
 | |
| gguf_convert_endian_entrypoint = import_module("scripts.gguf-convert-endian").main
 | |
| gguf_dump_entrypoint           = import_module("scripts.gguf-dump").main
 | |
| gguf_set_metadata_entrypoint   = import_module("scripts.gguf-set-metadata").main
 | |
| gguf_new_metadata_entrypoint   = import_module("scripts.gguf-new-metadata").main
 | |
| 
 | |
| del import_module, os
 |