mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-31 08:51:55 +00:00 
			
		
		
		
	wip : implement GGUF (#2397)
* Add LLAMA_DEFAULT_RMS_EPS so we can change the default (#2384) Co-authored-by: Iwan Kawrakow <iwan.kawrakow@gmail.com> * WIP: python class to write GGUF, incomplete C apı for reading --------- Co-authored-by: Kawrakow <48489457+ikawrakow@users.noreply.github.com> Co-authored-by: Iwan Kawrakow <iwan.kawrakow@gmail.com>
This commit is contained in:
		 M. Yusuf Sarıgöz
					M. Yusuf Sarıgöz
				
			
				
					committed by
					
						 Georgi Gerganov
						Georgi Gerganov
					
				
			
			
				
	
			
			
			 Georgi Gerganov
						Georgi Gerganov
					
				
			
						parent
						
							4d698495ea
						
					
				
				
					commit
					bae6b125f6
				
			
							
								
								
									
										32
									
								
								constants.py
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										32
									
								
								constants.py
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,32 @@ | ||||
| GGUF_MAGIC = 0x47475546 | ||||
| GGUF_VERSION = 1 | ||||
|  | ||||
| # general | ||||
| KEY_GENERAL_ARCHITECTURE = "general.architecture" | ||||
| KEY_GENERAL_QUANTIZATION_VERSION = "general.quantization_version" | ||||
| KEY_GENERAL_NAME = "general.name" | ||||
| KEY_GENERAL_AUTHOR = "general.author" | ||||
| KEY_GENERAL_URL = "general.url" | ||||
| KEY_GENERAL_DESCRIPTION = "general.description" | ||||
| KEY_GENERAL_FILE_TYPE = "general.file_type" | ||||
| KEY_GENERAL_LICENSE = "general.license" | ||||
| KEY_GENERAL_SOURCE_URL = "general.source.url" | ||||
| KEY_GENERAL_SOURCE_HF_REPO = "general.source.hugginface.repository" | ||||
|  | ||||
| # LLM | ||||
| KEY_LLM_CONTEXT_LENGTH = "{llm}.context_length" | ||||
| KEY_LLM_EMBEDDING_LENGTH = "{llm}.embedding_length" | ||||
| KEY_LLM_LAYER_COUNT = "{llm}.layer_count" | ||||
| KEY_LLM_FEED_FORWARD_LENGTH = "{llm}.feed_forward_length" | ||||
| KEY_LLM_USE_PARALLEL_RESIDUAL = "{llm}.use_parallel_residual" | ||||
| KEY_LLM_TENSOR_DATA_LAYOUT = "{llm}.tensor_data_layout" | ||||
|  | ||||
| # attention | ||||
| KEY_ATTENTION_HEAD_COUNT = "{llm}.attention.head_count" | ||||
| KEY_ATTENTION_HEAD_COUNT_KV = "{llm}.attention.head_count_kv" | ||||
| KEY_ATTENTION_MAX_ALIBI_BIAS = "{llm}.attention.max_alibi_bias" | ||||
| KEY_ATTENTION_CLAMP_KQV = "{llm}.attention.clamp_kqv" | ||||
|  | ||||
| # RoPE | ||||
| KEY_ROPE_DIMENSION_COUNT = "{llm}.rope.dimension_count" | ||||
| KEY_ROPE_SCALE = "{llm}.rope.scale" | ||||
		Reference in New Issue
	
	Block a user