mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-31 08:51:55 +00:00 
			
		
		
		
	convert : add custom attention mapping
This commit is contained in:
		| @@ -146,6 +146,7 @@ class TensorNameMap: | ||||
|         # Attention query | ||||
|         MODEL_TENSOR.ATTN_Q: ( | ||||
|             "model.layers.{bid}.self_attn.q_proj",                       # llama-hf nemotron olmoe olmo2 | ||||
|             "model.layers.{bid}.self_attn.q_proj_no_perm",               # llama-custom | ||||
|             "layers.{bid}.attention.wq",                                 # llama-pth | ||||
|             "encoder.layer.{bid}.attention.self.query",                  # bert | ||||
|             "transformer.h.{bid}.attn.q_proj",                           # gpt-j | ||||
| @@ -158,6 +159,7 @@ class TensorNameMap: | ||||
|         # Attention key | ||||
|         MODEL_TENSOR.ATTN_K: ( | ||||
|             "model.layers.{bid}.self_attn.k_proj",                     # llama-hf nemotron olmoe olmo2 | ||||
|             "model.layers.{bid}.self_attn.k_proj_no_perm",             # llama-custom | ||||
|             "layers.{bid}.attention.wk",                               # llama-pth | ||||
|             "encoder.layer.{bid}.attention.self.key",                  # bert | ||||
|             "transformer.h.{bid}.attn.k_proj",                         # gpt-j | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Georgi Gerganov
					Georgi Gerganov