mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-28 08:31:25 +00:00 
			
		
		
		
	convert : refactor rope_freqs generation (#9396)
* convert : refactor rope_freqs generation This should also fix vocab-only conversion for Phi-3. * convert : adapt MiniCPM3 to separate rope_freqs insertion MiniCPM3's tokenizer is treated as a SentencePiece tokenizer to avoid having to run its custom Python code which mixes tokenization in the same file as tool calls. gguf-py : add long and short RoPE factors to tensor mappings Empty, but the key names are used to populate the mappings.
This commit is contained in:
		| @@ -814,6 +814,8 @@ MODEL_TENSORS: dict[MODEL_ARCH, list[MODEL_TENSOR]] = { | ||||
|         MODEL_TENSOR.TOKEN_EMBD, | ||||
|         MODEL_TENSOR.OUTPUT_NORM, | ||||
|         MODEL_TENSOR.OUTPUT, | ||||
|         MODEL_TENSOR.ROPE_FACTORS_LONG, | ||||
|         MODEL_TENSOR.ROPE_FACTORS_SHORT, | ||||
|         MODEL_TENSOR.ATTN_NORM, | ||||
|         MODEL_TENSOR.ATTN_QKV, | ||||
|         MODEL_TENSOR.ATTN_Q, | ||||
| @@ -892,6 +894,8 @@ MODEL_TENSORS: dict[MODEL_ARCH, list[MODEL_TENSOR]] = { | ||||
|         MODEL_TENSOR.TOKEN_EMBD, | ||||
|         MODEL_TENSOR.OUTPUT_NORM, | ||||
|         MODEL_TENSOR.OUTPUT, | ||||
|         MODEL_TENSOR.ROPE_FACTORS_LONG, | ||||
|         MODEL_TENSOR.ROPE_FACTORS_SHORT, | ||||
|         MODEL_TENSOR.ATTN_NORM, | ||||
|         MODEL_TENSOR.ATTN_Q_A, | ||||
|         MODEL_TENSOR.ATTN_Q_B, | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 compilade
					compilade