mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-31 08:51:55 +00:00 
			
		
		
		
	llama : fix Gemma rope type (#5691)
This commit is contained in:
		| @@ -12089,7 +12089,6 @@ enum llama_rope_type llama_rope_type(const struct llama_model * model) { | ||||
|         case LLM_ARCH_ORION: | ||||
|         case LLM_ARCH_INTERNLM2: | ||||
|         case LLM_ARCH_MINICPM: | ||||
|         case LLM_ARCH_GEMMA: | ||||
|             return LLAMA_ROPE_TYPE_NORM; | ||||
|  | ||||
|         // the pairs of head values are offset by n_rot/2 | ||||
| @@ -12101,6 +12100,7 @@ enum llama_rope_type llama_rope_type(const struct llama_model * model) { | ||||
|         case LLM_ARCH_QWEN: | ||||
|         case LLM_ARCH_QWEN2: | ||||
|         case LLM_ARCH_PHI2: | ||||
|         case LLM_ARCH_GEMMA: | ||||
|             return LLAMA_ROPE_TYPE_NEOX; | ||||
|  | ||||
|         // all model arches should be listed explicitly here | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Georgi Gerganov
					Georgi Gerganov