mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-30 08:42:00 +00:00 
			
		
		
		
	llama : add 18-layer model type for Gemma 3-270m (#15319)
This commit adds support for the 18-layer model type in the Gemma3 series, which is the size of the Gemma3-270m model. The motivation for this commit is was the only change required for Gemma3-270m to be converted to GGUF format and used with llama.cpp. Once the model has been converted and uploaded to Huggingface it can be used like this: ```console $ ./build/bin/llama-cli -hf ggml-org/gemma-3-270m-GGUF:Q8_0 ```
This commit is contained in:
		| @@ -1095,6 +1095,7 @@ void llama_model::load_hparams(llama_model_loader & ml) { | ||||
|                 ml.get_key(LLM_KV_ATTENTION_LAYERNORM_RMS_EPS, hparams.f_norm_rms_eps); | ||||
|  | ||||
|                 switch (hparams.n_layer) { | ||||
|                     case 18: type = LLM_TYPE_537M; break; | ||||
|                     case 26: type = LLM_TYPE_1B; break; | ||||
|                     case 34: type = LLM_TYPE_4B; break; | ||||
|                     case 48: type = LLM_TYPE_12B; break; | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Daniel Bevenius
					Daniel Bevenius