mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-31 08:51:55 +00:00 
			
		
		
		
	hparams : add comment [no ci]
This commit is contained in:
		| @@ -36,7 +36,7 @@ struct llama_hparams { | ||||
|     uint32_t n_layer; | ||||
|     uint32_t n_rot; | ||||
|     uint32_t n_swa = 0; // sliding window attention (SWA) | ||||
|     uint32_t n_swa_pattern = 1; | ||||
|     uint32_t n_swa_pattern = 1; // by default, all layers use non-sliding-window attention | ||||
|     uint32_t n_embd_head_k; // dimension of keys (d_k). d_q is assumed to be the same, but there are n_head q heads, and only n_head_kv k-v heads | ||||
|     uint32_t n_embd_head_v; // dimension of values (d_v) aka n_embd_head | ||||
|     uint32_t n_expert = 0; | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Georgi Gerganov
					Georgi Gerganov