mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-31 08:51:55 +00:00 
			
		
		
		
	main : disable token count by default (#4874)
This commit is contained in:
		| @@ -64,7 +64,7 @@ struct gpt_params { | ||||
|     int32_t n_beams                         = 0;     // if non-zero then use beam search of given width. | ||||
|     int32_t grp_attn_n                      = 1;     // group-attention factor | ||||
|     int32_t grp_attn_w                      = 512;   // group-attention width | ||||
|     int32_t token_interval                  = 512;   // show token count every 512 tokens | ||||
|     int32_t token_interval                  = -1;    // show token count every 512 tokens (-1 = disabled) | ||||
|     float   rope_freq_base                  = 0.0f;  // RoPE base frequency | ||||
|     float   rope_freq_scale                 = 0.0f;  // RoPE frequency scaling factor | ||||
|     float   yarn_ext_factor                 = -1.0f; // YaRN extrapolation mix factor | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Georgi Gerganov
					Georgi Gerganov