mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-31 08:51:55 +00:00 
			
		
		
		
	server : bring back timings_per_token (#15879)
This commit is contained in:
		| @@ -308,6 +308,7 @@ struct server_task { | ||||
|  | ||||
|         // enabling this will output extra debug information in the HTTP responses from the server | ||||
|         params.verbose           = params_base.verbosity > 9; | ||||
|         params.timings_per_token = json_value(data, "timings_per_token", false); | ||||
|  | ||||
|         params.stream           = json_value(data, "stream",             false); | ||||
|         params.cache_prompt     = json_value(data, "cache_prompt",       true); | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Xuan-Son Nguyen
					Xuan-Son Nguyen