mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-30 08:42:00 +00:00 
			
		
		
		
	Server: Enable setting default sampling parameters via command-line (#8402)
* Load server sampling parameters from the server context by default. * Wordsmithing comment
This commit is contained in:
		| @@ -884,7 +884,8 @@ struct server_context { | |||||||
|  |  | ||||||
|     bool launch_slot_with_task(server_slot & slot, const server_task & task) { |     bool launch_slot_with_task(server_slot & slot, const server_task & task) { | ||||||
|         slot_params default_params; |         slot_params default_params; | ||||||
|         llama_sampling_params default_sparams; |         // Sampling parameter defaults are loaded from the global server context (but individual requests can still override them) | ||||||
|  |         llama_sampling_params default_sparams = params.sparams; | ||||||
|         auto & data = task.data; |         auto & data = task.data; | ||||||
|  |  | ||||||
|         if (data.count("__oaicompat") != 0) { |         if (data.count("__oaicompat") != 0) { | ||||||
|   | |||||||
		Reference in New Issue
	
	Block a user
	 Clint Herron
					Clint Herron