mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-30 08:42:00 +00:00 
			
		
		
		
	Server: Enable setting default sampling parameters via command-line (#8402)
* Load server sampling parameters from the server context by default. * Wordsmithing comment
This commit is contained in:
		| @@ -884,7 +884,8 @@ struct server_context { | ||||
|  | ||||
|     bool launch_slot_with_task(server_slot & slot, const server_task & task) { | ||||
|         slot_params default_params; | ||||
|         llama_sampling_params default_sparams; | ||||
|         // Sampling parameter defaults are loaded from the global server context (but individual requests can still override them) | ||||
|         llama_sampling_params default_sparams = params.sparams; | ||||
|         auto & data = task.data; | ||||
|  | ||||
|         if (data.count("__oaicompat") != 0) { | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Clint Herron
					Clint Herron