mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-29 08:41:22 +00:00 
			
		
		
		
	context : remove redundant explicit casting to the same type (#15948)
The function 'output_reserve' return type is 'uint32_t', so need to add explicit casting.
This commit is contained in:
		| @@ -181,7 +181,7 @@ llama_context::llama_context( | |||||||
|         // graph outputs buffer |         // graph outputs buffer | ||||||
|         { |         { | ||||||
|             // resized during inference when a batch uses more outputs |             // resized during inference when a batch uses more outputs | ||||||
|             if ((uint32_t) output_reserve(params.n_seq_max) < params.n_seq_max) { |             if (output_reserve(params.n_seq_max) < params.n_seq_max) { | ||||||
|                 throw std::runtime_error("failed to reserve initial output buffer"); |                 throw std::runtime_error("failed to reserve initial output buffer"); | ||||||
|             } |             } | ||||||
|  |  | ||||||
|   | |||||||
		Reference in New Issue
	
	Block a user
	 Haiyue Wang
					Haiyue Wang