mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-11-03 09:22:01 +00:00 
			
		
		
		
	common : refactor downloading system, handle mmproj with -hf option (#12694)
* (wip) refactor downloading system [no ci] * fix all examples * fix mmproj with -hf * gemma3: update readme * only handle mmproj in llava example * fix multi-shard download * windows: fix problem with std::min and std::max * fix 2
This commit is contained in:
		@@ -38,7 +38,7 @@ int main(int argc, char ** argv) {
 | 
			
		||||
 | 
			
		||||
    llama_model_params model_params = common_model_params_to_llama(params);
 | 
			
		||||
 | 
			
		||||
    llama_model * model = llama_model_load_from_file(params.model.c_str(), model_params);
 | 
			
		||||
    llama_model * model = llama_model_load_from_file(params.model.path.c_str(), model_params);
 | 
			
		||||
 | 
			
		||||
    if (model == NULL) {
 | 
			
		||||
        fprintf(stderr , "%s: error: unable to load model\n" , __func__);
 | 
			
		||||
 
 | 
			
		||||
		Reference in New Issue
	
	Block a user