mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-30 08:42:00 +00:00 
			
		
		
		
	llama : add option to override model tensor buffers (#11397)
* llama : add option to override tensor buffers * ggml : fix possible underflow in ggml_nbytes
This commit is contained in:
		| @@ -445,7 +445,8 @@ llama_model_loader::llama_model_loader( | ||||
|         std::vector<std::string> & splits, | ||||
|         bool use_mmap, | ||||
|         bool check_tensors, | ||||
|         const struct llama_model_kv_override * param_overrides_p) { | ||||
|         const llama_model_kv_override * param_overrides_p, | ||||
|         const llama_model_tensor_buft_override * param_tensor_buft_overrides_p) { | ||||
|     int trace = 0; | ||||
|     if (getenv("LLAMA_TRACE")) { | ||||
|         trace = atoi(getenv("LLAMA_TRACE")); | ||||
| @@ -457,6 +458,8 @@ llama_model_loader::llama_model_loader( | ||||
|         } | ||||
|     } | ||||
|  | ||||
|     tensor_buft_overrides = param_tensor_buft_overrides_p; | ||||
|  | ||||
|     // Load the main GGUF | ||||
|     struct ggml_context * ctx = NULL; | ||||
|     struct gguf_init_params params = { | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Diego Devesa
					Diego Devesa