mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-31 08:51:55 +00:00 
			
		
		
		
	rpc: fix register position (#11424)
Signed-off-by: thxCode <thxcode0824@gmail.com>
This commit is contained in:
		| @@ -1303,10 +1303,12 @@ bool llama_model::load_tensors(llama_model_loader & ml) { | ||||
|     const int act_gpu_layers = devices.empty() ? 0 : std::min(n_gpu_layers, (int)n_layer + 1); | ||||
|     auto get_layer_buft_list = [&](int il) -> llama_model::impl::layer_dev { | ||||
|         if (il < i_gpu_start || (il - i_gpu_start) >= act_gpu_layers) { | ||||
|             LLAMA_LOG_DEBUG("load_tensors: layer %3d assigned to device %s\n", il, ggml_backend_dev_name(cpu_dev)); | ||||
|             return {cpu_dev, &pimpl->cpu_buft_list}; | ||||
|         } | ||||
|         const int layer_gpu = std::upper_bound(splits.begin(), splits.begin() + n_devices(), float(il - i_gpu_start)/act_gpu_layers) - splits.begin(); | ||||
|         auto * dev = devices.at(layer_gpu); | ||||
|         LLAMA_LOG_DEBUG("load_tensors: layer %3d assigned to device %s\n", il, ggml_backend_dev_name(dev)); | ||||
|         return {dev, &pimpl->gpu_buft_list.at(dev)}; | ||||
|     }; | ||||
|  | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Frank Mai
					Frank Mai