mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-11-02 09:12:03 +00:00
Disable CUDA host buffers on integrated GPUs (#16308)
This commit is contained in:
@@ -231,7 +231,7 @@ static ggml_cuda_device_info ggml_cuda_init() {
|
|||||||
|
|
||||||
info.default_tensor_split[id] = total_vram;
|
info.default_tensor_split[id] = total_vram;
|
||||||
total_vram += prop.totalGlobalMem;
|
total_vram += prop.totalGlobalMem;
|
||||||
info.devices[id].integrated = prop.integrated;
|
info.devices[id].integrated = false; // Temporarily disabled due to issues with corrupted output (e.g. #15034)
|
||||||
info.devices[id].nsm = prop.multiProcessorCount;
|
info.devices[id].nsm = prop.multiProcessorCount;
|
||||||
info.devices[id].smpb = prop.sharedMemPerBlock;
|
info.devices[id].smpb = prop.sharedMemPerBlock;
|
||||||
info.devices[id].warp_size = prop.warpSize;
|
info.devices[id].warp_size = prop.warpSize;
|
||||||
|
|||||||
Reference in New Issue
Block a user