mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-11-03 09:22:01 +00:00 
			
		
		
		
	CUDA/HIP: add warp_size to cuda_device_info
This commit is contained in:
		@@ -520,6 +520,7 @@ struct ggml_cuda_device_info {
 | 
			
		||||
        bool    vmm;                // virtual memory support
 | 
			
		||||
        size_t  vmm_granularity;    // granularity of virtual memory
 | 
			
		||||
        size_t  total_vram;
 | 
			
		||||
        int     warp_size;          // Number of threads in a dispatch
 | 
			
		||||
    };
 | 
			
		||||
 | 
			
		||||
    cuda_device_info devices[GGML_CUDA_MAX_DEVICES] = {};
 | 
			
		||||
 
 | 
			
		||||
		Reference in New Issue
	
	Block a user