mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-31 08:51:55 +00:00 
			
		
		
		
	cuda : only use native when supported by cmake (#10389)
This commit is contained in:
		| @@ -12,7 +12,7 @@ if (CUDAToolkit_FOUND) | |||||||
|         # 61     == Pascal, __dp4a instruction (per-byte integer dot product) |         # 61     == Pascal, __dp4a instruction (per-byte integer dot product) | ||||||
|         # 70     == V100, FP16 tensor cores |         # 70     == V100, FP16 tensor cores | ||||||
|         # 75     == Turing, int8 tensor cores |         # 75     == Turing, int8 tensor cores | ||||||
|         if (GGML_NATIVE AND CUDAToolkit_VERSION VERSION_GREATER_EQUAL "11.6") |         if (GGML_NATIVE AND CUDAToolkit_VERSION VERSION_GREATER_EQUAL "11.6" AND CMAKE_VERSION VERSION_GREATER_EQUAL "3.24") | ||||||
|             set(CMAKE_CUDA_ARCHITECTURES "native") |             set(CMAKE_CUDA_ARCHITECTURES "native") | ||||||
|         elseif(GGML_CUDA_F16 OR GGML_CUDA_DMMV_F16) |         elseif(GGML_CUDA_F16 OR GGML_CUDA_DMMV_F16) | ||||||
|             set(CMAKE_CUDA_ARCHITECTURES "60;61;70;75") |             set(CMAKE_CUDA_ARCHITECTURES "60;61;70;75") | ||||||
|   | |||||||
		Reference in New Issue
	
	Block a user
	 Diego Devesa
					Diego Devesa