mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-10-29 08:41:22 +00:00 
			
		
		
		
	nix : enable curl (#8043)
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
This commit is contained in:
		| @@ -17,6 +17,7 @@ | ||||
|   rocmPackages, | ||||
|   vulkan-headers, | ||||
|   vulkan-loader, | ||||
|   curl, | ||||
|   useBlas ? builtins.all (x: !x) [ | ||||
|     useCuda | ||||
|     useMetalKit | ||||
| @@ -27,6 +28,7 @@ | ||||
|   useMetalKit ? stdenv.isAarch64 && stdenv.isDarwin, | ||||
|   useMpi ? false, # Increases the runtime closure size by ~700M | ||||
|   useRocm ? config.rocmSupport, | ||||
|   enableCurl ? true, | ||||
|   useVulkan ? false, | ||||
|   llamaVersion ? "0.0.0", # Arbitrary version, substituted by the flake | ||||
|  | ||||
| @@ -196,13 +198,15 @@ effectiveStdenv.mkDerivation ( | ||||
|       ++ optionals useMpi [ mpi ] | ||||
|       ++ optionals useRocm rocmBuildInputs | ||||
|       ++ optionals useBlas [ blas ] | ||||
|       ++ optionals useVulkan vulkanBuildInputs; | ||||
|       ++ optionals useVulkan vulkanBuildInputs | ||||
|       ++ optionals enableCurl [ curl ]; | ||||
|  | ||||
|     cmakeFlags = | ||||
|       [ | ||||
|         (cmakeBool "LLAMA_BUILD_SERVER" true) | ||||
|         (cmakeBool "BUILD_SHARED_LIBS" (!enableStatic)) | ||||
|         (cmakeBool "CMAKE_SKIP_BUILD_RPATH" true) | ||||
|         (cmakeBool "LLAMA_CURL" enableCurl) | ||||
|         (cmakeBool "GGML_NATIVE" false) | ||||
|         (cmakeBool "GGML_BLAS" useBlas) | ||||
|         (cmakeBool "GGML_CUDA" useCuda) | ||||
|   | ||||
		Reference in New Issue
	
	Block a user
	 Michael Francis
					Michael Francis