mirror of
				https://github.com/ggml-org/llama.cpp.git
				synced 2025-11-04 09:32:00 +00:00 
			
		
		
		
	main: target name -> llama-cli
This commit is contained in:
		@@ -23,7 +23,7 @@ RUN if [ "${LLAMA_SYCL_F16}" = "ON" ]; then \
 | 
			
		||||
        export OPT_SYCL_F16="-DLLAMA_SYCL_F16=ON"; \
 | 
			
		||||
    fi && \
 | 
			
		||||
    cmake -B build -DLLAMA_SYCL=ON -DCMAKE_C_COMPILER=icx -DCMAKE_CXX_COMPILER=icpx ${OPT_SYCL_F16} && \
 | 
			
		||||
    cmake --build build --config Release --target llama-bin
 | 
			
		||||
    cmake --build build --config Release --target llama-cli
 | 
			
		||||
 | 
			
		||||
FROM intel/oneapi-basekit:$ONEAPI_VERSION as runtime
 | 
			
		||||
 | 
			
		||||
 
 | 
			
		||||
@@ -15,7 +15,7 @@ RUN wget -qO - https://packages.lunarg.com/lunarg-signing-key-pub.asc | apt-key
 | 
			
		||||
WORKDIR /app
 | 
			
		||||
COPY . .
 | 
			
		||||
RUN cmake -B build -DLLAMA_VULKAN=1 && \
 | 
			
		||||
    cmake --build build --config Release --target llama-bin
 | 
			
		||||
    cmake --build build --config Release --target llama-cli
 | 
			
		||||
 | 
			
		||||
# Clean up
 | 
			
		||||
WORKDIR /
 | 
			
		||||
 
 | 
			
		||||
@@ -1,4 +1,4 @@
 | 
			
		||||
set(TARGET llama-bin)
 | 
			
		||||
set(TARGET llama-cli)
 | 
			
		||||
add_executable(${TARGET} main.cpp)
 | 
			
		||||
set_target_properties(${TARGET} PROPERTIES OUTPUT_NAME llama)
 | 
			
		||||
install(TARGETS ${TARGET} RUNTIME)
 | 
			
		||||
 
 | 
			
		||||
		Reference in New Issue
	
	Block a user