mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-10-27 08:21:30 +00:00
124 lines
3.6 KiB
Docker
124 lines
3.6 KiB
Docker
ARG GCC_VERSION=15.2.0
|
|
ARG UBUNTU_VERSION=24.04
|
|
|
|
### Build Llama.cpp stage
|
|
FROM gcc:${GCC_VERSION} AS build
|
|
|
|
RUN --mount=type=cache,target=/var/cache/apt,sharing=locked \
|
|
--mount=type=cache,target=/var/lib/apt/lists,sharing=locked \
|
|
apt update -y && \
|
|
apt upgrade -y && \
|
|
apt install -y --no-install-recommends \
|
|
git cmake ccache ninja-build \
|
|
# WARNING: Do not use libopenblas-openmp-dev. libopenblas-dev is faster.
|
|
libopenblas-dev libcurl4-openssl-dev && \
|
|
rm -rf /var/lib/apt/lists/*
|
|
|
|
WORKDIR /app
|
|
COPY . .
|
|
|
|
RUN --mount=type=cache,target=/root/.ccache \
|
|
--mount=type=cache,target=/app/build \
|
|
cmake -S . -B build -G Ninja \
|
|
-DCMAKE_BUILD_TYPE=Release \
|
|
-DCMAKE_C_COMPILER_LAUNCHER=ccache \
|
|
-DCMAKE_CXX_COMPILER_LAUNCHER=ccache \
|
|
-DLLAMA_BUILD_TESTS=OFF \
|
|
-DGGML_BACKEND_DL=OFF \
|
|
-DGGML_NATIVE=OFF \
|
|
-DGGML_BLAS=ON \
|
|
-DGGML_BLAS_VENDOR=OpenBLAS && \
|
|
cmake --build build --config Release -j $(nproc) && \
|
|
cmake --install build --prefix /opt/llama.cpp
|
|
|
|
COPY *.py /opt/llama.cpp/bin
|
|
COPY .devops/tools.sh /opt/llama.cpp/bin
|
|
|
|
COPY gguf-py /opt/llama.cpp/gguf-py
|
|
COPY requirements.txt /opt/llama.cpp/gguf-py
|
|
COPY requirements /opt/llama.cpp/gguf-py/requirements
|
|
|
|
|
|
### Collect all llama.cpp binaries, libraries and distro libraries
|
|
FROM scratch AS collector
|
|
|
|
# Copy llama.cpp binaries and libraries
|
|
COPY --from=build /opt/llama.cpp/bin /llama.cpp/bin
|
|
COPY --from=build /opt/llama.cpp/lib /llama.cpp/lib
|
|
COPY --from=build /opt/llama.cpp/gguf-py /llama.cpp/gguf-py
|
|
|
|
|
|
### Base image
|
|
FROM ubuntu:${UBUNTU_VERSION} AS base
|
|
|
|
RUN --mount=type=cache,target=/var/cache/apt,sharing=locked \
|
|
--mount=type=cache,target=/var/lib/apt/lists,sharing=locked \
|
|
apt update -y && \
|
|
apt install -y --no-install-recommends \
|
|
# WARNING: Do not use libopenblas-openmp-dev. libopenblas-dev is faster.
|
|
# See: https://github.com/ggml-org/llama.cpp/pull/15915#issuecomment-3317166506
|
|
curl libgomp1 libopenblas-dev && \
|
|
apt autoremove -y && \
|
|
apt clean -y && \
|
|
rm -rf /tmp/* /var/tmp/* && \
|
|
find /var/cache/apt/archives /var/lib/apt/lists -not -name lock -type f -delete && \
|
|
find /var/cache -type f -delete
|
|
|
|
# Copy llama.cpp libraries
|
|
COPY --from=collector /llama.cpp/lib /usr/lib/s390x-linux-gnu
|
|
|
|
|
|
### Full
|
|
FROM base AS full
|
|
|
|
ENV PATH="/root/.cargo/bin:${PATH}"
|
|
WORKDIR /app
|
|
|
|
RUN --mount=type=cache,target=/var/cache/apt,sharing=locked \
|
|
--mount=type=cache,target=/var/lib/apt/lists,sharing=locked \
|
|
apt update -y && \
|
|
apt install -y \
|
|
git cmake libjpeg-dev \
|
|
python3 python3-pip python3-dev && \
|
|
apt autoremove -y && \
|
|
apt clean -y && \
|
|
rm -rf /tmp/* /var/tmp/* && \
|
|
find /var/cache/apt/archives /var/lib/apt/lists -not -name lock -type f -delete && \
|
|
find /var/cache -type f -delete
|
|
|
|
RUN curl https://sh.rustup.rs -sSf | bash -s -- -y
|
|
|
|
COPY --from=collector /llama.cpp/bin /app
|
|
COPY --from=collector /llama.cpp/gguf-py /app/gguf-py
|
|
|
|
RUN pip install --no-cache-dir --break-system-packages \
|
|
-r /app/gguf-py/requirements.txt
|
|
|
|
ENTRYPOINT [ "/app/tools.sh" ]
|
|
|
|
|
|
### CLI Only
|
|
FROM base AS light
|
|
|
|
WORKDIR /llama.cpp/bin
|
|
|
|
# Copy llama.cpp binaries and libraries
|
|
COPY --from=collector /llama.cpp/bin/llama-cli /llama.cpp/bin
|
|
|
|
ENTRYPOINT [ "/llama.cpp/bin/llama-cli" ]
|
|
|
|
|
|
### Server
|
|
FROM base AS server
|
|
|
|
ENV LLAMA_ARG_HOST=0.0.0.0
|
|
|
|
WORKDIR /llama.cpp/bin
|
|
|
|
# Copy llama.cpp binaries and libraries
|
|
COPY --from=collector /llama.cpp/bin/llama-server /llama.cpp/bin
|
|
|
|
EXPOSE 8080
|
|
|
|
ENTRYPOINT [ "/llama.cpp/bin/llama-server" ]
|