mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2025-11-04 09:32:00 +00:00
devops: attempt at fixing model loading failure
Signed-off-by: Aaron Teo <aaron.teo1@ibm.com>
This commit is contained in:
@@ -22,6 +22,9 @@ RUN --mount=type=cache,target=/root/.ccache \
|
|||||||
&& cmake --build build --config Release -j $(nproc) \
|
&& cmake --build build --config Release -j $(nproc) \
|
||||||
&& cmake --install build --prefix /opt/llama.cpp
|
&& cmake --install build --prefix /opt/llama.cpp
|
||||||
|
|
||||||
|
RUN find /opt/llama.cpp/bin -name "*.so" -exec mv {} /opt/llama.cpp/lib \;
|
||||||
|
|
||||||
|
# DOUBLE CHECK ALL FILES ARE COPIED INTO COLLECTOR
|
||||||
RUN cp *.py /opt/llama.cpp \
|
RUN cp *.py /opt/llama.cpp \
|
||||||
&& cp -r gguf-py /opt/llama.cpp \
|
&& cp -r gguf-py /opt/llama.cpp \
|
||||||
&& cp -r requirements /opt/llama.cpp \
|
&& cp -r requirements /opt/llama.cpp \
|
||||||
|
|||||||
Reference in New Issue
Block a user