## Running MUSA CI in a Docker Container Assuming `$PWD` is the root of the `llama.cpp` repository, follow these steps to set up and run MUSA CI in a Docker container: ### 1. Create a local directory to store cached models, configuration files and venv: ```bash mkdir -p $HOME/llama.cpp/ci-cache ``` ### 2. Create a local directory to store CI run results: ```bash mkdir -p $HOME/llama.cpp/ci-results ``` ### 3. Start a Docker container and run the CI: ```bash docker run --privileged -it \ -v $HOME/llama.cpp/ci-cache:/ci-cache \ -v $HOME/llama.cpp/ci-results:/ci-results \ -v $PWD:/ws -w /ws \ mthreads/musa:rc4.3.0-devel-ubuntu22.04-amd64 ``` Inside the container, execute the following commands: ```bash apt update -y && apt install -y bc cmake ccache git python3.10-venv time unzip wget git config --global --add safe.directory /ws GG_BUILD_MUSA=1 bash ./ci/run.sh /ci-results /ci-cache ``` This setup ensures that the CI runs within an isolated Docker environment while maintaining cached files and results across runs.