From 307772fcda771c0bff42e312d090addf445c1509 Mon Sep 17 00:00:00 2001 From: ixgbe <1113177880@qq.com> Date: Fri, 14 Nov 2025 15:12:56 +0800 Subject: [PATCH] readme : add RVV,ZVFH,ZFH,ZICBOP support for RISC-V (#17259) Signed-off-by: Wang Yang --- .github/copilot-instructions.md | 2 +- README.md | 1 + 2 files changed, 2 insertions(+), 1 deletion(-) diff --git a/.github/copilot-instructions.md b/.github/copilot-instructions.md index 3250e3279e..ad13c6ea8d 100644 --- a/.github/copilot-instructions.md +++ b/.github/copilot-instructions.md @@ -9,7 +9,7 @@ llama.cpp is a large-scale C/C++ project for efficient LLM (Large Language Model - **Size**: ~200k+ lines of code across 1000+ files - **Architecture**: Modular design with main library (`libllama`) and 40+ executable tools/examples - **Core dependency**: ggml tensor library (vendored in `ggml/` directory) -- **Backends supported**: CPU (AVX/NEON optimized), CUDA, Metal, Vulkan, SYCL, ROCm, MUSA +- **Backends supported**: CPU (AVX/NEON/RVV optimized), CUDA, Metal, Vulkan, SYCL, ROCm, MUSA - **License**: MIT ## Build Instructions diff --git a/README.md b/README.md index 258963ac16..2962783585 100644 --- a/README.md +++ b/README.md @@ -61,6 +61,7 @@ range of hardware - locally and in the cloud. - Plain C/C++ implementation without any dependencies - Apple silicon is a first-class citizen - optimized via ARM NEON, Accelerate and Metal frameworks - AVX, AVX2, AVX512 and AMX support for x86 architectures +- RVV, ZVFH, ZFH and ZICBOP support for RISC-V architectures - 1.5-bit, 2-bit, 3-bit, 4-bit, 5-bit, 6-bit, and 8-bit integer quantization for faster inference and reduced memory use - Custom CUDA kernels for running LLMs on NVIDIA GPUs (support for AMD GPUs via HIP and Moore Threads GPUs via MUSA) - Vulkan and SYCL backend support