From: ixgbe Date: Fri, 14 Nov 2025 07:12:56 +0000 (+0800) Subject: readme : add RVV,ZVFH,ZFH,ZICBOP support for RISC-V (#17259) X-Git-Tag: upstream/0.0.7446~390 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=307772fcda771c0bff42e312d090addf445c1509;p=pkg%2Fggml%2Fsources%2Fllama.cpp readme : add RVV,ZVFH,ZFH,ZICBOP support for RISC-V (#17259) Signed-off-by: Wang Yang --- diff --git a/.github/copilot-instructions.md b/.github/copilot-instructions.md index 3250e327..ad13c6ea 100644 --- a/.github/copilot-instructions.md +++ b/.github/copilot-instructions.md @@ -9,7 +9,7 @@ llama.cpp is a large-scale C/C++ project for efficient LLM (Large Language Model - **Size**: ~200k+ lines of code across 1000+ files - **Architecture**: Modular design with main library (`libllama`) and 40+ executable tools/examples - **Core dependency**: ggml tensor library (vendored in `ggml/` directory) -- **Backends supported**: CPU (AVX/NEON optimized), CUDA, Metal, Vulkan, SYCL, ROCm, MUSA +- **Backends supported**: CPU (AVX/NEON/RVV optimized), CUDA, Metal, Vulkan, SYCL, ROCm, MUSA - **License**: MIT ## Build Instructions diff --git a/README.md b/README.md index 258963ac..29627835 100644 --- a/README.md +++ b/README.md @@ -61,6 +61,7 @@ range of hardware - locally and in the cloud. - Plain C/C++ implementation without any dependencies - Apple silicon is a first-class citizen - optimized via ARM NEON, Accelerate and Metal frameworks - AVX, AVX2, AVX512 and AMX support for x86 architectures +- RVV, ZVFH, ZFH and ZICBOP support for RISC-V architectures - 1.5-bit, 2-bit, 3-bit, 4-bit, 5-bit, 6-bit, and 8-bit integer quantization for faster inference and reduced memory use - Custom CUDA kernels for running LLMs on NVIDIA GPUs (support for AMD GPUs via HIP and Moore Threads GPUs via MUSA) - Vulkan and SYCL backend support