From 0a8b764d39e8c549a29c08f7763aadb4fc971c33 Mon Sep 17 00:00:00 2001 From: Xiaodong Ye Date: Tue, 26 Nov 2024 18:26:54 +0800 Subject: [PATCH] README.md: Add MUSA as supported backend Signed-off-by: Xiaodong Ye --- README.md | 13 +++++++++++-- 1 file changed, 11 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index dbaec5077..0dde5f5e9 100644 --- a/README.md +++ b/README.md @@ -200,7 +200,7 @@ CMAKE_ARGS="-DGGML_VULKAN=on" pip install llama-cpp-python To install with SYCL support, set the `GGML_SYCL=on` environment variable before installing: ```bash -source /opt/intel/oneapi/setvars.sh +source /opt/intel/oneapi/setvars.sh CMAKE_ARGS="-DGGML_SYCL=on -DCMAKE_C_COMPILER=icx -DCMAKE_CXX_COMPILER=icpx" pip install llama-cpp-python ``` @@ -211,11 +211,20 @@ CMAKE_ARGS="-DGGML_SYCL=on -DCMAKE_C_COMPILER=icx -DCMAKE_CXX_COMPILER=icpx" pip To install with RPC support, set the `GGML_RPC=on` environment variable before installing: ```bash -source /opt/intel/oneapi/setvars.sh +source /opt/intel/oneapi/setvars.sh CMAKE_ARGS="-DGGML_RPC=on" pip install llama-cpp-python ``` +
+MUSA + +To install with MUSA support, set the `GGML_MUSA=on` environment variable before installing: + +```bash +CMAKE_ARGS="-DGGML_MUSA=on" pip install llama-cpp-python +``` +
### Windows Notes