]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/commitdiff
Change Windows AMD example to release build to make inference much faster. (#6525)
authorMark Fairbairn <redacted>
Sun, 7 Apr 2024 18:52:19 +0000 (19:52 +0100)
committerGitHub <redacted>
Sun, 7 Apr 2024 18:52:19 +0000 (20:52 +0200)
README.md

index 20dd783dc91062318dffc6d663c0ef20b7b07069..2fbbb4a0c4282a9934d8efcda853bf066235032b 100644 (file)
--- a/README.md
+++ b/README.md
@@ -519,7 +519,7 @@ Building the program with BLAS support may lead to some performance improvements
     set PATH=%HIP_PATH%\bin;%PATH%
     mkdir build
     cd build
-    cmake -G Ninja -DAMDGPU_TARGETS=gfx1100 -DLLAMA_HIPBLAS=ON -DCMAKE_C_COMPILER=clang -DCMAKE_CXX_COMPILER=clang++ ..
+    cmake -G Ninja -DAMDGPU_TARGETS=gfx1100 -DLLAMA_HIPBLAS=ON -DCMAKE_C_COMPILER=clang -DCMAKE_CXX_COMPILER=clang++ -DCMAKE_BUILD_TYPE=Release ..
     cmake --build .
     ```
     Make sure that `AMDGPU_TARGETS` is set to the GPU arch you want to compile for. The above example uses `gfx1100` that corresponds to Radeon RX 7900XTX/XT/GRE. You can find a list of targets [here](https://llvm.org/docs/AMDGPUUsage.html#processors)