]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/commitdiff
FP16 is supported in CM=6.0 (#2177)
authorHoward Su <redacted>
Wed, 12 Jul 2023 12:18:40 +0000 (20:18 +0800)
committerGitHub <redacted>
Wed, 12 Jul 2023 12:18:40 +0000 (20:18 +0800)
* FP16 is supported in CM=6.0

* Building PTX code for both of 60 and 61

Co-authored-by: Johannes Gäßler <redacted>
CMakeLists.txt

index cf6cd34f18ec1118a275ca84982d0caefa3f9381..d9381dae1a10ae23dc5015a9bb64c09e47f5e7ba 100644 (file)
@@ -272,7 +272,7 @@ if (LLAMA_CUBLAS)
 
     if (NOT DEFINED CMAKE_CUDA_ARCHITECTURES)
         if (LLAMA_CUDA_DMMV_F16)
-            set(CMAKE_CUDA_ARCHITECTURES "61") # needed for f16 CUDA intrinsics
+            set(CMAKE_CUDA_ARCHITECTURES "60;61") # needed for f16 CUDA intrinsics
         else()
             set(CMAKE_CUDA_ARCHITECTURES "52;61") # lowest CUDA 12 standard + lowest for integer intrinsics
         endif()