From: Howard Su Date: Mon, 19 Jun 2023 15:10:37 +0000 (+0800) Subject: cmake : fix build shared ggml when CUDA is enabled (#1929) X-Git-Tag: gguf-v0.4.0~594 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=1e3abfcef073e73c2b31e8570cb06c5cb2fd1f55;p=pkg%2Fggml%2Fsources%2Fllama.cpp cmake : fix build shared ggml when CUDA is enabled (#1929) Co-authored-by: Georgi Gerganov --- diff --git a/CMakeLists.txt b/CMakeLists.txt index dc06365d..a598593b 100644 --- a/CMakeLists.txt +++ b/CMakeLists.txt @@ -469,6 +469,7 @@ add_library(ggml_static STATIC $) if (BUILD_SHARED_LIBS) set_target_properties(ggml PROPERTIES POSITION_INDEPENDENT_CODE ON) add_library(ggml_shared SHARED $) + target_link_libraries(ggml_shared PUBLIC Threads::Threads ${LLAMA_EXTRA_LIBS}) endif() add_library(llama @@ -500,6 +501,11 @@ if (GGML_SOURCES_CUDA) set_property(TARGET ggml_static PROPERTY CUDA_ARCHITECTURES "native") set_property(TARGET ggml_static PROPERTY CUDA_SELECT_NVCC_ARCH_FLAGS "Auto") + if (BUILD_SHARED_LIBS) + set_property(TARGET ggml_shared PROPERTY CUDA_ARCHITECTURES "native") + set_property(TARGET ggml_shared PROPERTY CUDA_SELECT_NVCC_ARCH_FLAGS "Auto") + endif() + set_property(TARGET llama PROPERTY CUDA_ARCHITECTURES "native") endif()