From: Jared Van Bortel Date: Mon, 29 Jan 2024 22:11:27 +0000 (-0500) Subject: kompute : fix fallback to CPU (#5201) X-Git-Tag: upstream/0.0.4488~2480 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=6daa69ee81851ab26ca8aefca1a4202941fc0262;p=pkg%2Fggml%2Fsources%2Fllama.cpp kompute : fix fallback to CPU (#5201) --- diff --git a/llama.cpp b/llama.cpp index 9631506c..796aaa89 100644 --- a/llama.cpp +++ b/llama.cpp @@ -4136,7 +4136,7 @@ static int llama_model_load(const std::string & fname, llama_model & model, llam } #ifdef GGML_USE_KOMPUTE - if (ggml_vk_has_device() && params.n_gpu_layers > 0 && ( + if (params.n_gpu_layers > 0 && ( !(model.arch == LLM_ARCH_LLAMA || model.arch == LLM_ARCH_FALCON) || !( model.ftype == LLAMA_FTYPE_ALL_F32 || @@ -4145,8 +4145,8 @@ static int llama_model_load(const std::string & fname, llama_model & model, llam model.ftype == LLAMA_FTYPE_MOSTLY_Q4_1 ) )) { - // disable Vulkan due to unsupported model architecture or quantization type // TODO(cebtenzzre): propagate this error outside of llama_load_model_from_file + LLAMA_LOG_WARN("%s: disabling Kompute due to unsupported model arch or quantization\n", __func__); params.n_gpu_layers = 0; } #endif