From: Diego Devesa Date: Fri, 16 May 2025 14:38:07 +0000 (-0700) Subject: llama : print hint when loading a model when no backends are loaded (#13589) X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=5364ae4ba53cc6367b8c8bf78876839122ca4e57;p=pkg%2Fggml%2Fsources%2Fllama.cpp llama : print hint when loading a model when no backends are loaded (#13589) --- diff --git a/src/llama.cpp b/src/llama.cpp index 9fdddf7b..2f06e0f8 100644 --- a/src/llama.cpp +++ b/src/llama.cpp @@ -140,6 +140,11 @@ static struct llama_model * llama_model_load_from_file_impl( struct llama_model_params params) { ggml_time_init(); + if (!params.vocab_only && ggml_backend_reg_count() == 0) { + LLAMA_LOG_ERROR("%s: no backends are loaded. hint: use ggml_backend_load() or ggml_backend_load_all() to load a backend before calling this function\n", __func__); + return nullptr; + } + unsigned cur_percentage = 0; if (params.progress_callback == NULL) { params.progress_callback_user_data = &cur_percentage;