From: Georgi Gerganov Date: Sat, 13 May 2023 13:55:14 +0000 (+0300) Subject: llama : fix unused warning X-Git-Tag: gguf-v0.4.0~762 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=5a5aeb1e91009c72bf816400b758bb8a305616d7;p=pkg%2Fggml%2Fsources%2Fllama.cpp llama : fix unused warning --- diff --git a/llama.cpp b/llama.cpp index 73b932a7..98f49abd 100644 --- a/llama.cpp +++ b/llama.cpp @@ -1053,6 +1053,8 @@ static void llama_model_load_internal( fprintf(stderr, "%s: [cublas] total VRAM used: %zu MB\n", __func__, vram_total / 1024 / 1024); } +#else + (void) n_gpu_layers; #endif // loading time will be recalculate after the first eval, so