]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/commitdiff
Document the new max GPU layers default in help (#15771)
authorEric Curtin <redacted>
Thu, 4 Sep 2025 09:49:44 +0000 (10:49 +0100)
committerGitHub <redacted>
Thu, 4 Sep 2025 09:49:44 +0000 (10:49 +0100)
This is a key change, just letting users know.

Signed-off-by: Eric Curtin <redacted>
common/arg.cpp

index fcee0c447007765494ea991f5d080a8ba3debc20..7507c81155071559ec1c4fc62a52555c1d6f3222 100644 (file)
@@ -2466,7 +2466,7 @@ common_params_context common_params_parser_init(common_params & params, llama_ex
     ).set_examples({LLAMA_EXAMPLE_SPECULATIVE, LLAMA_EXAMPLE_SERVER}).set_env("LLAMA_ARG_N_CPU_MOE_DRAFT"));
     add_opt(common_arg(
         {"-ngl", "--gpu-layers", "--n-gpu-layers"}, "N",
-        "number of layers to store in VRAM",
+        string_format("max. number of layers to store in VRAM (default: %d)", params.n_gpu_layers),
         [](common_params & params, int value) {
             params.n_gpu_layers = value;
             if (!llama_supports_gpu_offload()) {