From: Daniel Bevenius Date: Mon, 12 Jan 2026 12:47:58 +0000 (+0100) Subject: examples : add --kv-unified to batched example (#18774) X-Git-Tag: upstream/0.0.7721~8 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=4150da9a956d0587a3e8a08e69940f9a27f88e5c;p=pkg%2Fggml%2Fsources%2Fllama.cpp examples : add --kv-unified to batched example (#18774) This commit adds the --kv-unified flag to the batched example. This flag is currently specified in the README.md as required, but is currently not available as a command line option for the batched example. The motivation for this is that specifying this flag as the README instructs, will lead to an error about the flag not being recognized, and without this option the example fail with the following error: ```console split_equal: sequential split is not supported when there are coupled sequences in the input batch (you may need to use the -kvu flag) decode: failed to find a memory slot for batch of size 4 main: llama_decode() failed ``` --- diff --git a/common/arg.cpp b/common/arg.cpp index ec0a2f01..f2675f84 100644 --- a/common/arg.cpp +++ b/common/arg.cpp @@ -1295,7 +1295,7 @@ common_params_context common_params_parser_init(common_params & params, llama_ex [](common_params & params) { params.kv_unified = true; } - ).set_env("LLAMA_ARG_KV_UNIFIED").set_examples({LLAMA_EXAMPLE_SERVER, LLAMA_EXAMPLE_PERPLEXITY})); + ).set_env("LLAMA_ARG_KV_UNIFIED").set_examples({LLAMA_EXAMPLE_SERVER, LLAMA_EXAMPLE_PERPLEXITY, LLAMA_EXAMPLE_BATCHED})); add_opt(common_arg( {"--context-shift"}, {"--no-context-shift"}, diff --git a/common/common.h b/common/common.h index 7794c026..b3ac04c4 100644 --- a/common/common.h +++ b/common/common.h @@ -80,6 +80,7 @@ int32_t cpu_get_num_math(); // enum llama_example { + LLAMA_EXAMPLE_BATCHED, LLAMA_EXAMPLE_DEBUG, LLAMA_EXAMPLE_COMMON, LLAMA_EXAMPLE_SPECULATIVE, diff --git a/examples/batched/batched.cpp b/examples/batched/batched.cpp index 6b134b4f..68750387 100644 --- a/examples/batched/batched.cpp +++ b/examples/batched/batched.cpp @@ -21,7 +21,7 @@ int main(int argc, char ** argv) { params.prompt = "Hello my name is"; params.n_predict = 32; - if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_COMMON, print_usage)) { + if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_BATCHED, print_usage)) { return 1; }