From: Georgi Gerganov Date: Tue, 19 Aug 2025 05:45:26 +0000 (+0300) Subject: server : remove swa_full warning (#15399) X-Git-Tag: upstream/0.0.6527~326 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=9d262f4bad0d37838100133537aaf0a83835ed12;p=pkg%2Fggml%2Fsources%2Fllama.cpp server : remove swa_full warning (#15399) --- diff --git a/src/llama-context.cpp b/src/llama-context.cpp index 7d7abad5..1ebfc88a 100644 --- a/src/llama-context.cpp +++ b/src/llama-context.cpp @@ -145,11 +145,6 @@ llama_context::llama_context( __func__, n_ctx_per_seq, hparams.n_ctx_train); } - if (!params.swa_full && cparams.n_seq_max > 1 && hparams.is_swa_any()) { - LLAMA_LOG_WARN("%s: requested n_seq_max (%u) > 1, but swa_full is not enabled -- performance may be degraded: %s\n", - __func__, cparams.n_seq_max, "https://github.com/ggml-org/llama.cpp/pull/13845#issuecomment-2924800573"); - } - if (!hparams.vocab_only) { // GPU backends for (auto * dev : model.devices) {