From: Oleksandr Kuvshynov Date: Thu, 13 Feb 2025 06:25:34 +0000 (-0500) Subject: llama.cpp: fix warning message (#11839) X-Git-Tag: upstream/0.0.4719~15 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=e4376270d971cff7992bdb6c5412a739195b1459;p=pkg%2Fggml%2Fsources%2Fllama.cpp llama.cpp: fix warning message (#11839) There was a typo-like error, which would print the same number twice if request is received with n_predict > server-side config. Before the fix: ``` slot launch_slot_: id 0 | task 0 | n_predict = 4096 exceeds server configuration, setting to 4096 ``` After the fix: ``` slot launch_slot_: id 0 | task 0 | n_predict = 8192 exceeds server configuration, setting to 4096 ``` --- diff --git a/examples/server/server.cpp b/examples/server/server.cpp index d320e9d6..5a36cd2a 100644 --- a/examples/server/server.cpp +++ b/examples/server/server.cpp @@ -2073,8 +2073,8 @@ struct server_context { if (slot.n_predict > 0 && slot.params.n_predict > slot.n_predict) { // Might be better to reject the request with a 400 ? + SLT_WRN(slot, "n_predict = %d exceeds server configuration, setting to %d", slot.params.n_predict, slot.n_predict); slot.params.n_predict = slot.n_predict; - SLT_WRN(slot, "n_predict = %d exceeds server configuration, setting to %d", slot.n_predict, slot.n_predict); } if (slot.params.ignore_eos && has_eos_token) {