From: Guoliang Hua Date: Mon, 2 Sep 2024 12:36:43 +0000 (+0800) Subject: llama-cli : remove duplicated log message (#9275) X-Git-Tag: upstream/0.0.4488~833 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=b60074f1c26d8354053a952844ef3f7ebefd8803;p=pkg%2Fggml%2Fsources%2Fllama.cpp llama-cli : remove duplicated log message (#9275) --- diff --git a/examples/main/main.cpp b/examples/main/main.cpp index 2c05afb0..c55efbb6 100644 --- a/examples/main/main.cpp +++ b/examples/main/main.cpp @@ -386,8 +386,8 @@ int main(int argc, char ** argv) { } LOGLN( - "recalculate the cached logits (check): embd_inp.empty() %s, n_matching_session_tokens %zu, embd_inp.size() %zu, session_tokens.size() %zu, embd_inp.size() %zu", - log_tostr(embd_inp.empty()), n_matching_session_tokens, embd_inp.size(), session_tokens.size(), embd_inp.size()); + "recalculate the cached logits (check): embd_inp.empty() %s, n_matching_session_tokens %zu, embd_inp.size() %zu, session_tokens.size() %zu", + log_tostr(embd_inp.empty()), n_matching_session_tokens, embd_inp.size(), session_tokens.size()); // if we will use the cache for the full prompt without reaching the end of the cache, force // reevaluation of the last token to recalculate the cached logits