From: Sigbjørn Skjæret Date: Tue, 24 Jun 2025 07:31:00 +0000 (+0200) Subject: main : honor --verbose-prompt on interactive prompts (#14350) X-Git-Tag: upstream/0.0.5760~11 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=abf241045d09cad70dc797b0fba393ad09ee2cbe;p=pkg%2Fggml%2Fsources%2Fllama.cpp main : honor --verbose-prompt on interactive prompts (#14350) --- diff --git a/tools/main/main.cpp b/tools/main/main.cpp index 154b37cd..516bf096 100644 --- a/tools/main/main.cpp +++ b/tools/main/main.cpp @@ -917,10 +917,19 @@ int main(int argc, char ** argv) { embd_inp.insert(embd_inp.end(), line_inp.begin(), line_inp.end()); embd_inp.insert(embd_inp.end(), line_sfx.begin(), line_sfx.end()); + if (params.verbose_prompt) { + LOG_INF("%s: number of tokens in prompt = %zu\n", __func__, embd_inp.size() - original_size); + } + for (size_t i = original_size; i < embd_inp.size(); ++i) { const llama_token token = embd_inp[i]; + const std::string token_str = common_token_to_piece(ctx, token); output_tokens.push_back(token); - output_ss << common_token_to_piece(ctx, token); + output_ss << token_str; + + if (params.verbose_prompt) { + LOG_INF("%6d -> '%s'\n", token, token_str.c_str()); + } } // reset assistant message