From: Georgi Gerganov Date: Wed, 6 Dec 2023 18:21:59 +0000 (+0200) Subject: server : recognize cache_prompt parameter in OAI API (#4347) X-Git-Tag: upstream/0.0.4488~2870 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=05cd6e5036d72d0930de4d8f6be7bce09e8dda24;p=pkg%2Fggml%2Fsources%2Fllama.cpp server : recognize cache_prompt parameter in OAI API (#4347) --- diff --git a/examples/server/server.cpp b/examples/server/server.cpp index 911f7bbe..369f81a8 100644 --- a/examples/server/server.cpp +++ b/examples/server/server.cpp @@ -2387,6 +2387,7 @@ json oaicompat_completion_params_parse( // Map OpenAI parameters to llama.cpp parameters llama_params["prompt"] = format_chatml(body["messages"]); // OpenAI 'messages' to llama.cpp 'prompt' + llama_params["cache_prompt"] = json_value(body, "cache_prompt", false); llama_params["temperature"] = json_value(body, "temperature", 0.8); llama_params["top_k"] = json_value(body, "top_k", 40); llama_params["top_p"] = json_value(body, "top_p", 0.95);