]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/commit
server : dynamic token limit for prompt cache (#16560)
authorGeorgi Gerganov <redacted>
Tue, 14 Oct 2025 05:48:50 +0000 (08:48 +0300)
committerGitHub <redacted>
Tue, 14 Oct 2025 05:48:50 +0000 (08:48 +0300)
commitbc07349a7f87ba6eb31ed4b0ea9d9a7352185213
tree3a649626d2a47fdd883c690755ff47fba020ea3e
parente60f241eacec42d3bd7c9edd37d236ebf35132a8
server : dynamic token limit for prompt cache (#16560)

* server : dynamic token limit for prompt cache

* cont : print estimated token limit
tools/server/server.cpp