]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/commit
server: stop generation at `n_ctx_train` if `n_predict` is not set (#6638)
authorPierrick Hymbert <redacted>
Fri, 26 Apr 2024 10:15:30 +0000 (12:15 +0200)
committerGitHub <redacted>
Fri, 26 Apr 2024 10:15:30 +0000 (12:15 +0200)
commit7f5ff558eed0f732af8f25c2ab0645610bdec80c
treea382beee28990c97271ab001de9fa3834f58ed14
parent9e4e077ec50fde6049b128662c72d37a3c28e34b
server: stop generation at `n_ctx_train` if `n_predict` is not set (#6638)

* server: cap n_predict if not set to n_ctx_train

* server: fix infinite loop

* server: infinite loop, move in process_token
server: infinite loop: set stop limit to true

* minor: spaces

* minor: spaces

* server: include prompt tokens in the EOS limit
examples/server/server.cpp