]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/commit
llama : infill sampling handle very long tokens (#9924)
authorGeorgi Gerganov <redacted>
Thu, 17 Oct 2024 19:32:47 +0000 (22:32 +0300)
committerGitHub <redacted>
Thu, 17 Oct 2024 19:32:47 +0000 (22:32 +0300)
commit99bd4ac28c32cd17c0e337ff5601393b033dc5fc
tree33e073f53b9b595bf4c20357a1b2e9fd039482e6
parent3752217ed5a6a11864682fbf009bcb36afffd6bc
llama : infill sampling handle very long tokens (#9924)

* llama : infill sampling handle very long tokens

ggml-ci

* cont : better indices

ggml-ci
include/llama.h
src/llama-sampling.cpp
src/llama-vocab.cpp
src/llama.cpp