]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/commit
llama : add support for GLM-Edge and GLM-Edge-V series models (#10573)
authorpiDack <redacted>
Sun, 2 Feb 2025 07:48:46 +0000 (15:48 +0800)
committerGitHub <redacted>
Sun, 2 Feb 2025 07:48:46 +0000 (09:48 +0200)
commit0cec062a638700495673f5494d200b74340538be
treec7c2db2b39d126e13ed1ce2c653069b737da56c7
parent53debe6f3c9cca87e9520a83ee8c14d88977afa4
llama : add support for GLM-Edge and GLM-Edge-V series models (#10573)

* add glm edge chat model

* use config partial_rotary_factor as rope ratio

* support for glm edge model

* vision model support

* remove debug info

* fix format

* llava.cpp trailing whitespace

* remove unused AutoTokenizer

* Update src/llama.cpp for not contain <|end|> or </s>

Co-authored-by: Xuan Son Nguyen <redacted>
* add edge template

* fix chat template

* fix confict

* fix confict

* fix ci err

* fix format err

* fix template err

* 9b hf chat support

* format

* format clip.cpp

* fix format

* Apply suggestions from code review

* Apply suggestions from code review

* Update examples/llava/clip.cpp

* fix format

* minor : style

---------

Co-authored-by: liyuhang <redacted>
Co-authored-by: piDack <redacted>
Co-authored-by: Xuan Son Nguyen <redacted>
Co-authored-by: liyuhang <redacted>
Co-authored-by: Georgi Gerganov <redacted>
15 files changed:
README.md
convert_hf_to_gguf.py
examples/llava/README-glmedge.md [new file with mode: 0644]
examples/llava/clip.cpp
examples/llava/clip.h
examples/llava/glmedge-convert-image-encoder-to-gguf.py [new file with mode: 0644]
examples/llava/glmedge-surgery.py [new file with mode: 0644]
examples/llava/llava.cpp
gguf-py/gguf/constants.py
src/llama-arch.cpp
src/llama-chat.cpp
src/llama-chat.h
src/llama-model.cpp
src/llama.cpp
tests/test-chat-template.cpp