]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/commit
llama : add Qwen support (#4281)
authorShijie <redacted>
Fri, 1 Dec 2023 18:16:31 +0000 (02:16 +0800)
committerGitHub <redacted>
Fri, 1 Dec 2023 18:16:31 +0000 (20:16 +0200)
commit37c746d687d877bc11803e96b4dc5f378b83c0a0
tree00976a7933be847bcb58e24c54d8a22c5bb0125b
parent880f57973b8e0091d0f9f50eb5ab4cd4e31582ca
llama : add Qwen support (#4281)

* enable qwen to llama.cpp

* llama : do not GPU split bias tensors

---------

Co-authored-by: Georgi Gerganov <redacted>
convert-hf-to-gguf.py
gguf-py/gguf/constants.py
gguf-py/gguf/tensor_mapping.py
llama.cpp
prompts/chat-with-qwen.txt [new file with mode: 0644]