]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/commit
llama : add Command R Plus support (#6491)
authorCarolinabanana <redacted>
Tue, 9 Apr 2024 08:16:13 +0000 (09:16 +0100)
committerGitHub <redacted>
Tue, 9 Apr 2024 08:16:13 +0000 (11:16 +0300)
commit5dc9dd7152dedc6046b646855585bd070c91e8c8
treed2bae3652d91cdd9327e28fa85d167a67e050c53
parente11a8999b5690f810c2c99c14347f0834e68c524
llama : add Command R Plus support (#6491)

* Add Command R Plus GGUF

* Add Command R Plus GGUF

* Loading works up to LayerNorm2D

* Export new tensors in 1D so they are not quantized.

* Fix embedding layer based on Noeda's example

* Whitespace

* Add line

* Fix unexpected tokens on MPS. Re-add F16 fix. ((Noeda)

* dranger003: Fix block index overflow in CUDA dequantizing.

* Reverted blocked multiplication code as it still has issues and could affect other Llama arches

* export norms as f32

* fix overflow issues during quant and other cleanup

* Type convention

Co-authored-by: Georgi Gerganov <redacted>
* dranger003: Fix more int overflow during quant.

---------

Co-authored-by: S <redacted>
Co-authored-by: S <redacted>
Co-authored-by: slaren <redacted>
Co-authored-by: Georgi Gerganov <redacted>
16 files changed:
convert-hf-to-gguf.py
ggml-cuda.cu
ggml-cuda/common.cuh
ggml-cuda/convert.cu
ggml-cuda/convert.cuh
ggml-cuda/dequantize.cuh
ggml-cuda/dmmv.cu
ggml-cuda/quantize.cu
ggml-cuda/quantize.cuh
ggml-quants.c
ggml-quants.h
ggml.c
ggml.h
gguf-py/gguf/constants.py
gguf-py/gguf/tensor_mapping.py
llama.cpp