]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/commit
py : re-enable mmap in convert hf (#4732)
authorNam D. Tran <redacted>
Tue, 2 Jan 2024 09:23:38 +0000 (16:23 +0700)
committerGitHub <redacted>
Tue, 2 Jan 2024 09:23:38 +0000 (11:23 +0200)
commit26f3071d714f0b27ad7f021a46a66a1085480258
treebbbc9da48238a470bde8854e94ee6cdc2b27b19c
parent775ac8712a7b42cfead2585f42cec0dfd56644ab
py : re-enable mmap in convert hf (#4732)

* update: awq support llama-7b model

* update: change order

* update: benchmark results for llama2-7b

* update: mistral 7b v1 benchmark

* update: support 4 models

* fix: Readme

* update: ready for PR

* update: readme

* fix: readme

* update: change order import

* black

* format code

* update: work for bot mpt and awqmpt

* update: readme

* Rename to llm_build_ffn_mpt_awq

* Formatted other files

* Fixed params count

* fix: remove code

* update: more detail for mpt

* fix: readme

* fix: readme

* update: change folder architecture

* fix: common.cpp

* fix: readme

* fix: remove ggml_repeat

* update: cicd

* update: cicd

* uppdate: remove use_awq arg

* update: readme

* llama : adapt plamo to new ffn

ggml-ci

* fix: update torch version

---------

Co-authored-by: Trần Đức Nam <redacted>
Co-authored-by: Le Hoang Anh <redacted>
Co-authored-by: Georgi Gerganov <redacted>
awq-py/requirements.txt
convert-hf-to-gguf.py