]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/commit
llama : add StableLM2 12B (#6635)
authorAshish <redacted>
Tue, 16 Apr 2024 15:48:35 +0000 (08:48 -0700)
committerGitHub <redacted>
Tue, 16 Apr 2024 15:48:35 +0000 (18:48 +0300)
commitdbceec87c0221ec952e69448df6a71f1372a7487
tree3c8773f6eccea909c670c16cf5b3bbb8e65fe12c
parentf4dea7da1841a92d2788b0535063abf2f0e28461
llama : add StableLM2 12B (#6635)

* StableLM2 12B support for huggingface -> GGUF

* StableLM12 tensormapping and constants

* StableLM-2-12b model support

* fix

* Added 12B support

* Removed autoformatting; resolved bug where model_arch was not selecting StableLM2

* Formatting

* Do QK norm stacking in model conversion step

* Converge StableLM and StableLM2 code to simplify graph construction

* Fix accidental removal

* Removed warnings

* Revert formatter

* Move QK norm stack to private function so it's easier to read

* refactor stablelm graph builder to support 1.6, 3b and 12b more efficiently

* Proper check for None type for new_name to avoid crash; formatting; revert change to base class `write_tensors()`

* Format

* Formatting

* format

Co-authored-by: compilade <redacted>
* Fix incorrect check for K norm

* space after commas; Keep indentation multiple of 4 spaces

* Flake8 format

* Removed unnecessary conditional branches

* Removed unused comment

* Fixed incorrect tensor passing

* Format

---------

Co-authored-by: compilade <redacted>
convert-hf-to-gguf.py
gguf-py/gguf/constants.py
llama.cpp