]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/commit
llama : parameter conversion and loading fixes for PLaMo2 variants (#16075)
authorShunta Saito <redacted>
Wed, 1 Oct 2025 21:08:15 +0000 (06:08 +0900)
committerGitHub <redacted>
Wed, 1 Oct 2025 21:08:15 +0000 (23:08 +0200)
commitded67b94446ef4f7fd988dbde7a12deef9870c13
tree8e6287a947c276a319e5a14994d77aa1b8c10349
parent1fe4e38cc20af058ed320bd46cac934991190056
llama : parameter conversion and loading fixes for PLaMo2 variants (#16075)

* Fix to use hidden_size_per_head

* Fix num heads

* Fix array

* Fix loading weights

* Support old GGUF converted by the previous version of llama.cpp

* Update src/llama-model.cpp

Co-authored-by: Sigbjørn Skjæret <redacted>
* Move shared parameter definitions to the outside of loop

* Not calculating n_embd_head_k,v by n_embd / n_head

---------

Co-authored-by: Sigbjørn Skjæret <redacted>
convert_hf_to_gguf.py
src/llama-hparams.h
src/llama-model.cpp