]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/commit
llama: automatically set parameters not set by the user in such a way that maximizes...
authorJohannes Gäßler <redacted>
Mon, 15 Dec 2025 08:24:59 +0000 (09:24 +0100)
committerGitHub <redacted>
Mon, 15 Dec 2025 08:24:59 +0000 (09:24 +0100)
commitb1f3a6e5db7b782ef077bd0e8253ce03283b1f37
treeba0475a546443cc22e981d7f4056a5c6b7160466
parent4aced7a63156555911157d3002f9d3ddef4a1e55
llama: automatically set parameters not set by the user in such a way that maximizes GPU utilization (#16653)

* llama: automatically fit args to free memory

llama-fit-params tool

* fix CI

* hints for bug reports, ensure no reallocation

* fix segfault with Vulkan

* add llama-fit-params to CI

* fix CI

* fix CI

* fix CI

* minor adjustments

* fix assignment of 1 dense layer

* fix logger not being reset on model load failure

* remove --n-gpu-layer hint on model load failure

* fix llama-fit-params verbosity

* fix edge case

* fix typo [no ci]
26 files changed:
.github/ISSUE_TEMPLATE/011-bug-results.yml
ci/run.sh
common/arg.cpp
common/common.cpp
common/common.h
ggml/include/ggml-alloc.h
ggml/include/ggml-backend.h
ggml/include/ggml.h
ggml/src/ggml-alloc.c
ggml/src/ggml-backend.cpp
ggml/src/ggml.c
include/llama.h
src/llama-context.cpp
src/llama-context.h
src/llama-hparams.h
src/llama-impl.cpp
src/llama-kv-cache.cpp
src/llama-model-loader.cpp
src/llama-model-loader.h
src/llama-model.cpp
src/llama-quant.cpp
src/llama.cpp
tools/CMakeLists.txt
tools/fit-params/CMakeLists.txt [new file with mode: 0644]
tools/fit-params/README.md [new file with mode: 0644]
tools/fit-params/fit-params.cpp [new file with mode: 0644]