]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/commit
ggml : more perfo with llamafile tinyblas on x86_64 (#10714)
authorDjip007 <redacted>
Tue, 24 Dec 2024 17:54:49 +0000 (18:54 +0100)
committerGitHub <redacted>
Tue, 24 Dec 2024 17:54:49 +0000 (18:54 +0100)
commit2cd43f4900ba0e34124fdcbf02a7f9df25a10a3d
tree68007c7db007f7b21c3b68b59c96f31d6bf6a7c6
parent09fe2e76137dde850b13313f720e7ffa17efdefa
ggml : more perfo with llamafile tinyblas on x86_64 (#10714)

* more perfo with llamafile tinyblas on x86_64.

- add bf16 suport
- change dispache strategie (thanks:
https://github.com/ikawrakow/ik_llama.cpp/pull/71 )
- reduce memory bandwidth

simple tinyblas dispache and more cache freindly

* tinyblas dynamic dispaching

* sgemm: add M blocs.

* - git 2.47 use short id of len 9.
- show-progress is not part of GNU Wget2

* remove not stable test
examples/server/tests/unit/test_completion.py
ggml/src/ggml-cpu/ggml-cpu.c
ggml/src/ggml-cpu/llamafile/sgemm.cpp
ggml/src/ggml-cpu/llamafile/sgemm.h
scripts/compare-llama-bench.py
scripts/hf.sh