]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/commit
ggml: fix cuda kernel launch configuration for k_compute_batched_ptrs to support...
authorleejet <redacted>
Sun, 26 Oct 2025 18:13:31 +0000 (02:13 +0800)
committerGitHub <redacted>
Sun, 26 Oct 2025 18:13:31 +0000 (19:13 +0100)
commitbbac6a26b2bd7f7c1f0831cb1e7b52734c66673b
tree1486896776f7f294d69d98acbc17f0552dac792d
parent73a48c9790d320476b3e5ef75bda09f2f8269e6e
ggml: fix cuda kernel launch configuration for k_compute_batched_ptrs to support large batch (#16744)

* fix k_compute_batched_ptrs

* add backend ops test

* Update ggml/src/ggml-cuda/ggml-cuda.cu

Co-authored-by: Johannes Gäßler <redacted>
* reduce the batch size

---------

Co-authored-by: Johannes Gäßler <redacted>
ggml/src/ggml-cuda/ggml-cuda.cu
tests/test-backend-ops.cpp