]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/commit
CUDA: add bf16 and f32 support to cublas_mul_mat_batched (#14361)
authorAman Gupta <redacted>
Sat, 28 Jun 2025 17:30:53 +0000 (01:30 +0800)
committerGitHub <redacted>
Sat, 28 Jun 2025 17:30:53 +0000 (01:30 +0800)
commit27208bf657cfe7262791df473927225e48efe482
tree489644a597e720c29d69bfe2b9007cd616ae0bf1
parent63a7bb3c7e1c6b0a92d03b0a594d3cd501d6ed3e
CUDA: add bf16 and f32 support to cublas_mul_mat_batched (#14361)

* CUDA: add bf16 and f32 support to cublas_mul_mat_batched

* Review: add type traits and make function more generic

* Review: make check more explicit, add back comments, and fix formatting

* Review: fix formatting, remove useless type conversion, fix naming for bools
ggml/src/ggml-cuda/convert.cu
ggml/src/ggml-cuda/convert.cuh
ggml/src/ggml-cuda/ggml-cuda.cu
tests/test-backend-ops.cpp