]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/commit
ROCm Port (#1087)
authorHenri Vasserman <redacted>
Fri, 25 Aug 2023 09:09:42 +0000 (12:09 +0300)
committerGitHub <redacted>
Fri, 25 Aug 2023 09:09:42 +0000 (12:09 +0300)
commit6bbc598a632560cb45dd2c51ad403bda8723b629
tree53be13238531021865642158403fbf92c5a9ff58
parent3f460a2b723c8b936ac29ecfd02f244b3adeba55
ROCm Port (#1087)

* use hipblas based on cublas
* Update Makefile for the Cuda kernels
* Expand arch list and make it overrideable
* Fix multi GPU on multiple amd architectures with rocblas_initialize() (#5)
* add hipBLAS to README
* new build arg LLAMA_CUDA_MMQ_Y
* fix half2 decomposition
* Add intrinsics polyfills for AMD
* AMD assembly optimized __dp4a
* Allow overriding CC_TURING
* use "ROCm" instead of "CUDA"
* ignore all build dirs
* Add Dockerfiles
* fix llama-bench
* fix -nommq help for non CUDA/HIP

---------

Co-authored-by: YellowRoseCx <redacted>
Co-authored-by: ardfork <redacted>
Co-authored-by: funnbot <redacted>
Co-authored-by: Engininja2 <redacted>
Co-authored-by: Kerfuffle <redacted>
Co-authored-by: jammm <redacted>
Co-authored-by: jdecourval <redacted>
12 files changed:
.devops/full-rocm.Dockerfile [new file with mode: 0644]
.devops/main-rocm.Dockerfile [new file with mode: 0644]
.dockerignore
.gitignore
CMakeLists.txt
Makefile
README.md
common/common.cpp
examples/llama-bench/llama-bench.cpp
ggml-cuda.cu
ggml-cuda.h
llama.cpp