]> git.djapps.eu Git - pkg/ggml/sources/ggml/shortlog
pkg/ggml/sources/ggml
2024-07-20 Tony Wasserkavulkan : initialize vk_buffer_struct members to VK_NULL...
2024-07-20 Georgi Gerganovpy : update pacakges + fix yolo warning
2024-07-12 Borislav Stanimirovcmake : only enable GGML_NATIVE and x86 flags if not...
2024-07-08 Georgi Gerganovsync : whisper.cpp
2024-07-08 Georgi Gerganovexamples : fix compile warnings [no ci] (whisper/0)
2024-07-08 Daniel Beveniusggml : remove unnecessary UNUSED macro call (#880)
2024-07-08 Georgi Gerganovsync : llama.cpp
2024-07-08 Georgi Gerganovtests : fix whitespace (llama/0)
2024-07-08 Natsucmake : add GGML_BUILD and GGML_SHARED macro definition...
2024-07-08 Ouadie EL FAROUKIEnabled more data types for oneMKL gemm_batch (llama...
2024-07-08 Johannes GäßlerCUDA: MMQ support for iq4_nl, iq4_xs (llama/8278)
2024-07-08 DanieleCUDA: revert part of the RDNA1 optimizations (llama...
2024-07-08 Johannes GäßlerCUDA: fix MMQ stream-k rounding if ne00 % 128 != 0...
2024-07-08 luoyu-intelFix WARP_SIZE=16 bug of Intel GPU (llama/8266)
2024-07-08 Neo Zhang Jianyurm get_work_group_size() by local cache for performance...
2024-07-08 AidanBeltonSRemove unneeded semicolons (llama/8280)
2024-07-08 DanieleDefine and optimize RDNA1 (llama/8085)
2024-07-08 Juddfix typo (llama/8267)
2024-07-08 AidanBeltonSDequant improvements rebase (llama/8255)
2024-07-08 Clint HerronRemoves multiple newlines at the end of files that...
2024-07-08 slarencuda : update supports_op for matrix multiplication...
2024-07-08 luoyu-intelFix win build conflict of math library (llama/8230)
2024-07-08 luoyu-intelFix the sub group size of Intel (llama/8106)
2024-07-08 Johannes GäßlerCUDA: refactor and optimize IQ MMVQ (llama/8215)
2024-07-08 zhentaoyuUpdate SYCL-Rope op and Refactor (llama/8157)
2024-07-08 Johannes GäßlerCUDA: fix MMQ stream-k for --split-mode row (llama...
2024-07-02 slarenfix uses of GGML_USE_CUBLAS in tests and examples ...
2024-07-02 John Balisfeat: cuda implementation for `ggml_conv_transpose_1d...
2024-06-30 Yilong Guosycl : add build instruction (#870)
2024-06-30 John Balisupdate "Using cuBLAS" to use correct update cuda compil...
2024-06-26 Georgi Gerganovsync : whisper.cpp
2024-06-26 Georgi Gerganovwhisper : disable CUDA mel + fix FFMPEG
2024-06-26 Georgi Gerganovsync : llama.cpp
2024-06-26 slarenggml : add GGML_CUDA_USE_GRAPHS option, restore GGML_CU...
2024-06-26 Georgi Gerganovsync : llama.cpp, whisper.cpp
2024-06-26 Georgi Gerganovggml : reorganize source code + improve CMake (#865)
2024-06-21 Georgi Gerganovfiles : remove old (#0)
2024-06-18 Georgi Gerganovsync : whisper.cpp
2024-06-18 Georgi Gerganovwhisper : use ggml_backend_sched (whisper/2239)
2024-06-16 Georgi Gerganovsync : whisper.cpp
2024-06-16 Georgi Gerganovcuda : fix bounds check for src0 rows in MMVQ kernel...
2024-06-16 Borislav Stanimirovwhisper : remove `speed_up` and `phase_vocoder*` functi...
2024-06-16 William Tambelliniexamples : add support for decoding input with ffmpeg...
2024-06-16 Georgi Gerganovexamples : remove whisper (#860)
2024-06-16 slarenmove BLAS to a separate backend (cont) (llama/6210)
2024-06-16 Georgi Gerganovscripts : sync ggml-blas
2024-06-16 0cc4mVulkan Shader Refactor, Memory Debugging Option (llama...
2024-06-16 Georgi Gerganovggml : remove OpenCL (#0)
2024-06-16 Georgi Gerganovcmake : fix cuda vars (#0)
2024-06-16 Georgi Gerganovscripts : update sync
2024-06-16 Hong Bo PENGggml : fix and optimize ppc64le (#849)
2024-06-16 Daniel Beveniusggml : remove duplicate include of ggml-common.h (...
2024-06-16 Yilong Guosycl : remove global variables (cont) (llama/7710)
2024-06-16 Yilong Guoscripts : add ggml-sycl to sync scripts (#857)
2024-06-15 Georgi Gerganovci : add GG_BUILD_NO_DOWNLOAD
2024-06-15 Georgi Gerganovggml : remove opencl (#0)
2024-06-15 Georgi Gerganovcuda : update build (#0)
2024-06-15 Georgi Gerganovsync : llama.cpp
2024-06-15 Georgi Gerganovtests : adapt to changes (#0)
2024-06-15 Meng, Hengyuremove global variables (llama/7710)
2024-06-15 Johannes GäßlerCUDA: faster q2_K, q3_K MMQ + int8 tensor cores (llama...
2024-06-15 Georgi Gerganovmetal : utilize max shared memory for mul_mat_id (llama...
2024-06-15 Radoslav Gerganovrpc : fix ggml_backend_rpc_supports_buft() (llama/7918)
2024-06-15 slarenmove BLAS to a separate backend (llama/6210)
2024-06-15 Johannes GäßlerCUDA: fix broken oob check for FA vec f32 kernel (llama...
2024-06-15 Georgi Gerganovtests : add non-cont unary tests (llama/7857)
2024-06-15 Georgi Gerganovggml : improve ggml_is_contiguous logic (llama/7856)
2024-06-15 k.h.laivulkan: select only one device for single gpu with...
2024-06-15 0cc4mUpdate Vulkan RoPE implementation (llama/7818)
2024-06-15 Johannes GäßlerCUDA: int8 tensor cores for MMQ (q4_K, q5_K, q6_K)...
2024-06-15 Johannes GäßlerCUDA: use tensor cores for MMQ (llama/7676)
2024-06-15 Ben Ashbaughuse the correct SYCL context for host USM allocations...
2024-06-15 Johannes GäßlerCUDA: revise q8_1 data layout for mul_mat_q (llama...
2024-06-15 slarenvulkan : reuse parent extra for views (llama/7806)
2024-06-15 pengxin99fix softmax r2r result wrong issue (llama/7811)
2024-06-15 Johannes GäßlerCUDA: refactor mmq, dmmv, mmvq (llama/7716)
2024-06-15 Georgi Gerganovggml : refactor rope norm/neox (llama/7634)
2024-06-15 agray3Allow number of nodes in CUDA graph to change (llama...
2024-06-15 Georgi Gerganovggml : remove OpenCL (llama/7735)
2024-06-15 Georgi Gerganovggml : prevent builds with -ffinite-math-only (llama...
2024-06-15 Radoslav Gerganovllama : offload to RPC in addition to other backends...
2024-06-15 Masaya, Katoggml : use OpenMP as a thread pool (llama/7606)
2024-06-15 0cc4mVulkan Mixture of Experts (MoE) support (llama/7628)
2024-06-15 woachkkompute : implement op_getrows_f32 (llama/6403)
2024-06-15 Dave Airliefix bug introduced in using calloc (llama/7701)
2024-06-15 Johannes GäßlerFix FlashAttention debug test, FP32 assert (llama/7684)
2024-06-15 Johannes GäßlerCUDA: fix Pascal FA, deq. KV to FP16 for batch > 8...
2024-06-15 Johannes GäßlerCUDA: quantized KV support for FA vec (llama/7527)
2024-06-15 Georgi Gerganovggml : fix loongson compile warnings (llama/7537)
2024-06-15 Chris Elrodfaster avx512 exp implementation (llama/7551)
2024-06-15 junchao-loongsonggml : fix loongarch build (O2 issue) (llama/7636)
2024-06-15 Georgi Gerganovmetal : remove invalid asserts (llama/7617)
2024-06-15 Georgi Gerganovmetal : add missing asserts (llama/7617)
2024-06-15 Georgi Gerganovggml : fix YARN + add tests + add asserts (llama/7617)
2024-06-15 Georgi Gerganovcuda : non-cont concat support (llama/7610)
2024-06-15 Radoslav Gerganovllama-bench : add support for the RPC backend (llama...
2024-06-15 slarenggml : use atomic_flag for critical section (llama...
2024-06-05 Danielecmake : update HIPBLAS (#847)
2024-06-05 Emmanuel Durandzig : fix build (#840)
2024-05-29 Georgi Gerganovsync : llama.cpp
next