]> git.djapps.eu Git - pkg/ggml/sources/ggml/shortlog
pkg/ggml/sources/ggml
2025-08-14 Charles Xukleidiai: fix unsigned overflow bug (llama/15150)
2025-08-14 David Zhaocuda: refactored ssm_scan and use CUB (llama/13291)
2025-08-14 Aman GuptaCUDA: add attention sinks for tile and wmma (llama...
2025-08-14 compiladegguf-py : add Numpy MXFP4 de/quantization support ...
2025-08-14 AN Longggml : fix field name when new ggml_backend (llama...
2025-08-14 Johannes GäßlerCUDA: attention sinks for mma FlashAttention (llama...
2025-08-14 lhezopencl: support sink in `soft_max` (attn sinks) (llama...
2025-08-14 Jeff Bolzvulkan: support fattn sinks (llama/15126)
2025-08-14 Jeff Bolzvulkan: Add env var to disable host visible vidmem...
2025-08-14 uvosHIP: add cmake option to enable compiler output of...
2025-08-14 Christian Kastnerggml: Skip backend library linking code when GGML_BACKE...
2025-08-14 Johannes GäßlerCUDA: GEMM for FP32/FP16/BF16 and ne11 <= 16 (llama...
2025-08-14 rmatiffix profiling crash (llama/15072)
2025-08-14 lhezopencl: add `swiglu_oai` and `add_id` (llama/15121)
2025-08-14 Diego Devesaggml : fix fallback to CPU for ununsupported ops (llama...
2025-08-14 Chenguang LiCANN: add support for ACL Graph (llama/15065)
2025-08-14 Georgi Gerganovllama : add gpt-oss (llama/15091)
2025-08-14 Romain Biessysycl: fix mul_mat selection (llama/15092)
2025-08-14 Christian Kastnercmake: Add GGML_BACKEND_DIR option (llama/15074)
2025-08-14 Jeff Bolzvulkan: fix build when using glslang that does not...
2025-08-14 Jeff Bolzvulkan: Use coopmat2 for conv2d (llama/14982)
2025-08-14 lhezopencl: fix adreno compiler detection logic (llama...
2025-08-14 Johannes GäßlerCUDA: use mma FA kernel for gqa > 4 on RTX 4000 (llama...
2025-08-02 Georgi Gerganovsync : llama.cpp upstream/0.0.2404
2025-08-02 leejetcuda: make im2col a little faster (llama/15025)
2025-08-02 Georgi Gerganovcuda, sycl : fix batched gemm when ne02 == 1 && ne03...
2025-08-02 Jeff Bolzvulkan: coopmat2 mul_mat optimizations (llama/14934)
2025-08-02 Jeff Bolzvulkan: Support ne[3]>1 in noncontig matrix-vector...
2025-08-02 Jeff Bolzvulkan: optimizations for direct convolution (llama...
2025-08-02 Johannes GäßlerCUDA: fix MMQ nwarps for AMD with warp_size==32 (llama...
2025-08-02 lhezopencl: add f16 for `add`, `sub`, `mul`, `div` (llama...
2025-08-02 Srihari-mcwggml : Q2k interleaving implementation - x86/x64 SIMD...
2025-08-02 diannaodocker : add cann build pipline (llama/14591)
2025-08-02 Ruben OrtlamVulkan: Fix minor debug mode issues (llama/14899)
2025-08-02 hipuddingCANN: Improve loading efficiency after converting weigh...
2025-08-02 lhezopencl: add `mul_mat_f32_f32_l4_lm` and `mul_mat_f16_f3...
2025-08-02 uvosHIP: enable mfma mmq on gfx908 and gfx90a for select...
2025-08-02 Johannes GäßlerCUDA: skip masked KV slices for all FA kernels (llama...
2025-08-02 uvosHIP: remove the use of __HIP_PLATFORM_AMD__, explicitly...
2025-08-02 uvosHIP: add GGML_HIP_MMQ_MFMA option to allow disableing...
2025-08-02 uvosHIP: Ignore unsupported unroll transformation in fattn...
2025-08-02 hipuddingCANN: Add ggml_set_rows (llama/14943)
2025-08-02 Sigbjørn Skjæretcuda : add softcap fusion (llama/14907)
2025-08-02 Aman GuptaCUDA: add roll (llama/14919)
2025-08-02 Leonard Mosescutest-backend-ops : extend test case filtering (llama...
2025-08-02 xctanggml-cpu : deduplicate scalar implementations (llama...
2025-08-02 Akarshan BiswasSYCL: Add set_rows support for quantized types (llama...
2025-08-02 Johannes GäßlerCUDA: fix pointer incrementation in FA (llama/14916)
2025-08-02 Alberto Cabrera... sycl: refactor quantization to q8_1 (llama/14815)
2025-08-02 Kai Pastorci : Move msvc to matrix (#1318)
2025-08-02 AN Longsimple : fix typo (#1319)
2025-07-30 Georgi Gerganovsync : whisper.cpp
2025-07-30 Kai Pastorcmake : Fix BLAS link interface (#1316)
2025-07-30 Kai Pastorvulkan : fix 32-bit builds (#1313)
2025-07-28 Georgi Gerganovsync : llama.cpp
2025-07-28 Erik Scholzvulkan : add fp16 support for the conv_2d kernel (llama...
2025-07-28 Jeff Bolzvulkan: skip empty set_rows to avoid invalid API usage...
2025-07-28 Aman GuptaDocs: add instructions for adding backends (llama/14889)
2025-07-28 deepsekHIP: Enable Matrix cores for MMQ Kernels, Enable stream...
2025-07-28 hipuddingCANN: Implement GLU ops (llama/14884)
2025-07-28 R0CKSTARmusa: fix build warnings (unused variable) (llama/14869)
2025-07-28 Aaron Teoggml-cpu : disable GGML_NNPA by default due to instabil...
2025-07-28 Gabe Goodhartmetal: SSM_SCAN performance (llama/14743)
2025-07-28 lhezopencl: add fused `rms_norm_mul` (llama/14841)
2025-07-28 Oliver Simonsggml : remove invalid portPos specifiers from dot files...
2025-07-28 Chris Rohlfrpc : check for null buffers in get/set/copy tensor...
2025-07-28 Diego Devesasched : fix multiple evaluations of the same graph...
2025-07-28 R0CKSTARmusa: upgrade musa sdk to rc4.2.0 (llama/14498)
2025-07-25 Georgi Gerganovcontrib : recommend PRs to llama.cpp (#1312)
2025-07-24 Kai Pastorcmake : Indent ggml-config.cmake (#1310)
2025-07-24 Georgi Gerganovsync : llama.cpp
2025-07-24 Alberto Cabrera... sycl: fixed semantics of block offset calculation ...
2025-07-24 Georgi Gerganovmetal : fix fusion across different encoders (llama...
2025-07-24 Donghyeon Jeongsycl: fix undefined variable in work group size check...
2025-07-24 Johannes GäßlerCUDA: fix overflow in FA, tune performance (llama/14840)
2025-07-24 Johannes GäßlerCUDA: fix compilation with GGML_CUDA_F16 (llama/14837)
2025-07-24 Johannes GäßlerCUDA: fix quantized KV cache + multiple sequences ...
2025-07-24 Georgi Gerganovtests : add non-cont K,V FA tests
2025-07-24 lixing-starggml: fix loongarch quantize_row_q8_1 error (llama...
2025-07-24 chen fanCANN: weight format to NZ for Ascend310P3 (llama/14407)
2025-07-24 Aman GuptaCUDA: add fused rms norm (llama/14800)
2025-07-24 Jeff Bolzvulkan: fix rms_norm_mul to handle broadcasting dim0...
2025-07-24 Sigbjørn Skjæretcuda : implement bf16 cpy ops and enable bf16 cont...
2025-07-24 lhezopencl: remove unreachable `return` (llama/14806)
2025-07-24 R0CKSTARcuda: remove linking to cublasLt (llama/14790)
2025-07-24 Sigbjørn Skjæretopencl: fix `im2col` when `KW!=KH` (llama/14803)
2025-07-24 rmatifopencl: add conv2d kernel (llama/14403)
2025-07-24 Romain Biessysycl: Fix im2col (llama/14797)
2025-07-24 Charles Xukleidiai: add support for get_rows (llama/14676)
2025-07-24 Jeff Bolzvulkan/cuda: Fix im2col when KW!=KH (llama/14789)
2025-07-24 Ervin Áron... ggml: adds CONV_2D op and direct GEMM Vulkan implementa...
2025-07-24 Peter0x44vulkan: Add logging for bf16 features to ggml_vk_print_...
2025-07-24 0cc4mVulkan: Fix fprintf format-security warning (llama...
2025-07-23 Kai PastorCI: Test static build (#1307)
2025-07-22 Kai Pastorcmake : fix usage issues (#1257)
2025-07-21 Daniel Beveniusggml-cpu : remove stdlib include from repack.cpp (...
2025-07-19 Georgi Gerganovsync : llama.cpp
2025-07-19 Georgi Gerganovmetal : fuse add, mul + add tests (llama/14596)
2025-07-19 Oliver Simonscuda : Fix Gemma3n not executed as CUDA_GRAPH on NVGPUs...
2025-07-19 Aman GuptaCUDA: set_rows + cpy.cu refactor (llama/14712)
next