]> git.djapps.eu Git - pkg/ggml/sources/ggml/shortlog
pkg/ggml/sources/ggml
2025-08-02 Georgi Gerganovsync : llama.cpp upstream/0.0.2404
2025-08-02 leejetcuda: make im2col a little faster (llama/15025)
2025-08-02 Georgi Gerganovcuda, sycl : fix batched gemm when ne02 == 1 && ne03...
2025-08-02 Jeff Bolzvulkan: coopmat2 mul_mat optimizations (llama/14934)
2025-08-02 Jeff Bolzvulkan: Support ne[3]>1 in noncontig matrix-vector...
2025-08-02 Jeff Bolzvulkan: optimizations for direct convolution (llama...
2025-08-02 Johannes GäßlerCUDA: fix MMQ nwarps for AMD with warp_size==32 (llama...
2025-08-02 lhezopencl: add f16 for `add`, `sub`, `mul`, `div` (llama...
2025-08-02 Srihari-mcwggml : Q2k interleaving implementation - x86/x64 SIMD...
2025-08-02 diannaodocker : add cann build pipline (llama/14591)
2025-08-02 Ruben OrtlamVulkan: Fix minor debug mode issues (llama/14899)
2025-08-02 hipuddingCANN: Improve loading efficiency after converting weigh...
2025-08-02 lhezopencl: add `mul_mat_f32_f32_l4_lm` and `mul_mat_f16_f3...
2025-08-02 uvosHIP: enable mfma mmq on gfx908 and gfx90a for select...
2025-08-02 Johannes GäßlerCUDA: skip masked KV slices for all FA kernels (llama...
2025-08-02 uvosHIP: remove the use of __HIP_PLATFORM_AMD__, explicitly...
2025-08-02 uvosHIP: add GGML_HIP_MMQ_MFMA option to allow disableing...
2025-08-02 uvosHIP: Ignore unsupported unroll transformation in fattn...
2025-08-02 hipuddingCANN: Add ggml_set_rows (llama/14943)
2025-08-02 Sigbjørn Skjæretcuda : add softcap fusion (llama/14907)
2025-08-02 Aman GuptaCUDA: add roll (llama/14919)
2025-08-02 Leonard Mosescutest-backend-ops : extend test case filtering (llama...
2025-08-02 xctanggml-cpu : deduplicate scalar implementations (llama...
2025-08-02 Akarshan BiswasSYCL: Add set_rows support for quantized types (llama...
2025-08-02 Johannes GäßlerCUDA: fix pointer incrementation in FA (llama/14916)
2025-08-02 Alberto Cabrera... sycl: refactor quantization to q8_1 (llama/14815)
2025-08-02 Kai Pastorci : Move msvc to matrix (#1318)
2025-08-02 AN Longsimple : fix typo (#1319)
2025-07-30 Georgi Gerganovsync : whisper.cpp
2025-07-30 Kai Pastorcmake : Fix BLAS link interface (#1316)
2025-07-30 Kai Pastorvulkan : fix 32-bit builds (#1313)
2025-07-28 Georgi Gerganovsync : llama.cpp
2025-07-28 Erik Scholzvulkan : add fp16 support for the conv_2d kernel (llama...
2025-07-28 Jeff Bolzvulkan: skip empty set_rows to avoid invalid API usage...
2025-07-28 Aman GuptaDocs: add instructions for adding backends (llama/14889)
2025-07-28 deepsekHIP: Enable Matrix cores for MMQ Kernels, Enable stream...
2025-07-28 hipuddingCANN: Implement GLU ops (llama/14884)
2025-07-28 R0CKSTARmusa: fix build warnings (unused variable) (llama/14869)
2025-07-28 Aaron Teoggml-cpu : disable GGML_NNPA by default due to instabil...
2025-07-28 Gabe Goodhartmetal: SSM_SCAN performance (llama/14743)
2025-07-28 lhezopencl: add fused `rms_norm_mul` (llama/14841)
2025-07-28 Oliver Simonsggml : remove invalid portPos specifiers from dot files...
2025-07-28 Chris Rohlfrpc : check for null buffers in get/set/copy tensor...
2025-07-28 Diego Devesasched : fix multiple evaluations of the same graph...
2025-07-28 R0CKSTARmusa: upgrade musa sdk to rc4.2.0 (llama/14498)
2025-07-25 Georgi Gerganovcontrib : recommend PRs to llama.cpp (#1312)
2025-07-24 Kai Pastorcmake : Indent ggml-config.cmake (#1310)
2025-07-24 Georgi Gerganovsync : llama.cpp
2025-07-24 Alberto Cabrera... sycl: fixed semantics of block offset calculation ...
2025-07-24 Georgi Gerganovmetal : fix fusion across different encoders (llama...
2025-07-24 Donghyeon Jeongsycl: fix undefined variable in work group size check...
2025-07-24 Johannes GäßlerCUDA: fix overflow in FA, tune performance (llama/14840)
2025-07-24 Johannes GäßlerCUDA: fix compilation with GGML_CUDA_F16 (llama/14837)
2025-07-24 Johannes GäßlerCUDA: fix quantized KV cache + multiple sequences ...
2025-07-24 Georgi Gerganovtests : add non-cont K,V FA tests
2025-07-24 lixing-starggml: fix loongarch quantize_row_q8_1 error (llama...
2025-07-24 chen fanCANN: weight format to NZ for Ascend310P3 (llama/14407)
2025-07-24 Aman GuptaCUDA: add fused rms norm (llama/14800)
2025-07-24 Jeff Bolzvulkan: fix rms_norm_mul to handle broadcasting dim0...
2025-07-24 Sigbjørn Skjæretcuda : implement bf16 cpy ops and enable bf16 cont...
2025-07-24 lhezopencl: remove unreachable `return` (llama/14806)
2025-07-24 R0CKSTARcuda: remove linking to cublasLt (llama/14790)
2025-07-24 Sigbjørn Skjæretopencl: fix `im2col` when `KW!=KH` (llama/14803)
2025-07-24 rmatifopencl: add conv2d kernel (llama/14403)
2025-07-24 Romain Biessysycl: Fix im2col (llama/14797)
2025-07-24 Charles Xukleidiai: add support for get_rows (llama/14676)
2025-07-24 Jeff Bolzvulkan/cuda: Fix im2col when KW!=KH (llama/14789)
2025-07-24 Ervin Áron... ggml: adds CONV_2D op and direct GEMM Vulkan implementa...
2025-07-24 Peter0x44vulkan: Add logging for bf16 features to ggml_vk_print_...
2025-07-24 0cc4mVulkan: Fix fprintf format-security warning (llama...
2025-07-23 Kai PastorCI: Test static build (#1307)
2025-07-22 Kai Pastorcmake : fix usage issues (#1257)
2025-07-21 Daniel Beveniusggml-cpu : remove stdlib include from repack.cpp (...
2025-07-19 Georgi Gerganovsync : llama.cpp
2025-07-19 Georgi Gerganovmetal : fuse add, mul + add tests (llama/14596)
2025-07-19 Oliver Simonscuda : Fix Gemma3n not executed as CUDA_GRAPH on NVGPUs...
2025-07-19 Aman GuptaCUDA: set_rows + cpy.cu refactor (llama/14712)
2025-07-19 Neo Zhang Jianyuuse max work group size for device to replace the magic...
2025-07-19 Reese Levineggml: Add initial WebGPU backend (llama/14521)
2025-07-19 Georgi Gerganovllama : add high-throughput mode (llama/14363)
2025-07-19 Georgi Gerganovggml : add asserts (llama/14720)
2025-07-19 Jeff Bolzvulkan: fix noncontig check for mat_mul_id splitting...
2025-07-19 Jeff Bolzvulkan: add RTE variants for glu/add/sub/mul/div (llama...
2025-07-19 R0CKSTARcuda: fix build warnings in set-rows.cu (unused variabl...
2025-07-19 Anton Mitkovsycl: Hotfix for non dnnl codepath (llama/14677)
2025-07-19 shalinib-ibmggml : refactor llamafile_sgemm PPC code (llama/14673)
2025-07-19 Akarshan BiswasSYCL: use 1D kernel for set_rows (llama/14618)
2025-07-19 Anton Mitkovsycl: Batched mulmat rework for oneDNN dispatch (llama...
2025-07-19 Sigbjørn Skjæretcuda : add set rows for bf16 (llama/14664)
2025-07-19 Yavor Ivanovcuda : add ELU support (llama/14657)
2025-07-19 Georgi Gerganovggml : add build-time message to remind about ggml_set_...
2025-07-19 Yavor Ivanovmetal : Add missing unary ops Metal support (llama...
2025-07-19 Tarek Dakhrantests : cover lfm2 cases in test_ssm_conv (llama/14651)
2025-07-19 Aman GuptaCUDA: add set rows for f32 and f16 (llama/14551)
2025-07-12 Georgi Gerganovsync : whispper.cpp
2025-07-12 Georgi Gerganovgit : remove kompute submodule (#1300) upstream/0.0.2309
2025-07-12 Georgi Gerganovsync : resolve conflicts (#0)
2025-07-12 Georgi Gerganovsync : llama.cpp
2025-07-12 Jeff Bolzvulkan: support SET_ROWS (llama/14587)
2025-07-12 Jeff Bolzvulkan: optimizations for deepseek prompt processing...
next