]> git.djapps.eu Git - pkg/ggml/sources/ggml/shortlog
pkg/ggml/sources/ggml
2025-07-24 lixing-starggml: fix loongarch quantize_row_q8_1 error (llama...
2025-07-24 chen fanCANN: weight format to NZ for Ascend310P3 (llama/14407)
2025-07-24 Aman GuptaCUDA: add fused rms norm (llama/14800)
2025-07-24 Jeff Bolzvulkan: fix rms_norm_mul to handle broadcasting dim0...
2025-07-24 Sigbjørn Skjæretcuda : implement bf16 cpy ops and enable bf16 cont...
2025-07-24 lhezopencl: remove unreachable `return` (llama/14806)
2025-07-24 R0CKSTARcuda: remove linking to cublasLt (llama/14790)
2025-07-24 Sigbjørn Skjæretopencl: fix `im2col` when `KW!=KH` (llama/14803)
2025-07-24 rmatifopencl: add conv2d kernel (llama/14403)
2025-07-24 Romain Biessysycl: Fix im2col (llama/14797)
2025-07-24 Charles Xukleidiai: add support for get_rows (llama/14676)
2025-07-24 Jeff Bolzvulkan/cuda: Fix im2col when KW!=KH (llama/14789)
2025-07-24 Ervin Áron... ggml: adds CONV_2D op and direct GEMM Vulkan implementa...
2025-07-24 Peter0x44vulkan: Add logging for bf16 features to ggml_vk_print_...
2025-07-24 0cc4mVulkan: Fix fprintf format-security warning (llama...
2025-07-23 Kai PastorCI: Test static build (#1307)
2025-07-22 Kai Pastorcmake : fix usage issues (#1257)
2025-07-21 Daniel Beveniusggml-cpu : remove stdlib include from repack.cpp (...
2025-07-19 Georgi Gerganovsync : llama.cpp
2025-07-19 Georgi Gerganovmetal : fuse add, mul + add tests (llama/14596)
2025-07-19 Oliver Simonscuda : Fix Gemma3n not executed as CUDA_GRAPH on NVGPUs...
2025-07-19 Aman GuptaCUDA: set_rows + cpy.cu refactor (llama/14712)
2025-07-19 Neo Zhang Jianyuuse max work group size for device to replace the magic...
2025-07-19 Reese Levineggml: Add initial WebGPU backend (llama/14521)
2025-07-19 Georgi Gerganovllama : add high-throughput mode (llama/14363)
2025-07-19 Georgi Gerganovggml : add asserts (llama/14720)
2025-07-19 Jeff Bolzvulkan: fix noncontig check for mat_mul_id splitting...
2025-07-19 Jeff Bolzvulkan: add RTE variants for glu/add/sub/mul/div (llama...
2025-07-19 R0CKSTARcuda: fix build warnings in set-rows.cu (unused variabl...
2025-07-19 Anton Mitkovsycl: Hotfix for non dnnl codepath (llama/14677)
2025-07-19 shalinib-ibmggml : refactor llamafile_sgemm PPC code (llama/14673)
2025-07-19 Akarshan BiswasSYCL: use 1D kernel for set_rows (llama/14618)
2025-07-19 Anton Mitkovsycl: Batched mulmat rework for oneDNN dispatch (llama...
2025-07-19 Sigbjørn Skjæretcuda : add set rows for bf16 (llama/14664)
2025-07-19 Yavor Ivanovcuda : add ELU support (llama/14657)
2025-07-19 Georgi Gerganovggml : add build-time message to remind about ggml_set_...
2025-07-19 Yavor Ivanovmetal : Add missing unary ops Metal support (llama...
2025-07-19 Tarek Dakhrantests : cover lfm2 cases in test_ssm_conv (llama/14651)
2025-07-19 Aman GuptaCUDA: add set rows for f32 and f16 (llama/14551)
2025-07-12 Georgi Gerganovsync : whispper.cpp
2025-07-12 Georgi Gerganovgit : remove kompute submodule (#1300) upstream/0.0.2309
2025-07-12 Georgi Gerganovsync : resolve conflicts (#0)
2025-07-12 Georgi Gerganovsync : llama.cpp
2025-07-12 Jeff Bolzvulkan: support SET_ROWS (llama/14587)
2025-07-12 Jeff Bolzvulkan: optimizations for deepseek prompt processing...
2025-07-12 Tarek Dakhranmodel : support LiquidAI LFM2 hybrid family (llama...
2025-07-12 Slobodan JosicHIP : Add HIP 7.0+ compatibility for hipBLAS compute...
2025-07-12 rmatifopencl: add tiled mul_mat_f16_f32 (llama/14535)
2025-07-12 lhezopencl: add `set_rows` for `f16` and `f32` (llama/14547)
2025-07-12 Aman GuptaDocs: script to auto-generate ggml operations docs...
2025-07-12 Akarshan BiswasSYCL: Initial set_rows kernel implementation (llama...
2025-07-12 compiladecuda : support Falcon-H1 state size for SSM_SCAN (llama...
2025-07-12 Xuan-Son Nguyenggml : add ggml_scale_bias (llama/14417)
2025-07-12 Miaoqian Linggml : prevent integer overflow in gguf tensor size...
2025-07-12 Jeff Bolzvulkan: optimize flash attention split_k_reduce (llama...
2025-07-12 Jeff Bolzvulkan : fix rope with partial rotation and non-cont...
2025-07-12 Georgi Gerganovcuda : fix rope with partial rotation and non-cont...
2025-07-12 Aman GuptaCUDA: add bilinear interpolation for upscale (llama...
2025-07-12 R0CKSTARmusa: fix build warnings (unused variable) (llama/14561)
2025-07-12 Aman GuptaCUDA: add bf16 and i32 to getrows (llama/14529)
2025-07-12 Evevulkan: increase LOAD_VEC_A to 8 (IQ1/IQ2) or 4 (IQ3...
2025-07-12 Jeff Bolzvulkan: fix rms_norm+mul fusion (llama/14545)
2025-07-12 Jeff Bolzvulkan: Handle updated FA dim2/3 definition (llama...
2025-07-12 Sigbjørn Skjæretopencl: add GELU_ERF (llama/14476)
2025-07-12 R0CKSTARtest-backend-ops: add support for specifying output...
2025-07-12 Georgi Gerganovmetal : disable fast math in all quantize kernels ...
2025-07-12 luyhcsuCANN: Replace aclrtMemsetSync with aclnnInplaceZero...
2025-07-12 Sigbjørn Skjæretggml : implement GEGLU_ERF and GEGLU_QUICK ops (llama...
2025-07-12 lhezopencl : broadcast for soft_max (llama/14510)
2025-07-12 Jeff Bolzvulkan: support mixed/deepseekR1 FA head sizes (llama...
2025-07-12 Johannes Gäßlerggml: backward pass for split swiglu (llama/14483)
2025-07-12 Nicolò ScipioneFix conditional enabling following arch checks for...
2025-07-12 Georgi Gerganovkv-cache : use ggml_set_rows (llama/14285)
2025-07-12 Georgi Gerganovggml : fix FA mask dim 2 and 3 (llama/14505)
2025-07-12 Georgi Gerganovggml : remove kompute backend (llama/14501)
2025-07-12 Aman GuptaCUDA: add dynamic shared mem to softmax, refactor gener...
2025-07-12 compiladellama : initial Mamba-2 support (llama/9126)
2025-07-12 Aman GuptaCUDA: add softmax broadcast (llama/14475)
2025-07-12 Johannes GäßlerCUDA: broadcasting for FlashAttention mask (llama/14500)
2025-07-12 Jeff Bolzvulkan: support softmax/FA batch and broadcast (llama...
2025-07-12 Georgi Gerganovsync : llama.cpp
2025-07-12 Georgi Gerganovggml : support bcast ggml_soft_max_ext, ggml_flash_attn...
2025-07-12 zhouwgopencl : fix possible buffer overflow in dump_tensor...
2025-07-12 Eric Zhangopencl : skip empty nodes on cgraph compute (llama...
2025-07-12 lhezopencl : update upscale to support align corners (llama...
2025-07-12 Björn Gansterggml : Callback before abort (llama/14481)
2025-07-12 Georgi Gerganovci : disable fast-math for Metal GHA CI (llama/14478)
2025-07-12 Chenguang LiCANN: update aclnnGroupedMatmulV2 to aclnnGroupedMatmul...
2025-07-12 Jeff Bolzvulkan: Split large mul_mat_id to fit in shared memory...
2025-07-12 Sigbjørn Skjæretadd GELU_ERF (llama/14455)
2025-07-11 Kai Pastorci : simplify, switch to ninja (#1295)
2025-07-10 Kai Pastorexamples : Test installed CMake config package (#1294)
2025-07-03 Aclyvulkan : implement bilinear interpolation for ggml_upsc...
2025-07-03 Aclyvulkan : implement ggml_roll (#1290)
2025-07-02 Daniel Beveniusggml : add version function to get lib version (#1286)
2025-07-02 Georgi Gerganovsync : whisper.cpp
2025-07-01 Georgi Gerganovsync : llama.cpp
2025-07-01 Georgi Gerganovggml : remove trailing whitespace (llama/0)
2025-07-01 lhezopencl : add GEGLU, REGLU, SWIGLU (llama/14456)
2025-07-01 Aman GuptaAdd Conv2d for CPU (llama/14388)
next