]> git.djapps.eu Git - pkg/ggml/sources/ggml/shortlog
pkg/ggml/sources/ggml
2025-07-12 Georgi Gerganovsync : resolve conflicts (#0)
2025-07-12 Georgi Gerganovsync : llama.cpp
2025-07-12 Jeff Bolzvulkan: support SET_ROWS (llama/14587)
2025-07-12 Jeff Bolzvulkan: optimizations for deepseek prompt processing...
2025-07-12 Tarek Dakhranmodel : support LiquidAI LFM2 hybrid family (llama...
2025-07-12 Slobodan JosicHIP : Add HIP 7.0+ compatibility for hipBLAS compute...
2025-07-12 rmatifopencl: add tiled mul_mat_f16_f32 (llama/14535)
2025-07-12 lhezopencl: add `set_rows` for `f16` and `f32` (llama/14547)
2025-07-12 Aman GuptaDocs: script to auto-generate ggml operations docs...
2025-07-12 Akarshan BiswasSYCL: Initial set_rows kernel implementation (llama...
2025-07-12 compiladecuda : support Falcon-H1 state size for SSM_SCAN (llama...
2025-07-12 Xuan-Son Nguyenggml : add ggml_scale_bias (llama/14417)
2025-07-12 Miaoqian Linggml : prevent integer overflow in gguf tensor size...
2025-07-12 Jeff Bolzvulkan: optimize flash attention split_k_reduce (llama...
2025-07-12 Jeff Bolzvulkan : fix rope with partial rotation and non-cont...
2025-07-12 Georgi Gerganovcuda : fix rope with partial rotation and non-cont...
2025-07-12 Aman GuptaCUDA: add bilinear interpolation for upscale (llama...
2025-07-12 R0CKSTARmusa: fix build warnings (unused variable) (llama/14561)
2025-07-12 Aman GuptaCUDA: add bf16 and i32 to getrows (llama/14529)
2025-07-12 Evevulkan: increase LOAD_VEC_A to 8 (IQ1/IQ2) or 4 (IQ3...
2025-07-12 Jeff Bolzvulkan: fix rms_norm+mul fusion (llama/14545)
2025-07-12 Jeff Bolzvulkan: Handle updated FA dim2/3 definition (llama...
2025-07-12 Sigbjørn Skjæretopencl: add GELU_ERF (llama/14476)
2025-07-12 R0CKSTARtest-backend-ops: add support for specifying output...
2025-07-12 Georgi Gerganovmetal : disable fast math in all quantize kernels ...
2025-07-12 luyhcsuCANN: Replace aclrtMemsetSync with aclnnInplaceZero...
2025-07-12 Sigbjørn Skjæretggml : implement GEGLU_ERF and GEGLU_QUICK ops (llama...
2025-07-12 lhezopencl : broadcast for soft_max (llama/14510)
2025-07-12 Jeff Bolzvulkan: support mixed/deepseekR1 FA head sizes (llama...
2025-07-12 Johannes Gäßlerggml: backward pass for split swiglu (llama/14483)
2025-07-12 Nicolò ScipioneFix conditional enabling following arch checks for...
2025-07-12 Georgi Gerganovkv-cache : use ggml_set_rows (llama/14285)
2025-07-12 Georgi Gerganovggml : fix FA mask dim 2 and 3 (llama/14505)
2025-07-12 Georgi Gerganovggml : remove kompute backend (llama/14501)
2025-07-12 Aman GuptaCUDA: add dynamic shared mem to softmax, refactor gener...
2025-07-12 compiladellama : initial Mamba-2 support (llama/9126)
2025-07-12 Aman GuptaCUDA: add softmax broadcast (llama/14475)
2025-07-12 Johannes GäßlerCUDA: broadcasting for FlashAttention mask (llama/14500)
2025-07-12 Jeff Bolzvulkan: support softmax/FA batch and broadcast (llama...
2025-07-12 Georgi Gerganovsync : llama.cpp
2025-07-12 Georgi Gerganovggml : support bcast ggml_soft_max_ext, ggml_flash_attn...
2025-07-12 zhouwgopencl : fix possible buffer overflow in dump_tensor...
2025-07-12 Eric Zhangopencl : skip empty nodes on cgraph compute (llama...
2025-07-12 lhezopencl : update upscale to support align corners (llama...
2025-07-12 Björn Gansterggml : Callback before abort (llama/14481)
2025-07-12 Georgi Gerganovci : disable fast-math for Metal GHA CI (llama/14478)
2025-07-12 Chenguang LiCANN: update aclnnGroupedMatmulV2 to aclnnGroupedMatmul...
2025-07-12 Jeff Bolzvulkan: Split large mul_mat_id to fit in shared memory...
2025-07-12 Sigbjørn Skjæretadd GELU_ERF (llama/14455)
2025-07-11 Kai Pastorci : simplify, switch to ninja (#1295)
2025-07-10 Kai Pastorexamples : Test installed CMake config package (#1294)
2025-07-03 Aclyvulkan : implement bilinear interpolation for ggml_upsc...
2025-07-03 Aclyvulkan : implement ggml_roll (#1290)
2025-07-02 Daniel Beveniusggml : add version function to get lib version (#1286)
2025-07-02 Georgi Gerganovsync : whisper.cpp
2025-07-01 Georgi Gerganovsync : llama.cpp
2025-07-01 Georgi Gerganovggml : remove trailing whitespace (llama/0)
2025-07-01 lhezopencl : add GEGLU, REGLU, SWIGLU (llama/14456)
2025-07-01 Aman GuptaAdd Conv2d for CPU (llama/14388)
2025-07-01 Georgi Gerganovmetal : disable fast-math for some cpy kernels (llama...
2025-07-01 Romain Biessyggml-cpu: sycl: Re-enable exp f16 (llama/14462)
2025-07-01 Diego Devesatest-backend-ops : disable llama test (llama/14461)
2025-07-01 xiaobing318cmake : Remove redundant include path in CMakeLists...
2025-07-01 Vedran Miletićscripts : make the shell scripts cross-platform (llama...
2025-07-01 Akarshan BiswasSYCL: disable faulty fp16 exp kernel (llama/14395)
2025-07-01 Sigbjørn Skjæretggml : fix unmerged GGML_FPxx_TO_FPxx refactoring ...
2025-07-01 Sigbjørn Skjæretggml : implement REGLU/GEGLU/SWIGLU ops (llama/14158)
2025-07-01 Jeff Bolzvulkan: Add fusion support for RMS_NORM+MUL (llama...
2025-07-01 Aman GuptaCUDA: add bf16 and f32 support to cublas_mul_mat_batche...
2025-07-01 Jeff Bolzvulkan: handle noncontig in the final case of ggml_vk_g...
2025-07-01 Jeff Bolzvulkan: lock accesses of pinned_memory vector (llama...
2025-07-01 Xinpeng Doufix async_mode bug (llama/14432)
2025-07-01 Jeff Bolzvulkan: Fix GGML_VULKAN_SHADER_DEBUG_INFO (llama/14427)
2025-07-01 Radoslav Gerganovggml : add ggml_set_rows (llama/14274)
2025-07-01 bandoticmake: regen vulkan shaders when shaders-gen sources...
2025-07-01 Georgi Gerganovmetal : add special-case mat-vec mul for ne00 == 4...
2025-07-01 Georgi Gerganovmetal : batch rows copy in a single threadgroup (llama...
2025-07-01 R0CKSTARmusa: enable fp16 mma (all) and cublas on qy2 (llama...
2025-07-01 Aaron Teoggml-cpu: enable IBM NNPA Vector Intrinsics (llama...
2025-07-01 Sigbjørn Skjæretggml : do not output unprintable characters on GGUF...
2025-07-01 Anton Mitkovsycl: GGML_SYCL_DISABLE_OPT on by default for all Intel...
2025-07-01 lhezopencl: ref count `ggml_backend_opencl_context` and...
2025-07-01 uvosCUDA/HIP: optimize mmv paths taken for HIP devices...
2025-07-01 Johannes GäßlerCUDA: mul_mat_v support for batch sizes > 1 (llama...
2025-07-01 uvosHIP: enable vec fattn on RDNA4 (llama/14323)
2025-07-01 Aman GuptaCUDA: add mean operation (llama/14313)
2025-07-01 Markus TavenrathAdd support for VK_EXT_debug_utils to add labels to...
2025-07-01 Georgi Gerganovmetal : fix thread-safety (llama/14300)
2025-07-01 Aclyggml-cpu : "align corners" for bilinear upscale/downsca...
2025-06-25 Aclybuild : fix build with clang-cl on Windows (#1284)
2025-06-24 Daniel Beveniusggml-quants : rename best_mad to best_error (#1283)
2025-06-21 Georgi Gerganovtests : cleanup old tests (#1282)
2025-06-20 Georgi Gerganovsync : llama.cpp
2025-06-20 Aman GuptaCUDA: add conv_2d_transpose (llama/14287)
2025-06-20 Nicolò Scipionesycl: add usage of enqueue_functions extension (llama...
2025-06-20 Christian KastnerImplement GGML_CPU_ALL_VARIANTS for PowerPC (llama...
2025-06-20 Diego Devesacuda : synchronize graph capture and cublas handle...
2025-06-20 Georgi Gerganovggml : fix repack work size for mul_mat_id (llama/14292)
2025-06-20 Charles Xuggml: Update KleidiAI to v1.9.0 (llama/14277)
2025-06-20 Aman GuptaCUDA: add conv_2d_dw (llama/14265)
next