]> git.djapps.eu Git - pkg/ggml/sources/ggml/shortlog
pkg/ggml/sources/ggml
2025-10-14 Georgi Gerganovmetal : FA support F32 K and V and head size = 32 ...
2025-10-14 lhezopencl: fix build targeting CL 2 (llama/16554)
2025-10-14 Johannes GäßlerCUDA: fix numerical issues in tile FA kernel (llama...
2025-10-14 Jie Fu (傅杰)ggml : fix build broken with -march=armv9-a on MacOS...
2025-10-14 Chenguang LiCANN: fix CPU memory leak in CANN backend (llama/16549)
2025-10-14 Sam/Samuelmetal: add support for opt_step_sgd (llama/16539)
2025-10-14 Georgi Gerganovggml : fix scalar path for computing norm (llama/16558)
2025-10-14 hipuddingCANN: Update several operators to support FP16 data...
2025-10-14 Sam/Samuelmetal : add opt_step_adamw and op_sum (llama/16529)
2025-10-14 Neo Zhang Jianyufix UT fault cases: count-equal, argsort, pad OPs ...
2025-10-14 sirus20x6ggml : Fix FP16 ELU positive branch (llama/16519)
2025-10-14 sirus20x6ggml: Correct SVE implementation in ggml_vec_dot_f16_un...
2025-10-14 Johannes GäßlerCUDA: faster tile FA, add oob checks, more HSs (llama...
2025-10-12 Georgi Gerganovsync : llama.cpp
2025-10-12 Georgi Gerganovmetal : fix mul-mm condition + fix mul-mv permuted...
2025-10-12 Diego Devesacuda : avoid initializing unused devices (llama/16510)
2025-10-12 Prajwal B Mehendarkarcmake : Dont define XOPENSOURCE on AIX (llama/16481)
2025-10-12 dudutacpu : optimize the ggml NORM operation (llama/15953)
2025-10-12 Chenguang LiCANN: Improve ACL graph matching (llama/16166)
2025-10-12 Charles Xukleidiai: kernel interface refactoring (llama/16460)
2025-10-12 Neo Zhang Jianyurefactor soft_max, add soft_max_back (llama/16472)
2025-10-12 ai-fonsiDisable CUDA host buffers on integrated GPUs (llama...
2025-10-12 Georgi Gerganovmetal : mark FA blocks (llama/16372)
2025-10-12 Reese Levineggml webgpu: profiling, CI updates, reworking of comman...
2025-10-12 Georgi Gerganovmetal : add support for non-padded FA KV (llama/16148)
2025-10-12 Georgi Gerganovtests : add -INF blocks to the KQ mask in the FA tests...
2025-10-12 Georgi Gerganovmetal : various optimizations + refactoring (llama...
2025-10-12 Georgi Gerganovggml : fix unaligned access in AMX code (llama/16315)
2025-10-12 Daniel Beveniusggml-cpu : fix leftover handling in ggml_vec_scale_f32...
2025-10-12 Reese Levineggml webgpu: actually add softmax, fix rms_norm offset...
2025-10-12 Evevulkan: use a more appropriate amount of threads when...
2025-10-12 Radoslav Gerganovrpc : check src buffer when copying tensor (llama/16421)
2025-10-12 Radoslav Gerganovrpc : add support for multiple devices (llama/16276)
2025-10-12 Georgi Gerganovsync : llama.cpp
2025-10-12 Aclyvulkan : incremental shader builds (llama/16341)
2025-10-12 Georgi Gerganovsync : llama.cpp
2025-10-12 Georgi Gerganovmetal : fix loop bound in ggml_mem_ranges (llama/16412)
2025-10-12 Aclyggml : fix graph reallocation with multiple chunks...
2025-10-12 Jeff Bolzvulkan: Replace uses of maxMemoryAllocationSize and...
2025-10-12 Jeff Bolzvulkan: Fix FA coopmat1 invalid array indexing (llama...
2025-10-12 Jeff Bolzvulkan: in flash attention, bounds check against nem1...
2025-10-12 Reese Levineggml webgpu: add support for soft_max, optimize rms_nor...
2025-10-12 Piotr Wilkin... model : Apertus model implementation (llama/15852)
2025-10-12 R0CKSTARmusa: update compile flags (llama/16265)
2025-10-12 uvosHIP: Disable ROCWMMA fattn on CDNA when compiled agains...
2025-10-12 Evevulkan: make ggml_vk_default_dispatcher support older...
2025-10-12 lhezopencl: support pad_ext (llama/15888)
2025-10-12 Reese Levineggml webgpu: support for rope,div,sub,glu,scale,cont...
2025-10-12 lhezopencl: support ne3 in get_rows (llama/15866)
2025-09-30 Georgi Gerganovggml : bump version to 0.9.4 (#1363) upstream/0.9.4 v0.9.4
2025-09-30 Georgi Gerganovsync : whisper.cpp [no ci]
2025-09-30 Georgi Gerganovsync : llama.cpp
2025-09-30 anavp-nvidiacuda : Enable CUDA Graph usage for Nemotron Nano v2...
2025-09-30 Georgi Gerganovmetal : dynamic simdgroups for MV kernels (llama/16340)
2025-09-30 Charles Xukleidiai : fix work size and threads sync for fp16...
2025-09-30 Jeff Bolztests: override test_set_rows::max_nmse_err to allow...
2025-09-29 Georgi Gerganovsync : llama.cpp
2025-09-29 alex-spacemitggml: riscv: add riscv spacemit backend (llama/15288)
2025-09-29 Rafal Lewczukggml-backend : add root cause in error message if loadi...
2025-09-29 Georgi Gerganovsync : whisper.cpp (#1359)
2025-09-29 Georgi Gerganovci : print results [no ci] (#1358)
2025-09-29 Georgi Gerganovci : add self-hosted workflows (#1357)
2025-09-29 Georgi Gerganovcmake : remove metal flag (llama/0)
2025-09-29 Georgi Gerganovsync : llama.cpp
2025-09-29 Sigbjørn Skjæretggml : check cuda and metal argsort limits and add...
2025-09-29 Georgi Gerganovggml : fix dependencies for ggml_set_rows (llama/16318)
2025-09-29 Jeff Bolzvulkan: Fix validation failure in quantized flash atten...
2025-09-29 Sigbjørn Skjæretggml : fix GGML_F32_VEC_FMA argument order in ggml_vec_...
2025-09-29 Jeff Bolzvulkan: 64-bit im2col (llama/16135)
2025-09-29 Georgi Gerganovmetal : extend mat-mat multiplication support (llama...
2025-09-29 Georgi Gerganovmetal : fuse non-sequential nodes (llama/16102)
2025-09-29 Jeff Bolzvulkan: handle mat_mul with A matrix > 4GB (llama/16176)
2025-09-29 Jeff Bolzvulkan: support arbitrary KV dimension in flash attenti...
2025-09-29 Aclyvulkan : make the vulkan.hpp dynamic dispatcher instanc...
2025-09-29 Aman GuptaCUDA: mul_mat_id for mmf for bs <= 64 for f16 and bs...
2025-09-29 Johannes GäßlerCUDA: refactor and deduplicate vector FA kernels (llama...
2025-09-29 Dmytro Minochkinvulkan: throw system error instead of SIGABRT during...
2025-09-29 Jeff Bolzvulkan: support GET_ROWS for k-quants (llama/16235)
2025-09-29 Aaron Teodevops: add s390x & ppc64le CI (llama/15925)
2025-09-29 Georgi Gerganovmetal : report OOM errors (llama/16274)
2025-09-29 Adrien Gallouëtcommon : use cpp-httplib as a cURL alternative for...
2025-09-29 Aaron Teoggml-cpu: implement MXFP4 SIMD for s390x (llama/16193)
2025-09-29 R0CKSTARmusa: fix build warnings (llama/15611)
2025-09-29 Aman GuptaCUDA: add a fused top-K MoE kernel (llama/16130)
2025-09-29 junchao-zhaoggml : fix loongarch lsx compilation error (llama/15864)
2025-09-26 Daniel Beveniusggml : remove -dev suffix from release version (#1355)
2025-09-25 Christoph Reiterpkg-config: include the new GGML_VERSION as a version...
2025-09-25 hebangwenexamples : fix typo mismatch in gpt (#1349)
2025-09-25 Daniel Beveniusggml : bump version to 0.9.3 (#1353) v0.9.3
2025-09-25 Daniel Beveniusscripts : refactor release script into prepare and...
2025-09-25 Daniel Beveniusscripts : fix next dev version calculation [no ci]...
2025-09-25 Georgi Gerganovsync : llama.cpp
2025-09-25 Georgi Gerganovmetal : fuse NORM + MUL + ADD, support non-multiples...
2025-09-25 Georgi Gerganovmetal : relax reorder conditions (llama/16216)
2025-09-25 Georgi Gerganovmetal : restore im2col perf (llama/16219)
2025-09-25 Georgi Gerganovsync : llama.cpp
2025-09-25 Radoslav Gerganovrpc : use ggml logging facilities
2025-09-25 Eveci: run the x64 and arm ci on the github machines inste...
2025-09-25 Johannes Gäßlerllama: print memory breakdown on exit (llama/15860)
2025-09-25 Aclyggml : split graph allocations according to backend...
next