]> git.djapps.eu Git - pkg/ggml/sources/whisper.cpp/shortlog
pkg/ggml/sources/whisper.cpp
11 days ago Evevulkan: more mul mat optimizations (llama/18533)
11 days ago hipuddingCANN: Fix rename for get_env (llama/18652)
11 days ago Raul TorresCANN: Rename `get_env` to `get_env_as_lowercase` (llama...
11 days ago Max KrasnyanskyHexagon add support for f16/f32 flash attention, scale...
11 days ago Aadeshveer... ggml : optimize cuda ssm_scan using warp-level reductio...
11 days ago Jeff Bolzvulkan: support buffer_from_host_ptr (llama/18467)
11 days ago Aman Guptaggml-cuda: refactor cuda graph usage (llama/18637)
11 days ago Beinseziimmq.cu: tune mmq/rocblas switching for RDNA (llama...
11 days ago Adrien Gallouëtggml : fix avx512bf16 build (llama/18623)
11 days ago Raul TorresCANN: Make `valid_values` variable `static const` ...
11 days ago nwyinggml webgpu: add CEIL operation support (llama/18605)
11 days ago Johannes GäßlerCUDA: fix FA FP16 accumulator overflow for Granite...
11 days ago Aman Guptaggml-cuda: check for srcs outside the cgraph (llama...
11 days ago Jeff Bolzvulkan: fix topk_moe_sigmoid_norm_bias failures in...
11 days ago Jeff Bolzvulkan: handle quantize_q8_1 overflowing the max workgr...
11 days ago Chenguang LiCANN: add operator fusion support for ADD + RMS_NORM...
11 days ago Daniel Beveniussampling : add support for backend sampling (llama...
11 days ago Aman GuptaCUDA: disable cuda graph when using n-cpu-moe (llama...
11 days ago Aman Guptaggml-cuda: remove unused params in ggml_cuda_graph...
11 days ago Aman Guptaggml-cuda: fixes for concurrent streams (llama/18496)
11 days ago Johannes GäßlerCUDA: only allocate FA tmp buffer if needed (llama...
11 days ago pl752(Bugfix, ggml-cuda) Pool alloc count fix + small size...
11 days ago Shouyuggml-hexagon: optimize activation function (llama/18393)
11 days ago Jeff Bolzvulkan: Optimize GGML_OP_CUMSUM (llama/18417)
11 days ago Jeff Bolzvulkan: Implement mmvq for iq1_s/iq1_m (llama/18450)
11 days ago Georgi Gerganovmetal : adjust extra size for FA buffer to avoid reallo...
11 days ago Chris Rohlfrpc : use unordered_map::reserve and emplace (llama...
11 days ago MeeMincuda : fix copy of large tensors (ggml_nbytes <= INT_MA...
11 days ago Aman Guptaggml-cuda: remove unneccesary prints on ggml_cuda_init...
11 days ago Jeff Bolzvulkan: extend topk_moe to handle sigmoid w/exp_probs_b...
12 days ago Peter A.examples : fix executable example targets (#3600)
2026-01-05 KITAITI Makotoruby : fix segmentation fault (#3591)
2025-12-31 Georgi Gerganovsync : ggml
2025-12-31 Georgi Gerganovggml : bump version to 0.9.5 (ggml/1410)
2025-12-31 Georgi Gerganovtalk-llama : sync llama.cpp
2025-12-31 Georgi Gerganovsync : ggml
2025-12-31 gatbontonpcmetal : add count_equal op (llama/18314)
2025-12-31 Johannes GäßlerCUDA: fix KQ max calculation (llama/18487)
2025-12-31 Georgi Gerganovmetal : remove BF16 x F16 kernels (llama/18456)
2025-12-31 Aman Guptasycl: add newline at the end of CMakeLists.txt (llama...
2025-12-31 Rahul SatheWork around broken IntelSYCLConfig.cmake in Intel oneAP...
2025-12-31 Charles Xukleidiai: add and integrate SVE 256-bit vector-length...
2025-12-31 Aman GuptaCUDA: add log line when mxfp4 acceleration is used...
2025-12-31 Johannes GäßlerCUDA: fix replacment of bad archs in CMake (llama/18457)
2025-12-31 Johannes GäßlerCUDA: Blackwell features for non-native builds (llama...
2025-12-31 Aman Guptacuda: fix race condition in cumsum (llama/18448)
2025-12-31 uvosHIP: Use mmq on MFMA devices for MUL_MAT_ID in cases...
2025-12-31 Aman GuptaRevert "ggml-cuda: use CMAKE_CUDA_ARCHITECTURES if...
2025-12-31 o7sirpc: fix segfault on invalid endpoint format (llama...
2025-12-31 Boian Berberovcmake: Added more x86_64 CPU backends when building...
2025-12-31 QDeltaggml-cuda: use CMAKE_CUDA_ARCHITECTURES if set when...
2025-12-31 lhezopencl: allow resizing transpose buffers (llama/18384)
2025-12-31 Aman Guptaggml-cuda: Use same regex for GGML_NATIVE=OFF (llama...
2025-12-31 Jeff Bolzvulkan: preprocess mul_mat_id experts and discard workg...
2025-12-31 Jeff Bolzvulkan: optimize decodeFuncB in coopmat2 mul_mat_id...
2025-12-31 Jeff Bolzvulkan: Use BK=32 for coopmat2 mul_mat_id (llama/18332)
2025-12-31 Evevulkan: small dequantization improvements (llama/18380)
2025-12-31 Jeff Bolzvulkan: Support UPSCALE w/antialias (llama/18327)
2025-12-31 Jeff Bolzvulkan: handle rope with large number of rows (llama...
2025-12-31 0MarbleCANN: implement the SSM_CONV operator (llama/17737)
2025-12-31 Aman Guptaggml-cuda: fix regex for arch list (llama/18371)
2025-12-31 Aman Guptacuda: optimize cumsum cub path (llama/18362)
2025-12-31 Aman Guptaggml-cuda: fix blackwell native builds (llama/18361)
2025-12-31 Penglin CaiCANN: Add support for CONV_TRANSPOSE_1D when kernel...
2025-12-31 Aadeshveer... ggml : optimize cuda cumsum fallback kernel (llama...
2025-12-31 Aman GuptaCUDA: experimental native mxfp4 support for blackwell...
2025-12-31 Jeff Bolzvulkan: fix command buffer corruption in ggml_backend_v...
2025-12-31 Wang WeixuanCANN : refactor ACL graph cache (llama/17752)
2025-12-31 Ruben Ortlamvulkan: use fewer FA rows for small cache runs (llama...
2025-12-31 TianHao324CANN: Uses yarn_ramp cache in ROPE (llama/17725)
2025-12-31 Chris Rohlfrpc : add check for rpc buffer type (llama/18242)
2025-12-31 nullnameggml-hexagon: create generalized functions for cpu...
2025-12-31 Shouyuggml-hexagon: gelu optimization (llama/18151)
2025-12-31 Taimur Ahmadllamafile: add rvv support for sgemm kernels (llama...
2025-12-31 lhezopencl: unpack q4_0 for adreno in get_tensor (llama...
2025-12-31 Jeff Bolzvulkan: Extend rope fusions to allow mrope (llama/18264)
2025-12-31 Jeff Bolzvulkan: Implement set_tensor_async and the event interf...
2025-12-31 Johannes Gäßlerllama: fix RPC for -fit on (llama/18233)
2025-12-31 Jeff Bolzvulkan: fix im2col overflowing maxworkgroupcount (llama...
2025-12-31 Jeff Bolzvulkan/cuda: fix topk_moe with exp_probs_b (llama/18071)
2025-12-31 Jeff Bolzvulkan: support GGML_UNARY_OP_XIELU (llama/18062)
2025-12-31 Jeff Bolzvulkan: in graph_optimize, try to group ADD operations...
2025-12-31 lovedheartVulkan: some improvement on mul_mat_iq2_xs (llama/18031)
2025-12-31 Aadeshveer... Added comments explaining thread block size selection...
2025-12-31 Alfredggml-hexagon: Implement true Q8_0 quantization on Hexag...
2025-12-31 Jeff Bolzvulkan: Add perf logger mode with concurrency (llama...
2025-12-31 Xuan-Son Nguyenmodel : add ASR support for LFM2-Audio-1.5B (conformer...
2025-12-31 Taimur Ahmadggml-cpu: extend support for RVV floating-point kernels...
2025-12-31 yuloremove i_major_dual (llama/18157)
2025-12-31 Shouyuggml-hexagon: swiglu_oai operation (llama/18114)
2025-12-31 Shouyuggml-hexagon: gelu operation (llama/17921)
2025-12-31 Alberto Cabrera... ggml-cpu: ARM64: repack version of q8_0 (dotprod and...
2025-12-31 yuloHIP: Refactor mma for RDNA and CDNA (llama/17990)
2025-12-24 KITAITI Makotoruby : add Whisper::Token, fix model URI (#3575)
2025-12-18 Georgi Gerganovtalk-llama : sync llama.cpp
2025-12-18 Georgi Gerganovsync : ggml
2025-12-18 Naco Sirenllama.android : Rewrite Android binding (w/o cpu_featur...
2025-12-18 Aadeshveer... ggml : use WARP_SIZE/2 for argmax reduction offset...
2025-12-18 Shouyuggml-hexagon: mm for mtmd (llama/17894)
2025-12-18 Jeremy Demeulemetal: use shared buffers on eGPU (llama/17866)
next