]> git.djapps.eu Git - pkg/ggml/sources/whisper.cpp/shortlog
pkg/ggml/sources/whisper.cpp
2024-02-28 Kawrakowggml : make i-quants work with super-blocks of 64 ...
2024-02-28 KawrakowAttempt to fix android build (llama/5752)
2024-02-28 KawrakowIQ4_XS: a 4.25 bpw quantization (llama/5747)
2024-02-28 Engininja2cuda : replace remaining shfl_xor with calls to warp_re...
2024-02-28 Engininja2ggml-quants : fix avx2 iq1_s vec_dot when compiled...
2024-02-28 KawrakowAdding IQ2_S and IQ2_M to complete coverage of the...
2024-02-28 Johannes GäßlerCUDA: fix DEBUG_CUDA_MALLOC (llama/5729)
2024-02-28 AidanBeltonSAdd support for soft_max ALiBi (llama/5639)
2024-02-28 Radosław Grytaggml-quants : provide ggml_vqtbl1q_u8 for 64bit compati...
2024-02-28 slarenadd google magika inference example (ggml/748)
2024-02-26 Andrew Sstream.wasm : fix invalid memory access when no segment...
2024-02-25 Georgi Gerganovtalk-llama : sync llama.cpp
2024-02-25 Georgi Gerganovsync : ggml
2024-02-25 Georgi Gerganovsync : llama.cpp (ggml/0)
2024-02-25 Georgi Gerganovcode : normalize enum names (llama/5697)
2024-02-25 KawrakowIQ3_S: a much better alternative to Q3_K (llama/5676)
2024-02-25 UEXTM.comIntroduce backend GUIDs (ggml/743)
2024-02-24 Tamotsu Takahashitalk, talk-llama : pass text_to_speak as a file (#1865)
2024-02-23 Abhilash Majumderwhisper : add SYCL support (#1863)
2024-02-22 Georgi Gerganovtalk-llama : sync llama.cpp
2024-02-22 Georgi Gerganovsync : ggml
2024-02-22 Georgi Gerganovggml : always define ggml_fp16_t as uint16_t (llama...
2024-02-22 Georgi Gerganovci : fix whitespace
2024-02-22 Georgi Gerganovggml : 32-bit arm compat (#1891)
2024-02-22 Georgi Gerganovsync : ggml
2024-02-22 Georgi Gerganovsync : llama.cpp (ggml/0)
2024-02-22 Meng, Hengyuconext add name (llama/5624)
2024-02-22 AidanBeltonSUpdate ggml_sycl_op_mul_mat_vec_q (llama/5502)
2024-02-22 0cc4mRefactor validation and enumeration platform checks...
2024-02-22 0cc4mAdd check for VK_KHR_portability_enumeration for Molten...
2024-02-22 Mathijs de... Add preprocessor checks for Apple devices.
2024-02-22 Mathijs de... Resolve ErrorIncompatibleDriver with Vulkan on MacOS.
2024-02-22 Mathijs de... Allow for Vulkan build with Accelerate.
2024-02-22 slarencuda : ignore peer access already enabled errors (llama...
2024-02-22 Siddharth Ramakrishnanggml : compute forward no longer pass src tensors ...
2024-02-22 bssrdfggml : fix conv_2d batch mode (ggml/737)
2024-02-22 st-gropenvino : fix convert-whisper-to-openvino.py (#1890)
2024-02-22 Davidson Francismain : fix file existence check in main.cpp (#1889)
2024-02-20 Georgi Gerganovtalk-llama : sync llama.cpp
2024-02-20 LBluemake : fix CUBLAS link with WSL (#1878)
2024-02-19 Georgi Gerganovsync : ggml
2024-02-19 Georgi Gerganovggml : resolve merge conflicts (ggml/0)
2024-02-19 Georgi Gerganovcommon : add IQ1_S (ggml/0)
2024-02-19 Georgi Gerganovci : enable -Werror for CUDA builds (llama/5579)
2024-02-19 slarencuda, metal : fix nans in soft_max (llama/5574)
2024-02-19 bmwlggml : android and old glibc NUMA incompatibility bugfi...
2024-02-19 Georgi Gerganovggml : restore vec dot stride arg names (llama/5453)
2024-02-19 Georgi Gerganovci : fix wikitext url + compile warnings (llama/5569)
2024-02-19 Georgi Gerganovmetal : fix unused warnings (llama/0)
2024-02-19 Herman Semenovggml, common, examples, tests : fixed type arguments...
2024-02-19 Kawrakow1.5 bit quantization (llama/5453)
2024-02-19 Georgi Gerganovggml : add ALiBi support for ggml_soft_max_ext (llama...
2024-02-19 Ananta Bastolaci : add an option to fail on compile warning (llama...
2024-02-19 Georgi Gerganovcmake : fix VULKAN and ROCm builds (llama/5525)
2024-02-19 bmwlggml : add numa options (llama/5377)
2024-02-19 slarencuda : print message when initialization fails (llama...
2024-02-19 Neuman Vongvulkan: Find optimal memory type but with fallback...
2024-02-19 ATEarly return for zero size calls to get_tensor. (llama...
2024-02-19 Kawrakowggml-quants : fix compiler warnings (shadow variable...
2024-02-19 Abhilash Majumderggml-sycl: Replace 3d ops with macro (llama/5458)
2024-02-19 Georgi Gerganovbuild : update CBLAS flags + fix unused var warning...
2024-02-19 Davidson Francismain : check if input files exist before proceeding...
2024-02-19 Felixexamples : clean up common code (#1871)
2024-02-19 Jumper775models : fix openvino setup info (#1874)
2024-02-13 Georgi Gerganovmodels : add update py requirements
2024-02-12 Georgi Gerganovswift : package no longer use ggml dependency (#1861)
2024-02-12 Georgi Gerganovwhisper : fix external encoder (#1860)
2024-02-12 Georgi Gerganovsync : ggml
2024-02-12 slarenggml-alloc : allocate all leafs as if they were inputs...
2024-02-12 Georgi Gerganovtalk-llama : sync llama.cpp
2024-02-12 Georgi Gerganovsync : ggml
2024-02-12 Georgi Gerganovggml-backend : sync remnant
2024-02-12 Johannes GäßlerCUDA: mul_mat_vec_q tiling, refactor mul mat logic...
2024-02-12 Sergio Lópezvulkan: only use M-sized matmul on Apple GPUs (llama...
2024-02-12 Georgi Gerganovggml : fix compile warnings (unused vars) (llama/4966)
2024-02-12 snadampalggml : add mmla kernels for quantized GEMM (llama/4966)
2024-02-12 Ian Bullmetal : use autoreleasepool to avoid memory leaks ...
2024-02-12 slarenggml-alloc : v3 (ggml/727)
2024-02-12 dscripkaexamples : added audio_ctx argument to main and server...
2024-02-11 Didzis Goskometal : option to embed MSL source into compiled binary...
2024-02-11 Georgi Gerganovexamples : initialize context params properly (#1852)
2024-02-10 Georgi Gerganovtalk-llama : sync llama.cpp
2024-02-10 Georgi Gerganovsync : ggml
2024-02-10 Georgi Gerganovsrc : relocate new backend sources
2024-02-10 Michael Podvitskiyggml : fix `error C2078: too many initializers` for...
2024-02-10 Johannes GäßlerCUDA: more warps for mmvq on NVIDIA (llama/5394)
2024-02-10 Johannes GäßlerCUDA: fixed mmvq kernel for bs 2,3,4 and -sm row (llama...
2024-02-10 0cc4mBasic Vulkan Multi-GPU implementation (llama/5321)
2024-02-10 Johannes GäßlerCUDA: mul_mat_vec_q max. batch size 8 -> 4 (llama/5370)
2024-02-10 KawrakowSlight quantization improvement for Q4_K and Q5_K ...
2024-02-10 Johannes GäßlerCUDA: mul_mat_vec_q for batch sizes > 1 (llama/5351)
2024-02-10 Kawrakowggml : make use of ggml-quants.h possible in C++ code...
2024-02-10 Dr. Tom Murphy... ggml : avoid duplicating function calls using MIN/MAX...
2024-02-10 Kawrakowiq2_xxs: tune quantization (llama/5320)
2024-02-10 slarencuda : fix LLAMA_CUDA_F16 (llama/5262)
2024-02-10 Georgi Gerganovmetal : add im2col F32 dst support (llama/5132)
2024-02-10 JidongZhang-THUllava : add MobileVLM support (llama/5132)
2024-02-10 slarenggml : limit n_threads to the max n_tasks (llama/5238)
2024-02-10 Jared Van Bortelkompute : llama-bench support and ggml_cpu_has_kompute...
2024-02-10 Michael Podvitskiyggml : add abort_callback for cpu backend (ggml/725)
next