]> git.djapps.eu Git - pkg/ggml/sources/whisper.cpp/shortlog
pkg/ggml/sources/whisper.cpp
2024-02-19 Kawrakowggml-quants : fix compiler warnings (shadow variable...
2024-02-19 Abhilash Majumderggml-sycl: Replace 3d ops with macro (llama/5458)
2024-02-19 Georgi Gerganovbuild : update CBLAS flags + fix unused var warning...
2024-02-19 Davidson Francismain : check if input files exist before proceeding...
2024-02-19 Felixexamples : clean up common code (#1871)
2024-02-19 Jumper775models : fix openvino setup info (#1874)
2024-02-13 Georgi Gerganovmodels : add update py requirements
2024-02-12 Georgi Gerganovswift : package no longer use ggml dependency (#1861)
2024-02-12 Georgi Gerganovwhisper : fix external encoder (#1860)
2024-02-12 Georgi Gerganovsync : ggml
2024-02-12 slarenggml-alloc : allocate all leafs as if they were inputs...
2024-02-12 Georgi Gerganovtalk-llama : sync llama.cpp
2024-02-12 Georgi Gerganovsync : ggml
2024-02-12 Georgi Gerganovggml-backend : sync remnant
2024-02-12 Johannes GäßlerCUDA: mul_mat_vec_q tiling, refactor mul mat logic...
2024-02-12 Sergio Lópezvulkan: only use M-sized matmul on Apple GPUs (llama...
2024-02-12 Georgi Gerganovggml : fix compile warnings (unused vars) (llama/4966)
2024-02-12 snadampalggml : add mmla kernels for quantized GEMM (llama/4966)
2024-02-12 Ian Bullmetal : use autoreleasepool to avoid memory leaks ...
2024-02-12 slarenggml-alloc : v3 (ggml/727)
2024-02-12 dscripkaexamples : added audio_ctx argument to main and server...
2024-02-11 Didzis Goskometal : option to embed MSL source into compiled binary...
2024-02-11 Georgi Gerganovexamples : initialize context params properly (#1852)
2024-02-10 Georgi Gerganovtalk-llama : sync llama.cpp
2024-02-10 Georgi Gerganovsync : ggml
2024-02-10 Georgi Gerganovsrc : relocate new backend sources
2024-02-10 Michael Podvitskiyggml : fix `error C2078: too many initializers` for...
2024-02-10 Johannes GäßlerCUDA: more warps for mmvq on NVIDIA (llama/5394)
2024-02-10 Johannes GäßlerCUDA: fixed mmvq kernel for bs 2,3,4 and -sm row (llama...
2024-02-10 0cc4mBasic Vulkan Multi-GPU implementation (llama/5321)
2024-02-10 Johannes GäßlerCUDA: mul_mat_vec_q max. batch size 8 -> 4 (llama/5370)
2024-02-10 KawrakowSlight quantization improvement for Q4_K and Q5_K ...
2024-02-10 Johannes GäßlerCUDA: mul_mat_vec_q for batch sizes > 1 (llama/5351)
2024-02-10 Kawrakowggml : make use of ggml-quants.h possible in C++ code...
2024-02-10 Dr. Tom Murphy... ggml : avoid duplicating function calls using MIN/MAX...
2024-02-10 Kawrakowiq2_xxs: tune quantization (llama/5320)
2024-02-10 slarencuda : fix LLAMA_CUDA_F16 (llama/5262)
2024-02-10 Georgi Gerganovmetal : add im2col F32 dst support (llama/5132)
2024-02-10 JidongZhang-THUllava : add MobileVLM support (llama/5132)
2024-02-10 slarenggml : limit n_threads to the max n_tasks (llama/5238)
2024-02-10 Jared Van Bortelkompute : llama-bench support and ggml_cpu_has_kompute...
2024-02-10 Michael Podvitskiyggml : add abort_callback for cpu backend (ggml/725)
2024-02-10 Georgi Gerganovextra : update sync scripts
2024-02-09 Valentin Gosuserver : allow CORS request with authorization headers...
2024-02-09 Neuman Vongwhisper.android : how to build with CLBlast (#1809)
2024-02-09 Didzis Goskowhisper : expose CUDA device setting in public API...
2024-02-09 Didzis Goskomake : add macOS deployment target option (#1839)
2024-02-06 Georgi Gerganovtalk-llama : stream response (#1121)
2024-01-30 Georgi Gerganovsync : ggml (#0)
2024-01-30 Kawrakowggml : fix IQ3_XXS on Metal (llama/5219)
2024-01-30 Georgi Gerganovsync : ggml (llama/0)
2024-01-30 KawrakowFaster AVX2 dot product for IQ2_XS (llama/5187)
2024-01-30 KawrakowSOTA 3-bit quants (llama/5196)
2024-01-30 Paul Tsochantarisggml alloc: Fix for null dereference on alloc failure...
2024-01-30 Jared Van BortelNomic Vulkan backend (llama/4456)
2024-01-30 slarenggml : add max buffer sizes to opencl and metal backend...
2024-01-30 Paul Tsochantarismetal : free metal objects (llama/5161)
2024-01-30 Georgi Gerganovgguf : fix comparison (ggml/715)
2024-01-30 John Balis`ggml_cuda_cpy` support for 4d tensors and float16...
2024-01-30 Georgi Gerganovgguf : add input validation, prevent integer overflows...
2024-01-30 Georgi Gerganovci : fix yolo URLs + fix metal capture (ggml/712)
2024-01-30 Jack Mousseaumetal : add debug capture backend function (ggml/694)
2024-01-30 JacobLinCoolcommon : fix wav buffer detection (#1819)
2024-01-30 JacobLinCoolserver : add fields to `verbose_json` response (#1802)
2024-01-30 jwijffelsmake : update MSYS_NT (#1813)
2024-01-28 Georgi Gerganovtalk-llama : sync llama.cpp
2024-01-28 Georgi Gerganovsync : ggml
2024-01-28 0cc4mggml : add Vulkan backend (llama/2059)
2024-01-28 Abhilash Majumderggml : add unified SYCL backend for Intel GPUs (llama...
2024-01-28 Georgi Gerganovggml : minor type fix (int64_t -> size_t)
2024-01-27 Georgi Gerganovcommon : fix input buffer check (#1812)
2024-01-27 Georgi Gerganovtalk-llama : sync llama.cpp
2024-01-27 Georgi Gerganovsync : ggml
2024-01-27 0cc4mAdd OpenCL add kernel (llama/5151)
2024-01-27 slarencuda : fix tensor size calculation for non-split buffer...
2024-01-27 slarenggml-alloc : add 10% margin to the buffer sizes (llama...
2024-01-27 snadampalggml : update softmax n_task calculation (llama/5126)
2024-01-27 Paul Tsochantarismetal : remove unused `n_buffers` and `buffers` (llama...
2024-01-27 Georgi Gerganovmetal : show compile log messages
2024-01-27 Engininja2cuda : fix 2-bit quants on amd hip (llama/5105)
2024-01-27 slarenllama : pre-allocate input tensors in a separate buffer...
2024-01-27 Georgi Gerganovmetal : disable support for MUL_MAT F32 x F16
2024-01-27 Johannes GäßlerCUDA: more info when no device code (llama/5088)
2024-01-27 Georgi Gerganovminor : clean-up some warnings and style (llama/5094)
2024-01-27 Reinforce-IIggml : parallelize FP32 conversion when using BLAS...
2024-01-27 XiaotaoChenllava : MobileVLM support (llama/4954)
2024-01-27 slarenllama : run all KQV ops on the CPU with no KV offload...
2024-01-27 Kylincuda : fix compile error in jetson platform (llama...
2024-01-27 Juddggml : check ggml_add src1 type (ggml/708)
2024-01-26 Michael Rienstradocs : make model options / model install methods clear...
2024-01-22 trixirtcmake : make libwhisper.so position independent (#1792)
2024-01-22 Georgi Gerganovcmake : temporary remove VLA check (#1795)
2024-01-19 Neuman Vongwhisper.android : return output from benchmarks (#1785)
2024-01-18 Ryan Hitchmanserver : implement "verbose_json" format with token...
2024-01-18 Georgi Gerganovggml : sync ggml-metal.m
2024-01-17 Georgi Gerganovsync : llama.cpp
2024-01-17 Georgi Gerganovsync : ggml
2024-01-17 Georgi Gerganovggml : add IQ2 to test-backend-ops + refactoring (llama...
2024-01-17 Georgi Gerganovimatrix : offload to GPU support (llama/4957)
2024-01-17 Georgi Gerganovbackend : add eval callback (llama/4935)
next