]> git.djapps.eu Git - pkg/ggml/sources/whisper.cpp/shortlog
pkg/ggml/sources/whisper.cpp
2024-01-17 Georgi Gerganovimatrix : offload to GPU support (llama/4957)
2024-01-17 Georgi Gerganovbackend : add eval callback (llama/4935)
2024-01-17 Georgi Gerganovmetal : create autorelease pool during library build...
2024-01-17 Kawrakowggml : importance matrix support for legacy quants...
2024-01-17 Alex Azarovmetal : log `recommendedMaxWorkingSetSize` on iOS 16...
2024-01-17 Justine Tunneyggml : introduce GGML_CALL function annotation (llama...
2024-01-17 Georgi Gerganovcuda : fix dequantize kernel names (llama/4938)
2024-01-17 KawrakowCUDA: faster dequantize kernels for Q4_0 and Q4_1 ...
2024-01-17 KawrakowAdd ability to use importance matrix for all k-quants...
2024-01-16 Benjamin Heinigertalk-llama : optional wake-up command and audio confirm...
2024-01-15 Przemysław... server : fix building and simplify lib deps on Windows...
2024-01-14 Georgi Gerganovtalk-llama : sync llama.cpp
2024-01-14 Georgi Gerganovtalk-llama : llama.cpp
2024-01-14 Georgi Gerganovsync : ggml
2024-01-14 Alex Azarovmetal : correctly set SIMD support flags on iOS (llama...
2024-01-14 Kawrakow2-bit quantizations (llama/4897)
2024-01-14 Georgi Gerganovscripts : sync-ggml-am.sh add option to skip commits
2024-01-13 Georgi Gerganovtalk-llama : sync llama.cpp
2024-01-13 Georgi Gerganovsync : ggml
2024-01-13 Georgi Gerganovexamples : adapt to metal API
2024-01-13 Johannes Gäßlerggml: cache sin/cos for RoPE (llama/4908)
2024-01-13 Georgi Gerganovmetal : remove old API (llama/4919)
2024-01-13 Georgi Gerganovmetal : disable log for loaded kernels (llama/4794)
2024-01-13 texmex76gguf : fix potential infinite for-loop (llama/4600)
2024-01-13 Georgi Gerganovmetal : refactor kernel loading code (llama/4794)
2024-01-13 Johannes GäßlerCUDA: faster q8_0 -> f16 dequantization (llama/4895)
2024-01-13 RhinoDeveltalk-llama : add optional CLI arg to set the bot name...
2024-01-13 james wolfexamples : add python example for transcription (#1744)
2024-01-13 Georgi Gerganovwhisper : load the model into multiple buffers of max...
2024-01-12 Georgi Gerganovtalk-llama : sync llama.cpp
2024-01-12 Georgi Gerganovsync : ggml
2024-01-12 slarenbackend_sched : fix assignments
2024-01-12 slarenllama : ggml-backend integration (llama/4766)
2024-01-12 Johannes GäßlerCUDA: fix softmax compile for old CUDA versions (llama...
2024-01-12 KawrakowImportance Matrix calculation (llama/4861)
2024-01-12 Sơn Phan Trungmodels : make all scripts to be POSIX Compliant (#1725)
2024-01-12 Georgi Gerganovggml : fix 32-bit ARM compat for IQ2_XS (#1758)
2024-01-12 Boris Bliznioukovgo : add SetInitialPrompt method to bindings (#1753)
2024-01-12 George Hindleserver : add more parameters to server api (#1754)
2024-01-12 Georgi Gerganovwhisper : fix segment length with params.no_timestamps...
2024-01-12 George Hindleparams : don't compute timestamps when not printing...
2024-01-11 Georgi Gerganovtalk-llama : sync llama.cpp
2024-01-11 Georgi Gerganovswift : remove local ggml.h reference
2024-01-11 Georgi Gerganovswift : track ggml release branch
2024-01-11 Georgi Gerganovsync : ggml
2024-01-11 Georgi Gerganovsync : llama.cpp
2024-01-11 Kawrakowggml : SOTA 2-bit quants (add IQ2_XS) (llama/4856)
2024-01-11 Paul Tsochantarismetal : put encoder debug group behind a define (llama...
2024-01-11 Georgi Gerganovmetal : improve dequantize precision to match CPU ...
2024-01-11 Georgi Gerganovggml : fix vld1q_s8_x4 32-bit compat (llama/4828)
2024-01-11 Johannes GäßlerCUDA: faster softmax via shared memory + fp16 math...
2024-01-11 Georgi Gerganovmetal : fix deprecation warning (ggml/690)
2024-01-11 Timothy Croninggml : remove ggml_cpy_inplace and ggml_cont_inplace...
2024-01-11 Jack Mousseaumetal : wrap each operation in debug group (ggml/690)
2024-01-11 leejetggml : change GGML_MAX_NAME at compile time (ggml/682)
2024-01-11 Halalaluyafail3Fix execlp call (ggml/689)
2024-01-11 KawrakowSOTA 2-bit quants (llama/4773)
2024-01-11 Johannes GäßlerCUDA: fixed redundant value dequantization (llama/4809)
2024-01-11 Konstantin... ggml : use __builtin_amdgcn_sudot4 in __dp4a for gfx11...
2024-01-11 Georgi Gerganovggml : do not sched_yield when calling BLAS (llama...
2024-01-11 Georgi Gerganovggml : include stdlib.h before intrin.h (llama/4736)
2024-01-10 Alexandru Mariutiswift : checkout ggml commit instead of branch (#1750)
2024-01-10 RhinoDeveltalk-llama : add optional Piper TTS support (#1749)
2024-01-08 Emmanuel Schmidbauerserver : add request path option(#1741)
2024-01-08 Georgi Gerganovmain : add cli option to disable system prints (#1740)
2024-01-07 Georgi Gerganovserver : fix server temperature + add temperature_inc...
2024-01-06 Georgi Gerganovtalk-llama : sync latest llama.cpp
2024-01-05 Georgi Gerganovrelease : v1.5.4
2024-01-05 Erik Scholzfix : cuda order of synchronization when setting a...
2024-01-05 Georgi Gerganovmetal : switch back to default.metallib (ggml/681)
2024-01-05 Georgi Gerganovggml : fix q2_k bpw in comments (ggml/680)
2024-01-04 Yajing Tangcoreml : fix ANE optimized encoder (#1716)
2024-01-04 Georgi Gerganovwhisper.swiftui : add .gitignore
2024-01-04 Georgi Gerganovwhispser : reset the "batched" timings (#1721)
2024-01-03 Georgi Gerganovrelease : v1.5.3
2024-01-03 Ashraful Islamswift : update Package.swift to use ggml as package...
2024-01-03 Finn Voorheesggml : add error handling to graph_compute (#1714)
2024-01-03 Georgi Gerganovcuda : simplify expression
2024-01-03 Georgi Gerganovcuda : mark I16 and I32 ops as unsupported
2024-01-03 Georgi Gerganovmetal : add kernel_get_rows_i32
2024-01-03 Georgi Gerganovmetal : optimize ggml_mul_mat_id (faster Mixtral PP...
2024-01-03 Georgi Gerganovmetal : enable shader debugging (cmake option) (llama...
2024-01-03 Georgi Gerganovggml : add ggml_vdotq_s32 alias (llama/4715)
2024-01-03 Johannes GäßlerCUDA: fixed tensor cores not being used on RDNA3 (llama...
2024-01-03 automaticcatggml : add ggml_cpu_has_avx_vnni() (llama/4589)
2024-01-03 Johannes GäßlerCUDA: fix tensor core logic for Pascal and HIP (llama...
2024-01-03 hydaicuda: fix vmm oom issue on NVIDIA AGX Orin (llama/4687)
2024-01-03 Guillaume Wenzekggml : extend ggml_get_rows, ggml_repeat, ggml_concat...
2024-01-03 Georgi Gerganovscripts : fix sync order + metal sed
2024-01-02 Andreu Huguetexamples : fix WASM Stack Overflow (#1713)
2023-12-30 bobqianicdocker : fix the publishing of the CUDA Docker image...
2023-12-29 Georgi Gerganovscripts : do not sync commits from this repo
2023-12-29 Tamotsu Takahashici : build with CLBlast + ggml-opencl use GGML_API...
2023-12-29 bobqianicwhisper : replace `tensor->n_dims` with `ggml_n_dims...
2023-12-29 Georgi Gerganovsync : ggml (VMM, sync-ggml-am, dotprod ARM fixes,...
2023-12-29 Dimodownload : fix large q5 model name (#1695)
2023-12-23 bobqianicwhisper : Replace WHISPER_PRINT_DEBUG with WHISPER_LOG_...
2023-12-22 Georgi Gerganovsync : ggml (ggml_scale, ggml_row_size, etc.) (#1677)
2023-12-22 Chaoqundocker : Dockerize whisper.cpp (#1674)
2023-12-21 bobqianicCI : Add coverage for talk-llama when WHISPER_CUBLAS...
next