]> git.djapps.eu Git - pkg/ggml/sources/whisper.cpp/shortlog
pkg/ggml/sources/whisper.cpp
2024-08-08 Johannes GäßlerCUDA: fix partial offloading for ne0 % 256 != 0 (llama...
2024-08-08 65acmake : install all ggml public headers (llama/8480)
2024-08-08 hipuddingAdd Ascend NPU backend (llama/6035)
2024-08-08 Johannes Gäßlermake/cmake: add missing force MMQ/cuBLAS for HIP (llama...
2024-08-08 Xuan Son NguyenRefactor lora adapter support (llama/8332)
2024-08-08 Meng, Hengyuadd concat through dim 1/2 (llama/8483)
2024-08-08 0cc4mVulkan MMQ Fix (llama/8479)
2024-08-08 bandotivulkan : cmake integration (llama/8119)
2024-08-08 Georgi Gerganovmetal : template-ify some of the kernels (llama/8447)
2024-08-08 Georgi Gerganovggml : minor naming changes (llama/8433)
2024-08-08 Chen Xifix the mul_mat_id ut issues (llama/8427)
2024-08-08 Nicholai Tukanovggml : add NVPL BLAS support (ggml/8329) (llama/8425)
2024-08-08 Daniel Beveniuscuda : suppress 'noreturn' warn in no_device_code ...
2024-08-08 Johannes GäßlerCUDA: optimize and refactor MMQ (llama/8416)
2024-08-08 AidanBeltonSUse multi_ptr to clean up deprecated warnings (llama...
2024-08-08 Georgi Gerganovggml : move sgemm sources to llamafile subfolder (llama...
2024-08-08 Dibakar Gopeggml : add AArch64 optimized GEMV and GEMM Q4 kernels...
2024-08-08 Alberto Cabrera... sycl : Reenabled mmvq path for the SYCL Nvidia Backend...
2024-08-08 Alberto Cabrera... sycl : fix powf call in device code (llama/8368)
2024-08-08 Mahesh Madhavggml : loop tiling optimizations for scalar path (ggml...
2024-08-08 Ivan Filipovggml: add support for float16 input tensors in pooling...
2024-08-08 Tony Wasserkavulkan : initialize vk_buffer_struct members to VK_NULL...
2024-08-08 Borislav Stanimirovcmake : only enable GGML_NATIVE and x86 flags if not...
2024-08-08 Georgi Gerganovscripts : sync new files (#0)
2024-08-05 Daven Sanassycmake : fix compile in xcode (#2311)
2024-07-27 Georgi Gerganovwhisper : handle empty mel (#2324)
2024-07-16 Matt Stephensonwhisper : use vulkan as gpu backend when available...
2024-07-15 arizhihwhisper : fix DTW assert (#2299)
2024-07-09 Georgi Gerganovcmake : use WHISPER_EXTRA_FLAGS (#2294)
2024-07-09 Borislav Stanimirovcmake : allow external ggml
2024-07-08 Georgi Gerganovcmake : try to fix openvino build (#2281)
2024-07-08 Georgi Gerganovcmake : remove install of llama convert script [no...
2024-07-08 Georgi Gerganovmake : remove llama prints [no ci] (#2265)
2024-07-08 Georgi Gerganovtalk-llama : sync llama.cpp
2024-07-08 Georgi Gerganovexamples : fix compile warnings [no ci] (#0)
2024-07-08 Georgi Gerganovsync : ggml
2024-07-08 Georgi Gerganovggml : sync sycl (skip) (#0)
2024-07-08 Georgi Gerganovscripts : fix sync scripts
2024-07-08 Daniel Beveniusggml : remove unnecessary UNUSED macro call (ggml/880)
2024-07-08 Natsucmake : add GGML_BUILD and GGML_SHARED macro definition...
2024-07-08 Ouadie EL FAROUKIEnabled more data types for oneMKL gemm_batch (llama...
2024-07-08 Johannes GäßlerCUDA: MMQ support for iq4_nl, iq4_xs (llama/8278)
2024-07-08 DanieleCUDA: revert part of the RDNA1 optimizations (llama...
2024-07-08 Johannes GäßlerCUDA: fix MMQ stream-k rounding if ne00 % 128 != 0...
2024-07-08 luoyu-intelFix WARP_SIZE=16 bug of Intel GPU (llama/8266)
2024-07-08 Neo Zhang Jianyurm get_work_group_size() by local cache for performance...
2024-07-08 DanieleDefine and optimize RDNA1 (llama/8085)
2024-07-08 Juddfix typo (llama/8267)
2024-07-08 Clint HerronRemoves multiple newlines at the end of files that...
2024-07-08 slarencuda : update supports_op for matrix multiplication...
2024-07-08 luoyu-intelFix win build conflict of math library (llama/8230)
2024-07-08 luoyu-intelFix the sub group size of Intel (llama/8106)
2024-07-08 Johannes GäßlerCUDA: refactor and optimize IQ MMVQ (llama/8215)
2024-07-08 zhentaoyuUpdate SYCL-Rope op and Refactor (llama/8157)
2024-07-08 Johannes GäßlerCUDA: fix MMQ stream-k for --split-mode row (llama...
2024-07-08 John Balisfeat: cuda implementation for `ggml_conv_transpose_1d...
2024-07-08 Georgi Gerganovci : disable java build
2024-07-08 Emmanuel Schmidbauerserver : add inference path to make OAI API compatible...
2024-06-26 Georgi Gerganovsync : ggml + fix sync script
2024-06-26 Georgi Gerganovmake : disable CUDA graphs
2024-06-26 slarenggml : add GGML_CUDA_USE_GRAPHS option, restore GGML_CU...
2024-06-26 Georgi Gerganovmake : disable CUDA mel build
2024-06-26 Georgi Gerganovcmake : minor fixes
2024-06-26 Georgi Gerganovmake : fix missing -O3
2024-06-26 Georgi Gerganovwhisper : disable CUDA mel + fix FFMPEG
2024-06-26 Georgi Gerganovsync : ggml
2024-06-26 Georgi Gerganovwhisper : reorganize source code + improve CMake (...
2024-06-18 mky_coderwhisper : optimize fft() function (#2242)
2024-06-18 Georgi Gerganovtalk-llama : sync llama.cpp
2024-06-18 Georgi Gerganovwhisper : use ggml_backend_sched (#2239)
2024-06-18 Georgi Gerganovfix : remove extra files
2024-06-18 Georgi Gerganovscripts : sync ggml-blas
2024-06-18 Georgi Gerganovbuild : update make / cmake
2024-06-18 Georgi Gerganovsync : ggml
2024-06-18 slarenmove BLAS to a separate backend (cont) (llama/6210)
2024-06-18 0cc4mVulkan Shader Refactor, Memory Debugging Option (llama...
2024-06-18 Georgi Gerganovscripts : stop sync whisper example from ggml
2024-06-16 Georgi Gerganovcmake : fix sycl build (#0)
2024-06-16 Georgi Gerganovggml : remove OpenCL (#0)
2024-06-16 Georgi Gerganovsycl : sync (#0)
2024-06-16 Georgi Gerganovcuda : enable CUDA graphs (#0)
2024-06-16 Georgi Gerganovtalk-llama : sync llama.cpp
2024-06-16 Georgi Gerganovcmake : fix CUDA build (#0)
2024-06-16 Georgi Gerganovsync : ggml
2024-06-16 Hong Bo PENGggml : fix and optimize ppc64le (ggml/849)
2024-06-16 Daniel Beveniusggml : remove duplicate include of ggml-common.h (ggml...
2024-06-16 Meng, Hengyuremove global variables (llama/7710)
2024-06-16 Johannes GäßlerCUDA: faster q2_K, q3_K MMQ + int8 tensor cores (llama...
2024-06-16 Georgi Gerganovmetal : utilize max shared memory for mul_mat_id (llama...
2024-06-16 Radoslav Gerganovrpc : fix ggml_backend_rpc_supports_buft() (llama/7918)
2024-06-16 slarenmove BLAS to a separate backend (llama/6210)
2024-06-16 Johannes GäßlerCUDA: fix broken oob check for FA vec f32 kernel (llama...
2024-06-16 Georgi Gerganovtests : add non-cont unary tests (llama/7857)
2024-06-16 Georgi Gerganovggml : improve ggml_is_contiguous logic (llama/7856)
2024-06-16 k.h.laivulkan: select only one device for single gpu with...
2024-06-16 0cc4mUpdate Vulkan RoPE implementation (llama/7818)
2024-06-16 Johannes GäßlerCUDA: int8 tensor cores for MMQ (q4_K, q5_K, q6_K)...
2024-06-16 Johannes GäßlerCUDA: use tensor cores for MMQ (llama/7676)
2024-06-16 Ben Ashbaughuse the correct SYCL context for host USM allocations...
2024-06-16 Johannes GäßlerCUDA: revise q8_1 data layout for mul_mat_q (llama...
next