]> git.djapps.eu Git - pkg/ggml/sources/ggml/shortlog
pkg/ggml/sources/ggml
2024-03-27 Georgi Gerganovsync : whisper.cpp
2024-03-27 Georgi Gerganovwhisper : improve handling of prompts (whisper/1981)
2024-03-27 Sanchit Gandhiwhisper : improve support for distil-large-v3 (whisper...
2024-03-27 denerscwhisper : token-level timestamps with DTW (whisper...
2024-03-27 Jo Lissexamples : rename --audio-context to --audio-ctx per...
2024-03-27 Georgi Gerganovexamples : more CUDA leftovers (#0)
2024-03-27 Georgi Gerganovexamples : fix CUBLAS leftovers (#0)
2024-03-27 Georgi Gerganovsync : adapt to CUDA changes (#0)
2024-03-27 Georgi Gerganovsync : llama.cpp
2024-03-27 AidanBeltonSFix batched impl for NVidia GPU (llama/6164)
2024-03-27 KawrakowMake IQ1_M work for QK_K = 64 (llama/6327)
2024-03-27 compiladellama : greatly reduce output buffer memory usage ...
2024-03-27 KawrakowIQ1_M: 1.75 bpw quantization (llama/6302)
2024-03-27 slarencuda : rename build flag to LLAMA_CUDA (llama/6299)
2024-03-27 Kawrakowtests : include IQ2_XXS and IQ2_XS in test-quantize...
2024-03-27 slarencuda : refactor into multiple files (llama/6269)
2024-03-27 Justine Tunneyggml : support AVX512VNNI (llama/6280)
2024-03-27 Rick GFix heap corruption from wmode out-of-bound writes...
2024-03-27 Meng, Hengyuoffload op (llama/6217)
2024-03-27 Jared Van Borteluse _wfopen instead of fopen on Windows (llama/6248)
2024-03-27 slarencuda : add LLAMA_CUDA_NO_PEER_COPY to workaround broken...
2024-03-27 Georgi Gerganovmetal : proper assert for mat-mat memory alignment...
2024-03-27 Georgi Gerganovmetal : pad n_ctx by 32 (llama/6177)
2024-03-27 slarencuda : disable host register by default (llama/6206)
2024-03-27 slarencuda : fix LLAMA_CUDA_F16 build (llama/6197)
2024-03-27 Kawrakowggml : same IQ4_NL quantization for CPU/CUDA/Metal...
2024-03-27 KawrakowAdd ability to use Q5_0, Q5_1, and IQ4_NL for quantized...
2024-03-27 AidanBeltonSAdd nvidia and amd backends (llama/6157)
2024-03-27 slarencuda : fix conflict with std::swap (llama/6186)
2024-03-27 slarencuda : print the returned error when CUDA initializatio...
2024-03-27 slarencuda : refactor to remove global resources (llama/6170)
2024-03-27 Abhilash Majumderincrease igpu cluster limit (llama/6159)
2024-03-27 slarenbackend : set max split inputs to GGML_MAX_SRC (llama...
2024-03-27 slarenbackend : offload large batches to GPU (llama/6083)
2024-03-27 GainLeeggml:fix finding transfer queue family index error...
2024-03-27 AmirAli Mirianggml : add AVX512F SIMD (llama/6088)
2024-03-27 slarencuda : disable unused cudaLaunchHostFunc code (llama...
2024-03-27 Neo Zhang Jianyufix set main gpu error (llama/6073)
2024-03-27 AidanBeltonSFix non-intel device selection (llama/6042)
2024-03-27 Ondřej Čertíkgguf : add support for I64 and F64 arrays (llama/6062)
2024-03-22 Bryan Lozanocmake : bump to 3.12 (#768)
2024-03-22 Bryan Lozanoexamples : fix simple (#770)
2024-03-18 Georgi Gerganovsync : whisper.cpp
2024-03-18 Georgi Gerganovwhisper : set outputs from conv graph (whisper/1959)
2024-03-18 slarenalloc : fix allocation data of pre-allocated leafs
2024-03-15 Georgi Gerganovsync : whisper.cpp
2024-03-15 Georgi Gerganovspec : add GGUF diagram (#765)
2024-03-14 slarenupdate examples and tests
2024-03-14 Georgi Gerganovggml : add ggml-common.h
2024-03-14 Georgi Gerganovsync : llama.cpp
2024-03-14 Georgi Gerganovggml : designate enum vals for integer types (llama...
2024-03-14 Georgi Gerganovmetal : build metallib + fix embed path (llama/6015)
2024-03-14 slarenllama : add pipeline parallelism support (llama/6017)
2024-03-14 slarentest-backend-ops : skip CPU backend by default (llama...
2024-03-14 AidanBeltonSUpdate get version (llama/6025)
2024-03-14 Georgi Gerganovggml : reuse quantum structs across backends (llama...
2024-03-14 Georgi Gerganovggml : fix UB in IQ2_S and IQ3_S (llama/6012)
2024-03-14 Georgi Gerganovsycl : update IQ1_S kernels (WIP - not working!) (llama...
2024-03-14 Kawrakow1.5 bit: we can do even better (llama/5999)
2024-03-14 Michael Podvitskiyggml, ci : Windows ARM runner and build fixes (llama...
2024-03-14 KawrakowBetter 1.5 bit quantization (llama/5971)
2024-03-14 Abhilash MajumderAdd q3_s and q1_s (llama/5886)
2024-03-14 Georgi Gerganovmetal : move mm_id indices to shared mem (llama/5982)
2024-03-14 Georgi Gerganovggml : remove __constant__ specifier for CUDA tables...
2024-03-14 Georgi Gerganovggml : fix unnecessary f32 -> f16 -> f32 casts (mmla...
2024-03-14 Georgi Gerganovggml : remove old quantization functions (llama/5942)
2024-03-14 Georgi Gerganovggml : add ggml-common.h to deduplicate shared code...
2024-03-14 compiladellama : support Mamba Selective State Space Models...
2024-03-13 compiladegguf : add Mamba keys and tensors (#763)
2024-03-10 Georgi Gerganovsync : whisper.cpp
2024-03-10 Josh Bleecher... whisper : document whisper_batch.n_seq_id (whisper...
2024-03-10 Josh Bleecher... whisper : improve beam search candidate diversity ...
2024-03-10 Josh Bleecher... whisper : make beam candidate sort more stable (whisper...
2024-03-10 Georgi Gerganovggml : try fix 32-bit arm compat (whisper/1938)
2024-03-08 Georgi Gerganovsync : whisper.cpp
2024-03-08 F1L1Pexamples : Auto lowercase language parameter in main...
2024-03-08 zhouwgwhisper : fix typo (whisper/1925)
2024-03-08 Georgi Gerganovsync : llama.cpp
2024-03-08 Neo Zhang JianyuRevert "[SYCL] fix error when set main gpu to non-zero...
2024-03-08 Neo Zhang Jianyufix error when set main gpu to non-zero (llama/5901)
2024-03-08 Jared Van Bortelggml : use SYS_get_cpu if SYS_getcpu is not defined...
2024-03-08 bobqianicggml : use `uint8x16_t` return type for `ggml_vqtbl1q_u...
2024-03-08 Neo Zhang Jianyuadd wait() to make code stable (llama/5895)
2024-03-08 Jared Van Bortelquants : use MM256_SET_M128I consistently to fix gcc...
2024-03-08 0cc4mVulkan Improvements (llama/5835)
2024-03-08 Neo Zhang Jianyufix mul_mat fault in CI/unit-test (llama/5862)
2024-03-08 Georgi Gerganovggml : fix unknown status (llama/0)
2024-03-07 slarenmagika : fix ggml_backend_graph_compute status check...
2024-03-05 Georgi Gerganovwhisper : fix compute helper return (#750)
2024-03-04 Michael Podvitskiyggml : introduce ggml_status (#750)
2024-03-04 Georgi Gerganovsync : llama.cpp
2024-03-04 slarencuda : fix data race in soft max (llama/5853)
2024-03-04 Georgi Gerganovggml : fix IQ3_S AVX implementation (llama/5834)
2024-03-04 Kawrakowggml : IQ3_S improvements (llama/5829)
2024-03-04 Neo Zhang JianyuSupport multiple GPUs (split mode) on SYCL backend...
2024-03-04 ddpasaggml-vulkan: fix VULKAN_CHECK_RESULTS flag, which was...
2024-03-04 AidanBeltonSUse batched mul_mat pathway (llama/5591)
2024-03-04 Evemake portability_enumeration_ext apple only (llama...
2024-03-03 leejetadd some new ops, fix some operators and add batch...
2024-02-28 Steward Garciaggml : add simple example (#713)
next