]> git.djapps.eu Git - pkg/ggml/sources/ggml/shortlog
pkg/ggml/sources/ggml
2024-02-10 Georgi Gerganovmetal : add im2col F32 dst support (llama/5132)
2024-02-10 JidongZhang-THUllava : add MobileVLM support (llama/5132)
2024-02-10 Neo Zhang Jianyuformat license text, restore apache license by legal...
2024-02-10 slarenggml : limit n_threads to the max n_tasks (llama/5238)
2024-02-10 0cc4mVulkan Fixes (llama/5223)
2024-02-10 Jared Van Bortelkompute : llama-bench support and ggml_cpu_has_kompute...
2024-02-09 Michael Podvitskiyggml : add abort_callback for cpu backend (#725)
2024-01-30 Georgi Gerganovsync : whisper.cpp
2024-01-30 JacobLinCoolcommon : fix wav buffer detection (whisper/1819)
2024-01-30 Georgi Gerganovsync : llama.cpp
2024-01-30 Kawrakowggml : fix IQ3_XXS on Metal (llama/5219)
2024-01-30 Georgi Gerganovsync : ggml (llama/0)
2024-01-30 KawrakowFaster AVX2 dot product for IQ2_XS (llama/5187)
2024-01-30 KawrakowSOTA 3-bit quants (llama/5196)
2024-01-30 0cc4mVulkan Windows APU Memory Handling (llama/5199)
2024-01-30 Paul Tsochantarisggml alloc: Fix for null dereference on alloc failure...
2024-01-30 Jared Van BortelNomic Vulkan backend (llama/4456)
2024-01-30 slarenggml : add max buffer sizes to opencl and metal backend...
2024-01-30 Paul Tsochantarismetal : free metal objects (llama/5161)
2024-01-29 Georgi Gerganovgguf : fix comparison (#715)
2024-01-29 John Balis`ggml_cuda_cpy` support for 4d tensors and float16...
2024-01-29 Georgi Gerganovgguf : add input validation, prevent integer overflows...
2024-01-29 Georgi Gerganovci : fix yolo URLs + fix metal capture (#712)
2024-01-29 Jack Mousseaumetal : add debug capture backend function (#694)
2024-01-28 Georgi Gerganovsync : llama.cpp
2024-01-28 Georgi Gerganovsync : whisper.cpp
2024-01-28 Georgi Gerganovsync : llama.cpp
2024-01-28 0cc4mggml : add Vulkan backend (llama/2059)
2024-01-28 Abhilash Majumderggml : add unified SYCL backend for Intel GPUs (llama...
2024-01-28 Georgi Gerganovggml : minor type fix (int64_t -> size_t)
2024-01-27 Georgi Gerganovsync : whisper.cpp
2024-01-27 Georgi Gerganovcommon : fix input buffer check (whisper/1812)
2024-01-27 Ryan Hitchmanserver : implement "verbose_json" format with token...
2024-01-27 Georgi Gerganovsync : llama.cpp
2024-01-27 Michael KlimenkoRemove unused data and add fixes (llama/5154)
2024-01-27 0cc4mAdd OpenCL add kernel (llama/5151)
2024-01-27 slarencuda : fix tensor size calculation for non-split buffer...
2024-01-27 slarenggml-alloc : add 10% margin to the buffer sizes (llama...
2024-01-27 snadampalggml : update softmax n_task calculation (llama/5126)
2024-01-27 Paul Tsochantarismetal : remove unused `n_buffers` and `buffers` (llama...
2024-01-27 Georgi Gerganovmetal : show compile log messages
2024-01-27 Engininja2cuda : fix 2-bit quants on amd hip (llama/5105)
2024-01-27 slarenllama : pre-allocate input tensors in a separate buffer...
2024-01-27 Georgi Gerganovmetal : disable support for MUL_MAT F32 x F16
2024-01-27 Johannes GäßlerCUDA: more info when no device code (llama/5088)
2024-01-27 Georgi Gerganovminor : clean-up some warnings and style (llama/5094)
2024-01-27 Reinforce-IIggml : parallelize FP32 conversion when using BLAS...
2024-01-27 XiaotaoChenllava : MobileVLM support (llama/4954)
2024-01-27 slarenllama : run all KQV ops on the CPU with no KV offload...
2024-01-27 Kylincuda : fix compile error in jetson platform (llama...
2024-01-26 Neuman Vonggpt-2 : clarify instructions for CLBlast on Android...
2024-01-26 Juddggml : check ggml_add src1 type (#708)
2024-01-22 Jack Vialmnist : add tensorflow and keras to requirements.txt...
2024-01-18 Georgi Gerganovsync : whisper.cpp
2024-01-18 Georgi Gerganovsync : llama.cpp
2024-01-18 Paul Tsochantarismetal : fix memory leak, dangling pointer and unused...
2024-01-18 Georgi Gerganovggml : fix SPM package headers
2024-01-17 Juddreadme : add link (#699)
2024-01-17 Georgi Gerganovsync : llama.cpp
2024-01-17 Georgi Gerganovmetal : update ggml-metal.m from llama.cpp
2024-01-17 Georgi Gerganovggml : add IQ2 to test-backend-ops + refactoring (llama...
2024-01-17 Georgi Gerganovimatrix : offload to GPU support (llama/4957)
2024-01-17 Georgi Gerganovbackend : add eval callback (llama/4935)
2024-01-17 Georgi Gerganovmetal : create autorelease pool during library build...
2024-01-17 Kawrakowggml : importance matrix support for legacy quants...
2024-01-17 Alex Azarovmetal : log `recommendedMaxWorkingSetSize` on iOS 16...
2024-01-17 Justine Tunneyggml : introduce GGML_CALL function annotation (llama...
2024-01-17 Georgi Gerganovcuda : fix dequantize kernel names (llama/4938)
2024-01-17 KawrakowCUDA: faster dequantize kernels for Q4_0 and Q4_1 ...
2024-01-17 KawrakowAdd ability to use importance matrix for all k-quants...
2024-01-14 Georgi Gerganovsync : whisper.cpp
2024-01-14 Georgi Gerganovsync : llama.cpp
2024-01-14 Alex Azarovmetal : correctly set SIMD support flags on iOS (llama...
2024-01-14 Kawrakow2-bit quantizations (llama/4897)
2024-01-13 Georgi Gerganovsync : whisper.cpp
2024-01-13 Georgi Gerganovwhisper : load the model into multiple buffers of max...
2024-01-13 Georgi Gerganovsync : llama.cpp
2024-01-13 Georgi Gerganovexamples : adapt to metal API
2024-01-13 Johannes Gäßlerggml: cache sin/cos for RoPE (llama/4908)
2024-01-13 Georgi Gerganovmetal : remove old API (llama/4919)
2024-01-13 Georgi Gerganovmetal : disable log for loaded kernels (llama/4794)
2024-01-13 texmex76gguf : fix potential infinite for-loop (llama/4600)
2024-01-13 Georgi Gerganovmetal : refactor kernel loading code (llama/4794)
2024-01-13 Johannes GäßlerCUDA: faster q8_0 -> f16 dequantization (llama/4895)
2024-01-12 Georgi Gerganovsync : whisper.cpp
2024-01-12 Georgi Gerganovggml : fix 32-bit ARM compat for IQ2_XS (whisper/1758)
2024-01-12 Georgi Gerganovwhisper : fix segment length with params.no_timestamps...
2024-01-12 George Hindleparams : don't compute timestamps when not printing...
2024-01-12 slarenbackend_sched : fix assignments
2024-01-12 Georgi Gerganovexamples : remove obsolete starcoder mmap example
2024-01-12 slarenupdate gpt-2 example
2024-01-12 Georgi Gerganovsync : llama.cpp
2024-01-12 slarenllama : ggml-backend integration (llama/4766)
2024-01-12 Johannes GäßlerCUDA: fix softmax compile for old CUDA versions (llama...
2024-01-12 KawrakowImportance Matrix calculation (llama/4861)
2024-01-11 Georgi Gerganovsync : whisper.cpp
2024-01-11 Georgi Gerganovsync : ggml
2024-01-11 Georgi Gerganovmain : add cli option to disable system prints (whisper...
2024-01-11 Georgi Gerganovsync : llama.cpp
2024-01-11 Kawrakowggml : SOTA 2-bit quants (add IQ2_XS) (llama/4856)
next