]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/shortlog
pkg/ggml/sources/llama.cpp
2025-01-15 Junil Kimfix: ggml: fix vulkan-shaders-gen build (#10448) upstream/0.0.4488
2025-01-15 Johannes GäßlerRoPE: fix back, CUDA support for back + noncont. (...
2025-01-15 Daniel Beveniusexamples : add embd_to_audio to tts-outetts.py [no...
2025-01-15 Akarshan BiswasSYCL: Add gated linear attention kernel (#11175)
2025-01-14 Xuan Son Nguyenci : add -no-cnv for tests (#11238)
2025-01-14 Georgi Gerganovvocab : add dummy tokens for "no_vocab" type (#11231)
2025-01-14 ebraminioserver : Improve code snippets direction between RTL...
2025-01-14 Olivier ChafikRefactor test-chat-template.cpp (#11224)
2025-01-14 Georgi Gerganovsync : ggml
2025-01-14 Georgi Gerganovscripts : sync gguf (cont)
2025-01-14 Georgi Gerganovscripts : sync gguf
2025-01-14 Georgi Gerganovscripts : sync opencl
2025-01-13 ebraminioserver : (UI) Improve messages bubble shape in RTL...
2025-01-13 Xuan Son Nguyencli : auto activate conversation mode if chat template...
2025-01-13 Andreas Kieslingercuda : CUDA Graph Compute Function Refactor (precursor...
2025-01-13 Georgi Gerganovcontrib : add naming guidelines (cont) (#11177)
2025-01-13 ebraminioserver : (UI) Support for RTL text as models input...
2025-01-13 Georgi Gerganovcontrib : add naming guidelines (cont) (#11177)
2025-01-13 Xuan Son Nguyencommon : support tag-based --hf-repo like on ollama...
2025-01-13 Georgi Gerganovcontrib : add naming guidelines (#11177)
2025-01-13 Daniel Beveniusllama : remove 'd' from bad special token log (#11212)
2025-01-13 Radoslav Gerganovggml : do not define GGML_USE_CUDA when building with...
2025-01-12 Eric CurtinReset color before we exit (#11205)
2025-01-12 Xuan Son Nguyenllama : fix chat template gguf key (#11201)
2025-01-12 Georgi Gerganovllama : remove notion of CLS token (#11064)
2025-01-12 Georgi Gerganovllama : add `llama_vocab`, functions -> methods, naming...
2025-01-11 Vinesh Janarthanangguf-py: fixed local detection of gguf package (#11180)
2025-01-11 Daniel Beveniusconvert : sort print supported models [no ci] (#11179)
2025-01-10 Daniel Beveniusexamples : add README.md to tts example [no ci] (#11155)
2025-01-10 Daniel Beveniusconvert : add --print-supported-models option (#11172)
2025-01-10 0cc4mVulkan: Fix float16 use on devices without float16...
2025-01-10 Molly Sophiallama: add support for QRWKV6 model architecture (...
2025-01-10 Akarshan BiswasSYCL: Refactor ggml_sycl_compute_forward (#11121)
2025-01-09 Tei Homedoc: add cuda guide for fedora (#11135)
2025-01-09 Daniel Beveniusserver : add tooltips to settings and themes btn (...
2025-01-09 Pierrick Hymbertmodel: Add support for PhiMoE arch (#11003)
2025-01-09 Georgi Gerganovmedia : remove old img [no ci]
2025-01-09 Xuan Son Nguyenllama-chat : add phi 4 template (#11148)
2025-01-08 hydaifix: add missing msg in static_assert (#11143)
2025-01-08 Vinesh Janarthanangguf-py : move scripts directory (#11116)
2025-01-08 Eric CurtinEnhance user input handling for llama-run (#11138)
2025-01-08 Xuan Son Nguyenci : use actions from ggml-org (#11140)
2025-01-08 Xuan Son Nguyenlora : improve compat with `mergekit-extract-lora`...
2025-01-08 Georgi Gerganovllama : avoid hardcoded QK_K (#11061)
2025-01-08 Georgi Gerganovsync : ggml
2025-01-08 Radoslav Gerganovggml : allow loading backend with env variable (ggml...
2025-01-08 Xuan Son Nguyenci : pin dependency to specific version (#11137)
2025-01-08 Georgi Gerganovarg : option to exclude arguments from specific example...
2025-01-08 amritahs-ibmllamafile : ppc64le MMA INT8 implementation (#10912)
2025-01-08 Georgi Gerganovci : fix cmake option (#11125)
2025-01-08 Mathieu BaudierDisable GL_KHR_cooperative_matrix Vulkan extension...
2025-01-08 ag2s20150909fix: Vulkan shader gen binary path when Cross-compiling...
2025-01-07 Johannes GäßlerGGUF: C++ refactor, backend support, misc fixes (#11030)
2025-01-07 Diego Devesaggml-backend : only offload from host buffers (fix...
2025-01-07 Diego Devesaggml-backend : only offload from host buffers (#11120)
2025-01-07 Radoslav Gerganovrpc : code cleanup (#11107)
2025-01-07 Akarshan BiswasSYCL: Use get_multi_ptr instead of deprecated get_point...
2025-01-06 Eric Curtinllama-run : fix context size (#11094)
2025-01-06 Georgi Gerganovllama : remove unused headers (#11109)
2025-01-06 Xuan Son Nguyengithub : add cmd line field to bug report (#11090)
2025-01-06 Georgi Gerganovserver : fix extra BOS in infill endpoint (#11106)
2025-01-06 Xuan Son Nguyenllama : remove check flash_attn with lora (#11104)
2025-01-06 Asghar Ghorbanillama : prevent system info string accumulation across...
2025-01-06 Daniel Beveniusllama : rename missed batch params/vars to ubatch ...
2025-01-06 Georgi Gerganovllama : update llama_model API names (#11063)
2025-01-06 Georgi Gerganovtokenize : escape the prompt (#11058)
2025-01-06 Georgi Gerganovmmap : fix fileno macro clash (#11076)
2025-01-06 Georgi Gerganovllama : use LLAMA_TOKEN_NULL (#11062)
2025-01-06 Georgi Gerganovllama : use _impl suffix instead of _internal (#11060)
2025-01-06 Johannes GäßlerCUDA: add BF16 support (#11093)
2025-01-04 0cc4mVulkan: Add device-specific blacklist for coopmat for...
2025-01-04 fairydreamingllama : Add support for DeepSeek V3 (#11049)
2025-01-04 matt23654[GGML][RPC] Support for models with non-512-aligned...
2025-01-04 DAN™llama : add support for the cohere2 model architecture...
2025-01-04 Georgi Gerganovsync : ggml
2025-01-04 Georgi Gerganovggml : do not install metal source when embed library...
2025-01-04 Daniel Beveniusggml : improve inputs log sched_print_assignments ...
2025-01-04 Gilad S.fix: Vulkan shader gen binary path (#11037)
2025-01-03 Molly Sophiacommon : disable KV cache shifting automatically for...
2025-01-03 Georgi Gerganovmetal : avoid uint (#11019)
2025-01-03 Georgi Gerganovllama : refactor `src/llama.cpp` (#10902)
2025-01-02 Pierrick Hymbertserver: bench: minor fixes (#10765)
2025-01-02 Xuan Son Nguyenserver : allow using LoRA adapters per-request (#10994)
2025-01-02 Benson Wongreadme : add llama-swap to infrastructure section ...
2024-12-31 Srihari-mcwggml : fixes for AVXVNNI instruction set with MSVC...
2024-12-31 Xuan Son Nguyenserver : clean up built-in template detection (#11026)
2024-12-31 Xuan Son Nguyenserver : add OAI compat for /v1/completions (#10974)
2024-12-31 ymckiconvert : fix Llama-3_1-Nemotron-51B rope settings...
2024-12-31 Petercommon, examples, ggml : fix MSYS2 GCC compiler errors...
2024-12-30 Jeff Bolzvulkan: optimize mul_mat for small values of N (#10991)
2024-12-30 ag2s20150909android : fix llama_batch free (#11014)
2024-12-29 Jeff Bolzvulkan: im2col and matmul optimizations for stable...
2024-12-29 Jeff Bolzvulkan: Use push constant offset to handle misaligned...
2024-12-28 Isaac McFadyenserver: added more docs for response_fields field ...
2024-12-28 Alexey Parfenovserver : fix token duplication when streaming with...
2024-12-26 Evevulkan: multi-row k quants (#10846)
2024-12-26 Peterexamples, ggml : fix GCC compiler warnings (#10983)
2024-12-24 Reza Kakhkiserver : add support for "encoding_format": "base64...
2024-12-24 Djip007ggml : more perfo with llamafile tinyblas on x86_64...
2024-12-24 NeverLuckyserver: allow filtering llama server response fields...
next