]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/shortlog
pkg/ggml/sources/llama.cpp
2025-01-08 Georgi Gerganovllama : avoid hardcoded QK_K (#11061)
2025-01-08 Georgi Gerganovsync : ggml
2025-01-08 Radoslav Gerganovggml : allow loading backend with env variable (ggml...
2025-01-08 Xuan Son Nguyenci : pin dependency to specific version (#11137)
2025-01-08 Georgi Gerganovarg : option to exclude arguments from specific example...
2025-01-08 amritahs-ibmllamafile : ppc64le MMA INT8 implementation (#10912)
2025-01-08 Georgi Gerganovci : fix cmake option (#11125)
2025-01-08 Mathieu BaudierDisable GL_KHR_cooperative_matrix Vulkan extension...
2025-01-08 ag2s20150909fix: Vulkan shader gen binary path when Cross-compiling...
2025-01-07 Johannes GäßlerGGUF: C++ refactor, backend support, misc fixes (#11030)
2025-01-07 Diego Devesaggml-backend : only offload from host buffers (fix...
2025-01-07 Diego Devesaggml-backend : only offload from host buffers (#11120)
2025-01-07 Radoslav Gerganovrpc : code cleanup (#11107)
2025-01-07 Akarshan BiswasSYCL: Use get_multi_ptr instead of deprecated get_point...
2025-01-06 Eric Curtinllama-run : fix context size (#11094)
2025-01-06 Georgi Gerganovllama : remove unused headers (#11109)
2025-01-06 Xuan Son Nguyengithub : add cmd line field to bug report (#11090)
2025-01-06 Georgi Gerganovserver : fix extra BOS in infill endpoint (#11106)
2025-01-06 Xuan Son Nguyenllama : remove check flash_attn with lora (#11104)
2025-01-06 Asghar Ghorbanillama : prevent system info string accumulation across...
2025-01-06 Daniel Beveniusllama : rename missed batch params/vars to ubatch ...
2025-01-06 Georgi Gerganovllama : update llama_model API names (#11063)
2025-01-06 Georgi Gerganovtokenize : escape the prompt (#11058)
2025-01-06 Georgi Gerganovmmap : fix fileno macro clash (#11076)
2025-01-06 Georgi Gerganovllama : use LLAMA_TOKEN_NULL (#11062)
2025-01-06 Georgi Gerganovllama : use _impl suffix instead of _internal (#11060)
2025-01-06 Johannes GäßlerCUDA: add BF16 support (#11093)
2025-01-04 0cc4mVulkan: Add device-specific blacklist for coopmat for...
2025-01-04 fairydreamingllama : Add support for DeepSeek V3 (#11049)
2025-01-04 matt23654[GGML][RPC] Support for models with non-512-aligned...
2025-01-04 DAN™llama : add support for the cohere2 model architecture...
2025-01-04 Georgi Gerganovsync : ggml
2025-01-04 Georgi Gerganovggml : do not install metal source when embed library...
2025-01-04 Daniel Beveniusggml : improve inputs log sched_print_assignments ...
2025-01-04 Gilad S.fix: Vulkan shader gen binary path (#11037)
2025-01-03 Molly Sophiacommon : disable KV cache shifting automatically for...
2025-01-03 Georgi Gerganovmetal : avoid uint (#11019)
2025-01-03 Georgi Gerganovllama : refactor `src/llama.cpp` (#10902)
2025-01-02 Pierrick Hymbertserver: bench: minor fixes (#10765)
2025-01-02 Xuan Son Nguyenserver : allow using LoRA adapters per-request (#10994)
2025-01-02 Benson Wongreadme : add llama-swap to infrastructure section ...
2024-12-31 Srihari-mcwggml : fixes for AVXVNNI instruction set with MSVC...
2024-12-31 Xuan Son Nguyenserver : clean up built-in template detection (#11026)
2024-12-31 Xuan Son Nguyenserver : add OAI compat for /v1/completions (#10974)
2024-12-31 ymckiconvert : fix Llama-3_1-Nemotron-51B rope settings...
2024-12-31 Petercommon, examples, ggml : fix MSYS2 GCC compiler errors...
2024-12-30 Jeff Bolzvulkan: optimize mul_mat for small values of N (#10991)
2024-12-30 ag2s20150909android : fix llama_batch free (#11014)
2024-12-29 Jeff Bolzvulkan: im2col and matmul optimizations for stable...
2024-12-29 Jeff Bolzvulkan: Use push constant offset to handle misaligned...
2024-12-28 Isaac McFadyenserver: added more docs for response_fields field ...
2024-12-28 Alexey Parfenovserver : fix token duplication when streaming with...
2024-12-26 Evevulkan: multi-row k quants (#10846)
2024-12-26 Peterexamples, ggml : fix GCC compiler warnings (#10983)
2024-12-24 Reza Kakhkiserver : add support for "encoding_format": "base64...
2024-12-24 Djip007ggml : more perfo with llamafile tinyblas on x86_64...
2024-12-24 NeverLuckyserver: allow filtering llama server response fields...
2024-12-24 Georgi Gerganovllama : the WPM vocabs use the CLS token as BOS (#10930)
2024-12-24 Diego Devesaggml : use wstring for backend search paths (#10960)
2024-12-24 Diego Devesaggml : fix arm enabled features check (#10961)
2024-12-23 Diego Devesaggml : fix const usage in SSE path (#10962)
2024-12-23 Xuan Son Nguyenserver : fix missing model id in /model endpoint (...
2024-12-23 Xuan Son Nguyenserver : add system_fingerprint to chat/completion...
2024-12-23 Radoslav Gerganovrpc-server : add support for the SYCL backend (#10934)
2024-12-23 Yun Doullama : support InfiniAI Megrez 3b (#10893)
2024-12-23 ymckillama : support for Llama-3_1-Nemotron-51B (#10669)
2024-12-23 Eric Curtinllama-run : include temperature option (#10899)
2024-12-23 yuri@FreeBSDggml : fix run-time on FreeBSD in get_executable_path...
2024-12-22 Rudi Servodevops : add docker-multi-stage builds (#10832)
2024-12-22 Billel Mokeddemllama : add Falcon3 support (#10883)
2024-12-22 Jeff Bolzvulkan: build fixes for 32b (#10927)
2024-12-21 Georgi Gerganovconvert : add BertForMaskedLM (#10919)
2024-12-21 Jeff Bolzvulkan: optimize coopmat2 dequant functions (#10855)
2024-12-20 Adrien Gallouëtggml-cpu: replace NEON asm with intrinsics in ggml_gemv...
2024-12-20 Akarshan BiswasSYCL: Migrate away from deprecated ggml_tensor->backend...
2024-12-20 Xuan Son Nguyenserver : (UI) fix copy to clipboard function (#10916)
2024-12-20 Diego Devesaggml : add test for SVE and disable when it fails ...
2024-12-20 Molly Sophiaconvert : fix RWKV v6 model conversion (#10913)
2024-12-19 Georgi Gerganovclip : disable GPU support (#10896)
2024-12-19 Georgi Gerganovllama : minor grammar refactor (#10897)
2024-12-19 Georgi Gerganovtts : small QoL for easy model fetch (#10903)
2024-12-19 Xuan Son Nguyenserver : fix logprobs, make it OAI-compatible (#10783)
2024-12-19 Adrien Gallouëtggml: fix arm build with gcc (#10895)
2024-12-19 Sukriti Sharmallama : fix Roberta embeddings (#10856)
2024-12-19 fairydreamingconvert : Add support for Microsoft Phi-4 model (...
2024-12-19 Johannes Gäßlertests: disable GGUF test for bad value size (#10886)
2024-12-19 Eric Curtinllama-run : improve progress bar (#10821)
2024-12-18 Diego Devesaggml : fix arm build (#10890)
2024-12-18 Georgi Gerganovtts : add OuteTTS support (#10784)
2024-12-18 Gaetan Bissonserver: avoid overwriting Authorization header (#10878)
2024-12-18 Georgi Gerganovserver : output embeddings for all tokens when pooling...
2024-12-18 Georgi Gerganovserver : add "tokens" output (#10853)
2024-12-18 Xuan Son Nguyenserver : (embeddings) using same format for "input...
2024-12-18 redbearddocs: Fix HIP (née hipBLAS) in README (#10880)
2024-12-18 Diego DevesaRevert "llama : add Falcon3 support (#10864)" (#10876)
2024-12-17 DAN™Use model->gguf_kv for loading the template instead...
2024-12-17 Johannes Gäßlertests: add tests for GGUF (#10830)
2024-12-17 Georgi Gerganovsync : ggml
2024-12-17 Georgi Gerganovcmake : fix "amd64" processor string (whisper/2638)
2024-12-17 gn64vulkan : fix soft_max.comp division by zero (whisper...
next