]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/commitdiff
readme : update hot topics
authorGeorgi Gerganov <redacted>
Tue, 7 May 2024 18:43:13 +0000 (21:43 +0300)
committerGitHub <redacted>
Tue, 7 May 2024 18:43:13 +0000 (21:43 +0300)
README.md

index 885322e68dcd6b3dbadf7dcc5e36244ca29fa17a..75fc10a15abde45373f76899def92fc21735f127 100644 (file)
--- a/README.md
+++ b/README.md
@@ -20,7 +20,8 @@ Inference of Meta's [LLaMA](https://arxiv.org/abs/2302.13971) model (and others)
 
 ### Hot topics
 
-- **BPE pre-tokenization support has been added: https://github.com/ggerganov/llama.cpp/pull/6920**
+- **Initial Flash-Attention support: https://github.com/ggerganov/llama.cpp/pull/5021**
+- BPE pre-tokenization support has been added: https://github.com/ggerganov/llama.cpp/pull/6920
 - MoE memory layout has been updated - reconvert models for `mmap` support and regenerate `imatrix` https://github.com/ggerganov/llama.cpp/pull/6387
 - Model sharding instructions using `gguf-split` https://github.com/ggerganov/llama.cpp/discussions/6404
 - Fix major bug in Metal batched inference https://github.com/ggerganov/llama.cpp/pull/6225