From: Georgi Gerganov Date: Wed, 20 Sep 2023 17:48:22 +0000 (+0300) Subject: readme : update hot topics X-Git-Tag: gguf-v0.4.0~44 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=7eb41179edc56083ef4eb2df7967ac9ff38b34fb;p=pkg%2Fggml%2Fsources%2Fllama.cpp readme : update hot topics --- diff --git a/README.md b/README.md index d8fd8bc4..670e2e67 100644 --- a/README.md +++ b/README.md @@ -11,6 +11,8 @@ Inference of [LLaMA](https://arxiv.org/abs/2302.13971) model in pure C/C++ ### Hot topics +- Parallel decoding + continuous batching support incoming: [#3228](https://github.com/ggerganov/llama.cpp/pull/3228) \ + **Devs should become familiar with the new API** - Local Falcon 180B inference on Mac Studio https://github.com/ggerganov/llama.cpp/assets/1991296/98abd4e8-7077-464c-ae89-aebabca7757e