From: Georgi Gerganov Date: Sun, 17 Dec 2023 18:16:23 +0000 (+0200) Subject: readme : update hot topics X-Git-Tag: upstream/0.0.4488~2832 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=b1306c439490c7fa4ec33594500d980d1e9e15e6;p=pkg%2Fggml%2Fsources%2Fllama.cpp readme : update hot topics --- diff --git a/README.md b/README.md index edbe6ba5..01aef2af 100644 --- a/README.md +++ b/README.md @@ -10,11 +10,11 @@ Inference of [LLaMA](https://arxiv.org/abs/2302.13971) model in pure C/C++ ### Hot topics +- Collecting Apple Silicon performance stats: + - M-series: https://github.com/ggerganov/llama.cpp/discussions/4167 + - A-series: https://github.com/ggerganov/llama.cpp/discussions/4508 - Added Mixtral support: https://github.com/ggerganov/llama.cpp/pull/4406 -- **llama.h API change for handling KV cache offloading and data type: https://github.com/ggerganov/llama.cpp/pull/4309** -- Using `llama.cpp` with AWS instances: https://github.com/ggerganov/llama.cpp/discussions/4225 - Looking for contributions to improve and maintain the `server` example: https://github.com/ggerganov/llama.cpp/issues/4216 -- Collecting Apple Silicon performance stats: https://github.com/ggerganov/llama.cpp/discussions/4167 ----