]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/commitdiff
Update README.md
authorGeorgi Gerganov <redacted>
Fri, 10 Mar 2023 23:18:10 +0000 (01:18 +0200)
committerGitHub <redacted>
Fri, 10 Mar 2023 23:18:10 +0000 (01:18 +0200)
README.md

index e16dd0740c151c34509b740dbe3e5a0a4c93d030..e7e7cb2a9edce0d93e9343577639b363c926dee2 100644 (file)
--- a/README.md
+++ b/README.md
@@ -139,5 +139,5 @@ python3 convert-pth-to-ggml.py models/7B/ 1
   In general, it seems to work, but I think it fails for unicode character support. Hopefully, someone can help with that
 - I don't know yet how much the quantization affects the quality of the generated text
 - Probably the token sampling can be improved
-- x86 quantization support [not yet ready](https://github.com/ggerganov/ggml/pull/27). Basically, you want to run this on Apple Silicon
+- x86 quantization support [not yet ready](https://github.com/ggerganov/ggml/pull/27). Basically, you want to run this on Apple Silicon. For now, on Linux and Windows you can use the F16 `ggml-model-f16.bin` model, but it will be much slower.