From: Ujjawal Panchal Date: Thu, 25 Jul 2024 08:13:27 +0000 (+0530) Subject: docs : Quantum -> Quantized (#8666) X-Git-Tag: upstream/0.0.4488~1030 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=4b0eff3df58d8d86e47348fb73d54da3194d416d;p=pkg%2Fggml%2Fsources%2Fllama.cpp docs : Quantum -> Quantized (#8666) * docfix: imatrix readme, quantum models -> quantized models. * docfix: server readme: quantum models -> quantized models. --- diff --git a/examples/imatrix/README.md b/examples/imatrix/README.md index 29602881..bb5faec9 100644 --- a/examples/imatrix/README.md +++ b/examples/imatrix/README.md @@ -1,6 +1,6 @@ # llama.cpp/examples/imatrix -Compute an importance matrix for a model and given text dataset. Can be used during quantization to enchance the quality of the quantum models. +Compute an importance matrix for a model and given text dataset. Can be used during quantization to enchance the quality of the quantized models. More information is available here: https://github.com/ggerganov/llama.cpp/pull/4861 ## Usage diff --git a/examples/server/README.md b/examples/server/README.md index ff407451..33a2b95c 100644 --- a/examples/server/README.md +++ b/examples/server/README.md @@ -5,7 +5,7 @@ Fast, lightweight, pure C/C++ HTTP server based on [httplib](https://github.com/ Set of LLM REST APIs and a simple web front end to interact with llama.cpp. **Features:** - * LLM inference of F16 and quantum models on GPU and CPU + * LLM inference of F16 and quantized models on GPU and CPU * [OpenAI API](https://github.com/openai/openai-openapi) compatible chat completions and embeddings routes * Parallel decoding with multi-user support * Continuous batching