From: Aisuko Date: Fri, 16 Aug 2024 09:08:59 +0000 (+1000) Subject: Fix inference example lacks required parameters (#9035) X-Git-Tag: upstream/0.0.4488~891 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=c8ddce85606d9fb6e30745b6e4fe103eecadc73f;p=pkg%2Fggml%2Fsources%2Fllama.cpp Fix inference example lacks required parameters (#9035) Signed-off-by: Aisuko --- diff --git a/examples/quantize/README.md b/examples/quantize/README.md index 553c2701..5d1e11c6 100644 --- a/examples/quantize/README.md +++ b/examples/quantize/README.md @@ -34,7 +34,7 @@ Run the quantized model: ```bash # start inference on a gguf model -./llama-cli -m ./models/mymodel/ggml-model-Q4_K_M.gguf -n 128 +./llama-cli -m ./models/mymodel/ggml-model-Q4_K_M.gguf -cnv -p "You are a helpful assistant" ``` When running the larger models, make sure you have enough disk space to store all the intermediate files.