]> git.djapps.eu Git - pkg/ggml/sources/ggml/commitdiff
readme : fix gpt-neox docs (#410)
authorTom Bailey <redacted>
Sun, 23 Jul 2023 14:51:13 +0000 (23:51 +0900)
committerGitHub <redacted>
Sun, 23 Jul 2023 14:51:13 +0000 (17:51 +0300)
* Fix gpt neox example directory reference

* Fix gpt neox bin directory references

examples/gpt-neox/README.md

index f7a862ff7d576ca0892192e06acf7b3483c20d5a..64c6d7c623f5d9f0708c655163f26a04647eca97 100644 (file)
@@ -21,10 +21,10 @@ git clone https://huggingface.co/stabilityai/gpt_neox-base-alpha-3b
 python3 -m pip install -r ../requirements.txt
 
 # convert model to FP16
-python3 ../examples/gpt_neox/convert-h5-to-ggml.py ./stablelm-base-alpha-3b/ 1
+python3 ../examples/gpt-neox/convert-h5-to-ggml.py ./stablelm-base-alpha-3b/ 1
 
 # run inference using FP16 precision
-make -j && ./bin/gpt_neox -m ./stablelm-base-alpha-3b/ggml-model-f16.bin -p "I believe the meaning of life is" -t 8 -n 64
+make -j && ./bin/gpt-neox -m ./stablelm-base-alpha-3b/ggml-model-f16.bin -p "I believe the meaning of life is" -t 8 -n 64
 
 main: seed = 1681940611
 gpt_neox_model_load: loading model from 'models/stablelm-base-alpha-3b/ggml-model-f16.bin' - please wait ...
@@ -63,10 +63,10 @@ main:    total time =  6911.26 ms
 
 ```bash
 # quantize the model to 5-bits using Q5_0 quantization
-./bin/gpt_neox-quantize ./stablelm-base-alpha-3b/ggml-model-f16.bin ./stablelm-base-alpha-3b/ggml-model-q5_0.bin q5_0
+./bin/gpt-neox-quantize ./stablelm-base-alpha-3b/ggml-model-f16.bin ./stablelm-base-alpha-3b/ggml-model-q5_0.bin q5_0
 
 # run the quantized model
-./bin/gpt_neox -m ./stablelm-base-alpha-3b/ggml-model-q5_0.bin -p "I believe the meaning of life is" -t 8 -n 64
+./bin/gpt-neox -m ./stablelm-base-alpha-3b/ggml-model-q5_0.bin -p "I believe the meaning of life is" -t 8 -n 64
 
 main: seed = 1682021489
 gpt_neox_model_load: loading model from 'models/stablelm-base-alpha-3b/ggml-model-q5_0.bin' - please wait ...