(time cmake -DCMAKE_BUILD_TYPE=Release ${CMAKE_EXTRA} .. ) 2>&1 | tee -a $OUT/${ci}-cmake.log
(time make -j ) 2>&1 | tee -a $OUT/${ci}-make.log
- python3 ../convert-hf-to-gguf.py ${path_models} --outfile ${path_models}/ggml-model-f16.gguf
+ python3 ../convert_hf_to_gguf.py ${path_models} --outfile ${path_models}/ggml-model-f16.gguf
model_f16="${path_models}/ggml-model-f16.gguf"
model_q8_0="${path_models}/ggml-model-q8_0.gguf"
-r ./requirements/requirements-convert-legacy-llama.txt
--r ./requirements/requirements-convert-hf-to-gguf.txt
--r ./requirements/requirements-convert-hf-to-gguf-update.txt
--r ./requirements/requirements-convert-llama-ggml-to-gguf.txt
+-r ./requirements/requirements-convert_hf_to_gguf.txt
+-r ./requirements/requirements-convert_hf_to_gguf_update.txt
+-r ./requirements/requirements-convert_llama_ggml_to_gguf.txt
+++ /dev/null
--r ./requirements-convert-legacy-llama.txt
-torch~=2.2.1
+++ /dev/null
--r ./requirements-convert-legacy-llama.txt
-torch~=2.2.1
+++ /dev/null
--r ./requirements-convert-legacy-llama.txt
--- /dev/null
+-r ./requirements-convert-legacy-llama.txt
+torch~=2.2.1
--- /dev/null
+-r ./requirements-convert-legacy-llama.txt
+torch~=2.2.1
--- /dev/null
+-r ./requirements-convert-legacy-llama.txt
--- /dev/null
+-r ./requirements-convert.txt
+torch~=2.1.1
+
--- /dev/null
+-r ./requirements-convert.txt
+torch~=2.1.1
+
fi
check_convert_script examples/convert-legacy-llama.py
-for py in convert-*.py; do
+for py in convert_*.py; do
# skip convert-hf-to-gguf-update.py
# TODO: the check is failing for some reason:
# https://github.com/ggerganov/llama.cpp/actions/runs/8875330981/job/24364557177?pr=6920
- [[ $py == convert-hf-to-gguf-update.py ]] && continue
+ [[ $py == convert_hf_to_gguf_update.py ]] && continue
check_convert_script "$py"
done