From: Galunid Date: Fri, 31 May 2024 15:42:33 +0000 (+0200) Subject: convert-hf : Handle NotImplementedError in convert-hf-to-gguf (#7660) X-Git-Tag: upstream/0.0.4488~1427 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=0515ad93f48df63bbff204eddb0cac75e8585c65;p=pkg%2Fggml%2Fsources%2Fllama.cpp convert-hf : Handle NotImplementedError in convert-hf-to-gguf (#7660) --- diff --git a/convert-hf-to-gguf.py b/convert-hf-to-gguf.py index 9f29cda2..ad071b97 100755 --- a/convert-hf-to-gguf.py +++ b/convert-hf-to-gguf.py @@ -2840,7 +2840,12 @@ def main() -> None: hparams = Model.load_hparams(dir_model) with torch.inference_mode(): - model_class = Model.from_model_architecture(hparams["architectures"][0]) + try: + model_class = Model.from_model_architecture(hparams["architectures"][0]) + except NotImplementedError: + logger.error(f"Model {hparams['architectures'][0]} is not supported") + sys.exit(1) + model_instance = model_class(dir_model, ftype_map[args.outtype], fname_out, args.bigendian, args.use_temp_file, args.no_lazy) logger.info("Set model parameters")