From: CrispStrobe Date: Mon, 14 Oct 2024 07:46:33 +0000 (+0200) Subject: convert : handle max_target_positions (#2477) X-Git-Tag: upstream/1.7.4~327 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=06a1da9daff94c1bf1b1d38950628264fe443f76;p=pkg%2Fggml%2Fsources%2Fwhisper.cpp convert : handle max_target_positions (#2477) as needed eg for https://huggingface.co/primeline/whisper-large-v3-turbo-german/blob/main/config.json --- diff --git a/models/convert-h5-to-ggml.py b/models/convert-h5-to-ggml.py index 50836a21..5474d586 100644 --- a/models/convert-h5-to-ggml.py +++ b/models/convert-h5-to-ggml.py @@ -82,7 +82,11 @@ dir_out = Path(sys.argv[3]) encoder = json.load((dir_model / "vocab.json").open("r", encoding="utf8")) encoder_added = json.load((dir_model / "added_tokens.json").open( "r", encoding="utf8")) -hparams = json.load((dir_model / "config.json").open("r", encoding="utf8") ) +hparams = json.load((dir_model / "config.json").open("r", encoding="utf8")) + +# Add this block to handle missing 'max_length' +if "max_length" not in hparams: + hparams["max_length"] = hparams.get("max_target_positions", 448) model = WhisperForConditionalGeneration.from_pretrained(dir_model)