From: Sigbjørn Skjæret Date: Thu, 12 Mar 2026 09:04:20 +0000 (+0100) Subject: convert : better mtp check and fix return [no ci] (#20419) X-Git-Tag: upstream/0.0.8611~295 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=c3e3f9e5338e8eb3d5b13d20f20d38b53d9d19ce;p=pkg%2Fggml%2Fsources%2Fllama.cpp convert : better mtp check and fix return [no ci] (#20419) --- diff --git a/convert_hf_to_gguf.py b/convert_hf_to_gguf.py index 6e3d3ad1d..37834c78b 100755 --- a/convert_hf_to_gguf.py +++ b/convert_hf_to_gguf.py @@ -10092,9 +10092,9 @@ class NemotronHModel(GraniteHybridModel): # Skip Multi-Token Prediction (MTP) tensors. These are used for # for speculative decoding but we don't include them in this model # conversion. See https://github.com/ggml-org/llama.cpp/pull/18886 - if "mtp" in name: + if name.startswith("mtp."): logger.info(f"gguf: Skipping MTP (Speculative) layer: {name}") - return [] + return if name.endswith("mixer.gate.e_score_correction_bias"): new_name = name.replace("e_score_correction_bias", "e_score_correction.bias")