]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/commitdiff
convert : Qwerky : use lora_rank_tokenshift and lora_rank_decay if present (#12667)
authorSigbjørn Skjæret <redacted>
Mon, 31 Mar 2025 14:36:25 +0000 (16:36 +0200)
committerGitHub <redacted>
Mon, 31 Mar 2025 14:36:25 +0000 (16:36 +0200)
convert_hf_to_gguf.py

index 0919cd3f0d9a857ddb5d6a262b804fc5f826d2e3..37d6cc2627c134b46f737fd7faa5767b6f775fbe 100755 (executable)
@@ -3557,8 +3557,8 @@ class RWKV6Qwen2Model(Rwkv6Model):
         head_size = hidden_size // num_attention_heads
         rms_norm_eps = self.hparams["rms_norm_eps"]
         intermediate_size = self.hparams["intermediate_size"]
-        time_mix_extra_dim = 64 if hidden_size >= 4096 else 32
-        time_decay_extra_dim = 128 if hidden_size >= 4096 else 64
+        time_mix_extra_dim = self.hparams.get("lora_rank_tokenshift", 64 if hidden_size >= 4096 else 32)
+        time_decay_extra_dim = self.hparams.get("lora_rank_decay", 128 if hidden_size >= 4096 else 64)
 
         # RWKV isn't context limited
         self.gguf_writer.add_context_length(1048576)