]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/commitdiff
Fix aya-23 conversion scripts (#7539)
authorGalunid <redacted>
Sun, 26 May 2024 14:02:34 +0000 (16:02 +0200)
committerGitHub <redacted>
Sun, 26 May 2024 14:02:34 +0000 (16:02 +0200)
convert-hf-to-gguf.py

index bfccf8623a175637037e3014ca5ca9a829d30628..a342f6b1c1dbac31a5113d0f6720ec6232e4b003 100755 (executable)
@@ -2395,7 +2395,8 @@ class CommandR2Model(Model):
 
         # max_position_embeddings = 8192 in config.json but model was actually
         # trained on 128k context length
-        self.hparams["max_position_embeddings"] = self.hparams["model_max_length"]
+        # aya-23 models don't have model_max_length specified
+        self.hparams["max_position_embeddings"] = self.find_hparam(["model_max_length", "max_position_embeddings"])
 
     def set_gguf_parameters(self):
         super().set_gguf_parameters()