class JambaModel(TextModel):
model_arch = gguf.MODEL_ARCH.JAMBA
- def get_vocab_base_pre(self, tokenizer) -> str:
- del tokenizer # unused
-
- return "gpt-2"
-
def set_vocab(self):
if (self.dir_model / "tokenizer.model").is_file():
- # Using Jamba's tokenizer.json causes errors on model load
- # (something about "byte not found in vocab"),
- # but there's a working tokenizer.model
self._set_vocab_sentencepiece()
else:
- # Some Jamba models only have a tokenizer.json, which works.
- self._set_vocab_gpt2()
+ self._set_vocab_llama_hf()
+ self.gguf_writer.add_add_space_prefix(False)
def set_gguf_parameters(self):
d_model = self.find_hparam(["hidden_size", "mamba_d_model"])