From: Sigbjørn Skjæret Date: Thu, 26 Jun 2025 13:01:14 +0000 (+0200) Subject: llama : return mistral-v7-tekken as default template only (#14390) X-Git-Tag: upstream/0.0.5882~121 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=b25346221dadb9101aa9dda55431dde4d3596943;p=pkg%2Fggml%2Fsources%2Fllama.cpp llama : return mistral-v7-tekken as default template only (#14390) --- diff --git a/src/llama-model.cpp b/src/llama-model.cpp index 9b19da98..c2835ce6 100644 --- a/src/llama-model.cpp +++ b/src/llama-model.cpp @@ -14377,7 +14377,7 @@ const char * llama_model_chat_template(const llama_model * model, const char * n // do not extend this list unless absolutely necessary // Mistral-Small-2503 does not have built-in chat template llama_vocab_pre_type pre_type = model->vocab.get_pre_type(); - if (pre_type == LLAMA_VOCAB_PRE_TYPE_TEKKEN && model->layers.size() == 40) { + if (!name && pre_type == LLAMA_VOCAB_PRE_TYPE_TEKKEN && model->layers.size() == 40) { return "mistral-v7-tekken"; }