From: Eric Zhang Date: Wed, 10 Dec 2025 14:22:40 +0000 (+0800) Subject: model : Qwen3-Next-80B-A3B has 48 layers (#17898) X-Git-Tag: upstream/0.0.7446~99 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=b6777218198888ed4b18d839e805f887562a9be4;p=pkg%2Fggml%2Fsources%2Fllama.cpp model : Qwen3-Next-80B-A3B has 48 layers (#17898) * model : Qwen3-Next-80B-A3B has 48 layers * model : Add 80B-A3B type name --- diff --git a/src/llama-model.cpp b/src/llama-model.cpp index 04fccc97..fc337b04 100644 --- a/src/llama-model.cpp +++ b/src/llama-model.cpp @@ -120,6 +120,7 @@ const char * llm_type_name(llm_type type) { case LLM_TYPE_16B_A1B: return "16B.A1B"; case LLM_TYPE_21B_A3B: return "21B.A3B"; case LLM_TYPE_30B_A3B: return "30B.A3B"; + case LLM_TYPE_80B_A3B: return "80B.A3B"; case LLM_TYPE_100B_A6B: return "100B.A6B"; case LLM_TYPE_106B_A12B: return "106B.A12B"; case LLM_TYPE_230B_A10B: return "230B.A10B"; @@ -2257,7 +2258,7 @@ void llama_model::load_hparams(llama_model_loader & ml) { } switch (hparams.n_layer) { - case 80: type = LLM_TYPE_80B_A3B; break; + case 48: type = LLM_TYPE_80B_A3B; break; default: type = LLM_TYPE_UNKNOWN; } } break;