From: Sigbjørn Skjæret Date: Wed, 7 May 2025 10:49:27 +0000 (+0200) Subject: llama : deci : support ffn-free with attention (#13296) X-Git-Tag: upstream/0.0.5318~15 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=bc4e1128f78be0fbb4e2fa630adb6a04b969ac68;p=pkg%2Fggml%2Fsources%2Fllama.cpp llama : deci : support ffn-free with attention (#13296) --- diff --git a/src/llama-model.cpp b/src/llama-model.cpp index 774e343f..7b66e2a0 100644 --- a/src/llama-model.cpp +++ b/src/llama-model.cpp @@ -4792,7 +4792,7 @@ struct llm_build_deci : public llm_graph_context { } // FFN-free layer of Llama-3_1-Nemotron-Ultra-253B - if (n_head == 0 && n_ff == 0) { + if (n_ff == 0) { continue; }