From bc4e1128f78be0fbb4e2fa630adb6a04b969ac68 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Sigbj=C3=B8rn=20Skj=C3=A6ret?= Date: Wed, 7 May 2025 12:49:27 +0200 Subject: [PATCH] llama : deci : support ffn-free with attention (#13296) --- src/llama-model.cpp | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/llama-model.cpp b/src/llama-model.cpp index 774e343f..7b66e2a0 100644 --- a/src/llama-model.cpp +++ b/src/llama-model.cpp @@ -4792,7 +4792,7 @@ struct llm_build_deci : public llm_graph_context { } // FFN-free layer of Llama-3_1-Nemotron-Ultra-253B - if (n_head == 0 && n_ff == 0) { + if (n_ff == 0) { continue; }