From 082addead22ca197277fbce9248f8dccca686646 Mon Sep 17 00:00:00 2001 From: Kawrakow Date: Fri, 6 Mar 2026 16:06:34 +0100 Subject: [PATCH] Minor: do not do SILU on the whole convolution output (#1374) --- src/llama-delta-net.cpp | 5 ++--- 1 file changed, 2 insertions(+), 3 deletions(-) diff --git a/src/llama-delta-net.cpp b/src/llama-delta-net.cpp index 092a2908..47a6ba0e 100644 --- a/src/llama-delta-net.cpp +++ b/src/llama-delta-net.cpp @@ -328,9 +328,8 @@ ggml_tensor * delta_net::build_layer_attn_linear_core(ggml_context * ctx0, ggml_ ggml_tensor * conv_output_raw = ggml_ssm_conv(ctx0, conv_states, qkv_mixed, model.layers[il].ssm_conv1d, inp_s_seq_qnext); cb(conv_output_raw, "conv_output_raw", il); - //ggml_tensor * conv_output = ggml_view_2d(ctx0, conv_output_raw, conv_dim, n_tok, conv_dim * ggml_element_size(conv_output_raw), 0); - //ggml_tensor * conv_output_silu = ggml_silu(ctx0, conv_output); - ggml_tensor * conv_output_silu = ggml_silu(ctx0, conv_output_raw); + ggml_tensor * conv_output = ggml_view_2d(ctx0, conv_output_raw, conv_dim, n_tok, conv_dim * ggml_element_size(conv_output_raw), 0); + ggml_tensor * conv_output_silu = ggml_silu(ctx0, conv_output); cb(conv_output_silu, "conv_output_silu", il); // Calculate the total conv dimension