From 3469ffb9044611d467fab4af4f5bae7cdeaf2129 Mon Sep 17 00:00:00 2001 From: Kawrakow Date: Fri, 27 Feb 2026 15:08:18 +0000 Subject: [PATCH] Minor --- src/llama-delta-net.cpp | 4 +--- src/llama.cpp | 1 + 2 files changed, 2 insertions(+), 3 deletions(-) diff --git a/src/llama-delta-net.cpp b/src/llama-delta-net.cpp index 86d1bd5e..b6566d05 100644 --- a/src/llama-delta-net.cpp +++ b/src/llama-delta-net.cpp @@ -246,9 +246,7 @@ ggml_tensor * delta_net::build_layer_attn_linear_core(ggml_context * ctx0, ggml_ const int64_t n_seqs = 1; const int64_t n_seq_tokens = n_tok; - auto qkvz = build_qkvz(ctx0, cur, il, cb); - ggml_tensor * qkv_mixed = qkvz.first; - ggml_tensor * z = qkvz.second; + auto [qkv_mixed, z] = build_qkvz(ctx0, cur, il, cb); ggml_tensor *alpha, *beta; if (model.layers[il].ssm_beta_alpha) { diff --git a/src/llama.cpp b/src/llama.cpp index 0f94c58f..3ccae37a 100644 --- a/src/llama.cpp +++ b/src/llama.cpp @@ -1512,6 +1512,7 @@ static void llm_load_print_meta(llama_model_loader & ml, llama_model & model) { LLAMA_LOG_INFO("%s: ssm_d_inner = %u\n", __func__, hparams.ssm_d_inner); LLAMA_LOG_INFO("%s: ssm_d_state = %u\n", __func__, hparams.ssm_d_state); LLAMA_LOG_INFO("%s: ssm_dt_rank = %u\n", __func__, hparams.ssm_dt_rank); + LLAMA_LOG_INFO("%s: ssm_n_group = %u\n", __func__, hparams.ssm_n_group); } LLAMA_LOG_INFO("%s: model type = %s\n", __func__, llama_model_type_name(model.type));