From 494d70626f65fd35acd1fbc3abd73ec6cc002faf Mon Sep 17 00:00:00 2001 From: Kawrakow Date: Sun, 8 Feb 2026 08:59:39 +0000 Subject: [PATCH] Allow missing rope_frequency_base_swa in Step-3.5 models --- src/llama-hparams.cpp | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/src/llama-hparams.cpp b/src/llama-hparams.cpp index c6697d4a..8fe793e9 100644 --- a/src/llama-hparams.cpp +++ b/src/llama-hparams.cpp @@ -1130,7 +1130,7 @@ void llm_load_hparams( hparams.expert_gating_func = LLM_EXPERT_GATING_FUNC_SIGMOID; } ml.get_key(LLM_KV_ATTENTION_SLIDING_WINDOW, hparams.n_swa); - ml.get_key(LLM_KV_ROPE_FREQ_BASE_SWA, hparams.rope_freq_base_train_swa); + bool have_rfb_train_swa = ml.get_key(LLM_KV_ROPE_FREQ_BASE_SWA, hparams.rope_freq_base_train_swa, false); ml.get_key_or_arr(LLM_KV_ATTENTION_SLIDING_WINDOW_PATTERN, hparams.swa_layers, hparams.n_layer); if (!ml.get_key_or_arr(LLM_KV_ROPE_DIMENSION_COUNT_PER_LAYER, hparams.rope_dim_per_layer, hparams.n_layer, false)) { for (int i = 0; i < hparams.n_layer; ++i) { @@ -1155,6 +1155,7 @@ void llm_load_hparams( hparams.rope_scaling_apply_mask, false); hparams.has_rope_freq_base_per_layer = ml.get_key_or_arr(LLM_KV_ROPE_FREQ_BASE_PER_LAYER, hparams.rope_freq_base_per_layer, hparams.n_layer, false); + GGML_ASSERT(hparams.has_rope_freq_base_per_layer || have_rfb_train_swa); } break; default: (void)0; }