From 29be3e93c4afb5b2f6d079cfa2ed2758ae3465aa Mon Sep 17 00:00:00 2001 From: Kawrakow Date: Thu, 28 Aug 2025 14:09:59 +0300 Subject: [PATCH] Make yarn_log_multiplier optional (#738) Co-authored-by: Iwan Kawrakow --- src/llama.cpp | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/llama.cpp b/src/llama.cpp index 9a6ef293..8d7d3917 100644 --- a/src/llama.cpp +++ b/src/llama.cpp @@ -4105,7 +4105,7 @@ static void llm_load_hparams( // that have no expert_gating_func model parameter set hparams.expert_gating_func = LLM_EXPERT_GATING_FUNC_SOFTMAX; } - ml.get_key(LLM_KV_ROPE_SCALING_YARN_LOG_MUL, hparams.rope_yarn_log_mul); + ml.get_key(LLM_KV_ROPE_SCALING_YARN_LOG_MUL, hparams.rope_yarn_log_mul, false); switch (hparams.n_layer) { case 27: model.type = e_model::MODEL_16B; break;