From 2fcf407ab33d6ef53a28286db439d8410d672bad Mon Sep 17 00:00:00 2001 From: Iwan Kawrakow Date: Tue, 16 Dec 2025 18:06:30 +0000 Subject: [PATCH] Nah, it is not working --- src/llama.cpp | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/llama.cpp b/src/llama.cpp index 325311be..178b8f77 100644 --- a/src/llama.cpp +++ b/src/llama.cpp @@ -4773,7 +4773,7 @@ struct llama_context * llama_new_context_with_model( LLAMA_LOG_INFO("XXXXXXXXXXXXXXXXXXXXX Setting only active experts offload\n"); ggml_backend_sched_set_only_active_experts(ctx->sched, true); } - if (model->split_mode == LLAMA_SPLIT_MODE_GRAPH) { // && !model->has_tensor_overrides()) { + if (model->split_mode == LLAMA_SPLIT_MODE_GRAPH && !model->has_tensor_overrides()) { ggml_backend_sched_set_split_mode_graph(ctx->sched, true); ggml_backend_sched_set_max_extra_alloc(ctx->sched, params.max_extra_alloc); }