From 47e1e316c580ce6bf264cb069bffc10a50d3f167 Mon Sep 17 00:00:00 2001 From: comfyanonymous <121283862+comfyanonymous@users.noreply.github.com> Date: Thu, 12 Mar 2026 13:54:38 -0700 Subject: [PATCH] Lower kv cache memory usage. (#12909) --- comfy_extras/nodes_flux.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/comfy_extras/nodes_flux.py b/comfy_extras/nodes_flux.py index c366d0d5b..3a23c7d04 100644 --- a/comfy_extras/nodes_flux.py +++ b/comfy_extras/nodes_flux.py @@ -248,7 +248,7 @@ class KV_Attn_Input: self.set_cache = False return {"q": q, "k": torch.cat((k, kk), dim=2), "v": torch.cat((v, vv), dim=2)} - self.cache[cache_key] = (k[:, :, -ref_toks:], v[:, :, -ref_toks:]) + self.cache[cache_key] = (k[:, :, -ref_toks:].clone(), v[:, :, -ref_toks:].clone()) self.set_cache = True return {"q": q, "k": k, "v": v}