From e2ba73750cc708b433101e7738412c53ae04b4b8 Mon Sep 17 00:00:00 2001 From: likelovewant Date: Mon, 13 Jan 2025 14:39:16 +0800 Subject: [PATCH] fix RecursionError when enable and disable the teacache --- README.md | 5 ----- scripts/teacache.py | 5 +++++ 2 files changed, 5 insertions(+), 5 deletions(-) diff --git a/README.md b/README.md index a65c57e..63b2742 100644 --- a/README.md +++ b/README.md @@ -1,4 +1,3 @@ -<<<<<<< HEAD ## Sd-Forge-TeaCache: Speed up Your Diffusion Models **Introduction** @@ -57,7 +56,3 @@ For additional information and other integrations, explore: * [ComfyUI-TeaCache](https://github.com/welltop-cn/ComfyUI-TeaCache) -======= -# sd-forge-teacache -teacache adaption on forge webui ->>>>>>> 45804d1411ad9da74bda8e26d6c7a6d2068c579c diff --git a/scripts/teacache.py b/scripts/teacache.py index 9274194..f557b41 100644 --- a/scripts/teacache.py +++ b/scripts/teacache.py @@ -86,6 +86,8 @@ class TeaCache(scripts.Script): setattr(IntegratedFluxTransformer2DModel, "cnt", 0) if hasattr(IntegratedFluxTransformer2DModel, "accumulated_rel_l1_distance"): setattr(IntegratedFluxTransformer2DModel, "accumulated_rel_l1_distance", 0) + if hasattr(IntegratedFluxTransformer2DModel, "_teacache_enabled_printed"): + delattr(IntegratedFluxTransformer2DModel, "_teacache_enabled_printed") # Free GPU memory torch.cuda.empty_cache() print("Residual cache cleared and GPU memory freed.") @@ -151,6 +153,9 @@ class TeaCache(scripts.Script): # Restore the original forward method if hasattr(IntegratedFluxTransformer2DModel, "original_forward"): IntegratedFluxTransformer2DModel.forward = IntegratedFluxTransformer2DModel.original_forward + # Remove the patched forward method to avoid recursion + if hasattr(IntegratedFluxTransformer2DModel, "patched_forward"): + delattr(IntegratedFluxTransformer2DModel, "patched_forward") def patched_inner_forward(self, img, img_ids, txt, txt_ids, timesteps, y, guidance=None):