From d7865901ec7d2aff0a46276bff017498adbec835 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E3=81=AD=E3=81=8A=E3=82=93?= <240400715+neon-aiart@users.noreply.github.com> Date: Mon, 10 Nov 2025 02:28:05 +0900 Subject: [PATCH] Update get_token_counter.py --- scripts/physton_prompt/get_token_counter.py | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/scripts/physton_prompt/get_token_counter.py b/scripts/physton_prompt/get_token_counter.py index fb40117..7d96e50 100644 --- a/scripts/physton_prompt/get_token_counter.py +++ b/scripts/physton_prompt/get_token_counter.py @@ -4,6 +4,10 @@ from functools import partial, reduce def get_token_counter(text, steps): + # Check if the model is fully loaded to prevent TypeError during model switching + if sd_models.model_data.sd_model is None: + return {"token_count": 0, "max_length": 0} + # copy from modules.ui.py try: text, _ = extra_networks.parse_prompt(text) @@ -27,10 +31,6 @@ def get_token_counter(text, steps): prompts = [prompt_text for step, prompt_text in flat_prompts] if forge: - # Check if the model is fully loaded to prevent TypeError during model switching - if sd_models.model_data.sd_model is None: - return {"token_count": 0, "max_length": 0} - cond_stage_model = sd_models.model_data.sd_model.cond_stage_model token_count, max_length = max([model_hijack.get_prompt_lengths(prompt,cond_stage_model) for prompt in prompts], key=lambda args: args[0])