rename files

This commit is contained in:
layerdiffusion
2024-08-03 15:54:39 -07:00
parent a5265df7f4
commit 8c087f920e
21 changed files with 29 additions and 29 deletions

View File

@@ -2,7 +2,7 @@ import os
from modules.modelloader import load_file_from_url
from modules.upscaler import Upscaler, UpscalerData
from modules_forge.forge_util import prepare_free_memory
from modules_forge.utils import prepare_free_memory
from ldsr_model_arch import LDSR
from modules import shared, script_callbacks, errors
import sd_hijack_autoencoder # noqa: F401

View File

@@ -6,7 +6,7 @@ from PIL import Image
from modules import devices, modelloader, script_callbacks, shared, upscaler_utils
from modules.upscaler import Upscaler, UpscalerData
from modules_forge.forge_util import prepare_free_memory
from modules_forge.utils import prepare_free_memory
SWINIR_MODEL_URL = "https://github.com/JingyunLiang/SwinIR/releases/download/v0.0/003_realSR_BSRGAN_DFOWMFC_s64w8_SwinIR-L_x4_GAN.pth"

View File

@@ -7,7 +7,7 @@ import einops
from omegaconf import OmegaConf
from modules_forge.supported_preprocessor import Preprocessor, PreprocessorParameter
from modules_forge.forge_util import numpy_to_pytorch, resize_image_with_pad
from modules_forge.utils import numpy_to_pytorch, resize_image_with_pad
from modules_forge.shared import preprocessor_dir, add_supported_preprocessor
from modules.modelloader import load_file_from_url
from annotator.lama.saicinpainting.training.trainers import load_checkpoint

View File

@@ -1,6 +1,6 @@
from modules_forge.supported_preprocessor import Preprocessor, PreprocessorParameter
from modules_forge.shared import preprocessor_dir, add_supported_preprocessor
from modules_forge.forge_util import resize_image_with_pad
from modules_forge.utils import resize_image_with_pad
import os
@@ -10,7 +10,7 @@ import numpy as np
from marigold.model.marigold_pipeline import MarigoldPipeline
from huggingface_hub import snapshot_download
from modules_forge.diffusers_patcher import DiffusersModelPatcher
from modules_forge.forge_util import numpy_to_pytorch, HWC3
from modules_forge.utils import numpy_to_pytorch, HWC3
class PreprocessorMarigold(Preprocessor):

View File

@@ -1,6 +1,6 @@
from modules_forge.supported_preprocessor import Preprocessor, PreprocessorParameter
from modules_forge.shared import preprocessor_dir, add_supported_preprocessor
from modules_forge.forge_util import resize_image_with_pad
from modules_forge.utils import resize_image_with_pad
from modules.modelloader import load_file_from_url
import types

View File

@@ -17,7 +17,7 @@ from lib_controlnet.controlnet_ui.photopea import Photopea
from lib_controlnet.enums import InputMode, HiResFixOption
from modules import shared, script_callbacks
from modules.ui_components import FormRow
from modules_forge.forge_util import HWC3
from modules_forge.utils import HWC3
from lib_controlnet.external_code import UiControlNetUnit
from modules.ui_components import ToolButton
from gradio_rangeslider import RangeSlider

View File

@@ -3,7 +3,7 @@ from modules import processing
from lib_controlnet import external_code
from modules_forge.forge_util import HWC3
from modules_forge.utils import HWC3
from PIL import Image, ImageFilter, ImageOps
from lib_controlnet.lvminthin import lvmin_thin, nake_nms

View File

@@ -20,7 +20,7 @@ from lib_controlnet.logging import logger
from modules.processing import StableDiffusionProcessingImg2Img, StableDiffusionProcessingTxt2Img, \
StableDiffusionProcessing
from lib_controlnet.infotext import Infotext
from modules_forge.forge_util import HWC3, numpy_to_pytorch
from modules_forge.utils import HWC3, numpy_to_pytorch
from lib_controlnet.enums import HiResFixOption
from lib_controlnet.api import controlnet_api

View File

@@ -1,6 +1,6 @@
from modules_forge.supported_preprocessor import PreprocessorClipVision, Preprocessor, PreprocessorParameter
from modules_forge.shared import add_supported_preprocessor
from modules_forge.forge_util import numpy_to_pytorch
from modules_forge.utils import numpy_to_pytorch
from modules_forge.shared import add_supported_control_model
from modules_forge.supported_controlnet import ControlModelPatcher
from lib_ipadapter.IPAdapterPlus import IPAdapterApply, InsightFaceLoader

View File

@@ -4,7 +4,7 @@ from modules import modelloader, errors
from modules.shared import cmd_opts, opts
from modules.upscaler import Upscaler, UpscalerData
from modules.upscaler_utils import upscale_with_model
from modules_forge.forge_util import prepare_free_memory
from modules_forge.utils import prepare_free_memory
class UpscalerDAT(Upscaler):

View File

@@ -2,7 +2,7 @@ from modules import modelloader, devices, errors
from modules.shared import opts
from modules.upscaler import Upscaler, UpscalerData
from modules.upscaler_utils import upscale_with_model
from modules_forge.forge_util import prepare_free_memory
from modules_forge.utils import prepare_free_memory
class UpscalerESRGAN(Upscaler):

View File

@@ -10,7 +10,7 @@ import numpy as np
import torch
from modules import devices, errors, face_restoration, shared
from modules_forge.forge_util import prepare_free_memory
from modules_forge.utils import prepare_free_memory
if TYPE_CHECKING:
from facexlib.utils.face_restoration_helper import FaceRestoreHelper

View File

@@ -5,7 +5,7 @@ from modules import modelloader, devices
from modules.shared import opts
from modules.upscaler import Upscaler, UpscalerData
from modules.upscaler_utils import upscale_with_model
from modules_forge.forge_util import prepare_free_memory
from modules_forge.utils import prepare_free_memory
class UpscalerHAT(Upscaler):

View File

@@ -34,7 +34,7 @@ from ldm.models.diffusion.ddpm import LatentDepth2ImageDiffusion
from einops import repeat, rearrange
from blendmodes.blend import blendLayers, BlendType
from modules.sd_models import apply_token_merging
from modules_forge.forge_util import apply_circular_forge
from modules_forge.utils import apply_circular_forge
# some of those options should not be changed at all because they would break the model, so I removed them from options.

View File

@@ -4,7 +4,7 @@ from modules import modelloader, errors
from modules.shared import cmd_opts, opts
from modules.upscaler import Upscaler, UpscalerData
from modules.upscaler_utils import upscale_with_model
from modules_forge.forge_util import prepare_free_memory
from modules_forge.utils import prepare_free_memory
class UpscalerRealESRGAN(Upscaler):

View File

@@ -17,7 +17,7 @@ from modules import paths, shared, modelloader, devices, script_callbacks, sd_va
from modules.shared import opts
from modules.timer import Timer
import numpy as np
from modules_forge import forge_loader
from modules_forge import loader
from backend import memory_management
@@ -663,7 +663,7 @@ def load_model(checkpoint_info=None, already_loaded_state_dict=None):
# cache newly loaded model
checkpoints_loaded[checkpoint_info] = state_dict.copy()
sd_model = forge_loader.load_model_for_a1111(timer=timer, checkpoint_info=checkpoint_info, state_dict=state_dict)
sd_model = loader.load_model_for_a1111(timer=timer, checkpoint_info=checkpoint_info, state_dict=state_dict)
sd_model.filename = checkpoint_info.filename
if not SkipWritingToConfig.skip:

View File

@@ -9,7 +9,7 @@ from modules import devices, shared, prompt_parser
from modules import torch_utils
from backend import memory_management
from modules_forge.forge_clip import move_clip_to_gpu
from modules_forge.clip import move_clip_to_gpu
def get_learned_conditioning(self: sgm.models.diffusion.DiffusionEngine, batch: prompt_parser.SdConditioning | list[str]):

View File

@@ -13,7 +13,7 @@ from modules.shared import cmd_opts
from modules import sd_hijack
from modules.sd_models_xl import extend_sdxl
from ldm.util import instantiate_from_config
from modules_forge import forge_clip
from modules_forge import clip
from modules_forge.unet_patcher import UnetPatcher
from backend.loader import load_huggingface_components
from backend.modules.k_model import KModel
@@ -119,13 +119,13 @@ def load_model_for_a1111(timer, checkpoint_info=None, state_dict=None):
timer.record("forge solving config")
if hasattr(a1111_config.model.params, 'network_config'):
a1111_config.model.params.network_config.target = 'modules_forge.forge_loader.FakeObject'
a1111_config.model.params.network_config.target = 'modules_forge.loader.FakeObject'
if hasattr(a1111_config.model.params, 'unet_config'):
a1111_config.model.params.unet_config.target = 'modules_forge.forge_loader.FakeObject'
a1111_config.model.params.unet_config.target = 'modules_forge.loader.FakeObject'
if hasattr(a1111_config.model.params, 'first_stage_config'):
a1111_config.model.params.first_stage_config.target = 'modules_forge.forge_loader.FakeObject'
a1111_config.model.params.first_stage_config.target = 'modules_forge.loader.FakeObject'
with no_clip():
sd_model = instantiate_from_config(a1111_config.model)
@@ -161,7 +161,7 @@ def load_model_for_a1111(timer, checkpoint_info=None, state_dict=None):
model_embeddings = embedder.transformer.text_model.embeddings
model_embeddings.token_embedding = sd_hijack.EmbeddingsWithFixes(
model_embeddings.token_embedding, sd_hijack.model_hijack)
embedder = forge_clip.CLIP_SD_XL_L(embedder, sd_hijack.model_hijack)
embedder = clip.CLIP_SD_XL_L(embedder, sd_hijack.model_hijack)
conditioner.embedders[i] = embedder
text_cond_models.append(embedder)
elif typename == 'FrozenOpenCLIPEmbedder2': # SDXL Clip G
@@ -171,7 +171,7 @@ def load_model_for_a1111(timer, checkpoint_info=None, state_dict=None):
model_embeddings = embedder.transformer.text_model.embeddings
model_embeddings.token_embedding = sd_hijack.EmbeddingsWithFixes(
model_embeddings.token_embedding, sd_hijack.model_hijack, textual_inversion_key='clip_g')
embedder = forge_clip.CLIP_SD_XL_G(embedder, sd_hijack.model_hijack)
embedder = clip.CLIP_SD_XL_G(embedder, sd_hijack.model_hijack)
conditioner.embedders[i] = embedder
text_cond_models.append(embedder)
@@ -185,14 +185,14 @@ def load_model_for_a1111(timer, checkpoint_info=None, state_dict=None):
model_embeddings = sd_model.cond_stage_model.transformer.text_model.embeddings
model_embeddings.token_embedding = sd_hijack.EmbeddingsWithFixes(
model_embeddings.token_embedding, sd_hijack.model_hijack)
sd_model.cond_stage_model = forge_clip.CLIP_SD_15_L(sd_model.cond_stage_model, sd_hijack.model_hijack)
sd_model.cond_stage_model = clip.CLIP_SD_15_L(sd_model.cond_stage_model, sd_hijack.model_hijack)
elif type(sd_model.cond_stage_model).__name__ == 'FrozenOpenCLIPEmbedder': # SD21 Clip
sd_model.cond_stage_model.tokenizer = forge_objects.clip.tokenizer.clip_l
sd_model.cond_stage_model.transformer = forge_objects.clip.cond_stage_model.clip_l.transformer
model_embeddings = sd_model.cond_stage_model.transformer.text_model.embeddings
model_embeddings.token_embedding = sd_hijack.EmbeddingsWithFixes(
model_embeddings.token_embedding, sd_hijack.model_hijack)
sd_model.cond_stage_model = forge_clip.CLIP_SD_21_H(sd_model.cond_stage_model, sd_hijack.model_hijack)
sd_model.cond_stage_model = clip.CLIP_SD_21_H(sd_model.cond_stage_model, sd_hijack.model_hijack)
else:
raise NotImplementedError('Bad Clip Class Name:' + type(sd_model.cond_stage_model).__name__)

View File

@@ -5,9 +5,9 @@ from modules_forge.shared import add_supported_preprocessor, preprocessor_dir
from backend import memory_management
from backend.patcher.base import ModelPatcher
from backend.patcher import clipvision
from modules_forge.forge_util import resize_image_with_pad
from modules_forge.utils import resize_image_with_pad
from modules.modelloader import load_file_from_url
from modules_forge.forge_util import numpy_to_pytorch
from modules_forge.utils import numpy_to_pytorch
class PreprocessorParameter: