From 56b32e53f326f279e390304c847fb1569e9c2142 Mon Sep 17 00:00:00 2001 From: lllyasviel Date: Wed, 31 Jan 2024 10:28:27 -0800 Subject: [PATCH] move files --- .../scripts/preprocessor_clipvision.py | 27 ------------------- .../scripts/forge_ipadapter.py | 27 +++++++++++++++++++ 2 files changed, 27 insertions(+), 27 deletions(-) delete mode 100644 extensions-builtin/forge_preprocessor_clipvision/scripts/preprocessor_clipvision.py diff --git a/extensions-builtin/forge_preprocessor_clipvision/scripts/preprocessor_clipvision.py b/extensions-builtin/forge_preprocessor_clipvision/scripts/preprocessor_clipvision.py deleted file mode 100644 index bc425523..00000000 --- a/extensions-builtin/forge_preprocessor_clipvision/scripts/preprocessor_clipvision.py +++ /dev/null @@ -1,27 +0,0 @@ -from modules_forge.supported_preprocessor import PreprocessorClipVision -from modules_forge.shared import add_supported_preprocessor -from modules_forge.forge_util import numpy_to_pytorch - - -class PreprocessorClipVisionForIPAdapter(PreprocessorClipVision): - def __init__(self, name, url, filename): - super().__init__(name, url, filename) - self.tags = ['IP-Adapter'] - self.model_filename_filters = ['IP-Adapter', 'IP_Adapter'] - - def __call__(self, input_image, resolution, slider_1=None, slider_2=None, slider_3=None, **kwargs): - clipvision = self.load_clipvision() - return clipvision, numpy_to_pytorch(input_image) - - -add_supported_preprocessor(PreprocessorClipVisionForIPAdapter( - name='CLIP-ViT-H (IPAdapter)', - url='https://huggingface.co/h94/IP-Adapter/resolve/main/models/image_encoder/model.safetensors', - filename='CLIP-ViT-H-14.safetensors' -)) - -add_supported_preprocessor(PreprocessorClipVisionForIPAdapter( - name='CLIP-ViT-bigG (IPAdapter)', - url='https://huggingface.co/h94/IP-Adapter/resolve/main/sdxl_models/image_encoder/model.safetensors', - filename='CLIP-ViT-bigG.safetensors' -)) diff --git a/extensions-builtin/sd_forge_ipadapter/scripts/forge_ipadapter.py b/extensions-builtin/sd_forge_ipadapter/scripts/forge_ipadapter.py index 30488a5e..5cdef307 100644 --- a/extensions-builtin/sd_forge_ipadapter/scripts/forge_ipadapter.py +++ b/extensions-builtin/sd_forge_ipadapter/scripts/forge_ipadapter.py @@ -1,3 +1,6 @@ +from modules_forge.supported_preprocessor import PreprocessorClipVision +from modules_forge.shared import add_supported_preprocessor +from modules_forge.forge_util import numpy_to_pytorch from modules_forge.shared import add_supported_control_model from modules_forge.supported_controlnet import ControlModelPatcher from lib_ipadapter.IPAdapterPlus import IPAdapterApply @@ -6,6 +9,30 @@ from lib_ipadapter.IPAdapterPlus import IPAdapterApply opIPAdapterApply = IPAdapterApply().apply_ipadapter +class PreprocessorClipVisionForIPAdapter(PreprocessorClipVision): + def __init__(self, name, url, filename): + super().__init__(name, url, filename) + self.tags = ['IP-Adapter'] + self.model_filename_filters = ['IP-Adapter', 'IP_Adapter'] + + def __call__(self, input_image, resolution, slider_1=None, slider_2=None, slider_3=None, **kwargs): + clipvision = self.load_clipvision() + return clipvision, numpy_to_pytorch(input_image) + + +add_supported_preprocessor(PreprocessorClipVisionForIPAdapter( + name='CLIP-ViT-H (IPAdapter)', + url='https://huggingface.co/h94/IP-Adapter/resolve/main/models/image_encoder/model.safetensors', + filename='CLIP-ViT-H-14.safetensors' +)) + +add_supported_preprocessor(PreprocessorClipVisionForIPAdapter( + name='CLIP-ViT-bigG (IPAdapter)', + url='https://huggingface.co/h94/IP-Adapter/resolve/main/sdxl_models/image_encoder/model.safetensors', + filename='CLIP-ViT-bigG.safetensors' +)) + + class IPAdapterPatcher(ControlModelPatcher): @staticmethod def try_build_from_state_dict(state_dict, ckpt_path):