diff --git a/extensions-builtin/forge_preprocessor_clipvision/scripts/preprocessor_clipvision.py b/extensions-builtin/forge_preprocessor_clipvision/scripts/preprocessor_clipvision.py index 43bcdc15..0f8b062b 100644 --- a/extensions-builtin/forge_preprocessor_clipvision/scripts/preprocessor_clipvision.py +++ b/extensions-builtin/forge_preprocessor_clipvision/scripts/preprocessor_clipvision.py @@ -29,9 +29,8 @@ class PreprocessorClipVision(Preprocessor): ) self.clipvision = ldm_patched.modules.clip_vision.load(ckpt_path) - input_image = numpy_to_pytorch(input_image) - - return self.clipvision.encode_image(input_image) + # For IPAdapter Format + return self.clipvision, input_image add_supported_preprocessor(PreprocessorClipVision( diff --git a/extensions-builtin/sd_forge_ipadapter/scripts/forge_ipadapter.py b/extensions-builtin/sd_forge_ipadapter/scripts/forge_ipadapter.py index ffac24c6..75b10ee3 100644 --- a/extensions-builtin/sd_forge_ipadapter/scripts/forge_ipadapter.py +++ b/extensions-builtin/sd_forge_ipadapter/scripts/forge_ipadapter.py @@ -28,7 +28,7 @@ class IPAdapterPatcher(ControlModelPatcher): super().__init__(model_patcher) self.ipadapter = model_patcher return - + def process_before_every_sampling(self, process, cond, *args, **kwargs): a = 0 return