From 2db78088a4010f69d2eb038f6d49b2e2c0379cca Mon Sep 17 00:00:00 2001 From: pharmapsychotic Date: Sat, 9 Sep 2023 20:20:39 -0500 Subject: [PATCH] Switch to clip-interrogator 0.6.0 now that webui uses new transformers version --- install.py | 2 +- scripts/clip_interrogator_ext.py | 9 ++++----- 2 files changed, 5 insertions(+), 6 deletions(-) diff --git a/install.py b/install.py index a88362a..2c8d851 100644 --- a/install.py +++ b/install.py @@ -1,6 +1,6 @@ import launch -CI_VERSION = "0.5.4" +CI_VERSION = "0.6.0" needs_install = False try: diff --git a/scripts/clip_interrogator_ext.py b/scripts/clip_interrogator_ext.py index 4ed7a25..4f7ded1 100644 --- a/scripts/clip_interrogator_ext.py +++ b/scripts/clip_interrogator_ext.py @@ -17,7 +17,7 @@ from fastapi import FastAPI from fastapi.exceptions import HTTPException from io import BytesIO -__version__ = "0.1.6" +__version__ = "0.2.0" ci = None low_vram = False @@ -64,7 +64,6 @@ def load(clip_model_name): device=devices.get_optimal_device(), cache_path = 'models/clip-interrogator', clip_model_name=clip_model_name, - blip_model=shared.interrogator.load_blip_model().float() ) if low_vram: config.apply_low_vram_defaults() @@ -78,10 +77,10 @@ def unload(): global ci if ci is not None: print("Offloading CLIP Interrogator...") - ci.blip_model = ci.blip_model.to(devices.cpu) + ci.caption_model = ci.caption_model.to(devices.cpu) ci.clip_model = ci.clip_model.to(devices.cpu) - ci.blip_offloaded = True - ci.clip_offloaded = True + ci.caption_offload = True + ci.clip_offload = True devices.torch_gc() def image_analysis(image, clip_model_name):