mirror of
https://github.com/comfyanonymous/ComfyUI.git
synced 2026-04-23 07:59:07 +00:00
More progress on brainstorming code for asset management for models
This commit is contained in:
@@ -26,9 +26,10 @@ class ReturnedAssetABC(ABC):
|
||||
|
||||
|
||||
class ModelReturnedAsset(ReturnedAssetABC):
|
||||
def __init__(self, model: dict[str, str] | tuple[dict[str, str], dict[str, str]]):
|
||||
def __init__(self, state_dict: dict[str, str], metadata: dict[str, str]=None):
|
||||
super().__init__("model")
|
||||
self.model = model
|
||||
self.state_dict = state_dict
|
||||
self.metadata = metadata
|
||||
|
||||
|
||||
class AssetResolverABC(ABC):
|
||||
@@ -38,24 +39,30 @@ class AssetResolverABC(ABC):
|
||||
|
||||
|
||||
class LocalAssetResolver(AssetResolverABC):
|
||||
def resolve(self, asset_info: AssetInfo) -> ReturnedAssetABC:
|
||||
def resolve(self, asset_info: AssetInfo, cache_result: bool=True) -> ReturnedAssetABC:
|
||||
# currently only supports models - make sure models is in the tags
|
||||
if "models" not in asset_info.tags:
|
||||
return None
|
||||
# TODO: if hash exists, call model processor to try to get info about model:
|
||||
if asset_info.hash:
|
||||
...
|
||||
try:
|
||||
from app.model_processor import model_processor
|
||||
model_db = model_processor.retrieve_model_by_hash(asset_info.hash)
|
||||
full_path = model_db.path
|
||||
except Exception as e:
|
||||
logging.error(f"Could not get model by hash with error: {e}")
|
||||
# the good ol' bread and butter - folder_paths's keys as tags
|
||||
folder_keys = folder_paths.folder_names_and_paths.keys()
|
||||
parent_paths = []
|
||||
for tag in asset_info.tags:
|
||||
if tag in folder_keys:
|
||||
parent_paths.append(tag)
|
||||
# if subdir metadata and name exists, use that as the model name going forward
|
||||
if "subdir" in asset_info.metadata and asset_info.name:
|
||||
relative_path = os.path.join(asset_info.metadata["subdir"], asset_info.name)
|
||||
# the good ol' bread and butter - folder_paths's keys as tags
|
||||
folder_keys = folder_paths.folder_names_and_paths.keys()
|
||||
parent_paths = []
|
||||
for tag in asset_info.tags:
|
||||
if tag in folder_keys:
|
||||
parent_paths.append(tag)
|
||||
# if no matching parent paths, then something went wrong and should return None
|
||||
if len(parent_paths) == 0:
|
||||
return None
|
||||
relative_path = os.path.join(asset_info.metadata["subdir"], asset_info.name)
|
||||
# now we have the parent keys, we can try to get the local path
|
||||
chosen_parent = None
|
||||
full_path = None
|
||||
@@ -64,27 +71,40 @@ class LocalAssetResolver(AssetResolverABC):
|
||||
if full_path:
|
||||
chosen_parent = parent_path
|
||||
break
|
||||
logging.info(f"Resolved {asset_info.name} to {full_path} in {chosen_parent}")
|
||||
# we know the path, so load the model and return it
|
||||
model = comfy.utils.load_torch_file(full_path, safe_load=True, device=asset_info.metadata.get("device", None), return_metadata=asset_info.metadata.get("return_metadata", False))
|
||||
return ModelReturnedAsset(model)
|
||||
# TODO: if name exists, try to find model by name in all subdirs of parent paths
|
||||
if full_path is not None:
|
||||
logging.info(f"Resolved {asset_info.name} to {full_path} in {chosen_parent}")
|
||||
# we know the path, so load the model and return it
|
||||
state_dict, metadata = comfy.utils.load_torch_file(full_path, safe_load=True, device=asset_info.metadata.get("device", None), return_metadata=True)
|
||||
# TODO: handle caching
|
||||
return ModelReturnedAsset(state_dict, metadata)
|
||||
# if just name exists, try to find model by name in all subdirs of parent paths
|
||||
# TODO: this behavior should be configurable by user
|
||||
if asset_info.name:
|
||||
...
|
||||
# TODO: if download_url metadata exists, download the model and load it
|
||||
for parent_path in parent_paths:
|
||||
filelist = folder_paths.get_filename_list(parent_path)
|
||||
for file in filelist:
|
||||
if os.path.basename(file) == asset_info.name:
|
||||
full_path = folder_paths.get_full_path(parent_path, file)
|
||||
state_dict, metadata = comfy.utils.load_torch_file(full_path, safe_load=True, device=asset_info.metadata.get("device", None), return_metadata=True)
|
||||
# TODO: handle caching
|
||||
return ModelReturnedAsset(state_dict, metadata)
|
||||
# TODO: if download_url metadata exists, download the model and load it; this should be configurable by user
|
||||
if asset_info.metadata.get("download_url", None):
|
||||
...
|
||||
return None
|
||||
|
||||
|
||||
resolvers: list[AssetResolverABC] = []
|
||||
resolvers.append(LocalAssetResolver())
|
||||
|
||||
|
||||
def resolve(asset_info: AssetInfo) -> Any:
|
||||
global resolvers
|
||||
for resolver in resolvers:
|
||||
try:
|
||||
return resolver.resolve(asset_info)
|
||||
to_return = resolver.resolve(asset_info)
|
||||
if to_return is not None:
|
||||
return resolver.resolve(asset_info)
|
||||
except Exception as e:
|
||||
logging.error(f"Error resolving asset {asset_info.hash}: {e}")
|
||||
logging.error(f"Error resolving asset {asset_info.name} using {resolver.__class__.__name__}: {e}")
|
||||
return None
|
||||
|
||||
Reference in New Issue
Block a user