mirror of
https://git.datalinker.icu/ltdrdata/ComfyUI-Manager
synced 2025-12-08 21:54:26 +08:00
https://github.com/ltdrdata/ComfyUI-Manager/issues/1552 fixed: comfyui versions should be based on commit date https://github.com/ltdrdata/ComfyUI-Manager/issues/1566 fixed: invalid identifying of nightly node packs which has `git@github.com:...` url fixed: switch comfyui should be based on `master` branch instead of `main` branch fixed: switch_to_default_branch - more robust switching refactor: endpoints for policies
3303 lines
118 KiB
Python
3303 lines
118 KiB
Python
"""
|
|
description:
|
|
`manager_core` contains the core implementation of the management functions in ComfyUI-Manager.
|
|
"""
|
|
|
|
import json
|
|
import logging
|
|
import os
|
|
import sys
|
|
import subprocess
|
|
import re
|
|
import shutil
|
|
import configparser
|
|
import platform
|
|
from datetime import datetime
|
|
|
|
import git
|
|
from git.remote import RemoteProgress
|
|
from urllib.parse import urlparse
|
|
from tqdm.auto import tqdm
|
|
import time
|
|
import yaml
|
|
import zipfile
|
|
import traceback
|
|
from concurrent.futures import ThreadPoolExecutor, as_completed
|
|
|
|
orig_print = print
|
|
|
|
from rich import print
|
|
from packaging import version
|
|
|
|
import uuid
|
|
|
|
glob_path = os.path.join(os.path.dirname(__file__)) # ComfyUI-Manager/glob
|
|
sys.path.append(glob_path)
|
|
|
|
import cm_global
|
|
import cnr_utils
|
|
import manager_util
|
|
import git_utils
|
|
import manager_downloader
|
|
from node_package import InstalledNodePackage
|
|
|
|
|
|
version_code = [3, 25]
|
|
version_str = f"V{version_code[0]}.{version_code[1]}" + (f'.{version_code[2]}' if len(version_code) > 2 else '')
|
|
|
|
|
|
DEFAULT_CHANNEL = "https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main"
|
|
|
|
|
|
default_custom_nodes_path = None
|
|
|
|
|
|
def get_default_custom_nodes_path():
|
|
global default_custom_nodes_path
|
|
if default_custom_nodes_path is None:
|
|
try:
|
|
import folder_paths
|
|
default_custom_nodes_path = folder_paths.get_folder_paths("custom_nodes")[0]
|
|
except:
|
|
default_custom_nodes_path = os.path.abspath(os.path.join(manager_util.comfyui_manager_path, '..'))
|
|
|
|
return default_custom_nodes_path
|
|
|
|
|
|
def get_custom_nodes_paths():
|
|
try:
|
|
import folder_paths
|
|
return folder_paths.get_folder_paths("custom_nodes")
|
|
except:
|
|
custom_nodes_path = os.path.abspath(os.path.join(manager_util.comfyui_manager_path, '..'))
|
|
return [custom_nodes_path]
|
|
|
|
|
|
def get_comfyui_tag():
|
|
repo = git.Repo(comfy_path)
|
|
try:
|
|
return repo.git.describe('--tags')
|
|
except:
|
|
return None
|
|
|
|
|
|
def get_script_env():
|
|
new_env = os.environ.copy()
|
|
git_exe = get_config().get('git_exe')
|
|
if git_exe is not None:
|
|
new_env['GIT_EXE_PATH'] = git_exe
|
|
|
|
if 'COMFYUI_PATH' not in new_env:
|
|
new_env['COMFYUI_PATH'] = comfy_path
|
|
|
|
if 'COMFYUI_FOLDERS_BASE_PATH' not in new_env:
|
|
new_env['COMFYUI_FOLDERS_BASE_PATH'] = comfy_path
|
|
|
|
return new_env
|
|
|
|
|
|
invalid_nodes = {}
|
|
|
|
|
|
def extract_base_custom_nodes_dir(x:str):
|
|
if os.path.dirname(x).endswith('.disabled'):
|
|
return os.path.dirname(os.path.dirname(x))
|
|
elif x.endswith('.disabled'):
|
|
return os.path.dirname(x)
|
|
else:
|
|
return os.path.dirname(x)
|
|
|
|
|
|
def check_invalid_nodes():
|
|
global invalid_nodes
|
|
|
|
try:
|
|
import folder_paths
|
|
except:
|
|
try:
|
|
sys.path.append(comfy_path)
|
|
import folder_paths
|
|
except:
|
|
raise Exception(f"Invalid COMFYUI_FOLDERS_BASE_PATH: {comfy_path}")
|
|
|
|
def check(root):
|
|
global invalid_nodes
|
|
|
|
subdirs = [d for d in os.listdir(root) if os.path.isdir(os.path.join(root, d))]
|
|
for subdir in subdirs:
|
|
if subdir in ['.disabled', '__pycache__']:
|
|
continue
|
|
|
|
package = unified_manager.installed_node_packages.get(subdir)
|
|
if not package:
|
|
continue
|
|
|
|
if not package.isValid():
|
|
invalid_nodes[subdir] = package.fullpath
|
|
|
|
node_paths = folder_paths.get_folder_paths("custom_nodes")
|
|
for x in node_paths:
|
|
check(x)
|
|
|
|
disabled_dir = os.path.join(x, '.disabled')
|
|
if os.path.exists(disabled_dir):
|
|
check(disabled_dir)
|
|
|
|
if len(invalid_nodes):
|
|
print("\n-------------------- ComfyUI-Manager invalid nodes notice ----------------")
|
|
print("\nNodes requiring reinstallation have been detected:\n(Directly delete the corresponding path and reinstall.)\n")
|
|
|
|
for x in invalid_nodes.values():
|
|
print(x)
|
|
|
|
print("\n---------------------------------------------------------------------------\n")
|
|
|
|
|
|
# read env vars
|
|
comfy_path = os.environ.get('COMFYUI_PATH')
|
|
comfy_base_path = os.environ.get('COMFYUI_FOLDERS_BASE_PATH')
|
|
|
|
if comfy_path is None:
|
|
try:
|
|
import folder_paths
|
|
comfy_path = os.path.join(os.path.dirname(folder_paths.__file__))
|
|
except:
|
|
comfy_path = os.path.abspath(os.path.join(manager_util.comfyui_manager_path, '..', '..'))
|
|
|
|
if comfy_base_path is None:
|
|
comfy_base_path = comfy_path
|
|
|
|
|
|
channel_list_template_path = os.path.join(manager_util.comfyui_manager_path, 'channels.list.template')
|
|
git_script_path = os.path.join(manager_util.comfyui_manager_path, "git_helper.py")
|
|
|
|
manager_files_path = None
|
|
manager_config_path = None
|
|
manager_channel_list_path = None
|
|
manager_startup_script_path:str = None
|
|
manager_snapshot_path = None
|
|
manager_pip_overrides_path = None
|
|
manager_pip_blacklist_path = None
|
|
manager_components_path = None
|
|
|
|
def update_user_directory(user_dir):
|
|
global manager_files_path
|
|
global manager_config_path
|
|
global manager_channel_list_path
|
|
global manager_startup_script_path
|
|
global manager_snapshot_path
|
|
global manager_pip_overrides_path
|
|
global manager_pip_blacklist_path
|
|
global manager_components_path
|
|
|
|
manager_files_path = os.path.abspath(os.path.join(user_dir, 'default', 'ComfyUI-Manager'))
|
|
if not os.path.exists(manager_files_path):
|
|
os.makedirs(manager_files_path)
|
|
|
|
manager_snapshot_path = os.path.join(manager_files_path, "snapshots")
|
|
if not os.path.exists(manager_snapshot_path):
|
|
os.makedirs(manager_snapshot_path)
|
|
|
|
manager_startup_script_path = os.path.join(manager_files_path, "startup-scripts")
|
|
if not os.path.exists(manager_startup_script_path):
|
|
os.makedirs(manager_startup_script_path)
|
|
|
|
manager_config_path = os.path.join(manager_files_path, 'config.ini')
|
|
manager_channel_list_path = os.path.join(manager_files_path, 'channels.list')
|
|
manager_pip_overrides_path = os.path.join(manager_files_path, "pip_overrides.json")
|
|
manager_pip_blacklist_path = os.path.join(manager_files_path, "pip_blacklist.list")
|
|
manager_components_path = os.path.join(manager_files_path, "components")
|
|
manager_util.cache_dir = os.path.join(manager_files_path, "cache")
|
|
|
|
if not os.path.exists(manager_util.cache_dir):
|
|
os.makedirs(manager_util.cache_dir)
|
|
|
|
try:
|
|
import folder_paths
|
|
update_user_directory(folder_paths.get_user_directory())
|
|
|
|
except Exception:
|
|
# fallback:
|
|
# This case is only possible when running with cm-cli, and in practice, this case is not actually used.
|
|
update_user_directory(os.path.abspath(manager_util.comfyui_manager_path))
|
|
|
|
|
|
cached_config = None
|
|
js_path = None
|
|
|
|
comfy_ui_required_revision = 1930
|
|
comfy_ui_required_commit_datetime = datetime(2024, 1, 24, 0, 0, 0)
|
|
|
|
comfy_ui_revision = "Unknown"
|
|
comfy_ui_commit_datetime = datetime(1900, 1, 1, 0, 0, 0)
|
|
|
|
channel_dict = None
|
|
channel_list = None
|
|
|
|
|
|
def remap_pip_package(pkg):
|
|
if pkg in cm_global.pip_overrides:
|
|
res = cm_global.pip_overrides[pkg]
|
|
print(f"[ComfyUI-Manager] '{pkg}' is remapped to '{res}'")
|
|
return res
|
|
else:
|
|
return pkg
|
|
|
|
|
|
def is_blacklisted(name):
|
|
name = name.strip()
|
|
|
|
pattern = r'([^<>!~=]+)([<>!~=]=?)([^ ]*)'
|
|
match = re.search(pattern, name)
|
|
|
|
if match:
|
|
name = match.group(1)
|
|
|
|
if name in cm_global.pip_blacklist:
|
|
return True
|
|
|
|
if name in cm_global.pip_downgrade_blacklist:
|
|
pips = manager_util.get_installed_packages()
|
|
|
|
if match is None:
|
|
if name in pips:
|
|
return True
|
|
elif match.group(2) in ['<=', '==', '<', '~=']:
|
|
if name in pips:
|
|
if manager_util.StrictVersion(pips[name]) >= manager_util.StrictVersion(match.group(3)):
|
|
return True
|
|
|
|
return False
|
|
|
|
|
|
def is_installed(name):
|
|
name = name.strip()
|
|
|
|
if name.startswith('#'):
|
|
return True
|
|
|
|
pattern = r'([^<>!~=]+)([<>!~=]=?)([0-9.a-zA-Z]*)'
|
|
match = re.search(pattern, name)
|
|
|
|
if match:
|
|
name = match.group(1)
|
|
|
|
if name in cm_global.pip_blacklist:
|
|
return True
|
|
|
|
if name in cm_global.pip_downgrade_blacklist:
|
|
pips = manager_util.get_installed_packages()
|
|
|
|
if match is None:
|
|
if name in pips:
|
|
return True
|
|
elif match.group(2) in ['<=', '==', '<', '~=']:
|
|
if name in pips:
|
|
if manager_util.StrictVersion(pips[name]) >= manager_util.StrictVersion(match.group(3)):
|
|
print(f"[ComfyUI-Manager] skip black listed pip installation: '{name}'")
|
|
return True
|
|
|
|
pkg = manager_util.get_installed_packages().get(name.lower())
|
|
if pkg is None:
|
|
return False # update if not installed
|
|
|
|
if match is None:
|
|
return True # don't update if version is not specified
|
|
|
|
if match.group(2) in ['>', '>=']:
|
|
if manager_util.StrictVersion(pkg) < manager_util.StrictVersion(match.group(3)):
|
|
return False
|
|
elif manager_util.StrictVersion(pkg) > manager_util.StrictVersion(match.group(3)):
|
|
print(f"[SKIP] Downgrading pip package isn't allowed: {name.lower()} (cur={pkg})")
|
|
|
|
if match.group(2) == '==':
|
|
if manager_util.StrictVersion(pkg) < manager_util.StrictVersion(match.group(3)):
|
|
return False
|
|
|
|
if match.group(2) == '~=':
|
|
if manager_util.StrictVersion(pkg) == manager_util.StrictVersion(match.group(3)):
|
|
return False
|
|
|
|
return name.lower() in manager_util.get_installed_packages()
|
|
|
|
|
|
def normalize_channel(channel):
|
|
if channel == 'local':
|
|
return channel
|
|
elif channel is None:
|
|
return None
|
|
elif channel.startswith('https://'):
|
|
return channel
|
|
elif channel.startswith('http://') and get_config()['http_channel_enabled'] == True:
|
|
return channel
|
|
|
|
tmp_dict = get_channel_dict()
|
|
channel_url = tmp_dict.get(channel)
|
|
if channel_url:
|
|
return channel_url
|
|
|
|
raise Exception(f"Invalid channel name '{channel}'")
|
|
|
|
|
|
class ManagedResult:
|
|
def __init__(self, action):
|
|
self.action = action
|
|
self.items = []
|
|
self.result = True
|
|
self.to_path = None
|
|
self.msg = None
|
|
self.target = None
|
|
self.postinstall = lambda: True
|
|
self.ver = None
|
|
|
|
def append(self, item):
|
|
self.items.append(item)
|
|
|
|
def fail(self, msg):
|
|
self.result = False
|
|
self.msg = msg
|
|
return self
|
|
|
|
def with_target(self, target):
|
|
self.target = target
|
|
return self
|
|
|
|
def with_msg(self, msg):
|
|
self.msg = msg
|
|
return self
|
|
|
|
def with_postinstall(self, postinstall):
|
|
self.postinstall = postinstall
|
|
return self
|
|
|
|
def with_ver(self, ver):
|
|
self.ver = ver
|
|
return self
|
|
|
|
|
|
class UnifiedManager:
|
|
def __init__(self):
|
|
self.installed_node_packages: dict[str, InstalledNodePackage] = {}
|
|
|
|
self.cnr_inactive_nodes = {} # node_id -> node_version -> fullpath
|
|
self.nightly_inactive_nodes = {} # node_id -> fullpath
|
|
self.unknown_inactive_nodes = {} # node_id -> repo url * fullpath
|
|
self.active_nodes = {} # node_id -> node_version * fullpath
|
|
self.unknown_active_nodes = {} # node_id -> repo url * fullpath
|
|
self.cnr_map = {} # node_id -> cnr info
|
|
self.repo_cnr_map = {} # repo_url -> cnr info
|
|
self.custom_node_map_cache = {} # (channel, mode) -> augmented custom node list json
|
|
self.processed_install = set()
|
|
|
|
def get_module_name(self, x):
|
|
info = self.active_nodes.get(x)
|
|
if info is None:
|
|
for url, fullpath in self.unknown_active_nodes.values():
|
|
if url == x:
|
|
return os.path.basename(fullpath)
|
|
else:
|
|
return os.path.basename(info[1])
|
|
|
|
return None
|
|
|
|
def get_cnr_by_repo(self, url):
|
|
return self.repo_cnr_map.get(git_utils.normalize_url(url))
|
|
|
|
def resolve_unspecified_version(self, node_name, guess_mode=None):
|
|
if guess_mode == 'active':
|
|
# priority:
|
|
# 1. CNR/nightly active nodes
|
|
# 2. unknown
|
|
# 3. Fail
|
|
|
|
if node_name in self.cnr_map:
|
|
version_spec = self.get_from_cnr_active_nodes(node_name)
|
|
|
|
if version_spec is None:
|
|
if node_name in self.unknown_active_nodes:
|
|
version_spec = "unknown"
|
|
else:
|
|
return None
|
|
|
|
elif node_name in self.unknown_active_nodes:
|
|
version_spec = "unknown"
|
|
else:
|
|
return None
|
|
|
|
elif guess_mode == 'inactive':
|
|
# priority:
|
|
# 1. CNR latest in inactive
|
|
# 2. nightly
|
|
# 3. unknown
|
|
# 4. Fail
|
|
|
|
if node_name in self.cnr_map:
|
|
latest = self.get_from_cnr_inactive_nodes(node_name)
|
|
|
|
if latest is not None:
|
|
version_spec = str(latest[0])
|
|
else:
|
|
if node_name in self.nightly_inactive_nodes:
|
|
version_spec = "nightly"
|
|
else:
|
|
version_spec = "unknown"
|
|
|
|
elif node_name in self.unknown_inactive_nodes:
|
|
version_spec = "unknown"
|
|
else:
|
|
return None
|
|
|
|
else:
|
|
# priority:
|
|
# 1. CNR latest in world
|
|
# 2. unknown
|
|
|
|
if node_name in self.cnr_map:
|
|
version_spec = self.cnr_map[node_name]['latest_version']['version']
|
|
else:
|
|
version_spec = "unknown"
|
|
|
|
return version_spec
|
|
|
|
def resolve_node_spec(self, node_name, guess_mode=None):
|
|
"""
|
|
resolve to 'node_name, version_spec' from version string
|
|
|
|
version string:
|
|
node_name@latest
|
|
node_name@nightly
|
|
node_name@unknown
|
|
node_name@<version>
|
|
node_name
|
|
|
|
if guess_mode is 'active' or 'inactive'
|
|
return can be 'None' based on state check
|
|
otherwise
|
|
return 'unknown' version when failed to guess
|
|
"""
|
|
|
|
spec = node_name.split('@')
|
|
|
|
if len(spec) == 2:
|
|
node_name = spec[0]
|
|
version_spec = spec[1]
|
|
|
|
if version_spec == 'latest':
|
|
if node_name not in self.cnr_map:
|
|
print(f"ERROR: '{node_name}' is not a CNR node.")
|
|
return None
|
|
else:
|
|
version_spec = self.cnr_map[node_name]['latest_version']['version']
|
|
|
|
elif guess_mode in ['active', 'inactive']:
|
|
node_name = spec[0]
|
|
version_spec = self.resolve_unspecified_version(node_name, guess_mode=guess_mode)
|
|
if version_spec is None:
|
|
return None
|
|
else:
|
|
node_name = spec[0]
|
|
version_spec = self.resolve_unspecified_version(node_name)
|
|
if version_spec is None:
|
|
return None
|
|
|
|
return node_name, version_spec, len(spec) > 1
|
|
|
|
def resolve_from_path(self, fullpath):
|
|
url = git_utils.git_url(fullpath)
|
|
if url:
|
|
url = git_utils.normalize_url(url)
|
|
|
|
cnr = self.get_cnr_by_repo(url)
|
|
commit_hash = git_utils.get_commit_hash(fullpath)
|
|
if cnr:
|
|
cnr_utils.generate_cnr_id(fullpath, cnr['id'])
|
|
return {'id': cnr['id'], 'cnr': cnr, 'ver': 'nightly', 'hash': commit_hash}
|
|
else:
|
|
url = os.path.basename(url)
|
|
if url.endswith('.git'):
|
|
url = url[:-4]
|
|
return {'id': url, 'ver': 'unknown', 'hash': commit_hash}
|
|
else:
|
|
info = cnr_utils.read_cnr_info(fullpath)
|
|
|
|
if info:
|
|
cnr = self.cnr_map.get(info['id'])
|
|
if cnr:
|
|
# normalize version
|
|
# for example: 2.5 -> 2.5.0
|
|
ver = str(manager_util.StrictVersion(info['version']))
|
|
return {'id': cnr['id'], 'cnr': cnr, 'ver': ver}
|
|
else:
|
|
return None
|
|
else:
|
|
return None
|
|
|
|
def update_cache_at_path(self, fullpath):
|
|
node_package = InstalledNodePackage.from_fullpath(fullpath, self.resolve_from_path)
|
|
self.installed_node_packages[node_package.id] = node_package
|
|
|
|
if node_package.is_disabled and node_package.is_unknown:
|
|
url = git_utils.git_url(node_package.fullpath)
|
|
if url is not None:
|
|
url = git_utils.normalize_url(url)
|
|
self.unknown_inactive_nodes[node_package.id] = (url, node_package.fullpath)
|
|
|
|
if node_package.is_disabled and node_package.is_nightly:
|
|
self.nightly_inactive_nodes[node_package.id] = node_package.fullpath
|
|
|
|
if node_package.is_enabled and not node_package.is_unknown:
|
|
self.active_nodes[node_package.id] = node_package.version, node_package.fullpath
|
|
|
|
if node_package.is_enabled and node_package.is_unknown:
|
|
url = git_utils.git_url(node_package.fullpath)
|
|
if url is not None:
|
|
url = git_utils.normalize_url(url)
|
|
self.unknown_active_nodes[node_package.id] = (url, node_package.fullpath)
|
|
|
|
if node_package.is_from_cnr and node_package.is_disabled:
|
|
self.add_to_cnr_inactive_nodes(node_package.id, node_package.version, node_package.fullpath)
|
|
|
|
def is_updatable(self, node_id):
|
|
cur_ver = self.get_cnr_active_version(node_id)
|
|
latest_ver = self.cnr_map[node_id]['latest_version']['version']
|
|
|
|
if cur_ver and latest_ver:
|
|
return self.safe_version(latest_ver) > self.safe_version(cur_ver)
|
|
|
|
return False
|
|
|
|
def fetch_or_pull_git_repo(self, is_pull=False):
|
|
updated = set()
|
|
failed = set()
|
|
|
|
def check_update(node_name, fullpath, ver_spec):
|
|
try:
|
|
if is_pull:
|
|
is_updated, success = git_repo_update_check_with(fullpath, do_update=True)
|
|
else:
|
|
is_updated, success = git_repo_update_check_with(fullpath, do_fetch=True)
|
|
|
|
return f"{node_name}@{ver_spec}", is_updated, success
|
|
except Exception:
|
|
traceback.print_exc()
|
|
|
|
return f"{node_name}@{ver_spec}", False, False
|
|
|
|
with ThreadPoolExecutor() as executor:
|
|
futures = []
|
|
|
|
for k, v in self.unknown_active_nodes.items():
|
|
futures.append(executor.submit(check_update, k, v[1], 'unknown'))
|
|
|
|
for k, v in self.active_nodes.items():
|
|
if v[0] == 'nightly':
|
|
futures.append(executor.submit(check_update, k, v[1], 'nightly'))
|
|
|
|
for future in as_completed(futures):
|
|
item, is_updated, success = future.result()
|
|
|
|
if is_updated:
|
|
updated.add(item)
|
|
|
|
if not success:
|
|
failed.add(item)
|
|
|
|
return dict(updated=list(updated), failed=list(failed))
|
|
|
|
def is_enabled(self, node_id, version_spec=None):
|
|
"""
|
|
1. true if node_id@<specified_version> is enabled
|
|
2. true if node_id@<any> is enabled and version_spec==None
|
|
3. false otherwise
|
|
|
|
remark: latest version_spec is not allowed. Must be resolved before call.
|
|
"""
|
|
if version_spec == "cnr":
|
|
return self.get_cnr_active_version(node_id) not in [None, 'nightly']
|
|
elif version_spec == 'unknown' and self.is_unknown_active(node_id):
|
|
return True
|
|
elif version_spec is not None and self.get_cnr_active_version(node_id) == version_spec:
|
|
return True
|
|
elif version_spec is None and (node_id in self.active_nodes or node_id in self.unknown_active_nodes):
|
|
return True
|
|
return False
|
|
|
|
def is_disabled(self, node_id, version_spec=None):
|
|
"""
|
|
1. node_id@unknown is disabled if version_spec is @unknown
|
|
2. node_id@nightly is disabled if version_spec is @nightly
|
|
4. node_id@<specified_version> is disabled if version_spec is not None
|
|
5. not exists (active node_id) if version_spec is None
|
|
|
|
remark: latest version_spec is not allowed. Must be resolved before call.
|
|
"""
|
|
if version_spec == "unknown":
|
|
return node_id in self.unknown_inactive_nodes
|
|
elif version_spec == "nightly":
|
|
return node_id in self.nightly_inactive_nodes
|
|
elif version_spec == "cnr":
|
|
res = self.cnr_inactive_nodes.get(node_id, None)
|
|
if res is None:
|
|
return False
|
|
|
|
res = [x for x in res.keys() if x != 'nightly']
|
|
return len(res) > 0
|
|
elif version_spec is not None:
|
|
return version_spec in self.cnr_inactive_nodes.get(node_id, [])
|
|
|
|
if node_id in self.nightly_inactive_nodes:
|
|
return True
|
|
elif node_id in self.unknown_inactive_nodes:
|
|
return True
|
|
|
|
target = self.cnr_inactive_nodes.get(node_id, None)
|
|
if target is not None and target == version_spec:
|
|
return True
|
|
|
|
return False
|
|
|
|
def is_registered_in_cnr(self, node_id):
|
|
return node_id in self.cnr_map
|
|
|
|
def get_cnr_active_version(self, node_id):
|
|
res = self.active_nodes.get(node_id)
|
|
if res:
|
|
return res[0]
|
|
else:
|
|
return None
|
|
|
|
def is_unknown_active(self, node_id):
|
|
return node_id in self.unknown_active_nodes
|
|
|
|
def add_to_cnr_inactive_nodes(self, node_id, ver, fullpath):
|
|
ver_map = self.cnr_inactive_nodes.get(node_id)
|
|
if ver_map is None:
|
|
ver_map = {}
|
|
self.cnr_inactive_nodes[node_id] = ver_map
|
|
|
|
ver_map[ver] = fullpath
|
|
|
|
def get_from_cnr_active_nodes(self, node_id):
|
|
ver_path = self.active_nodes.get(node_id)
|
|
if ver_path is None:
|
|
return None
|
|
|
|
return ver_path[0]
|
|
|
|
def get_from_cnr_inactive_nodes(self, node_id, ver=None):
|
|
ver_map = self.cnr_inactive_nodes.get(node_id)
|
|
if ver_map is None:
|
|
return None
|
|
|
|
if ver is not None:
|
|
return ver_map.get(ver)
|
|
|
|
latest = None
|
|
for k, v in ver_map.items():
|
|
if latest is None:
|
|
latest = self.safe_version(k), v
|
|
continue
|
|
|
|
cur_ver = self.safe_version(k)
|
|
if cur_ver > latest[0]:
|
|
latest = cur_ver, v
|
|
|
|
return latest
|
|
|
|
async def reload(self, cache_mode, dont_wait=True):
|
|
self.custom_node_map_cache = {}
|
|
self.cnr_inactive_nodes = {} # node_id -> node_version -> fullpath
|
|
self.nightly_inactive_nodes = {} # node_id -> fullpath
|
|
self.unknown_inactive_nodes = {} # node_id -> repo url * fullpath
|
|
self.unknown_active_nodes = {} # node_id -> repo url * fullpath
|
|
self.active_nodes = {} # node_id -> node_version * fullpath
|
|
|
|
if get_config()['network_mode'] != 'public':
|
|
dont_wait = True
|
|
|
|
# reload 'cnr_map' and 'repo_cnr_map'
|
|
cnrs = await cnr_utils.get_cnr_data(cache_mode=cache_mode=='cache', dont_wait=dont_wait)
|
|
|
|
for x in cnrs:
|
|
self.cnr_map[x['id']] = x
|
|
if 'repository' in x:
|
|
normalized_url = git_utils.normalize_url(x['repository'])
|
|
self.repo_cnr_map[normalized_url] = x
|
|
|
|
# reload node status info from custom_nodes/*
|
|
for custom_nodes_path in folder_paths.get_folder_paths('custom_nodes'):
|
|
for x in os.listdir(custom_nodes_path):
|
|
fullpath = os.path.join(custom_nodes_path, x)
|
|
if os.path.isdir(fullpath):
|
|
if x not in ['__pycache__', '.disabled']:
|
|
self.update_cache_at_path(fullpath)
|
|
|
|
# reload node status info from custom_nodes/.disabled/*
|
|
for custom_nodes_path in folder_paths.get_folder_paths('custom_nodes'):
|
|
disabled_dir = os.path.join(custom_nodes_path, '.disabled')
|
|
if os.path.exists(disabled_dir):
|
|
for x in os.listdir(disabled_dir):
|
|
fullpath = os.path.join(disabled_dir, x)
|
|
if os.path.isdir(fullpath):
|
|
self.update_cache_at_path(fullpath)
|
|
|
|
@staticmethod
|
|
async def load_nightly(channel, mode):
|
|
res = {}
|
|
|
|
channel_url = normalize_channel(channel)
|
|
if channel_url:
|
|
if mode not in ['remote', 'local', 'cache']:
|
|
print(f"[bold red]ERROR: Invalid mode is specified `--mode {mode}`[/bold red]", file=sys.stderr)
|
|
return {}
|
|
|
|
json_obj = await get_data_by_mode(mode, 'custom-node-list.json', channel_url=channel_url)
|
|
for x in json_obj['custom_nodes']:
|
|
try:
|
|
for y in x['files']:
|
|
if 'github.com' in y and not (y.endswith('.py') or y.endswith('.js')):
|
|
repo_name = y.split('/')[-1]
|
|
res[repo_name] = (x, False)
|
|
|
|
if 'id' in x:
|
|
if x['id'] not in res:
|
|
res[x['id']] = (x, True)
|
|
except:
|
|
logging.error(f"[ComfyUI-Manager] broken item:{x}")
|
|
|
|
return res
|
|
|
|
async def get_custom_nodes(self, channel, mode):
|
|
# default_channel = normalize_channel('default')
|
|
# cache = self.custom_node_map_cache.get((default_channel, mode)) # CNR/nightly should always be based on the default channel.
|
|
|
|
channel = normalize_channel(channel)
|
|
cache = self.custom_node_map_cache.get((channel, mode)) # CNR/nightly should always be based on the default channel.
|
|
|
|
if cache is not None:
|
|
return cache
|
|
|
|
channel = normalize_channel(channel)
|
|
print(f"nightly_channel: {channel}/{mode}")
|
|
nodes = await self.load_nightly(channel, mode)
|
|
|
|
res = {}
|
|
added_cnr = set()
|
|
for v in nodes.values():
|
|
v = v[0]
|
|
if len(v['files']) == 1:
|
|
cnr = self.get_cnr_by_repo(v['files'][0])
|
|
if cnr:
|
|
if 'latest_version' not in cnr:
|
|
v['cnr_latest'] = '0.0.0'
|
|
else:
|
|
v['cnr_latest'] = cnr['latest_version']['version']
|
|
v['id'] = cnr['id']
|
|
v['author'] = cnr['publisher']['name']
|
|
v['title'] = cnr['name']
|
|
v['description'] = cnr['description']
|
|
v['health'] = '-'
|
|
if 'repository' in cnr:
|
|
v['repository'] = cnr['repository']
|
|
added_cnr.add(cnr['id'])
|
|
node_id = v['id']
|
|
else:
|
|
node_id = v['files'][0].split('/')[-1]
|
|
v['repository'] = v['files'][0]
|
|
res[node_id] = v
|
|
elif len(v['files']) > 1:
|
|
res[v['files'][0]] = v # A custom node composed of multiple url is treated as a single repository with one representative path
|
|
|
|
self.custom_node_map_cache[(channel, mode)] = res
|
|
return res
|
|
|
|
@staticmethod
|
|
def safe_version(ver_str):
|
|
try:
|
|
return version.parse(ver_str)
|
|
except:
|
|
return version.parse("0.0.0")
|
|
|
|
def execute_install_script(self, url, repo_path, instant_execution=False, lazy_mode=False, no_deps=False):
|
|
install_script_path = os.path.join(repo_path, "install.py")
|
|
requirements_path = os.path.join(repo_path, "requirements.txt")
|
|
|
|
if lazy_mode:
|
|
install_cmd = ["#LAZY-INSTALL-SCRIPT", sys.executable]
|
|
return try_install_script(url, repo_path, install_cmd)
|
|
else:
|
|
if os.path.exists(requirements_path) and not no_deps:
|
|
print("Install: pip packages")
|
|
pip_fixer = manager_util.PIPFixer(manager_util.get_installed_packages())
|
|
res = True
|
|
lines = manager_util.robust_readlines(requirements_path)
|
|
for line in lines:
|
|
package_name = remap_pip_package(line.strip())
|
|
if package_name and not package_name.startswith('#') and package_name not in self.processed_install:
|
|
self.processed_install.add(package_name)
|
|
install_cmd = manager_util.make_pip_cmd(["install", package_name])
|
|
if package_name.strip() != "" and not package_name.startswith('#'):
|
|
res = res and try_install_script(url, repo_path, install_cmd, instant_execution=instant_execution)
|
|
|
|
pip_fixer.fix_broken()
|
|
return res
|
|
|
|
if os.path.exists(install_script_path) and install_script_path not in self.processed_install:
|
|
self.processed_install.add(install_script_path)
|
|
print("Install: install script")
|
|
install_cmd = [sys.executable, "install.py"]
|
|
return try_install_script(url, repo_path, install_cmd, instant_execution=instant_execution)
|
|
|
|
return True
|
|
|
|
def reserve_cnr_switch(self, target, zip_url, from_path, to_path, no_deps):
|
|
script_path = os.path.join(manager_startup_script_path, "install-scripts.txt")
|
|
with open(script_path, "a") as file:
|
|
obj = [target, "#LAZY-CNR-SWITCH-SCRIPT", zip_url, from_path, to_path, no_deps, get_default_custom_nodes_path(), sys.executable]
|
|
file.write(f"{obj}\n")
|
|
|
|
print(f"Installation reserved: {target}")
|
|
|
|
return True
|
|
|
|
def reserve_migration(self, moves):
|
|
script_path = os.path.join(manager_startup_script_path, "install-scripts.txt")
|
|
with open(script_path, "a") as file:
|
|
obj = ["", "#LAZY-MIGRATION", moves]
|
|
file.write(f"{obj}\n")
|
|
|
|
return True
|
|
|
|
def unified_fix(self, node_id, version_spec, instant_execution=False, no_deps=False):
|
|
"""
|
|
fix dependencies
|
|
"""
|
|
|
|
result = ManagedResult('fix')
|
|
|
|
if version_spec == 'unknown':
|
|
info = self.unknown_active_nodes.get(node_id)
|
|
else:
|
|
info = self.active_nodes.get(node_id)
|
|
|
|
if info is None or not os.path.exists(info[1]):
|
|
return result.fail(f'not found: {node_id}@{version_spec}')
|
|
|
|
self.execute_install_script(node_id, info[1], instant_execution=instant_execution, no_deps=no_deps)
|
|
|
|
return result
|
|
|
|
def cnr_switch_version(self, node_id, version_spec=None, instant_execution=False, no_deps=False, return_postinstall=False):
|
|
if instant_execution:
|
|
return self.cnr_switch_version_instant(node_id, version_spec, instant_execution, no_deps, return_postinstall)
|
|
else:
|
|
return self.cnr_switch_version_lazy(node_id, version_spec, no_deps, return_postinstall)
|
|
|
|
def cnr_switch_version_lazy(self, node_id, version_spec=None, no_deps=False, return_postinstall=False):
|
|
"""
|
|
switch between cnr version (lazy mode)
|
|
"""
|
|
|
|
result = ManagedResult('switch-cnr')
|
|
|
|
node_info = cnr_utils.install_node(node_id, version_spec)
|
|
if node_info is None or not node_info.download_url:
|
|
return result.fail(f'not available node: {node_id}@{version_spec}')
|
|
|
|
version_spec = node_info.version
|
|
|
|
if self.active_nodes[node_id][0] == version_spec:
|
|
return ManagedResult('skip').with_msg("Up to date")
|
|
|
|
zip_url = node_info.download_url
|
|
from_path = self.active_nodes[node_id][1]
|
|
target = node_id
|
|
to_path = os.path.join(get_default_custom_nodes_path(), target)
|
|
|
|
def postinstall():
|
|
return self.reserve_cnr_switch(target, zip_url, from_path, to_path, no_deps)
|
|
|
|
if return_postinstall:
|
|
return result.with_postinstall(postinstall)
|
|
else:
|
|
if not postinstall():
|
|
return result.fail(f"Failed to execute install script: {node_id}@{version_spec}")
|
|
|
|
return result
|
|
|
|
def cnr_switch_version_instant(self, node_id, version_spec=None, instant_execution=True, no_deps=False, return_postinstall=False):
|
|
"""
|
|
switch between cnr version
|
|
"""
|
|
|
|
# 1. download
|
|
result = ManagedResult('switch-cnr')
|
|
|
|
node_info = cnr_utils.install_node(node_id, version_spec)
|
|
if node_info is None or not node_info.download_url:
|
|
return result.fail(f'not available node: {node_id}@{version_spec}')
|
|
|
|
version_spec = node_info.version
|
|
|
|
if self.active_nodes[node_id][0] == version_spec:
|
|
return ManagedResult('skip').with_msg("Up to date")
|
|
|
|
archive_name = f"CNR_temp_{str(uuid.uuid4())}.zip" # should be unpredictable name - security precaution
|
|
download_path = os.path.join(get_default_custom_nodes_path(), archive_name)
|
|
manager_downloader.basic_download_url(node_info.download_url, get_default_custom_nodes_path(), archive_name)
|
|
|
|
# 2. extract files into <node_id>
|
|
install_path = self.active_nodes[node_id][1]
|
|
extracted = manager_util.extract_package_as_zip(download_path, install_path)
|
|
os.remove(download_path)
|
|
|
|
if extracted is None:
|
|
if len(os.listdir(install_path)) == 0:
|
|
shutil.rmtree(install_path)
|
|
|
|
return result.fail(f'Empty archive file: {node_id}@{version_spec}')
|
|
|
|
# 3. calculate garbage files (.tracking - extracted)
|
|
tracking_info_file = os.path.join(install_path, '.tracking')
|
|
prev_files = set()
|
|
with open(tracking_info_file, 'r') as f:
|
|
for line in f:
|
|
prev_files.add(line.strip())
|
|
garbage = prev_files.difference(extracted)
|
|
garbage = [os.path.join(install_path, x) for x in garbage]
|
|
|
|
# 4-1. remove garbage files
|
|
for x in garbage:
|
|
if os.path.isfile(x):
|
|
os.remove(x)
|
|
|
|
# 4-2. remove garbage dir if empty
|
|
for x in garbage:
|
|
if os.path.isdir(x):
|
|
if not os.listdir(x):
|
|
os.rmdir(x)
|
|
|
|
# 5. create .tracking file
|
|
tracking_info_file = os.path.join(install_path, '.tracking')
|
|
with open(tracking_info_file, "w", encoding='utf-8') as file:
|
|
file.write('\n'.join(list(extracted)))
|
|
|
|
# 6. post install
|
|
result.target = version_spec
|
|
|
|
def postinstall():
|
|
res = self.execute_install_script(f"{node_id}@{version_spec}", install_path, instant_execution=instant_execution, no_deps=no_deps)
|
|
return res
|
|
|
|
if return_postinstall:
|
|
return result.with_postinstall(postinstall)
|
|
else:
|
|
if not postinstall():
|
|
return result.fail(f"Failed to execute install script: {node_id}@{version_spec}")
|
|
|
|
return result
|
|
|
|
def unified_enable(self, node_id, version_spec=None):
|
|
"""
|
|
priority if version_spec == None
|
|
1. CNR latest in disk
|
|
2. nightly
|
|
3. unknown
|
|
|
|
remark: latest version_spec is not allowed. Must be resolved before call.
|
|
"""
|
|
|
|
result = ManagedResult('enable')
|
|
|
|
if version_spec is None:
|
|
version_spec = self.resolve_unspecified_version(node_id, guess_mode='inactive')
|
|
if version is None:
|
|
return result.fail(f'Specified inactive node not exists: {node_id}')
|
|
|
|
if self.is_enabled(node_id, version_spec):
|
|
return ManagedResult('skip').with_msg('Already enabled')
|
|
|
|
if not self.is_disabled(node_id, version_spec):
|
|
return ManagedResult('skip').with_msg('Not installed')
|
|
|
|
from_path = None
|
|
to_path = None
|
|
|
|
if version_spec == 'unknown':
|
|
repo_and_path = self.unknown_inactive_nodes.get(node_id)
|
|
if repo_and_path is None:
|
|
return result.fail(f'Specified inactive node not exists: {node_id}@unknown')
|
|
from_path = repo_and_path[1]
|
|
|
|
base_path = extract_base_custom_nodes_dir(from_path)
|
|
to_path = os.path.join(base_path, node_id)
|
|
elif version_spec == 'nightly':
|
|
self.unified_disable(node_id, False)
|
|
from_path = self.nightly_inactive_nodes.get(node_id)
|
|
if from_path is None:
|
|
return result.fail(f'Specified inactive node not exists: {node_id}@nightly')
|
|
base_path = extract_base_custom_nodes_dir(from_path)
|
|
to_path = os.path.join(base_path, node_id)
|
|
elif version_spec is not None:
|
|
self.unified_disable(node_id, False)
|
|
cnr_info = self.cnr_inactive_nodes.get(node_id)
|
|
|
|
if cnr_info is None or len(cnr_info) == 0:
|
|
return result.fail(f'Specified inactive cnr node not exists: {node_id}')
|
|
|
|
if version_spec == "cnr":
|
|
version_spec = next(iter(cnr_info))
|
|
|
|
if version_spec not in cnr_info:
|
|
return result.fail(f'Specified inactive node not exists: {node_id}@{version_spec}')
|
|
|
|
from_path = cnr_info[version_spec]
|
|
base_path = extract_base_custom_nodes_dir(from_path)
|
|
to_path = os.path.join(base_path, node_id)
|
|
|
|
if from_path is None or not os.path.exists(from_path):
|
|
return result.fail(f'Specified inactive node path not exists: {from_path}')
|
|
|
|
# move from disk
|
|
shutil.move(from_path, to_path)
|
|
|
|
# update cache
|
|
if version_spec == 'unknown':
|
|
self.unknown_active_nodes[node_id] = self.unknown_inactive_nodes[node_id][0], to_path
|
|
del self.unknown_inactive_nodes[node_id]
|
|
return result.with_target(to_path)
|
|
elif version_spec == 'nightly':
|
|
del self.nightly_inactive_nodes[node_id]
|
|
else:
|
|
del self.cnr_inactive_nodes[node_id][version_spec]
|
|
|
|
self.active_nodes[node_id] = version_spec, to_path
|
|
return result.with_target(to_path)
|
|
|
|
def unified_disable(self, node_id, is_unknown):
|
|
result = ManagedResult('disable')
|
|
|
|
if is_unknown:
|
|
version_spec = 'unknown'
|
|
else:
|
|
version_spec = None
|
|
|
|
if not self.is_enabled(node_id, version_spec):
|
|
if not self.is_disabled(node_id, version_spec):
|
|
return ManagedResult('skip').with_msg('Not installed')
|
|
else:
|
|
return ManagedResult('skip').with_msg('Already disabled')
|
|
|
|
if is_unknown:
|
|
repo_and_path = self.unknown_active_nodes.get(node_id)
|
|
|
|
if repo_and_path is None or not os.path.exists(repo_and_path[1]):
|
|
return result.fail(f'Specified active node not exists: {node_id}')
|
|
|
|
base_path = extract_base_custom_nodes_dir(repo_and_path[1])
|
|
to_path = os.path.join(base_path, '.disabled', node_id)
|
|
|
|
shutil.move(repo_and_path[1], to_path)
|
|
result.append((repo_and_path[1], to_path))
|
|
|
|
self.unknown_inactive_nodes[node_id] = repo_and_path[0], to_path
|
|
del self.unknown_active_nodes[node_id]
|
|
|
|
return result
|
|
|
|
ver_and_path = self.active_nodes.get(node_id)
|
|
|
|
if ver_and_path is None or not os.path.exists(ver_and_path[1]):
|
|
return result.fail(f'Specified active node not exists: {node_id}')
|
|
|
|
base_path = extract_base_custom_nodes_dir(ver_and_path[1])
|
|
|
|
# NOTE: A disabled node may have multiple versions, so preserve it using the `@ suffix`.
|
|
to_path = os.path.join(base_path, '.disabled', f"{node_id}@{ver_and_path[0].replace('.', '_')}")
|
|
shutil.move(ver_and_path[1], to_path)
|
|
result.append((ver_and_path[1], to_path))
|
|
|
|
if ver_and_path[0] == 'nightly':
|
|
self.nightly_inactive_nodes[node_id] = to_path
|
|
else:
|
|
self.add_to_cnr_inactive_nodes(node_id, ver_and_path[0], to_path)
|
|
|
|
del self.active_nodes[node_id]
|
|
|
|
return result
|
|
|
|
def unified_uninstall(self, node_id: str, is_unknown: bool):
|
|
"""
|
|
Remove whole installed custom nodes including inactive nodes
|
|
"""
|
|
result = ManagedResult('uninstall')
|
|
|
|
if is_unknown:
|
|
# remove from actives
|
|
repo_and_path = self.unknown_active_nodes.get(node_id)
|
|
|
|
is_removed = False
|
|
|
|
if repo_and_path is not None and os.path.exists(repo_and_path[1]):
|
|
rmtree(repo_and_path[1])
|
|
result.append(repo_and_path[1])
|
|
del self.unknown_active_nodes[node_id]
|
|
|
|
is_removed = True
|
|
|
|
# remove from inactives
|
|
repo_and_path = self.unknown_inactive_nodes.get(node_id)
|
|
|
|
if repo_and_path is not None and os.path.exists(repo_and_path[1]):
|
|
rmtree(repo_and_path[1])
|
|
result.append(repo_and_path[1])
|
|
del self.unknown_inactive_nodes[node_id]
|
|
|
|
is_removed = True
|
|
|
|
if is_removed:
|
|
return result
|
|
else:
|
|
return ManagedResult('skip')
|
|
|
|
# remove from actives
|
|
ver_and_path = self.active_nodes.get(node_id)
|
|
|
|
if ver_and_path is not None and os.path.exists(ver_and_path[1]):
|
|
shutil.rmtree(ver_and_path[1])
|
|
result.items.append(ver_and_path)
|
|
del self.active_nodes[node_id]
|
|
|
|
# remove from nightly inactives
|
|
fullpath = self.nightly_inactive_nodes.get(node_id)
|
|
if fullpath is not None and os.path.exists(fullpath):
|
|
shutil.rmtree(fullpath)
|
|
result.items.append(('nightly', fullpath))
|
|
del self.nightly_inactive_nodes[node_id]
|
|
|
|
# remove from cnr inactives
|
|
ver_map = self.cnr_inactive_nodes.get(node_id)
|
|
if ver_map is not None:
|
|
for key, fullpath in ver_map.items():
|
|
shutil.rmtree(fullpath)
|
|
result.items.append((key, fullpath))
|
|
del self.cnr_inactive_nodes[node_id]
|
|
|
|
if len(result.items) == 0:
|
|
return ManagedResult('skip').with_msg('Not installed')
|
|
|
|
return result
|
|
|
|
def cnr_install(self, node_id, version_spec=None, instant_execution=False, no_deps=False, return_postinstall=False):
|
|
result = ManagedResult('install-cnr')
|
|
|
|
node_info = cnr_utils.install_node(node_id, version_spec)
|
|
if node_info is None or not node_info.download_url:
|
|
return result.fail(f'not available node: {node_id}@{version_spec}')
|
|
|
|
archive_name = f"CNR_temp_{str(uuid.uuid4())}.zip" # should be unpredictable name - security precaution
|
|
download_path = os.path.join(get_default_custom_nodes_path(), archive_name)
|
|
|
|
# re-download. I cannot trust existing file.
|
|
if os.path.exists(download_path):
|
|
os.remove(download_path)
|
|
|
|
# install_path
|
|
install_path = os.path.join(get_default_custom_nodes_path(), node_id)
|
|
if os.path.exists(install_path):
|
|
return result.fail(f'Install path already exists: {install_path}')
|
|
|
|
manager_downloader.download_url(node_info.download_url, get_default_custom_nodes_path(), archive_name)
|
|
os.makedirs(install_path, exist_ok=True)
|
|
extracted = manager_util.extract_package_as_zip(download_path, install_path)
|
|
os.remove(download_path)
|
|
result.to_path = install_path
|
|
|
|
if extracted is None:
|
|
shutil.rmtree(install_path)
|
|
return result.fail(f'Empty archive file: {node_id}@{version_spec}')
|
|
|
|
# create .tracking file
|
|
tracking_info_file = os.path.join(install_path, '.tracking')
|
|
with open(tracking_info_file, "w", encoding='utf-8') as file:
|
|
file.write('\n'.join(extracted))
|
|
|
|
result.target = version_spec
|
|
|
|
def postinstall():
|
|
return self.execute_install_script(node_id, install_path, instant_execution=instant_execution, no_deps=no_deps)
|
|
|
|
if return_postinstall:
|
|
return result.with_postinstall(postinstall)
|
|
else:
|
|
if not postinstall():
|
|
return result.fail(f"Failed to execute install script: {node_id}@{version_spec}")
|
|
|
|
return result
|
|
|
|
def repo_install(self, url, repo_path, instant_execution=False, no_deps=False, return_postinstall=False):
|
|
result = ManagedResult('install-git')
|
|
result.append(url)
|
|
|
|
if not is_valid_url(url):
|
|
return result.fail(f"Invalid git url: {url}")
|
|
|
|
if url.endswith("/"):
|
|
url = url[:-1]
|
|
try:
|
|
# Clone the repository from the remote URL
|
|
clone_url = git_utils.get_url_for_clone(url)
|
|
print(f"Download: git clone '{clone_url}'")
|
|
|
|
if not instant_execution and platform.system() == 'Windows':
|
|
res = manager_funcs.run_script([sys.executable, git_script_path, "--clone", get_default_custom_nodes_path(), clone_url, repo_path], cwd=get_default_custom_nodes_path())
|
|
if res != 0:
|
|
return result.fail(f"Failed to clone repo: {clone_url}")
|
|
else:
|
|
repo = git.Repo.clone_from(clone_url, repo_path, recursive=True, progress=GitProgress())
|
|
repo.git.clear_cache()
|
|
repo.close()
|
|
|
|
def postinstall():
|
|
return self.execute_install_script(url, repo_path, instant_execution=instant_execution, no_deps=no_deps)
|
|
|
|
if return_postinstall:
|
|
return result.with_postinstall(postinstall)
|
|
else:
|
|
if not postinstall():
|
|
return result.fail(f"Failed to execute install script: {url}")
|
|
|
|
except Exception as e:
|
|
traceback.print_exc()
|
|
return result.fail(f"Install(git-clone) error[2]: {url} / {e}")
|
|
|
|
print("Installation was successful.")
|
|
return result
|
|
|
|
def repo_update(self, repo_path, instant_execution=False, no_deps=False, return_postinstall=False):
|
|
result = ManagedResult('update-git')
|
|
|
|
if not os.path.exists(os.path.join(repo_path, '.git')):
|
|
return result.fail(f'Path not found: {repo_path}')
|
|
|
|
# version check
|
|
repo = git.Repo(repo_path)
|
|
|
|
if repo.head.is_detached:
|
|
if not switch_to_default_branch(repo):
|
|
return result.fail(f"Failed to switch to default branch: {repo_path}")
|
|
|
|
current_branch = repo.active_branch
|
|
branch_name = current_branch.name
|
|
|
|
if current_branch.tracking_branch() is None:
|
|
print(f"[ComfyUI-Manager] There is no tracking branch ({current_branch})")
|
|
remote_name = get_remote_name(repo)
|
|
else:
|
|
remote_name = current_branch.tracking_branch().remote_name
|
|
|
|
if remote_name is None:
|
|
return result.fail(f"Failed to get remote when installing: {repo_path}")
|
|
|
|
remote = repo.remote(name=remote_name)
|
|
|
|
try:
|
|
remote.fetch()
|
|
except Exception as e:
|
|
if 'detected dubious' in str(e):
|
|
print(f"[ComfyUI-Manager] Try fixing 'dubious repository' error on '{repo_path}' repository")
|
|
safedir_path = repo_path.replace('\\', '/')
|
|
subprocess.run(['git', 'config', '--global', '--add', 'safe.directory', safedir_path])
|
|
try:
|
|
remote.fetch()
|
|
except Exception:
|
|
print("\n[ComfyUI-Manager] Failed to fixing repository setup. Please execute this command on cmd: \n"
|
|
"-----------------------------------------------------------------------------------------\n"
|
|
f'git config --global --add safe.directory "{safedir_path}"\n'
|
|
"-----------------------------------------------------------------------------------------\n")
|
|
|
|
commit_hash = repo.head.commit.hexsha
|
|
if f'{remote_name}/{branch_name}' in repo.refs:
|
|
remote_commit_hash = repo.refs[f'{remote_name}/{branch_name}'].object.hexsha
|
|
else:
|
|
return result.fail(f"Not updatable branch: {branch_name}")
|
|
|
|
if commit_hash != remote_commit_hash:
|
|
git_pull(repo_path)
|
|
|
|
if len(repo.remotes) > 0:
|
|
url = repo.remotes[0].url
|
|
else:
|
|
url = "unknown repo"
|
|
|
|
def postinstall():
|
|
return self.execute_install_script(url, repo_path, instant_execution=instant_execution, no_deps=no_deps)
|
|
|
|
if return_postinstall:
|
|
return result.with_postinstall(postinstall)
|
|
else:
|
|
if not postinstall():
|
|
return result.fail(f"Failed to execute install script: {url}")
|
|
|
|
return result
|
|
else:
|
|
return ManagedResult('skip').with_msg('Up to date')
|
|
|
|
def unified_update(self, node_id, version_spec=None, instant_execution=False, no_deps=False, return_postinstall=False):
|
|
orig_print(f"\x1b[2K\rUpdating: {node_id}", end='')
|
|
|
|
if version_spec is None:
|
|
version_spec = self.resolve_unspecified_version(node_id, guess_mode='active')
|
|
|
|
if version_spec is None:
|
|
return ManagedResult('update').fail(f'Update not available: {node_id}@{version_spec}').with_ver(version_spec)
|
|
|
|
if version_spec == 'nightly':
|
|
return self.repo_update(self.active_nodes[node_id][1], instant_execution=instant_execution, no_deps=no_deps, return_postinstall=return_postinstall).with_target('nightly').with_ver('nightly')
|
|
elif version_spec == 'unknown':
|
|
return self.repo_update(self.unknown_active_nodes[node_id][1], instant_execution=instant_execution, no_deps=no_deps, return_postinstall=return_postinstall).with_target('unknown').with_ver('unknown')
|
|
else:
|
|
return self.cnr_switch_version(node_id, instant_execution=instant_execution, no_deps=no_deps, return_postinstall=return_postinstall).with_ver('cnr')
|
|
|
|
async def install_by_id(self, node_id, version_spec=None, channel=None, mode=None, instant_execution=False, no_deps=False, return_postinstall=False):
|
|
"""
|
|
priority if version_spec == None
|
|
1. CNR latest
|
|
2. unknown
|
|
|
|
remark: latest version_spec is not allowed. Must be resolved before call.
|
|
"""
|
|
|
|
repo_url = None
|
|
if version_spec is None:
|
|
if self.is_enabled(node_id):
|
|
return ManagedResult('skip')
|
|
elif self.is_disabled(node_id):
|
|
return self.unified_enable(node_id)
|
|
else:
|
|
version_spec = self.resolve_unspecified_version(node_id)
|
|
|
|
if version_spec == 'unknown' or version_spec == 'nightly':
|
|
custom_nodes = await self.get_custom_nodes(channel, mode)
|
|
the_node = custom_nodes.get(node_id)
|
|
if the_node is not None:
|
|
if version_spec == 'unknown':
|
|
repo_url = the_node['files'][0]
|
|
else: # nightly
|
|
repo_url = the_node['repository']
|
|
else:
|
|
result = ManagedResult('install')
|
|
return result.fail(f"Node '{node_id}@{version_spec}' not found in [{channel}, {mode}]")
|
|
|
|
if self.is_enabled(node_id, version_spec):
|
|
return ManagedResult('skip').with_target(f"{node_id}@{version_spec}")
|
|
|
|
elif self.is_disabled(node_id, version_spec):
|
|
return self.unified_enable(node_id, version_spec)
|
|
|
|
elif version_spec == 'unknown' or version_spec == 'nightly':
|
|
if version_spec == 'nightly':
|
|
# disable cnr nodes
|
|
if self.is_enabled(node_id, 'cnr'):
|
|
self.unified_disable(node_id, False)
|
|
|
|
to_path = os.path.abspath(os.path.join(get_default_custom_nodes_path(), node_id))
|
|
res = self.repo_install(repo_url, to_path, instant_execution=instant_execution, no_deps=no_deps, return_postinstall=return_postinstall)
|
|
if res.result:
|
|
if version_spec == 'unknown':
|
|
self.unknown_active_nodes[node_id] = repo_url, to_path
|
|
elif version_spec == 'nightly':
|
|
cnr_utils.generate_cnr_id(to_path, node_id)
|
|
self.active_nodes[node_id] = 'nightly', to_path
|
|
else:
|
|
return res
|
|
|
|
return res.with_target(version_spec)
|
|
|
|
if self.is_enabled(node_id, 'nightly'):
|
|
# disable nightly nodes
|
|
self.unified_disable(node_id, False) # NOTE: don't return from here
|
|
|
|
if self.is_disabled(node_id, version_spec):
|
|
# enable and return if specified version is disabled
|
|
return self.unified_enable(node_id, version_spec)
|
|
|
|
if self.is_disabled(node_id, "cnr"):
|
|
# enable and switch version if cnr is disabled (not specified version)
|
|
self.unified_enable(node_id, "cnr")
|
|
return self.cnr_switch_version(node_id, version_spec, no_deps=no_deps, return_postinstall=return_postinstall)
|
|
|
|
if self.is_enabled(node_id, "cnr"):
|
|
return self.cnr_switch_version(node_id, version_spec, no_deps=no_deps, return_postinstall=return_postinstall)
|
|
|
|
res = self.cnr_install(node_id, version_spec, instant_execution=instant_execution, no_deps=no_deps, return_postinstall=return_postinstall)
|
|
if res.result:
|
|
self.active_nodes[node_id] = version_spec, res.to_path
|
|
|
|
return res
|
|
|
|
async def migrate_unmanaged_nodes(self):
|
|
"""
|
|
fix path for nightly and unknown nodes of unmanaged nodes
|
|
"""
|
|
await self.reload('cache')
|
|
await self.get_custom_nodes('default', 'cache')
|
|
|
|
print("Migration: STAGE 1")
|
|
moves = []
|
|
|
|
# migrate nightly inactive
|
|
for x, v in self.nightly_inactive_nodes.items():
|
|
if v.endswith('@nightly'):
|
|
continue
|
|
|
|
new_path = os.path.join(get_default_custom_nodes_path(), '.disabled', f"{x}@nightly")
|
|
moves.append((v, new_path))
|
|
|
|
self.reserve_migration(moves)
|
|
|
|
print("DONE (Migration reserved)")
|
|
|
|
|
|
unified_manager = UnifiedManager()
|
|
|
|
|
|
def identify_node_pack_from_path(fullpath):
|
|
module_name = os.path.basename(fullpath)
|
|
if module_name.endswith('.git'):
|
|
module_name = module_name[:-4]
|
|
|
|
repo_url = git_utils.git_url(fullpath)
|
|
if repo_url is None:
|
|
# cnr
|
|
cnr = cnr_utils.read_cnr_info(fullpath)
|
|
if cnr is not None:
|
|
return module_name, cnr['version'], cnr['id']
|
|
|
|
return None
|
|
else:
|
|
# nightly or unknown
|
|
cnr_id = cnr_utils.read_cnr_id(fullpath)
|
|
commit_hash = git_utils.get_commit_hash(fullpath)
|
|
|
|
if cnr_id is not None:
|
|
return module_name, commit_hash, cnr_id
|
|
else:
|
|
return module_name, commit_hash, ''
|
|
|
|
|
|
def get_installed_node_packs():
|
|
res = {}
|
|
|
|
for x in get_custom_nodes_paths():
|
|
for y in os.listdir(x):
|
|
if y == '__pycache__' or y == '.disabled':
|
|
continue
|
|
|
|
fullpath = os.path.join(x, y)
|
|
info = identify_node_pack_from_path(fullpath)
|
|
if info is None:
|
|
continue
|
|
|
|
is_disabled = not y.endswith('.disabled')
|
|
|
|
res[info[0]] = { 'ver': info[1], 'cnr_id': info[2], 'enabled': is_disabled }
|
|
|
|
disabled_dirs = os.path.join(x, '.disabled')
|
|
if os.path.exists(disabled_dirs):
|
|
for y in os.listdir(disabled_dirs):
|
|
if y == '__pycache__':
|
|
continue
|
|
|
|
fullpath = os.path.join(disabled_dirs, y)
|
|
info = identify_node_pack_from_path(fullpath)
|
|
if info is None:
|
|
continue
|
|
|
|
res[info[0]] = { 'ver': info[1], 'cnr_id': info[2], 'enabled': False }
|
|
|
|
return res
|
|
|
|
|
|
def get_channel_dict():
|
|
global channel_dict
|
|
|
|
if channel_dict is None:
|
|
channel_dict = {}
|
|
|
|
if not os.path.exists(manager_channel_list_path):
|
|
shutil.copy(channel_list_template_path, manager_channel_list_path)
|
|
|
|
with open(manager_channel_list_path, 'r') as file:
|
|
channels = file.read()
|
|
for x in channels.split('\n'):
|
|
channel_info = x.split("::")
|
|
if len(channel_info) == 2:
|
|
channel_dict[channel_info[0]] = channel_info[1]
|
|
|
|
return channel_dict
|
|
|
|
|
|
def get_channel_list():
|
|
global channel_list
|
|
|
|
if channel_list is None:
|
|
channel_list = []
|
|
for k, v in get_channel_dict().items():
|
|
channel_list.append(f"{k}::{v}")
|
|
|
|
return channel_list
|
|
|
|
|
|
class ManagerFuncs:
|
|
def __init__(self):
|
|
pass
|
|
|
|
def get_current_preview_method(self):
|
|
return "none"
|
|
|
|
def run_script(self, cmd, cwd='.'):
|
|
if len(cmd) > 0 and cmd[0].startswith("#"):
|
|
print(f"[ComfyUI-Manager] Unexpected behavior: `{cmd}`")
|
|
return 0
|
|
|
|
subprocess.check_call(cmd, cwd=cwd, env=get_script_env())
|
|
|
|
return 0
|
|
|
|
|
|
manager_funcs = ManagerFuncs()
|
|
|
|
|
|
def write_config():
|
|
config = configparser.ConfigParser()
|
|
|
|
config['default'] = {
|
|
'preview_method': manager_funcs.get_current_preview_method(),
|
|
'git_exe': get_config()['git_exe'],
|
|
'use_uv': get_config()['use_uv'],
|
|
'channel_url': get_config()['channel_url'],
|
|
'share_option': get_config()['share_option'],
|
|
'bypass_ssl': get_config()['bypass_ssl'],
|
|
"file_logging": get_config()['file_logging'],
|
|
'component_policy': get_config()['component_policy'],
|
|
'update_policy': get_config()['update_policy'],
|
|
'windows_selector_event_loop_policy': get_config()['windows_selector_event_loop_policy'],
|
|
'model_download_by_agent': get_config()['model_download_by_agent'],
|
|
'downgrade_blacklist': get_config()['downgrade_blacklist'],
|
|
'security_level': get_config()['security_level'],
|
|
'skip_migration_check': get_config()['skip_migration_check'],
|
|
'always_lazy_install': get_config()['always_lazy_install'],
|
|
'network_mode': get_config()['network_mode']
|
|
}
|
|
|
|
directory = os.path.dirname(manager_config_path)
|
|
if not os.path.exists(directory):
|
|
os.makedirs(directory)
|
|
|
|
with open(manager_config_path, 'w') as configfile:
|
|
config.write(configfile)
|
|
|
|
|
|
def read_config():
|
|
try:
|
|
config = configparser.ConfigParser()
|
|
config.read(manager_config_path)
|
|
default_conf = config['default']
|
|
manager_util.use_uv = default_conf['use_uv'].lower() == 'true' if 'use_uv' in default_conf else False
|
|
|
|
def get_bool(key, default_value):
|
|
return default_conf[key].lower() == 'true' if key in default_conf else False
|
|
|
|
return {
|
|
'http_channel_enabled': get_bool('http_channel_enabled', False),
|
|
'preview_method': default_conf.get('preview_method', manager_funcs.get_current_preview_method()).lower(),
|
|
'git_exe': default_conf.get('git_exe', ''),
|
|
'use_uv': get_bool('use_uv', False),
|
|
'channel_url': default_conf.get('channel_url', DEFAULT_CHANNEL),
|
|
'default_cache_as_channel_url': get_bool('default_cache_as_channel_url', False),
|
|
'share_option': default_conf.get('share_option', 'all').lower(),
|
|
'bypass_ssl': get_bool('bypass_ssl', False),
|
|
'file_logging': get_bool('file_logging', True),
|
|
'component_policy': default_conf.get('component_policy', 'workflow').lower(),
|
|
'update_policy': default_conf.get('update_policy', 'stable-comfyui').lower(),
|
|
'windows_selector_event_loop_policy': get_bool('windows_selector_event_loop_policy', False),
|
|
'model_download_by_agent': get_bool('model_download_by_agent', False),
|
|
'downgrade_blacklist': default_conf.get('downgrade_blacklist', '').lower(),
|
|
'skip_migration_check': get_bool('skip_migration_check', False),
|
|
'always_lazy_install': get_bool('always_lazy_install', False),
|
|
'network_mode': default_conf.get('network_mode', 'public').lower(),
|
|
'security_level': default_conf.get('security_level', 'normal').lower(),
|
|
}
|
|
|
|
except Exception:
|
|
manager_util.use_uv = False
|
|
return {
|
|
'http_channel_enabled': False,
|
|
'preview_method': manager_funcs.get_current_preview_method(),
|
|
'git_exe': '',
|
|
'use_uv': False,
|
|
'channel_url': DEFAULT_CHANNEL,
|
|
'default_cache_as_channel_url': False,
|
|
'share_option': 'all',
|
|
'bypass_ssl': False,
|
|
'file_logging': True,
|
|
'component_policy': 'workflow',
|
|
'update_policy': 'stable-comfyui',
|
|
'windows_selector_event_loop_policy': False,
|
|
'model_download_by_agent': False,
|
|
'downgrade_blacklist': '',
|
|
'skip_migration_check': False,
|
|
'always_lazy_install': False,
|
|
'network_mode': 'public', # public | private | offline
|
|
'security_level': 'normal', # strong | normal | normal- | weak
|
|
}
|
|
|
|
|
|
def get_config():
|
|
global cached_config
|
|
|
|
if cached_config is None:
|
|
cached_config = read_config()
|
|
if cached_config['http_channel_enabled']:
|
|
print("[ComfyUI-Manager] Warning: http channel enabled, make sure server in secure env")
|
|
|
|
return cached_config
|
|
|
|
|
|
def get_remote_name(repo):
|
|
available_remotes = [remote.name for remote in repo.remotes]
|
|
if 'origin' in available_remotes:
|
|
return 'origin'
|
|
elif 'upstream' in available_remotes:
|
|
return 'upstream'
|
|
elif len(available_remotes) > 0:
|
|
return available_remotes[0]
|
|
|
|
if not available_remotes:
|
|
print(f"[ComfyUI-Manager] No remotes are configured for this repository: {repo.working_dir}")
|
|
else:
|
|
print(f"[ComfyUI-Manager] Available remotes in '{repo.working_dir}': ")
|
|
for remote in available_remotes:
|
|
print(f"- {remote}")
|
|
|
|
return None
|
|
|
|
|
|
def switch_to_default_branch(repo):
|
|
remote_name = get_remote_name(repo)
|
|
|
|
try:
|
|
if remote_name is None:
|
|
return False
|
|
|
|
default_branch = repo.git.symbolic_ref(f'refs/remotes/{remote_name}/HEAD').replace(f'refs/remotes/{remote_name}/', '')
|
|
repo.git.checkout(default_branch)
|
|
return True
|
|
except:
|
|
# try checkout master
|
|
# try checkout main if failed
|
|
try:
|
|
repo.git.checkout(repo.heads.master)
|
|
return True
|
|
except:
|
|
try:
|
|
if remote_name is not None:
|
|
repo.git.checkout('-b', 'master', f'{remote_name}/master')
|
|
return True
|
|
except:
|
|
try:
|
|
repo.git.checkout(repo.heads.main)
|
|
return True
|
|
except:
|
|
try:
|
|
if remote_name is not None:
|
|
repo.git.checkout('-b', 'main', f'{remote_name}/main')
|
|
return True
|
|
except:
|
|
pass
|
|
|
|
print("[ComfyUI Manager] Failed to switch to the default branch")
|
|
return False
|
|
|
|
|
|
def try_install_script(url, repo_path, install_cmd, instant_execution=False):
|
|
if not instant_execution and (
|
|
(len(install_cmd) > 0 and install_cmd[0].startswith('#')) or platform.system() == "Windows" or get_config()['always_lazy_install']
|
|
):
|
|
if not os.path.exists(manager_startup_script_path):
|
|
os.makedirs(manager_startup_script_path)
|
|
|
|
script_path = os.path.join(manager_startup_script_path, "install-scripts.txt")
|
|
with open(script_path, "a") as file:
|
|
obj = [repo_path] + install_cmd
|
|
file.write(f"{obj}\n")
|
|
|
|
return True
|
|
else:
|
|
if len(install_cmd) == 5 and install_cmd[2:4] == ['pip', 'install']:
|
|
if is_blacklisted(install_cmd[4]):
|
|
print(f"[ComfyUI-Manager] skip black listed pip installation: '{install_cmd[4]}'")
|
|
return True
|
|
elif len(install_cmd) == 6 and install_cmd[3:5] == ['pip', 'install']: # uv mode
|
|
if is_blacklisted(install_cmd[5]):
|
|
print(f"[ComfyUI-Manager] skip black listed pip installation: '{install_cmd[5]}'")
|
|
return True
|
|
|
|
print(f"\n## ComfyUI-Manager: EXECUTE => {install_cmd}")
|
|
code = manager_funcs.run_script(install_cmd, cwd=repo_path)
|
|
|
|
if platform.system() != "Windows":
|
|
try:
|
|
if not os.environ.get('__COMFYUI_DESKTOP_VERSION__') and comfy_ui_commit_datetime.date() < comfy_ui_required_commit_datetime.date():
|
|
print("\n\n###################################################################")
|
|
print(f"[WARN] ComfyUI-Manager: Your ComfyUI version ({comfy_ui_revision})[{comfy_ui_commit_datetime.date()}] is too old. Please update to the latest version.")
|
|
print("[WARN] The extension installation feature may not work properly in the current installed ComfyUI version on Windows environment.")
|
|
print("###################################################################\n\n")
|
|
except:
|
|
pass
|
|
|
|
if code != 0:
|
|
if url is None:
|
|
url = os.path.dirname(repo_path)
|
|
print(f"install script failed: {url}")
|
|
return False
|
|
|
|
return True
|
|
|
|
|
|
# use subprocess to avoid file system lock by git (Windows)
|
|
def __win_check_git_update(path, do_fetch=False, do_update=False):
|
|
if do_fetch:
|
|
command = [sys.executable, git_script_path, "--fetch", path]
|
|
elif do_update:
|
|
command = [sys.executable, git_script_path, "--pull", path]
|
|
else:
|
|
command = [sys.executable, git_script_path, "--check", path]
|
|
|
|
new_env = get_script_env()
|
|
process = subprocess.Popen(command, stdout=subprocess.PIPE, stderr=subprocess.PIPE, cwd=get_default_custom_nodes_path(), env=new_env)
|
|
output, _ = process.communicate()
|
|
output = output.decode('utf-8').strip()
|
|
|
|
if 'detected dubious' in output:
|
|
# fix and try again
|
|
safedir_path = path.replace('\\', '/')
|
|
try:
|
|
print(f"[ComfyUI-Manager] Try fixing 'dubious repository' error on '{safedir_path}' repo")
|
|
process = subprocess.Popen(['git', 'config', '--global', '--add', 'safe.directory', safedir_path], env=new_env, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
|
|
output, _ = process.communicate()
|
|
|
|
process = subprocess.Popen(command, env=new_env, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
|
|
output, _ = process.communicate()
|
|
output = output.decode('utf-8').strip()
|
|
except Exception:
|
|
print('[ComfyUI-Manager] failed to fixing')
|
|
|
|
if 'detected dubious' in output:
|
|
print(f'\n[ComfyUI-Manager] Failed to fixing repository setup. Please execute this command on cmd: \n'
|
|
f'-----------------------------------------------------------------------------------------\n'
|
|
f'git config --global --add safe.directory "{safedir_path}"\n'
|
|
f'-----------------------------------------------------------------------------------------\n')
|
|
|
|
if do_update:
|
|
if "CUSTOM NODE PULL: Success" in output:
|
|
process.wait()
|
|
print(f"\x1b[2K\rUpdated: {path}")
|
|
return True, True # updated
|
|
elif "CUSTOM NODE PULL: None" in output:
|
|
process.wait()
|
|
return False, True # there is no update
|
|
else:
|
|
print(f"\x1b[2K\rUpdate error: {path}")
|
|
process.wait()
|
|
return False, False # update failed
|
|
else:
|
|
if "CUSTOM NODE CHECK: True" in output:
|
|
process.wait()
|
|
return True, True
|
|
elif "CUSTOM NODE CHECK: False" in output:
|
|
process.wait()
|
|
return False, True
|
|
else:
|
|
print(f"\x1b[2K\rFetch error: {path}")
|
|
print(f"\n{output}\n")
|
|
process.wait()
|
|
return False, True
|
|
|
|
|
|
def __win_check_git_pull(path):
|
|
command = [sys.executable, git_script_path, "--pull", path]
|
|
process = subprocess.Popen(command, env=get_script_env(), cwd=get_default_custom_nodes_path())
|
|
process.wait()
|
|
|
|
|
|
def execute_install_script(url, repo_path, lazy_mode=False, instant_execution=False, no_deps=False):
|
|
# import ipdb; ipdb.set_trace()
|
|
install_script_path = os.path.join(repo_path, "install.py")
|
|
requirements_path = os.path.join(repo_path, "requirements.txt")
|
|
|
|
if lazy_mode:
|
|
install_cmd = ["#LAZY-INSTALL-SCRIPT", sys.executable]
|
|
try_install_script(url, repo_path, install_cmd)
|
|
else:
|
|
if os.path.exists(requirements_path) and not no_deps:
|
|
print("Install: pip packages")
|
|
pip_fixer = manager_util.PIPFixer(manager_util.get_installed_packages())
|
|
with open(requirements_path, "r") as requirements_file:
|
|
for line in requirements_file:
|
|
#handle comments
|
|
if '#' in line:
|
|
if line.strip()[0] == '#':
|
|
print("Line is comment...skipping")
|
|
continue
|
|
else:
|
|
line = line.split('#')[0].strip()
|
|
|
|
package_name = remap_pip_package(line.strip())
|
|
|
|
if package_name and not package_name.startswith('#'):
|
|
if '--index-url' in package_name:
|
|
s = package_name.split('--index-url')
|
|
install_cmd = manager_util.make_pip_cmd(["install", s[0].strip(), '--index-url', s[1].strip()])
|
|
else:
|
|
install_cmd = manager_util.make_pip_cmd(["install", package_name])
|
|
|
|
if package_name.strip() != "" and not package_name.startswith('#'):
|
|
try_install_script(url, repo_path, install_cmd, instant_execution=instant_execution)
|
|
pip_fixer.fix_broken()
|
|
|
|
if os.path.exists(install_script_path):
|
|
print("Install: install script")
|
|
install_cmd = [sys.executable, "install.py"]
|
|
try_install_script(url, repo_path, install_cmd, instant_execution=instant_execution)
|
|
|
|
return True
|
|
|
|
|
|
def git_repo_update_check_with(path, do_fetch=False, do_update=False, no_deps=False):
|
|
"""
|
|
|
|
perform update check for git custom node
|
|
and fetch or update if flag is on
|
|
|
|
:param path: path to git custom node
|
|
:param do_fetch: do fetch during check
|
|
:param do_update: do update during check
|
|
:param no_deps: don't install dependencies
|
|
:return: update state * success
|
|
"""
|
|
if do_fetch:
|
|
orig_print(f"\x1b[2K\rFetching: {path}", end='')
|
|
elif do_update:
|
|
orig_print(f"\x1b[2K\rUpdating: {path}", end='')
|
|
|
|
# Check if the path is a git repository
|
|
if not os.path.exists(os.path.join(path, '.git')):
|
|
raise ValueError(f'[ComfyUI-Manager] Not a valid git repository: {path}')
|
|
|
|
if platform.system() == "Windows":
|
|
updated, success = __win_check_git_update(path, do_fetch, do_update)
|
|
if updated and success:
|
|
execute_install_script(None, path, lazy_mode=True, no_deps=no_deps)
|
|
return updated, success
|
|
else:
|
|
# Fetch the latest commits from the remote repository
|
|
repo = git.Repo(path)
|
|
|
|
remote_name = get_remote_name(repo)
|
|
|
|
if remote_name is None:
|
|
raise ValueError(f"No remotes are configured for this repository: {path}")
|
|
|
|
remote = repo.remote(name=remote_name)
|
|
|
|
if not do_update and repo.head.is_detached:
|
|
if do_fetch:
|
|
remote.fetch()
|
|
|
|
return True, True # detached branch is treated as updatable
|
|
|
|
if repo.head.is_detached:
|
|
if not switch_to_default_branch(repo):
|
|
raise ValueError(f"Failed to switch detached branch to default branch: {path}")
|
|
|
|
current_branch = repo.active_branch
|
|
branch_name = current_branch.name
|
|
|
|
# Get the current commit hash
|
|
commit_hash = repo.head.commit.hexsha
|
|
|
|
if do_fetch or do_update:
|
|
remote.fetch()
|
|
|
|
if do_update:
|
|
if repo.is_dirty():
|
|
print(f"\nSTASH: '{path}' is dirty.")
|
|
repo.git.stash()
|
|
|
|
if f'{remote_name}/{branch_name}' not in repo.refs:
|
|
if not switch_to_default_branch(repo):
|
|
raise ValueError(f"Failed to switch to default branch while updating: {path}")
|
|
|
|
current_branch = repo.active_branch
|
|
branch_name = current_branch.name
|
|
|
|
if f'{remote_name}/{branch_name}' in repo.refs:
|
|
remote_commit_hash = repo.refs[f'{remote_name}/{branch_name}'].object.hexsha
|
|
else:
|
|
return False, False
|
|
|
|
if commit_hash == remote_commit_hash:
|
|
repo.close()
|
|
return False, True
|
|
|
|
try:
|
|
remote.pull()
|
|
repo.git.submodule('update', '--init', '--recursive')
|
|
new_commit_hash = repo.head.commit.hexsha
|
|
|
|
if commit_hash != new_commit_hash:
|
|
execute_install_script(None, path, no_deps=no_deps)
|
|
print(f"\x1b[2K\rUpdated: {path}")
|
|
return True, True
|
|
else:
|
|
return False, False
|
|
|
|
except Exception as e:
|
|
print(f"\nUpdating failed: {path}\n{e}", file=sys.stderr)
|
|
return False, False
|
|
|
|
if repo.head.is_detached:
|
|
repo.close()
|
|
return True, True
|
|
|
|
# Get commit hash of the remote branch
|
|
current_branch = repo.active_branch
|
|
branch_name = current_branch.name
|
|
|
|
if f'{remote_name}/{branch_name}' in repo.refs:
|
|
remote_commit_hash = repo.refs[f'{remote_name}/{branch_name}'].object.hexsha
|
|
else:
|
|
return True, True # Assuming there's an update if it's not the default branch.
|
|
|
|
# Compare the commit hashes to determine if the local repository is behind the remote repository
|
|
if commit_hash != remote_commit_hash:
|
|
# Get the commit dates
|
|
commit_date = repo.head.commit.committed_datetime
|
|
remote_commit_date = repo.refs[f'{remote_name}/{branch_name}'].object.committed_datetime
|
|
|
|
# Compare the commit dates to determine if the local repository is behind the remote repository
|
|
if commit_date < remote_commit_date:
|
|
repo.close()
|
|
return True, True
|
|
|
|
repo.close()
|
|
|
|
return False, True
|
|
|
|
|
|
class GitProgress(RemoteProgress):
|
|
def __init__(self):
|
|
super().__init__()
|
|
self.pbar = tqdm()
|
|
|
|
def update(self, op_code, cur_count, max_count=None, message=''):
|
|
self.pbar.total = max_count
|
|
self.pbar.n = cur_count
|
|
self.pbar.pos = 0
|
|
self.pbar.refresh()
|
|
|
|
|
|
def is_valid_url(url):
|
|
try:
|
|
# Check for HTTP/HTTPS URL format
|
|
result = urlparse(url)
|
|
if all([result.scheme, result.netloc]):
|
|
return True
|
|
finally:
|
|
# Check for SSH git URL format
|
|
pattern = re.compile(r"^(.+@|ssh://).+:.+$")
|
|
if pattern.match(url):
|
|
return True
|
|
return False
|
|
|
|
|
|
async def gitclone_install(url, instant_execution=False, msg_prefix='', no_deps=False):
|
|
await unified_manager.reload('cache')
|
|
await unified_manager.get_custom_nodes('default', 'cache')
|
|
|
|
print(f"{msg_prefix}Install: {url}")
|
|
|
|
result = ManagedResult('install-git')
|
|
|
|
if not is_valid_url(url):
|
|
return result.fail(f"Invalid git url: '{url}'")
|
|
|
|
if url.endswith("/"):
|
|
url = url[:-1]
|
|
try:
|
|
cnr = unified_manager.get_cnr_by_repo(url)
|
|
if cnr:
|
|
cnr_id = cnr['id']
|
|
return await unified_manager.install_by_id(cnr_id, version_spec='nightly')
|
|
else:
|
|
repo_name = os.path.splitext(os.path.basename(url))[0]
|
|
|
|
# NOTE: Keep original name as possible if unknown node
|
|
# node_dir = f"{repo_name}@unknown"
|
|
node_dir = repo_name
|
|
|
|
repo_path = os.path.join(get_default_custom_nodes_path(), node_dir)
|
|
|
|
if os.path.exists(repo_path):
|
|
return result.fail(f"Already exists: '{repo_path}'")
|
|
|
|
for custom_nodes_dir in get_custom_nodes_paths():
|
|
disabled_repo_path1 = os.path.join(custom_nodes_dir, '.disabled', node_dir)
|
|
disabled_repo_path2 = os.path.join(custom_nodes_dir, repo_name+'.disabled') # old style
|
|
|
|
if os.path.exists(disabled_repo_path1):
|
|
return result.fail(f"Already exists (disabled): '{disabled_repo_path1}'")
|
|
|
|
if os.path.exists(disabled_repo_path2):
|
|
return result.fail(f"Already exists (disabled): '{disabled_repo_path2}'")
|
|
|
|
print(f"CLONE into '{repo_path}'")
|
|
|
|
# Clone the repository from the remote URL
|
|
clone_url = git_utils.get_url_for_clone(url)
|
|
|
|
if not instant_execution and platform.system() == 'Windows':
|
|
res = manager_funcs.run_script([sys.executable, git_script_path, "--clone", get_default_custom_nodes_path(), clone_url, repo_path], cwd=get_default_custom_nodes_path())
|
|
if res != 0:
|
|
return result.fail(f"Failed to clone '{clone_url}' into '{repo_path}'")
|
|
else:
|
|
repo = git.Repo.clone_from(clone_url, repo_path, recursive=True, progress=GitProgress())
|
|
repo.git.clear_cache()
|
|
repo.close()
|
|
|
|
execute_install_script(url, repo_path, instant_execution=instant_execution, no_deps=no_deps)
|
|
print("Installation was successful.")
|
|
return result.with_target(repo_path)
|
|
|
|
except Exception as e:
|
|
traceback.print_exc()
|
|
print(f"Install(git-clone) error[1]: {url} / {e}", file=sys.stderr)
|
|
return result.fail(f"Install(git-clone)[1] error: {url} / {e}")
|
|
|
|
|
|
def git_pull(path):
|
|
# Check if the path is a git repository
|
|
if not os.path.exists(os.path.join(path, '.git')):
|
|
raise ValueError('Not a git repository')
|
|
|
|
# Pull the latest changes from the remote repository
|
|
if platform.system() == "Windows":
|
|
return __win_check_git_pull(path)
|
|
else:
|
|
repo = git.Repo(path)
|
|
|
|
if repo.is_dirty():
|
|
print(f"STASH: '{path}' is dirty.")
|
|
repo.git.stash()
|
|
|
|
if repo.head.is_detached:
|
|
if not switch_to_default_branch(repo):
|
|
raise ValueError(f"Failed to switch to default branch while pulling: {path}")
|
|
|
|
current_branch = repo.active_branch
|
|
remote_name = current_branch.tracking_branch().remote_name
|
|
remote = repo.remote(name=remote_name)
|
|
|
|
remote.pull()
|
|
repo.git.submodule('update', '--init', '--recursive')
|
|
|
|
repo.close()
|
|
|
|
return True
|
|
|
|
|
|
async def get_data_by_mode(mode, filename, channel_url=None):
|
|
if channel_url in get_channel_dict():
|
|
channel_url = get_channel_dict()[channel_url]
|
|
|
|
try:
|
|
local_uri = os.path.join(manager_util.comfyui_manager_path, filename)
|
|
|
|
if mode == "local":
|
|
json_obj = await manager_util.get_data(local_uri)
|
|
else:
|
|
if channel_url is None:
|
|
uri = get_config()['channel_url'] + '/' + filename
|
|
else:
|
|
uri = channel_url + '/' + filename
|
|
|
|
cache_uri = str(manager_util.simple_hash(uri))+'_'+filename
|
|
cache_uri = os.path.join(manager_util.cache_dir, cache_uri)
|
|
|
|
if get_config()['network_mode'] == 'offline':
|
|
# offline network mode
|
|
if os.path.exists(cache_uri):
|
|
json_obj = await manager_util.get_data(cache_uri)
|
|
else:
|
|
local_uri = os.path.join(manager_util.comfyui_manager_path, filename)
|
|
if os.path.exists(local_uri):
|
|
json_obj = await manager_util.get_data(local_uri)
|
|
else:
|
|
json_obj = {} # fallback
|
|
else:
|
|
# public network mode
|
|
if mode == "cache" and manager_util.is_file_created_within_one_day(cache_uri):
|
|
json_obj = await manager_util.get_data(cache_uri)
|
|
else:
|
|
json_obj = await manager_util.get_data(uri)
|
|
with manager_util.cache_lock:
|
|
with open(cache_uri, "w", encoding='utf-8') as file:
|
|
json.dump(json_obj, file, indent=4, sort_keys=True)
|
|
except Exception as e:
|
|
print(f"[ComfyUI-Manager] Due to a network error, switching to local mode.\n=> {filename}\n=> {e}")
|
|
uri = os.path.join(manager_util.comfyui_manager_path, filename)
|
|
json_obj = await manager_util.get_data(uri)
|
|
|
|
return json_obj
|
|
|
|
|
|
def gitclone_fix(files, instant_execution=False, no_deps=False):
|
|
print(f"Try fixing: {files}")
|
|
for url in files:
|
|
if not is_valid_url(url):
|
|
print(f"Invalid git url: '{url}'")
|
|
return False
|
|
|
|
if url.endswith("/"):
|
|
url = url[:-1]
|
|
try:
|
|
repo_name = os.path.splitext(os.path.basename(url))[0]
|
|
repo_path = os.path.join(get_default_custom_nodes_path(), repo_name)
|
|
|
|
if os.path.exists(repo_path+'.disabled'):
|
|
repo_path = repo_path+'.disabled'
|
|
|
|
if not execute_install_script(url, repo_path, instant_execution=instant_execution, no_deps=no_deps):
|
|
return False
|
|
|
|
except Exception as e:
|
|
print(f"Fix(git-clone) error: {url} / {e}", file=sys.stderr)
|
|
return False
|
|
|
|
print(f"Attempt to fixing '{files}' is done.")
|
|
return True
|
|
|
|
|
|
def pip_install(packages):
|
|
install_cmd = ['#FORCE'] + manager_util.make_pip_cmd(["install", '-U']) + packages
|
|
try_install_script('pip install via manager', '..', install_cmd)
|
|
|
|
|
|
def rmtree(path):
|
|
retry_count = 3
|
|
|
|
while True:
|
|
try:
|
|
retry_count -= 1
|
|
|
|
if platform.system() == "Windows":
|
|
manager_funcs.run_script(['attrib', '-R', path + '\\*', '/S'])
|
|
shutil.rmtree(path)
|
|
|
|
return True
|
|
|
|
except Exception as ex:
|
|
print(f"ex: {ex}")
|
|
time.sleep(3)
|
|
|
|
if retry_count < 0:
|
|
raise ex
|
|
|
|
print(f"Uninstall retry({retry_count})")
|
|
|
|
|
|
def gitclone_uninstall(files):
|
|
import os
|
|
|
|
print(f"Uninstall: {files}")
|
|
for url in files:
|
|
if url.endswith("/"):
|
|
url = url[:-1]
|
|
try:
|
|
for custom_nodes_dir in get_custom_nodes_paths():
|
|
dir_name = os.path.splitext(os.path.basename(url))[0].replace(".git", "")
|
|
dir_path = os.path.join(custom_nodes_dir, dir_name)
|
|
|
|
# safety check
|
|
if dir_path == '/' or dir_path[1:] == ":/" or dir_path == '':
|
|
print(f"Uninstall(git-clone) error: invalid path '{dir_path}' for '{url}'")
|
|
return False
|
|
|
|
install_script_path = os.path.join(dir_path, "uninstall.py")
|
|
disable_script_path = os.path.join(dir_path, "disable.py")
|
|
if os.path.exists(install_script_path):
|
|
uninstall_cmd = [sys.executable, "uninstall.py"]
|
|
code = manager_funcs.run_script(uninstall_cmd, cwd=dir_path)
|
|
|
|
if code != 0:
|
|
print(f"An error occurred during the execution of the uninstall.py script. Only the '{dir_path}' will be deleted.")
|
|
elif os.path.exists(disable_script_path):
|
|
disable_script = [sys.executable, "disable.py"]
|
|
code = manager_funcs.run_script(disable_script, cwd=dir_path)
|
|
if code != 0:
|
|
print(f"An error occurred during the execution of the disable.py script. Only the '{dir_path}' will be deleted.")
|
|
|
|
if os.path.exists(dir_path):
|
|
rmtree(dir_path)
|
|
elif os.path.exists(dir_path + ".disabled"):
|
|
rmtree(dir_path + ".disabled")
|
|
except Exception as e:
|
|
print(f"Uninstall(git-clone) error: {url} / {e}", file=sys.stderr)
|
|
return False
|
|
|
|
print("Uninstallation was successful.")
|
|
return True
|
|
|
|
|
|
def gitclone_set_active(files, is_disable):
|
|
import os
|
|
|
|
if is_disable:
|
|
action_name = "Disable"
|
|
else:
|
|
action_name = "Enable"
|
|
|
|
print(f"{action_name}: {files}")
|
|
for url in files:
|
|
if url.endswith("/"):
|
|
url = url[:-1]
|
|
try:
|
|
for custom_nodes_dir in get_custom_nodes_paths():
|
|
dir_name = os.path.splitext(os.path.basename(url))[0].replace(".git", "")
|
|
dir_path = os.path.join(custom_nodes_dir, dir_name)
|
|
|
|
# safety check
|
|
if dir_path == '/' or dir_path[1:] == ":/" or dir_path == '':
|
|
print(f"{action_name}(git-clone) error: invalid path '{dir_path}' for '{url}'")
|
|
return False
|
|
|
|
if is_disable:
|
|
current_path = dir_path
|
|
base_path = extract_base_custom_nodes_dir(current_path)
|
|
new_path = os.path.join(base_path, ".disabled", dir_name)
|
|
|
|
if not os.path.exists(current_path):
|
|
continue
|
|
else:
|
|
current_path1 = os.path.join(get_default_custom_nodes_path(), ".disabled", dir_name)
|
|
current_path2 = dir_path + ".disabled"
|
|
|
|
if os.path.exists(current_path1):
|
|
current_path = current_path1
|
|
elif os.path.exists(current_path2):
|
|
current_path = current_path2
|
|
else:
|
|
continue
|
|
|
|
base_path = extract_base_custom_nodes_dir(current_path)
|
|
new_path = os.path.join(base_path, dir_name)
|
|
|
|
shutil.move(current_path, new_path)
|
|
|
|
if is_disable:
|
|
if os.path.exists(os.path.join(new_path, "disable.py")):
|
|
disable_script = [sys.executable, "disable.py"]
|
|
try_install_script(url, new_path, disable_script)
|
|
else:
|
|
if os.path.exists(os.path.join(new_path, "enable.py")):
|
|
enable_script = [sys.executable, "enable.py"]
|
|
try_install_script(url, new_path, enable_script)
|
|
|
|
break # for safety
|
|
|
|
except Exception as e:
|
|
print(f"{action_name}(git-clone) error: {url} / {e}", file=sys.stderr)
|
|
return False
|
|
|
|
print(f"{action_name} was successful.")
|
|
return True
|
|
|
|
|
|
def gitclone_update(files, instant_execution=False, skip_script=False, msg_prefix="", no_deps=False):
|
|
import os
|
|
|
|
print(f"{msg_prefix}Update: {files}")
|
|
for url in files:
|
|
if url.endswith("/"):
|
|
url = url[:-1]
|
|
try:
|
|
for custom_nodes_dir in get_default_custom_nodes_path():
|
|
repo_name = os.path.splitext(os.path.basename(url))[0].replace(".git", "")
|
|
repo_path = os.path.join(custom_nodes_dir, repo_name)
|
|
|
|
if os.path.exists(repo_path+'.disabled'):
|
|
repo_path = repo_path+'.disabled'
|
|
|
|
elif os.path.exists(os.path.join(get_default_custom_nodes_path(), "disabled", repo_name)):
|
|
repo_path = os.path.join(get_default_custom_nodes_path(), "disabled", repo_name)
|
|
|
|
if not os.path.exists(repo_path):
|
|
continue
|
|
|
|
git_pull(repo_path)
|
|
|
|
if not skip_script:
|
|
if instant_execution:
|
|
if not execute_install_script(url, repo_path, lazy_mode=False, instant_execution=True, no_deps=no_deps):
|
|
return False
|
|
else:
|
|
if not execute_install_script(url, repo_path, lazy_mode=True, no_deps=no_deps):
|
|
return False
|
|
|
|
break # for safety
|
|
|
|
except Exception as e:
|
|
print(f"Update(git-clone) error: {url} / {e}", file=sys.stderr)
|
|
return False
|
|
|
|
if not skip_script:
|
|
print("Update was successful.")
|
|
return True
|
|
|
|
|
|
def update_to_stable_comfyui(repo_path):
|
|
try:
|
|
repo = git.Repo(repo_path)
|
|
repo.git.checkout(repo.heads.master)
|
|
versions, current_tag, _ = get_comfyui_versions(repo)
|
|
|
|
if len(versions) == 0 or (len(versions) == 1 and versions[0] == 'nightly'):
|
|
logging.info("[ComfyUI-Manager] Unable to update to the stable ComfyUI version.")
|
|
return "fail", None
|
|
|
|
if versions[0] == 'nightly':
|
|
latest_tag = versions[1]
|
|
else:
|
|
latest_tag = versions[0]
|
|
|
|
if current_tag == latest_tag:
|
|
return "skip", None
|
|
else:
|
|
logging.info(f"[ComfyUI-Manager] Updating ComfyUI: {current_tag} -> {latest_tag}")
|
|
repo.git.checkout(latest_tag)
|
|
return 'updated', latest_tag
|
|
except:
|
|
traceback.print_exc()
|
|
return "fail", None
|
|
|
|
|
|
def update_path(repo_path, instant_execution=False, no_deps=False):
|
|
if not os.path.exists(os.path.join(repo_path, '.git')):
|
|
return "fail"
|
|
|
|
# version check
|
|
repo = git.Repo(repo_path)
|
|
|
|
is_switched = False
|
|
if repo.head.is_detached:
|
|
if not switch_to_default_branch(repo):
|
|
return "fail"
|
|
else:
|
|
is_switched = True
|
|
|
|
current_branch = repo.active_branch
|
|
branch_name = current_branch.name
|
|
|
|
if current_branch.tracking_branch() is None:
|
|
print(f"[ComfyUI-Manager] There is no tracking branch ({current_branch})")
|
|
remote_name = get_remote_name(repo)
|
|
else:
|
|
remote_name = current_branch.tracking_branch().remote_name
|
|
remote = repo.remote(name=remote_name)
|
|
|
|
try:
|
|
remote.fetch()
|
|
except Exception as e:
|
|
if 'detected dubious' in str(e):
|
|
print(f"[ComfyUI-Manager] Try fixing 'dubious repository' error on '{repo_path}' repository")
|
|
safedir_path = repo_path.replace('\\', '/')
|
|
subprocess.run(['git', 'config', '--global', '--add', 'safe.directory', safedir_path])
|
|
try:
|
|
remote.fetch()
|
|
except Exception:
|
|
print(f"\n[ComfyUI-Manager] Failed to fixing repository setup. Please execute this command on cmd: \n"
|
|
f"-----------------------------------------------------------------------------------------\n"
|
|
f'git config --global --add safe.directory "{safedir_path}"\n'
|
|
f"-----------------------------------------------------------------------------------------\n")
|
|
return "fail"
|
|
|
|
commit_hash = repo.head.commit.hexsha
|
|
|
|
if f'{remote_name}/{branch_name}' in repo.refs:
|
|
remote_commit_hash = repo.refs[f'{remote_name}/{branch_name}'].object.hexsha
|
|
else:
|
|
return "fail"
|
|
|
|
if commit_hash != remote_commit_hash:
|
|
git_pull(repo_path)
|
|
execute_install_script("ComfyUI", repo_path, instant_execution=instant_execution, no_deps=no_deps)
|
|
return "updated"
|
|
elif is_switched:
|
|
return "updated"
|
|
else:
|
|
return "skipped"
|
|
|
|
|
|
def lookup_customnode_by_url(data, target):
|
|
for x in data['custom_nodes']:
|
|
if target in x['files']:
|
|
for custom_nodes_dir in get_custom_nodes_paths():
|
|
dir_name = os.path.splitext(os.path.basename(target))[0].replace(".git", "")
|
|
dir_path = os.path.join(custom_nodes_dir, dir_name)
|
|
if os.path.exists(dir_path):
|
|
x['installed'] = 'True'
|
|
else:
|
|
disabled_path1 = os.path.join(custom_nodes_dir, '.disabled', dir_name)
|
|
disabled_path2 = dir_path + ".disabled"
|
|
|
|
if os.path.exists(disabled_path1) or os.path.exists(disabled_path2):
|
|
x['installed'] = 'Disabled'
|
|
else:
|
|
continue
|
|
|
|
return x
|
|
|
|
return None
|
|
|
|
|
|
def lookup_installed_custom_nodes_legacy(repo_name):
|
|
base_paths = get_custom_nodes_paths()
|
|
|
|
for base_path in base_paths:
|
|
repo_path = os.path.join(base_path, repo_name)
|
|
if os.path.exists(repo_path):
|
|
return True, repo_path
|
|
elif os.path.exists(repo_path + '.disabled'):
|
|
return False, repo_path
|
|
|
|
return None
|
|
|
|
|
|
def simple_check_custom_node(url):
|
|
dir_name = os.path.splitext(os.path.basename(url))[0].replace(".git", "")
|
|
dir_path = os.path.join(get_default_custom_nodes_path(), dir_name)
|
|
if os.path.exists(dir_path):
|
|
return 'installed'
|
|
elif os.path.exists(dir_path+'.disabled'):
|
|
return 'disabled'
|
|
|
|
return 'not-installed'
|
|
|
|
|
|
def check_state_of_git_node_pack_single(item, do_fetch=False, do_update_check=True, do_update=False):
|
|
if item['version'] == 'unknown':
|
|
dir_path = unified_manager.unknown_active_nodes.get(item['id'])[1]
|
|
elif item['version'] == 'nightly':
|
|
dir_path = unified_manager.active_nodes.get(item['id'])[1]
|
|
else:
|
|
# skip CNR nodes
|
|
dir_path = None
|
|
|
|
if dir_path and os.path.exists(dir_path):
|
|
if do_update_check:
|
|
try:
|
|
update_state, success = git_repo_update_check_with(dir_path, do_fetch, do_update)
|
|
if (do_update_check or do_update) and update_state:
|
|
item['update-state'] = 'true'
|
|
elif do_update and not success:
|
|
item['update-state'] = 'fail'
|
|
except Exception:
|
|
print(f"[ComfyUI-Manager] Failed to check state of the git node pack: {dir_path}")
|
|
|
|
|
|
def get_installed_pip_packages():
|
|
# extract pip package infos
|
|
cmd = manager_util.make_pip_cmd(['freeze'])
|
|
pips = subprocess.check_output(cmd, text=True).split('\n')
|
|
|
|
res = {}
|
|
for x in pips:
|
|
if x.strip() == "":
|
|
continue
|
|
|
|
if ' @ ' in x:
|
|
spec_url = x.split(' @ ')
|
|
res[spec_url[0]] = spec_url[1]
|
|
else:
|
|
res[x] = ""
|
|
|
|
return res
|
|
|
|
|
|
async def get_current_snapshot(custom_nodes_only = False):
|
|
await unified_manager.reload('cache')
|
|
await unified_manager.get_custom_nodes('default', 'cache')
|
|
|
|
# Get ComfyUI hash
|
|
repo_path = comfy_path
|
|
|
|
if not os.path.exists(os.path.join(repo_path, '.git')):
|
|
print("ComfyUI update fail: The installed ComfyUI does not have a Git repository.")
|
|
return {}
|
|
|
|
comfyui_commit_hash = None
|
|
if not custom_nodes_only:
|
|
repo = git.Repo(repo_path)
|
|
comfyui_commit_hash = repo.head.commit.hexsha
|
|
|
|
git_custom_nodes = {}
|
|
cnr_custom_nodes = {}
|
|
file_custom_nodes = []
|
|
|
|
# Get custom nodes hash
|
|
for custom_nodes_dir in get_custom_nodes_paths():
|
|
paths = os.listdir(custom_nodes_dir)
|
|
|
|
disabled_path = os.path.join(custom_nodes_dir, '.disabled')
|
|
if os.path.exists(disabled_path):
|
|
for x in os.listdir(disabled_path):
|
|
paths.append(os.path.join(disabled_path, x))
|
|
|
|
for path in paths:
|
|
if path in ['.disabled', '__pycache__']:
|
|
continue
|
|
|
|
fullpath = os.path.join(custom_nodes_dir, path)
|
|
|
|
if os.path.isdir(fullpath):
|
|
is_disabled = path.endswith(".disabled") or os.path.basename(os.path.dirname(fullpath)) == ".disabled"
|
|
|
|
try:
|
|
info = unified_manager.resolve_from_path(fullpath)
|
|
|
|
if info is None:
|
|
continue
|
|
|
|
if info['ver'] not in ['nightly', 'latest', 'unknown']:
|
|
if is_disabled:
|
|
continue # don't restore disabled state of CNR node.
|
|
|
|
cnr_custom_nodes[info['id']] = info['ver']
|
|
else:
|
|
repo = git.Repo(fullpath)
|
|
|
|
if repo.head.is_detached:
|
|
remote_name = get_remote_name(repo)
|
|
else:
|
|
current_branch = repo.active_branch
|
|
|
|
if current_branch.tracking_branch() is None:
|
|
remote_name = get_remote_name(repo)
|
|
else:
|
|
remote_name = current_branch.tracking_branch().remote_name
|
|
|
|
commit_hash = repo.head.commit.hexsha
|
|
|
|
url = repo.remotes[remote_name].url
|
|
|
|
git_custom_nodes[url] = dict(hash=commit_hash, disabled=is_disabled)
|
|
except:
|
|
print(f"Failed to extract snapshots for the custom node '{path}'.")
|
|
|
|
elif path.endswith('.py'):
|
|
is_disabled = path.endswith(".py.disabled")
|
|
filename = os.path.basename(path)
|
|
item = {
|
|
'filename': filename,
|
|
'disabled': is_disabled
|
|
}
|
|
|
|
file_custom_nodes.append(item)
|
|
|
|
pip_packages = None if custom_nodes_only else get_installed_pip_packages()
|
|
|
|
return {
|
|
'comfyui': comfyui_commit_hash,
|
|
'git_custom_nodes': git_custom_nodes,
|
|
'cnr_custom_nodes': cnr_custom_nodes,
|
|
'file_custom_nodes': file_custom_nodes,
|
|
'pips': pip_packages,
|
|
}
|
|
|
|
|
|
async def save_snapshot_with_postfix(postfix, path=None, custom_nodes_only = False):
|
|
if path is None:
|
|
now = datetime.now()
|
|
|
|
date_time_format = now.strftime("%Y-%m-%d_%H-%M-%S")
|
|
file_name = f"{date_time_format}_{postfix}"
|
|
|
|
path = os.path.join(manager_snapshot_path, f"{file_name}.json")
|
|
else:
|
|
file_name = path.replace('\\', '/').split('/')[-1]
|
|
file_name = file_name.split('.')[-2]
|
|
|
|
snapshot = await get_current_snapshot(custom_nodes_only)
|
|
if path.endswith('.json'):
|
|
with open(path, "w") as json_file:
|
|
json.dump(snapshot, json_file, indent=4)
|
|
|
|
return file_name + '.json'
|
|
|
|
elif path.endswith('.yaml'):
|
|
with open(path, "w") as yaml_file:
|
|
snapshot = {'custom_nodes': snapshot}
|
|
yaml.dump(snapshot, yaml_file, allow_unicode=True)
|
|
|
|
return path
|
|
|
|
|
|
async def extract_nodes_from_workflow(filepath, mode='local', channel_url='default'):
|
|
# prepare json data
|
|
workflow = None
|
|
if filepath.endswith('.json'):
|
|
with open(filepath, "r", encoding="UTF-8", errors="ignore") as json_file:
|
|
try:
|
|
workflow = json.load(json_file)
|
|
except:
|
|
print(f"Invalid workflow file: {filepath}")
|
|
exit(-1)
|
|
|
|
elif filepath.endswith('.png'):
|
|
from PIL import Image
|
|
with Image.open(filepath) as img:
|
|
if 'workflow' not in img.info:
|
|
print(f"The specified .png file doesn't have a workflow: {filepath}")
|
|
exit(-1)
|
|
else:
|
|
try:
|
|
workflow = json.loads(img.info['workflow'])
|
|
except:
|
|
print(f"This is not a valid .png file containing a ComfyUI workflow: {filepath}")
|
|
exit(-1)
|
|
|
|
if workflow is None:
|
|
print(f"Invalid workflow file: {filepath}")
|
|
exit(-1)
|
|
|
|
# extract nodes
|
|
used_nodes = set()
|
|
|
|
def extract_nodes(sub_workflow):
|
|
for x in sub_workflow['nodes']:
|
|
node_name = x.get('type')
|
|
|
|
# skip virtual nodes
|
|
if node_name in ['Reroute', 'Note']:
|
|
continue
|
|
|
|
if node_name is not None and not (node_name.startswith('workflow/') or node_name.startswith('workflow>')):
|
|
used_nodes.add(node_name)
|
|
|
|
if 'nodes' in workflow:
|
|
extract_nodes(workflow)
|
|
|
|
if 'extra' in workflow:
|
|
if 'groupNodes' in workflow['extra']:
|
|
for x in workflow['extra']['groupNodes'].values():
|
|
extract_nodes(x)
|
|
|
|
# lookup dependent custom nodes
|
|
ext_map = await get_data_by_mode(mode, 'extension-node-map.json', channel_url)
|
|
|
|
rext_map = {}
|
|
preemption_map = {}
|
|
patterns = []
|
|
for k, v in ext_map.items():
|
|
if k == 'https://github.com/comfyanonymous/ComfyUI':
|
|
for x in v[0]:
|
|
if x not in preemption_map:
|
|
preemption_map[x] = []
|
|
|
|
preemption_map[x] = k
|
|
continue
|
|
|
|
for x in v[0]:
|
|
if x not in rext_map:
|
|
rext_map[x] = []
|
|
|
|
rext_map[x].append(k)
|
|
|
|
if 'preemptions' in v[1]:
|
|
for x in v[1]['preemptions']:
|
|
if x not in preemption_map:
|
|
preemption_map[x] = []
|
|
|
|
preemption_map[x] = k
|
|
|
|
if 'nodename_pattern' in v[1]:
|
|
patterns.append((v[1]['nodename_pattern'], k))
|
|
|
|
# identify used extensions
|
|
used_exts = set()
|
|
unknown_nodes = set()
|
|
|
|
for node_name in used_nodes:
|
|
ext = preemption_map.get(node_name)
|
|
|
|
if ext is None:
|
|
ext = rext_map.get(node_name)
|
|
if ext is not None:
|
|
ext = ext[0]
|
|
|
|
if ext is None:
|
|
for pat_ext in patterns:
|
|
if re.search(pat_ext[0], node_name):
|
|
ext = pat_ext[1]
|
|
break
|
|
|
|
if ext == 'https://github.com/comfyanonymous/ComfyUI':
|
|
pass
|
|
elif ext is not None:
|
|
used_exts.add(ext)
|
|
else:
|
|
unknown_nodes.add(node_name)
|
|
|
|
return used_exts, unknown_nodes
|
|
|
|
|
|
def unzip(model_path):
|
|
if not os.path.exists(model_path):
|
|
print(f"[ComfyUI-Manager] unzip: File not found: {model_path}")
|
|
return False
|
|
|
|
base_dir = os.path.dirname(model_path)
|
|
filename = os.path.basename(model_path)
|
|
target_dir = os.path.join(base_dir, filename[:-4])
|
|
|
|
os.makedirs(target_dir, exist_ok=True)
|
|
|
|
with zipfile.ZipFile(model_path, 'r') as zip_ref:
|
|
zip_ref.extractall(target_dir)
|
|
|
|
# Check if there's only one directory inside the target directory
|
|
contents = os.listdir(target_dir)
|
|
if len(contents) == 1 and os.path.isdir(os.path.join(target_dir, contents[0])):
|
|
nested_dir = os.path.join(target_dir, contents[0])
|
|
# Move each file and sub-directory in the nested directory up to the target directory
|
|
for item in os.listdir(nested_dir):
|
|
shutil.move(os.path.join(nested_dir, item), os.path.join(target_dir, item))
|
|
# Remove the now empty nested directory
|
|
os.rmdir(nested_dir)
|
|
|
|
os.remove(model_path)
|
|
return True
|
|
|
|
|
|
def map_to_unified_keys(json_obj):
|
|
res = {}
|
|
for k, v in json_obj.items():
|
|
cnr = unified_manager.get_cnr_by_repo(k)
|
|
if cnr:
|
|
res[cnr['id']] = v
|
|
else:
|
|
res[k] = v
|
|
|
|
return res
|
|
|
|
|
|
async def get_unified_total_nodes(channel, mode, regsitry_cache_mode='cache'):
|
|
await unified_manager.reload(regsitry_cache_mode)
|
|
|
|
res = await unified_manager.get_custom_nodes(channel, mode)
|
|
|
|
# collect pure cnr ids (i.e. not exists in custom-node-list.json)
|
|
# populate state/updatable field to non-pure cnr nodes
|
|
cnr_ids = set(unified_manager.cnr_map.keys())
|
|
for k, v in res.items():
|
|
# resolve cnr_id from repo url
|
|
files_in_json = v.get('files', [])
|
|
cnr_id = None
|
|
if len(files_in_json) == 1:
|
|
cnr = unified_manager.get_cnr_by_repo(files_in_json[0])
|
|
if cnr:
|
|
cnr_id = cnr['id']
|
|
|
|
if cnr_id is not None:
|
|
# cnr or nightly version
|
|
cnr_ids.remove(cnr_id)
|
|
updatable = False
|
|
cnr = unified_manager.cnr_map[cnr_id]
|
|
|
|
if cnr_id in invalid_nodes:
|
|
v['invalid-installation'] = True
|
|
|
|
if cnr_id in unified_manager.active_nodes:
|
|
# installed
|
|
v['state'] = 'enabled'
|
|
if unified_manager.active_nodes[cnr_id][0] != 'nightly':
|
|
updatable = unified_manager.is_updatable(cnr_id)
|
|
else:
|
|
updatable = False
|
|
v['active_version'] = unified_manager.active_nodes[cnr_id][0]
|
|
v['version'] = v['active_version']
|
|
|
|
if cm_global.try_call(api="cm.is_import_failed_extension", name=unified_manager.active_nodes[cnr_id][1]):
|
|
v['import-fail'] = True
|
|
|
|
elif cnr_id in unified_manager.cnr_inactive_nodes:
|
|
# disabled
|
|
v['state'] = 'disabled'
|
|
cnr_ver = unified_manager.get_from_cnr_inactive_nodes(cnr_id)
|
|
if cnr_ver is not None:
|
|
v['version'] = str(cnr_ver[0])
|
|
else:
|
|
v['version'] = '0'
|
|
|
|
elif cnr_id in unified_manager.nightly_inactive_nodes:
|
|
# disabled
|
|
v['state'] = 'disabled'
|
|
v['version'] = 'nightly'
|
|
else:
|
|
# not installed
|
|
v['state'] = 'not-installed'
|
|
|
|
if 'version' not in v:
|
|
v['version'] = cnr['latest_version']['version']
|
|
|
|
v['update-state'] = 'true' if updatable else 'false'
|
|
else:
|
|
# unknown version
|
|
v['version'] = 'unknown'
|
|
|
|
if unified_manager.is_enabled(k, 'unknown'):
|
|
v['state'] = 'enabled'
|
|
v['active_version'] = 'unknown'
|
|
|
|
if cm_global.try_call(api="cm.is_import_failed_extension", name=unified_manager.unknown_active_nodes[k][1]):
|
|
v['import-fail'] = True
|
|
|
|
elif unified_manager.is_disabled(k, 'unknown'):
|
|
v['state'] = 'disabled'
|
|
else:
|
|
v['state'] = 'not-installed'
|
|
|
|
# add items for pure cnr nodes
|
|
if normalize_channel(channel) == DEFAULT_CHANNEL:
|
|
# Don't show CNR nodes unless default channel
|
|
for cnr_id in cnr_ids:
|
|
cnr = unified_manager.cnr_map[cnr_id]
|
|
author = cnr['publisher']['name']
|
|
title = cnr['name']
|
|
reference = f"https://registry.comfy.org/nodes/{cnr['id']}"
|
|
repository = cnr.get('repository', '')
|
|
install_type = "cnr"
|
|
description = cnr.get('description', '')
|
|
|
|
ver = None
|
|
active_version = None
|
|
updatable = False
|
|
import_fail = None
|
|
if cnr_id in unified_manager.active_nodes:
|
|
# installed
|
|
state = 'enabled'
|
|
updatable = unified_manager.is_updatable(cnr_id)
|
|
active_version = unified_manager.active_nodes[cnr['id']][0]
|
|
ver = active_version
|
|
|
|
if cm_global.try_call(api="cm.is_import_failed_extension", name=unified_manager.active_nodes[cnr_id][1]):
|
|
import_fail = True
|
|
|
|
elif cnr['id'] in unified_manager.cnr_inactive_nodes:
|
|
# disabled
|
|
state = 'disabled'
|
|
elif cnr['id'] in unified_manager.nightly_inactive_nodes:
|
|
# disabled
|
|
state = 'disabled'
|
|
ver = 'nightly'
|
|
else:
|
|
# not installed
|
|
state = 'not-installed'
|
|
|
|
if ver is None:
|
|
ver = cnr['latest_version']['version']
|
|
|
|
item = dict(author=author, title=title, reference=reference, repository=repository, install_type=install_type,
|
|
description=description, state=state, updatable=updatable, version=ver)
|
|
|
|
if active_version:
|
|
item['active_version'] = active_version
|
|
|
|
if import_fail:
|
|
item['import-fail'] = True
|
|
|
|
res[cnr_id] = item
|
|
|
|
return res
|
|
|
|
|
|
def populate_github_stats(node_packs, json_obj_github):
|
|
for k, v in node_packs.items():
|
|
try:
|
|
url = v['reference']
|
|
if url in json_obj_github:
|
|
v['stars'] = json_obj_github[url]['stars']
|
|
v['last_update'] = json_obj_github[url]['last_update']
|
|
v['trust'] = json_obj_github[url]['author_account_age_days'] > 600
|
|
else:
|
|
v['stars'] = -1
|
|
v['last_update'] = -1
|
|
v['trust'] = False
|
|
except:
|
|
logging.error(f"[ComfyUI-Manager] DB item is broken:\n{v}")
|
|
|
|
|
|
def populate_favorites(node_packs, json_obj_extras):
|
|
favorites = set(json_obj_extras['favorites'])
|
|
|
|
for k, v in node_packs.items():
|
|
if v.get('version') != 'unknown':
|
|
if k in favorites:
|
|
v['is_favorite'] = True
|
|
|
|
|
|
async def restore_snapshot(snapshot_path, git_helper_extras=None):
|
|
cloned_repos = []
|
|
checkout_repos = []
|
|
enabled_repos = []
|
|
disabled_repos = []
|
|
skip_node_packs = []
|
|
|
|
await unified_manager.reload('cache')
|
|
await unified_manager.get_custom_nodes('default', 'cache')
|
|
|
|
cnr_repo_map = {}
|
|
for k, v in unified_manager.repo_cnr_map.items():
|
|
cnr_repo_map[v['id']] = k
|
|
|
|
print("Restore snapshot.")
|
|
|
|
postinstalls = []
|
|
|
|
with open(snapshot_path, 'r', encoding="UTF-8") as snapshot_file:
|
|
if snapshot_path.endswith('.json'):
|
|
info = json.load(snapshot_file)
|
|
elif snapshot_path.endswith('.yaml'):
|
|
info = yaml.load(snapshot_file, Loader=yaml.SafeLoader)
|
|
info = info['custom_nodes']
|
|
|
|
# for cnr restore
|
|
cnr_info = info.get('cnr_custom_nodes')
|
|
if cnr_info is not None:
|
|
# disable not listed cnr nodes
|
|
todo_disable = []
|
|
todo_checkout = []
|
|
|
|
for k, v in unified_manager.active_nodes.items():
|
|
if 'comfyui-manager' in k:
|
|
continue
|
|
|
|
if v[0] != 'nightly':
|
|
if k not in cnr_info:
|
|
todo_disable.append(k)
|
|
else:
|
|
cnr_ver = cnr_info[k]
|
|
if v[1] != cnr_ver:
|
|
todo_checkout.append((k, cnr_ver))
|
|
else:
|
|
skip_node_packs.append(k)
|
|
|
|
for x in todo_disable:
|
|
unified_manager.unified_disable(x, False)
|
|
disabled_repos.append(x)
|
|
|
|
for x in todo_checkout:
|
|
unified_manager.cnr_switch_version(x[0], x[1], instant_execution=True, no_deps=True, return_postinstall=False)
|
|
checkout_repos.append(x[1])
|
|
|
|
# install listed cnr nodes
|
|
for k, v in cnr_info.items():
|
|
if 'comfyui-manager' in k:
|
|
continue
|
|
|
|
ps = await unified_manager.install_by_id(k, version_spec=v, instant_execution=True, return_postinstall=True)
|
|
cloned_repos.append(k)
|
|
if ps is not None and ps.result:
|
|
if hasattr(ps, 'postinstall'):
|
|
postinstalls.append(ps.postinstall)
|
|
else:
|
|
print("cm-cli: unexpected [0001]")
|
|
|
|
# for nightly restore
|
|
_git_info = info.get('git_custom_nodes')
|
|
git_info = {}
|
|
|
|
# normalize github repo
|
|
for k, v in _git_info.items():
|
|
norm_k = git_utils.normalize_url(k)
|
|
git_info[norm_k] = v
|
|
|
|
if git_info is not None:
|
|
todo_disable = []
|
|
todo_enable = []
|
|
todo_checkout = []
|
|
processed_urls = []
|
|
|
|
for k, v in unified_manager.active_nodes.items():
|
|
if 'comfyui-manager' in k:
|
|
continue
|
|
|
|
if v[0] == 'nightly' and cnr_repo_map.get(k):
|
|
repo_url = cnr_repo_map.get(k)
|
|
normalized_url = git_utils.normalize_url(repo_url)
|
|
|
|
if normalized_url not in git_info:
|
|
todo_disable.append(k)
|
|
else:
|
|
commit_hash = git_info[normalized_url]['hash']
|
|
todo_checkout.append((v[1], commit_hash))
|
|
|
|
for k, v in unified_manager.nightly_inactive_nodes.items():
|
|
if 'comfyui-manager' in k:
|
|
continue
|
|
|
|
if cnr_repo_map.get(k):
|
|
repo_url = cnr_repo_map.get(k)
|
|
normalized_url = git_utils.normalize_url(repo_url)
|
|
|
|
if normalized_url in git_info:
|
|
commit_hash = git_info[normalized_url]['hash']
|
|
todo_enable.append((k, commit_hash))
|
|
processed_urls.append(normalized_url)
|
|
|
|
for x in todo_disable:
|
|
unified_manager.unified_disable(x, False)
|
|
disabled_repos.append(x)
|
|
|
|
for x in todo_enable:
|
|
res = unified_manager.unified_enable(x, 'nightly')
|
|
|
|
is_switched = False
|
|
if res and res.target:
|
|
is_switched = repo_switch_commit(res.target, x[1])
|
|
|
|
if is_switched:
|
|
checkout_repos.append(x)
|
|
else:
|
|
enabled_repos.append(x)
|
|
|
|
for x in todo_checkout:
|
|
is_switched = repo_switch_commit(x[0], x[1])
|
|
|
|
if is_switched:
|
|
checkout_repos.append(x)
|
|
else:
|
|
skip_node_packs.append(x[0])
|
|
|
|
for x in git_info.keys():
|
|
normalized_url = git_utils.normalize_url(x)
|
|
cnr = unified_manager.repo_cnr_map.get(normalized_url)
|
|
if cnr is not None:
|
|
pack_id = cnr['id']
|
|
await unified_manager.install_by_id(pack_id, 'nightly', instant_execution=True, no_deps=False, return_postinstall=False)
|
|
cloned_repos.append(pack_id)
|
|
processed_urls.append(x)
|
|
|
|
for x in processed_urls:
|
|
if x in git_info:
|
|
del git_info[x]
|
|
|
|
# remained nightly will be installed and migrated
|
|
|
|
# for unknown restore
|
|
todo_disable = []
|
|
todo_enable = []
|
|
todo_checkout = []
|
|
processed_urls = []
|
|
|
|
for k2, v2 in unified_manager.unknown_active_nodes.items():
|
|
repo_url = resolve_giturl_from_path(v2[1])
|
|
|
|
if repo_url is None:
|
|
continue
|
|
|
|
normalized_url = git_utils.normalize_url(repo_url)
|
|
|
|
if normalized_url not in git_info:
|
|
todo_disable.append(k2)
|
|
else:
|
|
commit_hash = git_info[normalized_url]['hash']
|
|
todo_checkout.append((k2, commit_hash))
|
|
processed_urls.append(normalized_url)
|
|
|
|
for k2, v2 in unified_manager.unknown_inactive_nodes.items():
|
|
repo_url = resolve_giturl_from_path(v2[1])
|
|
|
|
if repo_url is None:
|
|
continue
|
|
|
|
normalized_url = git_utils.normalize_url(repo_url)
|
|
|
|
if normalized_url in git_info:
|
|
commit_hash = git_info[normalized_url]['hash']
|
|
todo_enable.append((k2, commit_hash))
|
|
processed_urls.append(normalized_url)
|
|
|
|
for x in todo_disable:
|
|
unified_manager.unified_disable(x, True)
|
|
disabled_repos.append(x)
|
|
|
|
for x in todo_enable:
|
|
res = unified_manager.unified_enable(x[0], 'unknown')
|
|
|
|
is_switched = False
|
|
if res and res.target:
|
|
is_switched = repo_switch_commit(res.target, x[1])
|
|
|
|
if is_switched:
|
|
checkout_repos.append(x)
|
|
else:
|
|
enabled_repos.append(x)
|
|
|
|
for x in todo_checkout:
|
|
is_switched = repo_switch_commit(x[0], x[1])
|
|
|
|
if is_switched:
|
|
checkout_repos.append(x)
|
|
else:
|
|
skip_node_packs.append(x[0])
|
|
|
|
for x in processed_urls:
|
|
if x in git_info:
|
|
del git_info[x]
|
|
|
|
for repo_url in git_info.keys():
|
|
repo_name = os.path.basename(repo_url)
|
|
if repo_name.endswith('.git'):
|
|
repo_name = repo_name[:-4]
|
|
|
|
to_path = os.path.join(get_default_custom_nodes_path(), repo_name)
|
|
unified_manager.repo_install(repo_url, to_path, instant_execution=True, no_deps=False, return_postinstall=False)
|
|
cloned_repos.append(repo_name)
|
|
|
|
# reload
|
|
await unified_manager.migrate_unmanaged_nodes()
|
|
|
|
# print summary
|
|
for x in cloned_repos:
|
|
print(f"[ INSTALLED ] {x}")
|
|
for x in checkout_repos:
|
|
print(f"[ CHECKOUT ] {x}")
|
|
for x in enabled_repos:
|
|
print(f"[ ENABLED ] {x}")
|
|
for x in disabled_repos:
|
|
print(f"[ DISABLED ] {x}")
|
|
for x in skip_node_packs:
|
|
print(f"[ SKIPPED ] {x}")
|
|
|
|
# if is_failed:
|
|
# print("[bold red]ERROR: Failed to restore snapshot.[/bold red]")
|
|
|
|
|
|
# check need to migrate
|
|
need_to_migrate = False
|
|
|
|
|
|
async def check_need_to_migrate():
|
|
global need_to_migrate
|
|
|
|
await unified_manager.reload('cache')
|
|
await unified_manager.load_nightly(channel='default', mode='cache')
|
|
|
|
legacy_custom_nodes = []
|
|
|
|
for x in unified_manager.active_nodes.values():
|
|
if x[0] == 'nightly' and not x[1].endswith('@nightly'):
|
|
legacy_custom_nodes.append(x[1])
|
|
|
|
for x in unified_manager.nightly_inactive_nodes.values():
|
|
if not x.endswith('@nightly'):
|
|
legacy_custom_nodes.append(x)
|
|
|
|
if len(legacy_custom_nodes) > 0:
|
|
print("\n--------------------- ComfyUI-Manager migration notice --------------------")
|
|
print("The following custom nodes were installed using the old management method and require migration:\n")
|
|
print("\n".join(legacy_custom_nodes))
|
|
print("---------------------------------------------------------------------------\n")
|
|
need_to_migrate = True
|
|
|
|
|
|
def get_comfyui_versions(repo=None):
|
|
if repo is None:
|
|
repo = git.Repo(comfy_path)
|
|
|
|
try:
|
|
remote = get_remote_name(repo)
|
|
repo.remotes[remote].fetch()
|
|
except:
|
|
logging.error("[ComfyUI-Manager] Failed to fetch ComfyUI")
|
|
|
|
versions = [x.name for x in repo.tags if x.name.startswith('v')]
|
|
|
|
# nearest tag
|
|
versions = sorted(versions, key=lambda v: repo.git.log('-1', '--format=%ct', v), reverse=True)
|
|
versions = versions[:4]
|
|
|
|
current_tag = repo.git.describe('--tags')
|
|
|
|
if current_tag not in versions:
|
|
versions = sorted(versions + [current_tag], key=lambda v: repo.git.log('-1', '--format=%ct', v), reverse=True)
|
|
versions = versions[:4]
|
|
|
|
main_branch = repo.heads.master
|
|
latest_commit = main_branch.commit
|
|
latest_tag = repo.git.describe('--tags', latest_commit.hexsha)
|
|
|
|
if latest_tag != versions[0]:
|
|
versions.insert(0, 'nightly')
|
|
else:
|
|
versions[0] = 'nightly'
|
|
current_tag = 'nightly'
|
|
|
|
return versions, current_tag, latest_tag
|
|
|
|
|
|
def switch_comfyui(tag):
|
|
repo = git.Repo(comfy_path)
|
|
|
|
if tag == 'nightly':
|
|
repo.git.checkout('master')
|
|
repo.remotes.origin.pull()
|
|
print("[ComfyUI-Manager] ComfyUI version is switched to the latest 'master' version")
|
|
else:
|
|
repo.git.checkout(tag)
|
|
print(f"[ComfyUI-Manager] ComfyUI version is switched to '{tag}'")
|
|
|
|
|
|
def resolve_giturl_from_path(fullpath):
|
|
"""
|
|
resolve giturl path of unclassified custom node based on remote url in .git/config
|
|
"""
|
|
git_config_path = os.path.join(fullpath, '.git', 'config')
|
|
|
|
if not os.path.exists(git_config_path):
|
|
return "unknown"
|
|
|
|
config = configparser.ConfigParser()
|
|
config.read(git_config_path)
|
|
|
|
for k, v in config.items():
|
|
if k.startswith('remote ') and 'url' in v:
|
|
return v['url'].replace("git@github.com:", "https://github.com/")
|
|
|
|
return None
|
|
|
|
|
|
def repo_switch_commit(repo_path, commit_hash):
|
|
try:
|
|
repo = git.Repo(repo_path)
|
|
if repo.head.commit.hexsha == commit_hash:
|
|
return False
|
|
|
|
repo.git.checkout(commit_hash)
|
|
return True
|
|
except:
|
|
return None
|