| import itertools |
| from typing import Sequence, Mapping, Dict |
| from comfy_execution.graph import DynamicPrompt |
|
|
| import nodes |
|
|
| from comfy_execution.graph_utils import is_link |
|
|
| NODE_CLASS_CONTAINS_UNIQUE_ID: Dict[str, bool] = {} |
|
|
|
|
| def include_unique_id_in_input(class_type: str) -> bool: |
| if class_type in NODE_CLASS_CONTAINS_UNIQUE_ID: |
| return NODE_CLASS_CONTAINS_UNIQUE_ID[class_type] |
| class_def = nodes.NODE_CLASS_MAPPINGS[class_type] |
| NODE_CLASS_CONTAINS_UNIQUE_ID[class_type] = "UNIQUE_ID" in class_def.INPUT_TYPES().get("hidden", {}).values() |
| return NODE_CLASS_CONTAINS_UNIQUE_ID[class_type] |
|
|
| class CacheKeySet: |
| def __init__(self, dynprompt, node_ids, is_changed_cache): |
| self.keys = {} |
| self.subcache_keys = {} |
|
|
| def add_keys(self, node_ids): |
| raise NotImplementedError() |
|
|
| def all_node_ids(self): |
| return set(self.keys.keys()) |
|
|
| def get_used_keys(self): |
| return self.keys.values() |
|
|
| def get_used_subcache_keys(self): |
| return self.subcache_keys.values() |
|
|
| def get_data_key(self, node_id): |
| return self.keys.get(node_id, None) |
|
|
| def get_subcache_key(self, node_id): |
| return self.subcache_keys.get(node_id, None) |
|
|
| class Unhashable: |
| def __init__(self): |
| self.value = float("NaN") |
|
|
| def to_hashable(obj): |
| |
| |
| if isinstance(obj, (int, float, str, bool, type(None))): |
| return obj |
| elif isinstance(obj, Mapping): |
| return frozenset([(to_hashable(k), to_hashable(v)) for k, v in sorted(obj.items())]) |
| elif isinstance(obj, Sequence): |
| return frozenset(zip(itertools.count(), [to_hashable(i) for i in obj])) |
| else: |
| |
| return Unhashable() |
|
|
| class CacheKeySetID(CacheKeySet): |
| def __init__(self, dynprompt, node_ids, is_changed_cache): |
| super().__init__(dynprompt, node_ids, is_changed_cache) |
| self.dynprompt = dynprompt |
| self.add_keys(node_ids) |
|
|
| def add_keys(self, node_ids): |
| for node_id in node_ids: |
| if node_id in self.keys: |
| continue |
| if not self.dynprompt.has_node(node_id): |
| continue |
| node = self.dynprompt.get_node(node_id) |
| self.keys[node_id] = (node_id, node["class_type"]) |
| self.subcache_keys[node_id] = (node_id, node["class_type"]) |
|
|
| class CacheKeySetInputSignature(CacheKeySet): |
| def __init__(self, dynprompt, node_ids, is_changed_cache): |
| super().__init__(dynprompt, node_ids, is_changed_cache) |
| self.dynprompt = dynprompt |
| self.is_changed_cache = is_changed_cache |
| self.add_keys(node_ids) |
|
|
| def include_node_id_in_input(self) -> bool: |
| return False |
|
|
| def add_keys(self, node_ids): |
| for node_id in node_ids: |
| if node_id in self.keys: |
| continue |
| if not self.dynprompt.has_node(node_id): |
| continue |
| node = self.dynprompt.get_node(node_id) |
| self.keys[node_id] = self.get_node_signature(self.dynprompt, node_id) |
| self.subcache_keys[node_id] = (node_id, node["class_type"]) |
|
|
| def get_node_signature(self, dynprompt, node_id): |
| signature = [] |
| ancestors, order_mapping = self.get_ordered_ancestry(dynprompt, node_id) |
| signature.append(self.get_immediate_node_signature(dynprompt, node_id, order_mapping)) |
| for ancestor_id in ancestors: |
| signature.append(self.get_immediate_node_signature(dynprompt, ancestor_id, order_mapping)) |
| return to_hashable(signature) |
|
|
| def get_immediate_node_signature(self, dynprompt, node_id, ancestor_order_mapping): |
| if not dynprompt.has_node(node_id): |
| |
| return [float("NaN")] |
| node = dynprompt.get_node(node_id) |
| class_type = node["class_type"] |
| class_def = nodes.NODE_CLASS_MAPPINGS[class_type] |
| signature = [class_type, self.is_changed_cache.get(node_id)] |
| if self.include_node_id_in_input() or (hasattr(class_def, "NOT_IDEMPOTENT") and class_def.NOT_IDEMPOTENT) or include_unique_id_in_input(class_type): |
| signature.append(node_id) |
| inputs = node["inputs"] |
| for key in sorted(inputs.keys()): |
| if is_link(inputs[key]): |
| (ancestor_id, ancestor_socket) = inputs[key] |
| ancestor_index = ancestor_order_mapping[ancestor_id] |
| signature.append((key,("ANCESTOR", ancestor_index, ancestor_socket))) |
| else: |
| signature.append((key, inputs[key])) |
| return signature |
|
|
| |
| |
| def get_ordered_ancestry(self, dynprompt, node_id): |
| ancestors = [] |
| order_mapping = {} |
| self.get_ordered_ancestry_internal(dynprompt, node_id, ancestors, order_mapping) |
| return ancestors, order_mapping |
|
|
| def get_ordered_ancestry_internal(self, dynprompt, node_id, ancestors, order_mapping): |
| if not dynprompt.has_node(node_id): |
| return |
| inputs = dynprompt.get_node(node_id)["inputs"] |
| input_keys = sorted(inputs.keys()) |
| for key in input_keys: |
| if is_link(inputs[key]): |
| ancestor_id = inputs[key][0] |
| if ancestor_id not in order_mapping: |
| ancestors.append(ancestor_id) |
| order_mapping[ancestor_id] = len(ancestors) - 1 |
| self.get_ordered_ancestry_internal(dynprompt, ancestor_id, ancestors, order_mapping) |
|
|
| class BasicCache: |
| def __init__(self, key_class): |
| self.key_class = key_class |
| self.initialized = False |
| self.dynprompt: DynamicPrompt |
| self.cache_key_set: CacheKeySet |
| self.cache = {} |
| self.subcaches = {} |
|
|
| def set_prompt(self, dynprompt, node_ids, is_changed_cache): |
| self.dynprompt = dynprompt |
| self.cache_key_set = self.key_class(dynprompt, node_ids, is_changed_cache) |
| self.is_changed_cache = is_changed_cache |
| self.initialized = True |
|
|
| def all_node_ids(self): |
| assert self.initialized |
| node_ids = self.cache_key_set.all_node_ids() |
| for subcache in self.subcaches.values(): |
| node_ids = node_ids.union(subcache.all_node_ids()) |
| return node_ids |
|
|
| def _clean_cache(self): |
| preserve_keys = set(self.cache_key_set.get_used_keys()) |
| to_remove = [] |
| for key in self.cache: |
| if key not in preserve_keys: |
| to_remove.append(key) |
| for key in to_remove: |
| del self.cache[key] |
|
|
| def _clean_subcaches(self): |
| preserve_subcaches = set(self.cache_key_set.get_used_subcache_keys()) |
|
|
| to_remove = [] |
| for key in self.subcaches: |
| if key not in preserve_subcaches: |
| to_remove.append(key) |
| for key in to_remove: |
| del self.subcaches[key] |
|
|
| def clean_unused(self): |
| assert self.initialized |
| self._clean_cache() |
| self._clean_subcaches() |
|
|
| def _set_immediate(self, node_id, value): |
| assert self.initialized |
| cache_key = self.cache_key_set.get_data_key(node_id) |
| self.cache[cache_key] = value |
|
|
| def _get_immediate(self, node_id): |
| if not self.initialized: |
| return None |
| cache_key = self.cache_key_set.get_data_key(node_id) |
| if cache_key in self.cache: |
| return self.cache[cache_key] |
| else: |
| return None |
|
|
| def _ensure_subcache(self, node_id, children_ids): |
| subcache_key = self.cache_key_set.get_subcache_key(node_id) |
| subcache = self.subcaches.get(subcache_key, None) |
| if subcache is None: |
| subcache = BasicCache(self.key_class) |
| self.subcaches[subcache_key] = subcache |
| subcache.set_prompt(self.dynprompt, children_ids, self.is_changed_cache) |
| return subcache |
|
|
| def _get_subcache(self, node_id): |
| assert self.initialized |
| subcache_key = self.cache_key_set.get_subcache_key(node_id) |
| if subcache_key in self.subcaches: |
| return self.subcaches[subcache_key] |
| else: |
| return None |
|
|
| def recursive_debug_dump(self): |
| result = [] |
| for key in self.cache: |
| result.append({"key": key, "value": self.cache[key]}) |
| for key in self.subcaches: |
| result.append({"subcache_key": key, "subcache": self.subcaches[key].recursive_debug_dump()}) |
| return result |
|
|
| class HierarchicalCache(BasicCache): |
| def __init__(self, key_class): |
| super().__init__(key_class) |
|
|
| def _get_cache_for(self, node_id): |
| assert self.dynprompt is not None |
| parent_id = self.dynprompt.get_parent_node_id(node_id) |
| if parent_id is None: |
| return self |
|
|
| hierarchy = [] |
| while parent_id is not None: |
| hierarchy.append(parent_id) |
| parent_id = self.dynprompt.get_parent_node_id(parent_id) |
|
|
| cache = self |
| for parent_id in reversed(hierarchy): |
| cache = cache._get_subcache(parent_id) |
| if cache is None: |
| return None |
| return cache |
|
|
| def get(self, node_id): |
| cache = self._get_cache_for(node_id) |
| if cache is None: |
| return None |
| return cache._get_immediate(node_id) |
|
|
| def set(self, node_id, value): |
| cache = self._get_cache_for(node_id) |
| assert cache is not None |
| cache._set_immediate(node_id, value) |
|
|
| def ensure_subcache_for(self, node_id, children_ids): |
| cache = self._get_cache_for(node_id) |
| assert cache is not None |
| return cache._ensure_subcache(node_id, children_ids) |
|
|
| class LRUCache(BasicCache): |
| def __init__(self, key_class, max_size=100): |
| super().__init__(key_class) |
| self.max_size = max_size |
| self.min_generation = 0 |
| self.generation = 0 |
| self.used_generation = {} |
| self.children = {} |
|
|
| def set_prompt(self, dynprompt, node_ids, is_changed_cache): |
| super().set_prompt(dynprompt, node_ids, is_changed_cache) |
| self.generation += 1 |
| for node_id in node_ids: |
| self._mark_used(node_id) |
|
|
| def clean_unused(self): |
| while len(self.cache) > self.max_size and self.min_generation < self.generation: |
| self.min_generation += 1 |
| to_remove = [key for key in self.cache if self.used_generation[key] < self.min_generation] |
| for key in to_remove: |
| del self.cache[key] |
| del self.used_generation[key] |
| if key in self.children: |
| del self.children[key] |
| self._clean_subcaches() |
|
|
| def get(self, node_id): |
| self._mark_used(node_id) |
| return self._get_immediate(node_id) |
|
|
| def _mark_used(self, node_id): |
| cache_key = self.cache_key_set.get_data_key(node_id) |
| if cache_key is not None: |
| self.used_generation[cache_key] = self.generation |
|
|
| def set(self, node_id, value): |
| self._mark_used(node_id) |
| return self._set_immediate(node_id, value) |
|
|
| def ensure_subcache_for(self, node_id, children_ids): |
| |
| super()._ensure_subcache(node_id, children_ids) |
|
|
| self.cache_key_set.add_keys(children_ids) |
| self._mark_used(node_id) |
| cache_key = self.cache_key_set.get_data_key(node_id) |
| self.children[cache_key] = [] |
| for child_id in children_ids: |
| self._mark_used(child_id) |
| self.children[cache_key].append(self.cache_key_set.get_data_key(child_id)) |
| return self |
|
|
|
|