|
import itertools |
|
from typing import Sequence, Mapping, Dict |
|
from comfy_execution.graph import DynamicPrompt |
|
|
|
import nodes |
|
|
|
from comfy_execution.graph_utils import is_link |
|
|
|
NODE_CLASS_CONTAINS_UNIQUE_ID: Dict[str, bool] = {} |
|
|
|
|
|
def include_unique_id_in_input(class_type: str) -> bool: |
|
if class_type in NODE_CLASS_CONTAINS_UNIQUE_ID: |
|
return NODE_CLASS_CONTAINS_UNIQUE_ID[class_type] |
|
class_def = nodes.NODE_CLASS_MAPPINGS[class_type] |
|
NODE_CLASS_CONTAINS_UNIQUE_ID[class_type] = "UNIQUE_ID" in class_def.INPUT_TYPES().get("hidden", {}).values() |
|
return NODE_CLASS_CONTAINS_UNIQUE_ID[class_type] |
|
|
|
class CacheKeySet: |
|
def __init__(self, dynprompt, node_ids, is_changed_cache): |
|
self.keys = {} |
|
self.subcache_keys = {} |
|
|
|
def add_keys(self, node_ids): |
|
raise NotImplementedError() |
|
|
|
def all_node_ids(self): |
|
return set(self.keys.keys()) |
|
|
|
def get_used_keys(self): |
|
return self.keys.values() |
|
|
|
def get_used_subcache_keys(self): |
|
return self.subcache_keys.values() |
|
|
|
def get_data_key(self, node_id): |
|
return self.keys.get(node_id, None) |
|
|
|
def get_subcache_key(self, node_id): |
|
return self.subcache_keys.get(node_id, None) |
|
|
|
class Unhashable: |
|
def __init__(self): |
|
self.value = float("NaN") |
|
|
|
def to_hashable(obj): |
|
|
|
|
|
if isinstance(obj, (int, float, str, bool, type(None))): |
|
return obj |
|
elif isinstance(obj, Mapping): |
|
return frozenset([(to_hashable(k), to_hashable(v)) for k, v in sorted(obj.items())]) |
|
elif isinstance(obj, Sequence): |
|
return frozenset(zip(itertools.count(), [to_hashable(i) for i in obj])) |
|
else: |
|
|
|
return Unhashable() |
|
|
|
class CacheKeySetID(CacheKeySet): |
|
def __init__(self, dynprompt, node_ids, is_changed_cache): |
|
super().__init__(dynprompt, node_ids, is_changed_cache) |
|
self.dynprompt = dynprompt |
|
self.add_keys(node_ids) |
|
|
|
def add_keys(self, node_ids): |
|
for node_id in node_ids: |
|
if node_id in self.keys: |
|
continue |
|
if not self.dynprompt.has_node(node_id): |
|
continue |
|
node = self.dynprompt.get_node(node_id) |
|
self.keys[node_id] = (node_id, node["class_type"]) |
|
self.subcache_keys[node_id] = (node_id, node["class_type"]) |
|
|
|
class CacheKeySetInputSignature(CacheKeySet): |
|
def __init__(self, dynprompt, node_ids, is_changed_cache): |
|
super().__init__(dynprompt, node_ids, is_changed_cache) |
|
self.dynprompt = dynprompt |
|
self.is_changed_cache = is_changed_cache |
|
self.add_keys(node_ids) |
|
|
|
def include_node_id_in_input(self) -> bool: |
|
return False |
|
|
|
def add_keys(self, node_ids): |
|
for node_id in node_ids: |
|
if node_id in self.keys: |
|
continue |
|
if not self.dynprompt.has_node(node_id): |
|
continue |
|
node = self.dynprompt.get_node(node_id) |
|
self.keys[node_id] = self.get_node_signature(self.dynprompt, node_id) |
|
self.subcache_keys[node_id] = (node_id, node["class_type"]) |
|
|
|
def get_node_signature(self, dynprompt, node_id): |
|
signature = [] |
|
ancestors, order_mapping = self.get_ordered_ancestry(dynprompt, node_id) |
|
signature.append(self.get_immediate_node_signature(dynprompt, node_id, order_mapping)) |
|
for ancestor_id in ancestors: |
|
signature.append(self.get_immediate_node_signature(dynprompt, ancestor_id, order_mapping)) |
|
return to_hashable(signature) |
|
|
|
def get_immediate_node_signature(self, dynprompt, node_id, ancestor_order_mapping): |
|
if not dynprompt.has_node(node_id): |
|
|
|
return [float("NaN")] |
|
node = dynprompt.get_node(node_id) |
|
class_type = node["class_type"] |
|
class_def = nodes.NODE_CLASS_MAPPINGS[class_type] |
|
signature = [class_type, self.is_changed_cache.get(node_id)] |
|
if self.include_node_id_in_input() or (hasattr(class_def, "NOT_IDEMPOTENT") and class_def.NOT_IDEMPOTENT) or include_unique_id_in_input(class_type): |
|
signature.append(node_id) |
|
inputs = node["inputs"] |
|
for key in sorted(inputs.keys()): |
|
if is_link(inputs[key]): |
|
(ancestor_id, ancestor_socket) = inputs[key] |
|
ancestor_index = ancestor_order_mapping[ancestor_id] |
|
signature.append((key,("ANCESTOR", ancestor_index, ancestor_socket))) |
|
else: |
|
signature.append((key, inputs[key])) |
|
return signature |
|
|
|
|
|
|
|
def get_ordered_ancestry(self, dynprompt, node_id): |
|
ancestors = [] |
|
order_mapping = {} |
|
self.get_ordered_ancestry_internal(dynprompt, node_id, ancestors, order_mapping) |
|
return ancestors, order_mapping |
|
|
|
def get_ordered_ancestry_internal(self, dynprompt, node_id, ancestors, order_mapping): |
|
if not dynprompt.has_node(node_id): |
|
return |
|
inputs = dynprompt.get_node(node_id)["inputs"] |
|
input_keys = sorted(inputs.keys()) |
|
for key in input_keys: |
|
if is_link(inputs[key]): |
|
ancestor_id = inputs[key][0] |
|
if ancestor_id not in order_mapping: |
|
ancestors.append(ancestor_id) |
|
order_mapping[ancestor_id] = len(ancestors) - 1 |
|
self.get_ordered_ancestry_internal(dynprompt, ancestor_id, ancestors, order_mapping) |
|
|
|
class BasicCache: |
|
def __init__(self, key_class): |
|
self.key_class = key_class |
|
self.initialized = False |
|
self.dynprompt: DynamicPrompt |
|
self.cache_key_set: CacheKeySet |
|
self.cache = {} |
|
self.subcaches = {} |
|
|
|
def set_prompt(self, dynprompt, node_ids, is_changed_cache): |
|
self.dynprompt = dynprompt |
|
self.cache_key_set = self.key_class(dynprompt, node_ids, is_changed_cache) |
|
self.is_changed_cache = is_changed_cache |
|
self.initialized = True |
|
|
|
def all_node_ids(self): |
|
assert self.initialized |
|
node_ids = self.cache_key_set.all_node_ids() |
|
for subcache in self.subcaches.values(): |
|
node_ids = node_ids.union(subcache.all_node_ids()) |
|
return node_ids |
|
|
|
def _clean_cache(self): |
|
preserve_keys = set(self.cache_key_set.get_used_keys()) |
|
to_remove = [] |
|
for key in self.cache: |
|
if key not in preserve_keys: |
|
to_remove.append(key) |
|
for key in to_remove: |
|
del self.cache[key] |
|
|
|
def _clean_subcaches(self): |
|
preserve_subcaches = set(self.cache_key_set.get_used_subcache_keys()) |
|
|
|
to_remove = [] |
|
for key in self.subcaches: |
|
if key not in preserve_subcaches: |
|
to_remove.append(key) |
|
for key in to_remove: |
|
del self.subcaches[key] |
|
|
|
def clean_unused(self): |
|
assert self.initialized |
|
self._clean_cache() |
|
self._clean_subcaches() |
|
|
|
def _set_immediate(self, node_id, value): |
|
assert self.initialized |
|
cache_key = self.cache_key_set.get_data_key(node_id) |
|
self.cache[cache_key] = value |
|
|
|
def _get_immediate(self, node_id): |
|
if not self.initialized: |
|
return None |
|
cache_key = self.cache_key_set.get_data_key(node_id) |
|
if cache_key in self.cache: |
|
return self.cache[cache_key] |
|
else: |
|
return None |
|
|
|
def _ensure_subcache(self, node_id, children_ids): |
|
subcache_key = self.cache_key_set.get_subcache_key(node_id) |
|
subcache = self.subcaches.get(subcache_key, None) |
|
if subcache is None: |
|
subcache = BasicCache(self.key_class) |
|
self.subcaches[subcache_key] = subcache |
|
subcache.set_prompt(self.dynprompt, children_ids, self.is_changed_cache) |
|
return subcache |
|
|
|
def _get_subcache(self, node_id): |
|
assert self.initialized |
|
subcache_key = self.cache_key_set.get_subcache_key(node_id) |
|
if subcache_key in self.subcaches: |
|
return self.subcaches[subcache_key] |
|
else: |
|
return None |
|
|
|
def recursive_debug_dump(self): |
|
result = [] |
|
for key in self.cache: |
|
result.append({"key": key, "value": self.cache[key]}) |
|
for key in self.subcaches: |
|
result.append({"subcache_key": key, "subcache": self.subcaches[key].recursive_debug_dump()}) |
|
return result |
|
|
|
class HierarchicalCache(BasicCache): |
|
def __init__(self, key_class): |
|
super().__init__(key_class) |
|
|
|
def _get_cache_for(self, node_id): |
|
assert self.dynprompt is not None |
|
parent_id = self.dynprompt.get_parent_node_id(node_id) |
|
if parent_id is None: |
|
return self |
|
|
|
hierarchy = [] |
|
while parent_id is not None: |
|
hierarchy.append(parent_id) |
|
parent_id = self.dynprompt.get_parent_node_id(parent_id) |
|
|
|
cache = self |
|
for parent_id in reversed(hierarchy): |
|
cache = cache._get_subcache(parent_id) |
|
if cache is None: |
|
return None |
|
return cache |
|
|
|
def get(self, node_id): |
|
cache = self._get_cache_for(node_id) |
|
if cache is None: |
|
return None |
|
return cache._get_immediate(node_id) |
|
|
|
def set(self, node_id, value): |
|
cache = self._get_cache_for(node_id) |
|
assert cache is not None |
|
cache._set_immediate(node_id, value) |
|
|
|
def ensure_subcache_for(self, node_id, children_ids): |
|
cache = self._get_cache_for(node_id) |
|
assert cache is not None |
|
return cache._ensure_subcache(node_id, children_ids) |
|
|
|
class LRUCache(BasicCache): |
|
def __init__(self, key_class, max_size=100): |
|
super().__init__(key_class) |
|
self.max_size = max_size |
|
self.min_generation = 0 |
|
self.generation = 0 |
|
self.used_generation = {} |
|
self.children = {} |
|
|
|
def set_prompt(self, dynprompt, node_ids, is_changed_cache): |
|
super().set_prompt(dynprompt, node_ids, is_changed_cache) |
|
self.generation += 1 |
|
for node_id in node_ids: |
|
self._mark_used(node_id) |
|
|
|
def clean_unused(self): |
|
while len(self.cache) > self.max_size and self.min_generation < self.generation: |
|
self.min_generation += 1 |
|
to_remove = [key for key in self.cache if self.used_generation[key] < self.min_generation] |
|
for key in to_remove: |
|
del self.cache[key] |
|
del self.used_generation[key] |
|
if key in self.children: |
|
del self.children[key] |
|
self._clean_subcaches() |
|
|
|
def get(self, node_id): |
|
self._mark_used(node_id) |
|
return self._get_immediate(node_id) |
|
|
|
def _mark_used(self, node_id): |
|
cache_key = self.cache_key_set.get_data_key(node_id) |
|
if cache_key is not None: |
|
self.used_generation[cache_key] = self.generation |
|
|
|
def set(self, node_id, value): |
|
self._mark_used(node_id) |
|
return self._set_immediate(node_id, value) |
|
|
|
def ensure_subcache_for(self, node_id, children_ids): |
|
|
|
super()._ensure_subcache(node_id, children_ids) |
|
|
|
self.cache_key_set.add_keys(children_ids) |
|
self._mark_used(node_id) |
|
cache_key = self.cache_key_set.get_data_key(node_id) |
|
self.children[cache_key] = [] |
|
for child_id in children_ids: |
|
self._mark_used(child_id) |
|
self.children[cache_key].append(self.cache_key_set.get_data_key(child_id)) |
|
return self |
|
|
|
|