|
""" |
|
This file is part of ComfyUI. |
|
Copyright (C) 2024 Comfy |
|
|
|
This program is free software: you can redistribute it and/or modify |
|
it under the terms of the GNU General Public License as published by |
|
the Free Software Foundation, either version 3 of the License, or |
|
(at your option) any later version. |
|
|
|
This program is distributed in the hope that it will be useful, |
|
but WITHOUT ANY WARRANTY; without even the implied warranty of |
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
|
GNU General Public License for more details. |
|
|
|
You should have received a copy of the GNU General Public License |
|
along with this program. If not, see <https://www.gnu.org/licenses/>. |
|
""" |
|
|
|
import torch |
|
from . import model_base |
|
from . import utils |
|
from . import latent_formats |
|
|
|
class ClipTarget: |
|
def __init__(self, tokenizer, clip): |
|
self.clip = clip |
|
self.tokenizer = tokenizer |
|
self.params = {} |
|
|
|
class BASE: |
|
unet_config = {} |
|
unet_extra_config = { |
|
"num_heads": -1, |
|
"num_head_channels": 64, |
|
} |
|
|
|
required_keys = {} |
|
|
|
clip_prefix = [] |
|
clip_vision_prefix = None |
|
noise_aug_config = None |
|
sampling_settings = {} |
|
latent_format = latent_formats.LatentFormat |
|
vae_key_prefix = ["first_stage_model."] |
|
text_encoder_key_prefix = ["cond_stage_model."] |
|
supported_inference_dtypes = [torch.float16, torch.bfloat16, torch.float32] |
|
|
|
memory_usage_factor = 2.0 |
|
|
|
manual_cast_dtype = None |
|
custom_operations = None |
|
scaled_fp8 = None |
|
optimizations = {"fp8": False} |
|
|
|
@classmethod |
|
def matches(s, unet_config, state_dict=None): |
|
for k in s.unet_config: |
|
if k not in unet_config or s.unet_config[k] != unet_config[k]: |
|
return False |
|
if state_dict is not None: |
|
for k in s.required_keys: |
|
if k not in state_dict: |
|
return False |
|
return True |
|
|
|
def model_type(self, state_dict, prefix=""): |
|
return model_base.ModelType.EPS |
|
|
|
def inpaint_model(self): |
|
return self.unet_config["in_channels"] > 4 |
|
|
|
def __init__(self, unet_config): |
|
self.unet_config = unet_config.copy() |
|
self.sampling_settings = self.sampling_settings.copy() |
|
self.latent_format = self.latent_format() |
|
self.optimizations = self.optimizations.copy() |
|
for x in self.unet_extra_config: |
|
self.unet_config[x] = self.unet_extra_config[x] |
|
|
|
def get_model(self, state_dict, prefix="", device=None): |
|
if self.noise_aug_config is not None: |
|
out = model_base.SD21UNCLIP(self, self.noise_aug_config, model_type=self.model_type(state_dict, prefix), device=device) |
|
else: |
|
out = model_base.BaseModel(self, model_type=self.model_type(state_dict, prefix), device=device) |
|
if self.inpaint_model(): |
|
out.set_inpaint() |
|
return out |
|
|
|
def process_clip_state_dict(self, state_dict): |
|
state_dict = utils.state_dict_prefix_replace(state_dict, {k: "" for k in self.text_encoder_key_prefix}, filter_keys=True) |
|
return state_dict |
|
|
|
def process_unet_state_dict(self, state_dict): |
|
return state_dict |
|
|
|
def process_vae_state_dict(self, state_dict): |
|
return state_dict |
|
|
|
def process_clip_state_dict_for_saving(self, state_dict): |
|
replace_prefix = {"": self.text_encoder_key_prefix[0]} |
|
return utils.state_dict_prefix_replace(state_dict, replace_prefix) |
|
|
|
def process_clip_vision_state_dict_for_saving(self, state_dict): |
|
replace_prefix = {} |
|
if self.clip_vision_prefix is not None: |
|
replace_prefix[""] = self.clip_vision_prefix |
|
return utils.state_dict_prefix_replace(state_dict, replace_prefix) |
|
|
|
def process_unet_state_dict_for_saving(self, state_dict): |
|
replace_prefix = {"": "model.diffusion_model."} |
|
return utils.state_dict_prefix_replace(state_dict, replace_prefix) |
|
|
|
def process_vae_state_dict_for_saving(self, state_dict): |
|
replace_prefix = {"": self.vae_key_prefix[0]} |
|
return utils.state_dict_prefix_replace(state_dict, replace_prefix) |
|
|
|
def set_inference_dtype(self, dtype, manual_cast_dtype): |
|
self.unet_config['dtype'] = dtype |
|
self.manual_cast_dtype = manual_cast_dtype |
|
|