WhiteAiZ's picture
Upload 1420 files
0070fce verified
# 1st edit by https://github.com/comfyanonymous/ComfyUI
# 2nd edit by Forge Official
import ldm_patched.modules.supported_models
import ldm_patched.modules.supported_models_base
def count_blocks(state_dict_keys, prefix_string):
count = 0
while True:
c = False
for k in state_dict_keys:
if k.startswith(prefix_string.format(count)):
c = True
break
if c == False:
break
count += 1
return count
def calculate_transformer_depth(prefix, state_dict_keys, state_dict):
context_dim = None
use_linear_in_transformer = False
transformer_prefix = prefix + "1.transformer_blocks."
transformer_keys = sorted(
list(filter(lambda a: a.startswith(transformer_prefix), state_dict_keys))
)
if len(transformer_keys) > 0:
last_transformer_depth = count_blocks(
state_dict_keys, transformer_prefix + "{}"
)
context_dim = state_dict[
"{}0.attn2.to_k.weight".format(transformer_prefix)
].shape[1]
use_linear_in_transformer = (
len(state_dict["{}1.proj_in.weight".format(prefix)].shape) == 2
)
time_stack = (
"{}1.time_stack.0.attn1.to_q.weight".format(prefix) in state_dict
or "{}1.time_mix_blocks.0.attn1.to_q.weight".format(prefix) in state_dict
)
return (
last_transformer_depth,
context_dim,
use_linear_in_transformer,
time_stack,
)
return None
def detect_unet_config(state_dict, key_prefix, dtype):
state_dict_keys = list(state_dict.keys())
unet_config = {
"use_checkpoint": False,
"image_size": 32,
"use_spatial_transformer": True,
"legacy": False,
}
y_input = "{}label_emb.0.0.weight".format(key_prefix)
if y_input in state_dict_keys:
unet_config["num_classes"] = "sequential"
unet_config["adm_in_channels"] = state_dict[y_input].shape[1]
else:
unet_config["adm_in_channels"] = None
unet_config["dtype"] = dtype
model_channels = state_dict["{}input_blocks.0.0.weight".format(key_prefix)].shape[0]
in_channels = state_dict["{}input_blocks.0.0.weight".format(key_prefix)].shape[1]
out_key = "{}out.2.weight".format(key_prefix)
if out_key in state_dict:
out_channels = state_dict[out_key].shape[0]
else:
out_channels = 4
num_res_blocks = []
channel_mult = []
attention_resolutions = []
transformer_depth = []
transformer_depth_output = []
context_dim = None
use_linear_in_transformer = False
video_model = False
current_res = 1
count = 0
last_res_blocks = 0
last_channel_mult = 0
input_block_count = count_blocks(
state_dict_keys, "{}input_blocks".format(key_prefix) + ".{}."
)
for count in range(input_block_count):
prefix = "{}input_blocks.{}.".format(key_prefix, count)
prefix_output = "{}output_blocks.{}.".format(
key_prefix, input_block_count - count - 1
)
block_keys = sorted(
list(filter(lambda a: a.startswith(prefix), state_dict_keys))
)
if len(block_keys) == 0:
break
block_keys_output = sorted(
list(filter(lambda a: a.startswith(prefix_output), state_dict_keys))
)
if "{}0.op.weight".format(prefix) in block_keys: # new layer
num_res_blocks.append(last_res_blocks)
channel_mult.append(last_channel_mult)
current_res *= 2
last_res_blocks = 0
last_channel_mult = 0
out = calculate_transformer_depth(
prefix_output, state_dict_keys, state_dict
)
if out is not None:
transformer_depth_output.append(out[0])
else:
transformer_depth_output.append(0)
else:
res_block_prefix = "{}0.in_layers.0.weight".format(prefix)
if res_block_prefix in block_keys:
last_res_blocks += 1
last_channel_mult = (
state_dict["{}0.out_layers.3.weight".format(prefix)].shape[0]
// model_channels
)
out = calculate_transformer_depth(prefix, state_dict_keys, state_dict)
if out is not None:
transformer_depth.append(out[0])
if context_dim is None:
context_dim = out[1]
use_linear_in_transformer = out[2]
video_model = out[3]
else:
transformer_depth.append(0)
res_block_prefix = "{}0.in_layers.0.weight".format(prefix_output)
if res_block_prefix in block_keys_output:
out = calculate_transformer_depth(
prefix_output, state_dict_keys, state_dict
)
if out is not None:
transformer_depth_output.append(out[0])
else:
transformer_depth_output.append(0)
num_res_blocks.append(last_res_blocks)
channel_mult.append(last_channel_mult)
if "{}middle_block.1.proj_in.weight".format(key_prefix) in state_dict_keys:
transformer_depth_middle = count_blocks(
state_dict_keys,
"{}middle_block.1.transformer_blocks.".format(key_prefix) + "{}",
)
else:
transformer_depth_middle = -1
unet_config["in_channels"] = in_channels
unet_config["out_channels"] = out_channels
unet_config["model_channels"] = model_channels
unet_config["num_res_blocks"] = num_res_blocks
unet_config["transformer_depth"] = transformer_depth
unet_config["transformer_depth_output"] = transformer_depth_output
unet_config["channel_mult"] = channel_mult
unet_config["transformer_depth_middle"] = transformer_depth_middle
unet_config["use_linear_in_transformer"] = use_linear_in_transformer
unet_config["context_dim"] = context_dim
if video_model:
unet_config["extra_ff_mix_layer"] = True
unet_config["use_spatial_context"] = True
unet_config["merge_strategy"] = "learned_with_images"
unet_config["merge_factor"] = 0.0
unet_config["video_kernel_size"] = [3, 1, 1]
unet_config["use_temporal_resblock"] = True
unet_config["use_temporal_attention"] = True
else:
unet_config["use_temporal_resblock"] = False
unet_config["use_temporal_attention"] = False
return unet_config
def model_config_from_unet_config(unet_config):
for model_config in ldm_patched.modules.supported_models.models:
if model_config.matches(unet_config):
return model_config(unet_config)
print("no match", unet_config)
return None
def model_config_from_unet(
state_dict, unet_key_prefix, dtype, use_base_if_no_match=False
):
unet_config = detect_unet_config(state_dict, unet_key_prefix, dtype)
model_config = model_config_from_unet_config(unet_config)
if model_config is None and use_base_if_no_match:
return ldm_patched.modules.supported_models_base.BASE(unet_config)
else:
return model_config
def convert_config(unet_config):
new_config = unet_config.copy()
num_res_blocks = new_config.get("num_res_blocks", None)
channel_mult = new_config.get("channel_mult", None)
if isinstance(num_res_blocks, int):
num_res_blocks = len(channel_mult) * [num_res_blocks]
if "attention_resolutions" in new_config:
attention_resolutions = new_config.pop("attention_resolutions")
transformer_depth = new_config.get("transformer_depth", None)
transformer_depth_middle = new_config.get("transformer_depth_middle", None)
if isinstance(transformer_depth, int):
transformer_depth = len(channel_mult) * [transformer_depth]
if transformer_depth_middle is None:
transformer_depth_middle = transformer_depth[-1]
t_in = []
t_out = []
s = 1
for i in range(len(num_res_blocks)):
res = num_res_blocks[i]
d = 0
if s in attention_resolutions:
d = transformer_depth[i]
t_in += [d] * res
t_out += [d] * (res + 1)
s *= 2
transformer_depth = t_in
transformer_depth_output = t_out
new_config["transformer_depth"] = t_in
new_config["transformer_depth_output"] = t_out
new_config["transformer_depth_middle"] = transformer_depth_middle
new_config["num_res_blocks"] = num_res_blocks
return new_config
def unet_config_from_diffusers_unet(state_dict, dtype):
match = {}
transformer_depth = []
attn_res = 1
down_blocks = count_blocks(state_dict, "down_blocks.{}")
for i in range(down_blocks):
attn_blocks = count_blocks(
state_dict, "down_blocks.{}.attentions.".format(i) + "{}"
)
for ab in range(attn_blocks):
transformer_count = count_blocks(
state_dict,
"down_blocks.{}.attentions.{}.transformer_blocks.".format(i, ab) + "{}",
)
transformer_depth.append(transformer_count)
if transformer_count > 0:
match["context_dim"] = state_dict[
"down_blocks.{}.attentions.{}.transformer_blocks.0.attn2.to_k.weight".format(
i, ab
)
].shape[1]
attn_res *= 2
if attn_blocks == 0:
transformer_depth.append(0)
transformer_depth.append(0)
match["transformer_depth"] = transformer_depth
match["model_channels"] = state_dict["conv_in.weight"].shape[0]
match["in_channels"] = state_dict["conv_in.weight"].shape[1]
match["adm_in_channels"] = None
if "class_embedding.linear_1.weight" in state_dict:
match["adm_in_channels"] = state_dict["class_embedding.linear_1.weight"].shape[
1
]
elif "add_embedding.linear_1.weight" in state_dict:
match["adm_in_channels"] = state_dict["add_embedding.linear_1.weight"].shape[1]
SDXL = {
"use_checkpoint": False,
"image_size": 32,
"out_channels": 4,
"use_spatial_transformer": True,
"legacy": False,
"num_classes": "sequential",
"adm_in_channels": 2816,
"dtype": dtype,
"in_channels": 4,
"model_channels": 320,
"num_res_blocks": [2, 2, 2],
"transformer_depth": [0, 0, 2, 2, 10, 10],
"channel_mult": [1, 2, 4],
"transformer_depth_middle": 10,
"use_linear_in_transformer": True,
"context_dim": 2048,
"num_head_channels": 64,
"transformer_depth_output": [0, 0, 0, 2, 2, 2, 10, 10, 10],
"use_temporal_attention": False,
"use_temporal_resblock": False,
}
SDXL_refiner = {
"use_checkpoint": False,
"image_size": 32,
"out_channels": 4,
"use_spatial_transformer": True,
"legacy": False,
"num_classes": "sequential",
"adm_in_channels": 2560,
"dtype": dtype,
"in_channels": 4,
"model_channels": 384,
"num_res_blocks": [2, 2, 2, 2],
"transformer_depth": [0, 0, 4, 4, 4, 4, 0, 0],
"channel_mult": [1, 2, 4, 4],
"transformer_depth_middle": 4,
"use_linear_in_transformer": True,
"context_dim": 1280,
"num_head_channels": 64,
"transformer_depth_output": [0, 0, 0, 4, 4, 4, 4, 4, 4, 0, 0, 0],
"use_temporal_attention": False,
"use_temporal_resblock": False,
}
SD15 = {
"use_checkpoint": False,
"image_size": 32,
"out_channels": 4,
"use_spatial_transformer": True,
"legacy": False,
"adm_in_channels": None,
"dtype": dtype,
"in_channels": 4,
"model_channels": 320,
"num_res_blocks": [2, 2, 2, 2],
"transformer_depth": [1, 1, 1, 1, 1, 1, 0, 0],
"channel_mult": [1, 2, 4, 4],
"transformer_depth_middle": 1,
"use_linear_in_transformer": False,
"context_dim": 768,
"num_heads": 8,
"transformer_depth_output": [1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0],
"use_temporal_attention": False,
"use_temporal_resblock": False,
}
SDXL_mid_cnet = {
"use_checkpoint": False,
"image_size": 32,
"out_channels": 4,
"use_spatial_transformer": True,
"legacy": False,
"num_classes": "sequential",
"adm_in_channels": 2816,
"dtype": dtype,
"in_channels": 4,
"model_channels": 320,
"num_res_blocks": [2, 2, 2],
"transformer_depth": [0, 0, 0, 0, 1, 1],
"channel_mult": [1, 2, 4],
"transformer_depth_middle": 1,
"use_linear_in_transformer": True,
"context_dim": 2048,
"num_head_channels": 64,
"transformer_depth_output": [0, 0, 0, 0, 0, 0, 1, 1, 1],
"use_temporal_attention": False,
"use_temporal_resblock": False,
}
SDXL_small_cnet = {
"use_checkpoint": False,
"image_size": 32,
"out_channels": 4,
"use_spatial_transformer": True,
"legacy": False,
"num_classes": "sequential",
"adm_in_channels": 2816,
"dtype": dtype,
"in_channels": 4,
"model_channels": 320,
"num_res_blocks": [2, 2, 2],
"transformer_depth": [0, 0, 0, 0, 0, 0],
"channel_mult": [1, 2, 4],
"transformer_depth_middle": 0,
"use_linear_in_transformer": True,
"num_head_channels": 64,
"context_dim": 1,
"transformer_depth_output": [0, 0, 0, 0, 0, 0, 0, 0, 0],
"use_temporal_attention": False,
"use_temporal_resblock": False,
}
SDXL_diffusers_inpaint = {
"use_checkpoint": False,
"image_size": 32,
"out_channels": 4,
"use_spatial_transformer": True,
"legacy": False,
"num_classes": "sequential",
"adm_in_channels": 2816,
"dtype": dtype,
"in_channels": 9,
"model_channels": 320,
"num_res_blocks": [2, 2, 2],
"transformer_depth": [0, 0, 2, 2, 10, 10],
"channel_mult": [1, 2, 4],
"transformer_depth_middle": 10,
"use_linear_in_transformer": True,
"context_dim": 2048,
"num_head_channels": 64,
"transformer_depth_output": [0, 0, 0, 2, 2, 2, 10, 10, 10],
"use_temporal_attention": False,
"use_temporal_resblock": False,
}
SSD_1B = {
"use_checkpoint": False,
"image_size": 32,
"out_channels": 4,
"use_spatial_transformer": True,
"legacy": False,
"num_classes": "sequential",
"adm_in_channels": 2816,
"dtype": dtype,
"in_channels": 4,
"model_channels": 320,
"num_res_blocks": [2, 2, 2],
"transformer_depth": [0, 0, 2, 2, 4, 4],
"transformer_depth_output": [0, 0, 0, 1, 1, 2, 10, 4, 4],
"channel_mult": [1, 2, 4],
"transformer_depth_middle": -1,
"use_linear_in_transformer": True,
"context_dim": 2048,
"num_head_channels": 64,
"use_temporal_attention": False,
"use_temporal_resblock": False,
}
Segmind_Vega = {
"use_checkpoint": False,
"image_size": 32,
"out_channels": 4,
"use_spatial_transformer": True,
"legacy": False,
"num_classes": "sequential",
"adm_in_channels": 2816,
"dtype": dtype,
"in_channels": 4,
"model_channels": 320,
"num_res_blocks": [2, 2, 2],
"transformer_depth": [0, 0, 1, 1, 2, 2],
"transformer_depth_output": [0, 0, 0, 1, 1, 1, 2, 2, 2],
"channel_mult": [1, 2, 4],
"transformer_depth_middle": -1,
"use_linear_in_transformer": True,
"context_dim": 2048,
"num_head_channels": 64,
"use_temporal_attention": False,
"use_temporal_resblock": False,
}
supported_models = [
SDXL,
SDXL_refiner,
SD15,
SDXL_mid_cnet,
SDXL_small_cnet,
SDXL_diffusers_inpaint,
SSD_1B,
Segmind_Vega,
]
for unet_config in supported_models:
matches = True
for k in match:
if match[k] != unet_config[k]:
matches = False
break
if matches:
return convert_config(unet_config)
return None
def model_config_from_diffusers_unet(state_dict, dtype):
unet_config = unet_config_from_diffusers_unet(state_dict, dtype)
if unet_config is not None:
return model_config_from_unet_config(unet_config)
return None