# 1st edit by https://github.com/comfyanonymous/ComfyUI # 2nd edit by Forge Official import ldm_patched.modules.supported_models import ldm_patched.modules.supported_models_base def count_blocks(state_dict_keys, prefix_string): count = 0 while True: c = False for k in state_dict_keys: if k.startswith(prefix_string.format(count)): c = True break if c == False: break count += 1 return count def calculate_transformer_depth(prefix, state_dict_keys, state_dict): context_dim = None use_linear_in_transformer = False transformer_prefix = prefix + "1.transformer_blocks." transformer_keys = sorted( list(filter(lambda a: a.startswith(transformer_prefix), state_dict_keys)) ) if len(transformer_keys) > 0: last_transformer_depth = count_blocks( state_dict_keys, transformer_prefix + "{}" ) context_dim = state_dict[ "{}0.attn2.to_k.weight".format(transformer_prefix) ].shape[1] use_linear_in_transformer = ( len(state_dict["{}1.proj_in.weight".format(prefix)].shape) == 2 ) time_stack = ( "{}1.time_stack.0.attn1.to_q.weight".format(prefix) in state_dict or "{}1.time_mix_blocks.0.attn1.to_q.weight".format(prefix) in state_dict ) return ( last_transformer_depth, context_dim, use_linear_in_transformer, time_stack, ) return None def detect_unet_config(state_dict, key_prefix, dtype): state_dict_keys = list(state_dict.keys()) unet_config = { "use_checkpoint": False, "image_size": 32, "use_spatial_transformer": True, "legacy": False, } y_input = "{}label_emb.0.0.weight".format(key_prefix) if y_input in state_dict_keys: unet_config["num_classes"] = "sequential" unet_config["adm_in_channels"] = state_dict[y_input].shape[1] else: unet_config["adm_in_channels"] = None unet_config["dtype"] = dtype model_channels = state_dict["{}input_blocks.0.0.weight".format(key_prefix)].shape[0] in_channels = state_dict["{}input_blocks.0.0.weight".format(key_prefix)].shape[1] out_key = "{}out.2.weight".format(key_prefix) if out_key in state_dict: out_channels = state_dict[out_key].shape[0] else: out_channels = 4 num_res_blocks = [] channel_mult = [] attention_resolutions = [] transformer_depth = [] transformer_depth_output = [] context_dim = None use_linear_in_transformer = False video_model = False current_res = 1 count = 0 last_res_blocks = 0 last_channel_mult = 0 input_block_count = count_blocks( state_dict_keys, "{}input_blocks".format(key_prefix) + ".{}." ) for count in range(input_block_count): prefix = "{}input_blocks.{}.".format(key_prefix, count) prefix_output = "{}output_blocks.{}.".format( key_prefix, input_block_count - count - 1 ) block_keys = sorted( list(filter(lambda a: a.startswith(prefix), state_dict_keys)) ) if len(block_keys) == 0: break block_keys_output = sorted( list(filter(lambda a: a.startswith(prefix_output), state_dict_keys)) ) if "{}0.op.weight".format(prefix) in block_keys: # new layer num_res_blocks.append(last_res_blocks) channel_mult.append(last_channel_mult) current_res *= 2 last_res_blocks = 0 last_channel_mult = 0 out = calculate_transformer_depth( prefix_output, state_dict_keys, state_dict ) if out is not None: transformer_depth_output.append(out[0]) else: transformer_depth_output.append(0) else: res_block_prefix = "{}0.in_layers.0.weight".format(prefix) if res_block_prefix in block_keys: last_res_blocks += 1 last_channel_mult = ( state_dict["{}0.out_layers.3.weight".format(prefix)].shape[0] // model_channels ) out = calculate_transformer_depth(prefix, state_dict_keys, state_dict) if out is not None: transformer_depth.append(out[0]) if context_dim is None: context_dim = out[1] use_linear_in_transformer = out[2] video_model = out[3] else: transformer_depth.append(0) res_block_prefix = "{}0.in_layers.0.weight".format(prefix_output) if res_block_prefix in block_keys_output: out = calculate_transformer_depth( prefix_output, state_dict_keys, state_dict ) if out is not None: transformer_depth_output.append(out[0]) else: transformer_depth_output.append(0) num_res_blocks.append(last_res_blocks) channel_mult.append(last_channel_mult) if "{}middle_block.1.proj_in.weight".format(key_prefix) in state_dict_keys: transformer_depth_middle = count_blocks( state_dict_keys, "{}middle_block.1.transformer_blocks.".format(key_prefix) + "{}", ) else: transformer_depth_middle = -1 unet_config["in_channels"] = in_channels unet_config["out_channels"] = out_channels unet_config["model_channels"] = model_channels unet_config["num_res_blocks"] = num_res_blocks unet_config["transformer_depth"] = transformer_depth unet_config["transformer_depth_output"] = transformer_depth_output unet_config["channel_mult"] = channel_mult unet_config["transformer_depth_middle"] = transformer_depth_middle unet_config["use_linear_in_transformer"] = use_linear_in_transformer unet_config["context_dim"] = context_dim if video_model: unet_config["extra_ff_mix_layer"] = True unet_config["use_spatial_context"] = True unet_config["merge_strategy"] = "learned_with_images" unet_config["merge_factor"] = 0.0 unet_config["video_kernel_size"] = [3, 1, 1] unet_config["use_temporal_resblock"] = True unet_config["use_temporal_attention"] = True else: unet_config["use_temporal_resblock"] = False unet_config["use_temporal_attention"] = False return unet_config def model_config_from_unet_config(unet_config): for model_config in ldm_patched.modules.supported_models.models: if model_config.matches(unet_config): return model_config(unet_config) print("no match", unet_config) return None def model_config_from_unet( state_dict, unet_key_prefix, dtype, use_base_if_no_match=False ): unet_config = detect_unet_config(state_dict, unet_key_prefix, dtype) model_config = model_config_from_unet_config(unet_config) if model_config is None and use_base_if_no_match: return ldm_patched.modules.supported_models_base.BASE(unet_config) else: return model_config def convert_config(unet_config): new_config = unet_config.copy() num_res_blocks = new_config.get("num_res_blocks", None) channel_mult = new_config.get("channel_mult", None) if isinstance(num_res_blocks, int): num_res_blocks = len(channel_mult) * [num_res_blocks] if "attention_resolutions" in new_config: attention_resolutions = new_config.pop("attention_resolutions") transformer_depth = new_config.get("transformer_depth", None) transformer_depth_middle = new_config.get("transformer_depth_middle", None) if isinstance(transformer_depth, int): transformer_depth = len(channel_mult) * [transformer_depth] if transformer_depth_middle is None: transformer_depth_middle = transformer_depth[-1] t_in = [] t_out = [] s = 1 for i in range(len(num_res_blocks)): res = num_res_blocks[i] d = 0 if s in attention_resolutions: d = transformer_depth[i] t_in += [d] * res t_out += [d] * (res + 1) s *= 2 transformer_depth = t_in transformer_depth_output = t_out new_config["transformer_depth"] = t_in new_config["transformer_depth_output"] = t_out new_config["transformer_depth_middle"] = transformer_depth_middle new_config["num_res_blocks"] = num_res_blocks return new_config def unet_config_from_diffusers_unet(state_dict, dtype): match = {} transformer_depth = [] attn_res = 1 down_blocks = count_blocks(state_dict, "down_blocks.{}") for i in range(down_blocks): attn_blocks = count_blocks( state_dict, "down_blocks.{}.attentions.".format(i) + "{}" ) for ab in range(attn_blocks): transformer_count = count_blocks( state_dict, "down_blocks.{}.attentions.{}.transformer_blocks.".format(i, ab) + "{}", ) transformer_depth.append(transformer_count) if transformer_count > 0: match["context_dim"] = state_dict[ "down_blocks.{}.attentions.{}.transformer_blocks.0.attn2.to_k.weight".format( i, ab ) ].shape[1] attn_res *= 2 if attn_blocks == 0: transformer_depth.append(0) transformer_depth.append(0) match["transformer_depth"] = transformer_depth match["model_channels"] = state_dict["conv_in.weight"].shape[0] match["in_channels"] = state_dict["conv_in.weight"].shape[1] match["adm_in_channels"] = None if "class_embedding.linear_1.weight" in state_dict: match["adm_in_channels"] = state_dict["class_embedding.linear_1.weight"].shape[ 1 ] elif "add_embedding.linear_1.weight" in state_dict: match["adm_in_channels"] = state_dict["add_embedding.linear_1.weight"].shape[1] SDXL = { "use_checkpoint": False, "image_size": 32, "out_channels": 4, "use_spatial_transformer": True, "legacy": False, "num_classes": "sequential", "adm_in_channels": 2816, "dtype": dtype, "in_channels": 4, "model_channels": 320, "num_res_blocks": [2, 2, 2], "transformer_depth": [0, 0, 2, 2, 10, 10], "channel_mult": [1, 2, 4], "transformer_depth_middle": 10, "use_linear_in_transformer": True, "context_dim": 2048, "num_head_channels": 64, "transformer_depth_output": [0, 0, 0, 2, 2, 2, 10, 10, 10], "use_temporal_attention": False, "use_temporal_resblock": False, } SDXL_refiner = { "use_checkpoint": False, "image_size": 32, "out_channels": 4, "use_spatial_transformer": True, "legacy": False, "num_classes": "sequential", "adm_in_channels": 2560, "dtype": dtype, "in_channels": 4, "model_channels": 384, "num_res_blocks": [2, 2, 2, 2], "transformer_depth": [0, 0, 4, 4, 4, 4, 0, 0], "channel_mult": [1, 2, 4, 4], "transformer_depth_middle": 4, "use_linear_in_transformer": True, "context_dim": 1280, "num_head_channels": 64, "transformer_depth_output": [0, 0, 0, 4, 4, 4, 4, 4, 4, 0, 0, 0], "use_temporal_attention": False, "use_temporal_resblock": False, } SD15 = { "use_checkpoint": False, "image_size": 32, "out_channels": 4, "use_spatial_transformer": True, "legacy": False, "adm_in_channels": None, "dtype": dtype, "in_channels": 4, "model_channels": 320, "num_res_blocks": [2, 2, 2, 2], "transformer_depth": [1, 1, 1, 1, 1, 1, 0, 0], "channel_mult": [1, 2, 4, 4], "transformer_depth_middle": 1, "use_linear_in_transformer": False, "context_dim": 768, "num_heads": 8, "transformer_depth_output": [1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0], "use_temporal_attention": False, "use_temporal_resblock": False, } SDXL_mid_cnet = { "use_checkpoint": False, "image_size": 32, "out_channels": 4, "use_spatial_transformer": True, "legacy": False, "num_classes": "sequential", "adm_in_channels": 2816, "dtype": dtype, "in_channels": 4, "model_channels": 320, "num_res_blocks": [2, 2, 2], "transformer_depth": [0, 0, 0, 0, 1, 1], "channel_mult": [1, 2, 4], "transformer_depth_middle": 1, "use_linear_in_transformer": True, "context_dim": 2048, "num_head_channels": 64, "transformer_depth_output": [0, 0, 0, 0, 0, 0, 1, 1, 1], "use_temporal_attention": False, "use_temporal_resblock": False, } SDXL_small_cnet = { "use_checkpoint": False, "image_size": 32, "out_channels": 4, "use_spatial_transformer": True, "legacy": False, "num_classes": "sequential", "adm_in_channels": 2816, "dtype": dtype, "in_channels": 4, "model_channels": 320, "num_res_blocks": [2, 2, 2], "transformer_depth": [0, 0, 0, 0, 0, 0], "channel_mult": [1, 2, 4], "transformer_depth_middle": 0, "use_linear_in_transformer": True, "num_head_channels": 64, "context_dim": 1, "transformer_depth_output": [0, 0, 0, 0, 0, 0, 0, 0, 0], "use_temporal_attention": False, "use_temporal_resblock": False, } SDXL_diffusers_inpaint = { "use_checkpoint": False, "image_size": 32, "out_channels": 4, "use_spatial_transformer": True, "legacy": False, "num_classes": "sequential", "adm_in_channels": 2816, "dtype": dtype, "in_channels": 9, "model_channels": 320, "num_res_blocks": [2, 2, 2], "transformer_depth": [0, 0, 2, 2, 10, 10], "channel_mult": [1, 2, 4], "transformer_depth_middle": 10, "use_linear_in_transformer": True, "context_dim": 2048, "num_head_channels": 64, "transformer_depth_output": [0, 0, 0, 2, 2, 2, 10, 10, 10], "use_temporal_attention": False, "use_temporal_resblock": False, } SSD_1B = { "use_checkpoint": False, "image_size": 32, "out_channels": 4, "use_spatial_transformer": True, "legacy": False, "num_classes": "sequential", "adm_in_channels": 2816, "dtype": dtype, "in_channels": 4, "model_channels": 320, "num_res_blocks": [2, 2, 2], "transformer_depth": [0, 0, 2, 2, 4, 4], "transformer_depth_output": [0, 0, 0, 1, 1, 2, 10, 4, 4], "channel_mult": [1, 2, 4], "transformer_depth_middle": -1, "use_linear_in_transformer": True, "context_dim": 2048, "num_head_channels": 64, "use_temporal_attention": False, "use_temporal_resblock": False, } Segmind_Vega = { "use_checkpoint": False, "image_size": 32, "out_channels": 4, "use_spatial_transformer": True, "legacy": False, "num_classes": "sequential", "adm_in_channels": 2816, "dtype": dtype, "in_channels": 4, "model_channels": 320, "num_res_blocks": [2, 2, 2], "transformer_depth": [0, 0, 1, 1, 2, 2], "transformer_depth_output": [0, 0, 0, 1, 1, 1, 2, 2, 2], "channel_mult": [1, 2, 4], "transformer_depth_middle": -1, "use_linear_in_transformer": True, "context_dim": 2048, "num_head_channels": 64, "use_temporal_attention": False, "use_temporal_resblock": False, } supported_models = [ SDXL, SDXL_refiner, SD15, SDXL_mid_cnet, SDXL_small_cnet, SDXL_diffusers_inpaint, SSD_1B, Segmind_Vega, ] for unet_config in supported_models: matches = True for k in match: if match[k] != unet_config[k]: matches = False break if matches: return convert_config(unet_config) return None def model_config_from_diffusers_unet(state_dict, dtype): unet_config = unet_config_from_diffusers_unet(state_dict, dtype) if unet_config is not None: return model_config_from_unet_config(unet_config) return None