id
int64
0
190k
prompt
stringlengths
21
13.4M
docstring
stringlengths
1
12k
151,826
import os import time input_directory = os.path.join(os.path.dirname(os.path.realpath(__file__)), "input") def set_input_directory(input_dir): global input_directory input_directory = input_dir
null
151,828
import os import time if not os.path.exists(input_directory): try: os.makedirs(input_directory) except: print("Failed to create input directory") def get_input_directory(): global input_directory return input_directory def annotated_filepath(name): if name.endswith("[output]"): ...
null
151,829
import os import time if not os.path.exists(input_directory): try: os.makedirs(input_directory) except: print("Failed to create input directory") def get_input_directory(): global input_directory return input_directory def annotated_filepath(name): if name.endswith("[output]"): ...
null
151,831
import os import time folder_names_and_paths = {} folder_names_and_paths["checkpoints"] = ([os.path.join(models_dir, "checkpoints")], supported_pt_extensions) folder_names_and_paths["configs"] = ([os.path.join(models_dir, "configs")], [".yaml"]) folder_names_and_paths["loras"] = ([os.path.join(models_dir, "loras")], su...
null
151,832
import os import time folder_names_and_paths = {} folder_names_and_paths["checkpoints"] = ([os.path.join(models_dir, "checkpoints")], supported_pt_extensions) folder_names_and_paths["configs"] = ([os.path.join(models_dir, "configs")], [".yaml"]) folder_names_and_paths["loras"] = ([os.path.join(models_dir, "loras")], su...
null
151,833
import os import time filename_list_cache = {} def get_filename_list_(folder_name): def cached_filename_list_(folder_name): def get_filename_list(folder_name): out = cached_filename_list_(folder_name) if out is None: out = get_filename_list_(folder_name) global filename_list_cache filen...
null
151,834
import os import time if not os.path.exists(input_directory): try: os.makedirs(input_directory) except: print("Failed to create input directory") def get_save_image_path(filename_prefix, output_dir, image_width=0, image_height=0): def map_filename(filename): prefix_len = len(os.path...
null
151,835
import os import importlib.util from comfy.cli_args import args import subprocess def get_gpu_names(): if os.name == 'nt': import ctypes # Define necessary C structures and types class DISPLAY_DEVICEA(ctypes.Structure): _fields_ = [ ('cb', ctypes.c_ulong), ...
null
151,836
import os import sys import asyncio import traceback import nodes import folder_paths import execution import uuid import urllib import json import glob import struct from PIL import Image, ImageOps from PIL.PngImagePlugin import PngInfo from io import BytesIO import aiohttp from aiohttp import web import logging impor...
null
151,837
import os import sys import asyncio import traceback import nodes import folder_paths import execution import uuid import urllib import json import glob import struct from PIL import Image, ImageOps from PIL.PngImagePlugin import PngInfo from io import BytesIO import aiohttp from aiohttp import web import logging impor...
null
151,838
import os import sys import asyncio import traceback import nodes import folder_paths import execution import uuid import urllib import json import glob import struct from PIL import Image, ImageOps from PIL.PngImagePlugin import PngInfo from io import BytesIO import aiohttp from aiohttp import web import logging impor...
null
151,839
from .k_diffusion import sampling as k_diffusion_sampling from .extra_samplers import uni_pc import torch import collections from comfy import model_management import math import logging def calc_cond_uncond_batch(model, cond, uncond, x_in, timestep, model_options): def sampling_function(model, x, timestep, uncond, co...
null
151,840
from .k_diffusion import sampling as k_diffusion_sampling from .extra_samplers import uni_pc import torch import collections from comfy import model_management import math import logging def simple_scheduler(model, steps): s = model.model_sampling sigs = [] ss = len(s.sigmas) / steps for x in range(step...
null
151,841
from .k_diffusion import sampling as k_diffusion_sampling from .extra_samplers import uni_pc import torch import collections from comfy import model_management import math import logging class KSAMPLER(Sampler): def __init__(self, sampler_function, extra_options={}, inpaint_options={}): self.sampler_functio...
null
151,842
import re import torch import logging unet_conversion_map = [ # (stable-diffusion, HF Diffusers) ("time_embed.0.weight", "time_embedding.linear_1.weight"), ("time_embed.0.bias", "time_embedding.linear_1.bias"), ("time_embed.2.weight", "time_embedding.linear_2.weight"), ("time_embed.2.bias", "time_em...
null
151,843
import re import torch import logging vae_conversion_map = [ # (stable-diffusion, HF Diffusers) ("nin_shortcut", "conv_shortcut"), ("norm_out", "conv_norm_out"), ("mid.attn_1.", "mid_block.attentions.0."), ] vae_conversion_map_attn = [ # (stable-diffusion, HF Diffusers) ("norm.", "group_norm."),...
null
151,844
import re import torch import logging protected = {re.escape(x[1]): x[0] for x in textenc_conversion_lst} textenc_pattern = re.compile("|".join(protected.keys())) code2idx = {"q": 0, "k": 1, "v": 2} def convert_text_enc_state_dict_v20(text_enc_dict, prefix=""): new_state_dict = {} capture_qkv_weight = {} c...
null
151,845
import re import torch import logging def convert_text_enc_state_dict(text_enc_dict): return text_enc_dict
null
151,846
import torch import math import struct import comfy.checkpoint_pickle import safetensors.torch import numpy as np from PIL import Image import logging def state_dict_key_replace(state_dict, keys_to_replace): for x in keys_to_replace: if x in state_dict: state_dict[keys_to_replace[x]] = state_di...
null
151,847
import torch import math import struct import comfy.checkpoint_pickle import safetensors.torch import numpy as np from PIL import Image import logging def resize_to_batch_size(tensor, batch_size): in_batch_size = tensor.shape[0] if in_batch_size == batch_size: return tensor if batch_size <= 1: ...
null
151,848
import torch import math import struct import comfy.checkpoint_pickle import safetensors.torch import numpy as np from PIL import Image import logging def convert_sd_to(state_dict, dtype): keys = list(state_dict.keys()) for k in keys: state_dict[k] = state_dict[k].to(dtype) return state_dict
null
151,849
import torch import math import struct import comfy.checkpoint_pickle import safetensors.torch import numpy as np from PIL import Image import logging def safetensors_header(safetensors_path, max_size=100*1024*1024): with open(safetensors_path, "rb") as f: header = f.read(8) length_of_header = stru...
null
151,850
import torch import math import struct import comfy.checkpoint_pickle import safetensors.torch import numpy as np from PIL import Image import logging def set_attr(obj, attr, value): attrs = attr.split(".") for name in attrs[:-1]: obj = getattr(obj, name) prev = getattr(obj, attrs[-1]) setattr(o...
null
151,851
import torch import math import struct import comfy.checkpoint_pickle import safetensors.torch import numpy as np from PIL import Image import logging def copy_to_param(obj, attr, value): # inplace update tensor instead of replacing it attrs = attr.split(".") for name in attrs[:-1]: obj = getattr(o...
null
151,852
import torch import math import struct import comfy.checkpoint_pickle import safetensors.torch import numpy as np from PIL import Image import logging def get_attr(obj, attr): attrs = attr.split(".") for name in attrs: obj = getattr(obj, name) return obj
null
151,853
import torch import math import struct import comfy.checkpoint_pickle import safetensors.torch import numpy as np from PIL import Image import logging def get_tiled_scale_steps(width, height, tile_x, tile_y, overlap): return math.ceil((height / (tile_y - overlap))) * math.ceil((width / (tile_x - overlap)))
null
151,854
import torch import math import struct import comfy.checkpoint_pickle import safetensors.torch import numpy as np from PIL import Image import logging def tiled_scale(samples, function, tile_x=64, tile_y=64, overlap = 8, upscale_amount = 4, out_channels = 3, output_device="cpu", pbar = None): output = torch.empty(...
null
151,855
import torch import math import struct import comfy.checkpoint_pickle import safetensors.torch import numpy as np from PIL import Image import logging PROGRESS_BAR_ENABLED = True def set_progress_bar_enabled(enabled): global PROGRESS_BAR_ENABLED PROGRESS_BAR_ENABLED = enabled
null
151,856
import torch import math import comfy.utils def lcm(a, b): #TODO: eventually replace by math.lcm (added in python3.9) return abs(a*b) // math.gcd(a, b)
null
151,857
import torch import logging from comfy.ldm.modules.diffusionmodules.openaimodel import UNetModel, Timestep from comfy.ldm.cascade.stage_c import StageC from comfy.ldm.cascade.stage_b import StageB from comfy.ldm.modules.encoders.noise_aug_modules import CLIPEmbeddingNoiseAugmentation from comfy.ldm.modules.diffusionmod...
null
151,858
import torch import logging from comfy.ldm.modules.diffusionmodules.openaimodel import UNetModel, Timestep from comfy.ldm.cascade.stage_c import StageC from comfy.ldm.cascade.stage_b import StageB from comfy.ldm.modules.encoders.noise_aug_modules import CLIPEmbeddingNoiseAugmentation from comfy.ldm.modules.diffusionmod...
null
151,859
import psutil import logging from enum import Enum from comfy.cli_args import args import comfy.utils import torch import sys directml_enabled = False try: if torch.backends.mps.is_available(): cpu_state = CPUState.MPS import torch.mps except: pass def is_intel_xpu(): global cpu_state gl...
null
151,860
import psutil import logging from enum import Enum from comfy.cli_args import args import comfy.utils import torch import sys import threading def module_size(module): module_mem = 0 sd = module.state_dict() for k in sd: t = sd[k] module_mem += t.nelement() * t.element_size() return mod...
null
151,861
import psutil import logging from enum import Enum from comfy.cli_args import args import comfy.utils import torch import sys current_loaded_models = [] import threading def cleanup_models(): to_delete = [] for i in range(len(current_loaded_models)): if sys.getrefcount(current_loaded_models[i].model) <...
null
151,862
import psutil import logging from enum import Enum from comfy.cli_args import args import comfy.utils import torch import sys if args.deterministic: logging.info("Using deterministic algorithms for pytorch") torch.use_deterministic_algorithms(True, warn_only=True) if args.directml is not None: import torch_...
null
151,863
import psutil import logging from enum import Enum from comfy.cli_args import args import comfy.utils import torch import sys class VRAMState(Enum): DISABLED = 0 #No vram present: no need to move models to vram NO_VRAM = 1 #Very low vram: enable all the options to save vram LOW_VRAM = 2 NORMAL_VR...
null
151,864
import psutil import logging from enum import Enum from comfy.cli_args import args import comfy.utils import torch import sys if args.deterministic: logging.info("Using deterministic algorithms for pytorch") torch.use_deterministic_algorithms(True, warn_only=True) if args.directml is not None: import torch_...
null
151,865
import psutil import logging from enum import Enum from comfy.cli_args import args import comfy.utils import torch import sys if args.deterministic: logging.info("Using deterministic algorithms for pytorch") torch.use_deterministic_algorithms(True, warn_only=True) if args.directml is not None: import torch_...
null
151,866
import psutil import logging from enum import Enum from comfy.cli_args import args import comfy.utils import torch import sys if args.deterministic: logging.info("Using deterministic algorithms for pytorch") torch.use_deterministic_algorithms(True, warn_only=True) if args.directml is not None: import torch_...
null
151,867
import psutil import logging from enum import Enum from comfy.cli_args import args import comfy.utils import torch import sys VAE_DTYPE = torch.float32 import threading def vae_dtype(): global VAE_DTYPE return VAE_DTYPE
null
151,868
import psutil import logging from enum import Enum from comfy.cli_args import args import comfy.utils import torch import sys import threading def get_autocast_device(dev): if hasattr(dev, 'type'): return dev.type return "cuda"
null
151,869
import psutil import logging from enum import Enum from comfy.cli_args import args import comfy.utils import torch import sys try: if torch.backends.mps.is_available(): cpu_state = CPUState.MPS import torch.mps except: pass def is_device_cpu(device): return is_device_type(device, 'cpu') impo...
null
151,870
import psutil import logging from enum import Enum from comfy.cli_args import args import comfy.utils import torch import sys try: if torch.backends.mps.is_available(): cpu_state = CPUState.MPS import torch.mps except: pass def is_intel_xpu(): global cpu_state global xpu_available if...
null
151,871
import psutil import logging from enum import Enum from comfy.cli_args import args import comfy.utils import torch import sys XFORMERS_ENABLED_VAE = True def xformers_enabled(): global directml_enabled global cpu_state if cpu_state != CPUState.GPU: return False if is_intel_xpu(): return ...
null
151,872
import psutil import logging from enum import Enum from comfy.cli_args import args import comfy.utils import torch import sys def is_nvidia(): global cpu_state if cpu_state == CPUState.GPU: if torch.version.cuda: return True return False ENABLE_PYTORCH_ATTENTION = False try: if is_nv...
null
151,873
import psutil import logging from enum import Enum from comfy.cli_args import args import comfy.utils import torch import sys def is_device_type(device, type): if hasattr(device, 'type'): if (device.type == type): return True return False import threading def is_device_cuda(device): ret...
null
151,874
import psutil import logging from enum import Enum from comfy.cli_args import args import comfy.utils import torch import sys import threading def resolve_lowvram_weight(weight, model, key): #TODO: remove return weight
null
151,875
import psutil import logging from enum import Enum from comfy.cli_args import args import comfy.utils import torch import sys import threading interrupt_processing_mutex = threading.RLock() interrupt_processing = False def processing_interrupted(): global interrupt_processing global interrupt_processing_mutex ...
null
151,876
import torch import comfy.model_management import comfy.samplers import comfy.conds import comfy.utils import math import numpy as np def get_models_from_cond(cond, model_type): models = [] for c in cond: if model_type in c: models += [c[model_type]] return models def cleanup_additional_...
null
151,877
from .utils import load_torch_file, transformers_convert, state_dict_prefix_replace import os import torch import json import logging import comfy.ops import comfy.model_patcher import comfy.model_management import comfy.utils import comfy.clip_model def clip_preprocess(image, size=224): mean = torch.tensor([ 0.48...
null
151,878
from .utils import load_torch_file, transformers_convert, state_dict_prefix_replace import os import torch import json import logging import comfy.ops import comfy.model_patcher import comfy.model_management import comfy.utils import comfy.clip_model def load_clipvision_from_sd(sd, prefix="", convert_keys=False): i...
null
151,896
import math from scipy import integrate import torch from torch import nn import torchsde from tqdm.auto import trange, tqdm from . import utils def default_noise_sampler(x): return lambda sigma, sigma_next: torch.randn_like(x) def sample_lcm(model, x, sigmas, extra_args=None, callback=None, disable=None, noise_sa...
null
151,897
import math from scipy import integrate import torch from torch import nn import torchsde from tqdm.auto import trange, tqdm from . import utils def to_d(x, sigma, denoised): """Converts a denoiser output to a Karras ODE derivative.""" return (x - denoised) / utils.append_dims(sigma, x.ndim) def sample_heunpp2...
null
151,909
import torch from torch import nn from .ldm.modules.attention import CrossAttention from inspect import isfunction import comfy.ops def uniq(arr): return{el: True for el in arr}.keys()
null
151,910
import torch from torch import nn from .ldm.modules.attention import CrossAttention from inspect import isfunction import comfy.ops def exists(val): def default(val, d): if exists(val): return val return d() if isfunction(d) else d
null
151,911
import torch from torch import nn from .ldm.modules.attention import CrossAttention from inspect import isfunction import comfy.ops class GatedSelfAttentionDense(nn.Module): def __init__(self, query_dim, context_dim, n_heads, d_head): super().__init__() # we need a linear projection since we need ca...
null
151,912
import torch import comfy.model_management def cast_bias_weight(s, input): bias = None non_blocking = comfy.model_management.device_supports_non_blocking(input.device) if s.bias is not None: bias = s.bias.to(device=input.device, dtype=input.dtype, non_blocking=non_blocking) if s.bias_functi...
null
151,913
import torch import torch.nn as nn import comfy.utils import comfy.ops def conv(n_in, n_out, **kwargs): class Block(nn.Module): def __init__(self, n_in, n_out): def forward(self, x): def Encoder(): return nn.Sequential( conv(3, 64), Block(64, 64), conv(64, 64, stride=2, bias=False), Block...
null
151,914
import torch import torch.nn as nn import comfy.utils import comfy.ops def conv(n_in, n_out, **kwargs): return comfy.ops.disable_weight_init.Conv2d(n_in, n_out, 3, padding=1, **kwargs) class Clamp(nn.Module): def forward(self, x): return torch.tanh(x / 3) * 3 class Block(nn.Module): def __init__(sel...
null
151,915
import torch from enum import Enum import logging from comfy import model_management from .ldm.models.autoencoder import AutoencoderKL, AutoencodingEngine from .ldm.cascade.stage_a import StageA from .ldm.cascade.stage_c_coder import StageC_coder import yaml import comfy.utils from . import clip_vision from . import gl...
null
151,916
import torch from enum import Enum import logging from comfy import model_management from .ldm.models.autoencoder import AutoencoderKL, AutoencodingEngine from .ldm.cascade.stage_a import StageA from .ldm.cascade.stage_c_coder import StageC_coder import yaml import comfy.utils from . import clip_vision from . import gl...
null
151,917
import torch from enum import Enum import logging from comfy import model_management from .ldm.models.autoencoder import AutoencoderKL, AutoencodingEngine from .ldm.cascade.stage_a import StageA from .ldm.cascade.stage_c_coder import StageC_coder import yaml import comfy.utils from . import clip_vision from . import gl...
null
151,918
import torch from enum import Enum import logging from comfy import model_management from .ldm.models.autoencoder import AutoencoderKL, AutoencodingEngine from .ldm.cascade.stage_a import StageA from .ldm.cascade.stage_c_coder import StageC_coder import yaml import comfy.utils from . import clip_vision from . import gl...
null
151,919
import torch from enum import Enum import logging from comfy import model_management from .ldm.models.autoencoder import AutoencoderKL, AutoencodingEngine from .ldm.cascade.stage_a import StageA from .ldm.cascade.stage_c_coder import StageC_coder import yaml import comfy.utils from . import clip_vision from . import gl...
null
151,920
import torch from enum import Enum import logging from comfy import model_management from .ldm.models.autoencoder import AutoencoderKL, AutoencodingEngine from .ldm.cascade.stage_a import StageA from .ldm.cascade.stage_c_coder import StageC_coder import yaml import comfy.utils from . import clip_vision from . import gl...
null
151,922
import os import math import torch import torch.nn as nn import numpy as np from einops import repeat, rearrange from comfy.ldm.util import instantiate_from_config def betas_for_alpha_bar(num_diffusion_timesteps, alpha_bar, max_beta=0.999): """ Create a beta schedule that discretizes the given alpha_t_bar funct...
null
151,923
import os import math import torch import torch.nn as nn import numpy as np from einops import repeat, rearrange from comfy.ldm.util import instantiate_from_config def make_ddim_timesteps(ddim_discr_method, num_ddim_timesteps, num_ddpm_timesteps, verbose=True): if ddim_discr_method == 'uniform': c = num_dd...
null
151,924
import os import math import torch import torch.nn as nn import numpy as np from einops import repeat, rearrange from comfy.ldm.util import instantiate_from_config def make_ddim_sampling_parameters(alphacums, ddim_timesteps, eta, verbose=True): # select alphas for computing the variance schedule alphas = alpha...
null
151,925
import os import math import torch import torch.nn as nn import numpy as np from einops import repeat, rearrange from comfy.ldm.util import instantiate_from_config def extract_into_tensor(a, t, x_shape): b, *_ = t.shape out = a.gather(-1, t) return out.reshape(b, *((1,) * (len(x_shape) - 1)))
null
151,926
import os import math import torch import torch.nn as nn import numpy as np from einops import repeat, rearrange from comfy.ldm.util import instantiate_from_config class CheckpointFunction(torch.autograd.Function): def forward(ctx, run_function, length, *args): ctx.run_function = run_function ctx.in...
Evaluate a function without caching intermediate activations, allowing for reduced memory at the expense of extra compute in the backward pass. :param func: the function to evaluate. :param inputs: the argument sequence to pass to `func`. :param params: a sequence of parameters `func` depends on but does not explicitly...
151,927
import os import math import torch import torch.nn as nn import numpy as np from einops import repeat, rearrange from comfy.ldm.util import instantiate_from_config The provided code snippet includes necessary dependencies for implementing the `timestep_embedding` function. Write a Python function `def timestep_embeddi...
Create sinusoidal timestep embeddings. :param timesteps: a 1-D Tensor of N indices, one per batch element. These may be fractional. :param dim: the dimension of the output. :param max_period: controls the minimum frequency of the embeddings. :return: an [N x dim] Tensor of positional embeddings.
151,928
import os import math import torch import torch.nn as nn import numpy as np from einops import repeat, rearrange from comfy.ldm.util import instantiate_from_config The provided code snippet includes necessary dependencies for implementing the `zero_module` function. Write a Python function `def zero_module(module)` to...
Zero out the parameters of a module and return it.
151,929
import os import math import torch import torch.nn as nn import numpy as np from einops import repeat, rearrange from comfy.ldm.util import instantiate_from_config The provided code snippet includes necessary dependencies for implementing the `scale_module` function. Write a Python function `def scale_module(module, s...
Scale the parameters of a module and return it.
151,930
import os import math import torch import torch.nn as nn import numpy as np from einops import repeat, rearrange from comfy.ldm.util import instantiate_from_config The provided code snippet includes necessary dependencies for implementing the `mean_flat` function. Write a Python function `def mean_flat(tensor)` to sol...
Take the mean over all non-batch dimensions.
151,931
import os import math import torch import torch.nn as nn import numpy as np from einops import repeat, rearrange from comfy.ldm.util import instantiate_from_config The provided code snippet includes necessary dependencies for implementing the `avg_pool_nd` function. Write a Python function `def avg_pool_nd(dims, *args...
Create a 1D, 2D, or 3D average pooling module.
151,932
import os import math import torch import torch.nn as nn import numpy as np from einops import repeat, rearrange from comfy.ldm.util import instantiate_from_config def noise_like(shape, device, repeat=False): repeat_noise = lambda: torch.randn((1, *shape[1:]), device=device).repeat(shape[0], *((1,) * (len(shape) -...
null
151,933
import math import torch import torch.nn as nn import numpy as np from einops import rearrange from typing import Optional, Any import logging from comfy import model_management import comfy.ops The provided code snippet includes necessary dependencies for implementing the `get_timestep_embedding` function. Write a Py...
This matches the implementation in Denoising Diffusion Probabilistic Models: From Fairseq. Build sinusoidal embeddings. This matches the implementation in tensor2tensor, but differs slightly from the description in Section 3.5 of "Attention Is All You Need".
151,934
import math import torch import torch.nn as nn import numpy as np from einops import rearrange from typing import Optional, Any import logging from comfy import model_management import comfy.ops def nonlinearity(x): # swish return x*torch.sigmoid(x)
null
151,935
import math import torch import torch.nn as nn import numpy as np from einops import rearrange from typing import Optional, Any import logging from comfy import model_management import comfy.ops ops = comfy.ops.disable_weight_init def Normalize(in_channels, num_groups=32): return ops.GroupNorm(num_groups=num_group...
null
151,936
import math import torch import torch.nn as nn import numpy as np from einops import rearrange from typing import Optional, Any import logging from comfy import model_management import comfy.ops def slice_attention(q, k, v): r1 = torch.zeros_like(k, device=q.device) scale = (int(q.shape[-1])**(-0.5)) mem_fr...
null
151,937
import math import torch import torch.nn as nn import numpy as np from einops import rearrange from typing import Optional, Any import logging from comfy import model_management import comfy.ops ops = comfy.ops.disable_weight_init def slice_attention(q, k, v): r1 = torch.zeros_like(k, device=q.device) scale = (...
null
151,938
import math import torch import torch.nn as nn import numpy as np from einops import rearrange from typing import Optional, Any import logging from comfy import model_management import comfy.ops if model_management.xformers_enabled_vae(): import xformers import xformers.ops def slice_attention(q, k, v): r1 ...
null
151,939
import math import torch import torch.nn as nn import numpy as np from einops import rearrange from typing import Optional, Any import logging from comfy import model_management import comfy.ops class AttnBlock(nn.Module): def __init__(self, in_channels): def forward(self, x): def make_attn(in_channels, attn...
null
151,940
from abc import abstractmethod import torch as th import torch.nn as nn import torch.nn.functional as F from einops import rearrange import logging from .util import ( checkpoint, avg_pool_nd, zero_module, timestep_embedding, AlphaBlender, ) from ..attention import SpatialTransformer, SpatialVideoTr...
null
151,941
from abc import abstractmethod import torch as th import torch.nn as nn import torch.nn.functional as F from einops import rearrange import logging from .util import ( checkpoint, avg_pool_nd, zero_module, timestep_embedding, AlphaBlender, ) from ..attention import SpatialTransformer, SpatialVideoTr...
null
151,942
import math import torch import torch.nn.functional as F from torch import nn, einsum from einops import rearrange, repeat from typing import Optional, Any import logging from .diffusionmodules.util import checkpoint, AlphaBlender, timestep_embedding from .sub_quadratic_attention import efficient_dot_product_attention ...
null
151,943
import math import torch import torch.nn.functional as F from torch import nn, einsum from einops import rearrange, repeat from typing import Optional, Any import logging from .diffusionmodules.util import checkpoint, AlphaBlender, timestep_embedding from .sub_quadratic_attention import efficient_dot_product_attention ...
null
151,944
import math import torch import torch.nn.functional as F from torch import nn, einsum from einops import rearrange, repeat from typing import Optional, Any import logging from .diffusionmodules.util import checkpoint, AlphaBlender, timestep_embedding from .sub_quadratic_attention import efficient_dot_product_attention ...
null
151,945
import math import torch import torch.nn.functional as F from torch import nn, einsum from einops import rearrange, repeat from typing import Optional, Any import logging from .diffusionmodules.util import checkpoint, AlphaBlender, timestep_embedding from .sub_quadratic_attention import efficient_dot_product_attention ...
null
151,946
import math import torch import torch.nn.functional as F from torch import nn, einsum from einops import rearrange, repeat from typing import Optional, Any import logging from .diffusionmodules.util import checkpoint, AlphaBlender, timestep_embedding from .sub_quadratic_attention import efficient_dot_product_attention ...
null
151,947
import math import torch import torch.nn.functional as F from torch import nn, einsum from einops import rearrange, repeat from typing import Optional, Any import logging from .diffusionmodules.util import checkpoint, AlphaBlender, timestep_embedding from .sub_quadratic_attention import efficient_dot_product_attention ...
null
151,948
import math import torch import torch.nn.functional as F from torch import nn, einsum from einops import rearrange, repeat from typing import Optional, Any import logging from .diffusionmodules.util import checkpoint, AlphaBlender, timestep_embedding from .sub_quadratic_attention import efficient_dot_product_attention ...
null
151,949
import functools from typing import Callable, Iterable, Union import torch from einops import rearrange, repeat import comfy.ops from .diffusionmodules.model import ( AttnBlock, Decoder, ResnetBlock, ) from .diffusionmodules.openaimodel import ResBlock, timestep_embedding from .attention import BasicTransfo...
null
151,958
import torch import torch.nn as nn from comfy.ldm.modules.attention import optimized_attention def LayerNorm2d_op(operations): class LayerNorm2d(operations.LayerNorm): def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) def forward(self, x): return super(...
null
151,959
import torch import math import os import logging import comfy.utils import comfy.model_management import comfy.model_detection import comfy.model_patcher import comfy.ops import comfy.cldm.cldm import comfy.t2i_adapter.adapter import comfy.ldm.cascade.controlnet def broadcast_image_to(tensor, target_batch_size, batch...
null
151,960
import torch import math import os import logging import comfy.utils import comfy.model_management import comfy.model_detection import comfy.model_patcher import comfy.ops import comfy.cldm.cldm import comfy.t2i_adapter.adapter import comfy.ldm.cascade.controlnet class ControlNet(ControlBase): def __init__(self, co...
null
151,961
import os import comfy.sd def first_file(path, filenames): for f in filenames: p = os.path.join(path, f) if os.path.exists(p): return p return None def load_diffusers(model_path, output_vae=True, output_clip=True, embedding_directory=None): diffusion_model_names = ["diffusion_py...
null
151,962
import os from transformers import CLIPTokenizer import comfy.ops import torch import traceback import zipfile from . import model_management import comfy.clip_model import json import logging def gen_empty_tokens(special_tokens, length): start_token = special_tokens.get("start", None) end_token = special_toke...
null
151,963
import os from transformers import CLIPTokenizer import comfy.ops import torch import traceback import zipfile from . import model_management import comfy.clip_model import json import logging def parse_parentheses(string): result = [] current_item = "" nesting_level = 0 for char in string: if c...
null
151,964
import os from transformers import CLIPTokenizer import comfy.ops import torch import traceback import zipfile from . import model_management import comfy.clip_model import json import logging def escape_important(text): text = text.replace("\\)", "\0\1") text = text.replace("\\(", "\0\2") return text
null