id
int64
0
190k
prompt
stringlengths
21
13.4M
docstring
stringlengths
1
12k
2,660
import cv2 import numpy as np import torch from torch.nn import functional as F The provided code snippet includes necessary dependencies for implementing the `usm_sharp` function. Write a Python function `def usm_sharp(img, weight=0.5, radius=50, threshold=10)` to solve the following problem: USM sharpening. Input im...
USM sharpening. Input image: I; Blurry image: B. 1. sharp = I + weight * (I - B) 2. Mask = 1 if abs(I - B) > threshold, else: 0 3. Blur mask: 4. Out = Mask * sharp + (1 - Mask) * I Args: img (Numpy array): Input image, HWC, BGR; float32, [0, 1]. weight (float): Sharp weight. Default: 1. radius (float): Kernel size of G...
2,661
import itertools import numpy as np import torch import torch.nn as nn from torch.nn import functional as F The provided code snippet includes necessary dependencies for implementing the `diff_round` function. Write a Python function `def diff_round(x)` to solve the following problem: Differentiable rounding function ...
Differentiable rounding function
2,662
import itertools import numpy as np import torch import torch.nn as nn from torch.nn import functional as F The provided code snippet includes necessary dependencies for implementing the `quality_to_factor` function. Write a Python function `def quality_to_factor(quality)` to solve the following problem: Calculate fac...
Calculate factor corresponding to quality Args: quality(float): Quality for jpeg compression. Returns: float: Compression factor.
2,663
import numpy as np import torch def _convert_input_type_range(img): """Convert the type and range of the input image. It converts the input image to np.float32 type and range of [0, 1]. It is mainly used for pre-processing the input image in colorspace conversion functions such as rgb2ycbcr and ycbcr2rg...
Convert a RGB image to YCbCr image. This function produces the same results as Matlab's `rgb2ycbcr` function. It implements the ITU-R BT.601 conversion for standard-definition television. See more details in https://en.wikipedia.org/wiki/YCbCr#ITU-R_BT.601_conversion. It differs from a similar function in cv2.cvtColor:...
2,664
import numpy as np import torch def _convert_input_type_range(img): """Convert the type and range of the input image. It converts the input image to np.float32 type and range of [0, 1]. It is mainly used for pre-processing the input image in colorspace conversion functions such as rgb2ycbcr and ycbcr2rg...
Convert a BGR image to YCbCr image. The bgr version of rgb2ycbcr. It implements the ITU-R BT.601 conversion for standard-definition television. See more details in https://en.wikipedia.org/wiki/YCbCr#ITU-R_BT.601_conversion. It differs from a similar function in cv2.cvtColor: `BGR <-> YCrCb`. In OpenCV, it implements a...
2,665
import numpy as np import torch def _convert_input_type_range(img): """Convert the type and range of the input image. It converts the input image to np.float32 type and range of [0, 1]. It is mainly used for pre-processing the input image in colorspace conversion functions such as rgb2ycbcr and ycbcr2rg...
Convert a YCbCr image to RGB image. This function produces the same results as Matlab's ycbcr2rgb function. It implements the ITU-R BT.601 conversion for standard-definition television. See more details in https://en.wikipedia.org/wiki/YCbCr#ITU-R_BT.601_conversion. It differs from a similar function in cv2.cvtColor: `...
2,666
import numpy as np import torch def _convert_input_type_range(img): """Convert the type and range of the input image. It converts the input image to np.float32 type and range of [0, 1]. It is mainly used for pre-processing the input image in colorspace conversion functions such as rgb2ycbcr and ycbcr2rg...
Convert a YCbCr image to BGR image. The bgr version of ycbcr2rgb. It implements the ITU-R BT.601 conversion for standard-definition television. See more details in https://en.wikipedia.org/wiki/YCbCr#ITU-R_BT.601_conversion. It differs from a similar function in cv2.cvtColor: `YCrCb <-> BGR`. In OpenCV, it implements a...
2,667
import datetime import logging import time from .dist_util import get_dist_info, master_only def init_tb_logger(log_dir): from torch.utils.tensorboard import SummaryWriter tb_logger = SummaryWriter(log_dir=log_dir) return tb_logger
null
2,668
import datetime import logging import time from .dist_util import get_dist_info, master_only def get_root_logger(logger_name='basicsr', log_level=logging.INFO, log_file=None): """Get the root logger. The logger will be initialized if it has not been initialized. By default a StreamHandler will be added. If ...
We now only use wandb to sync tensorboard log.
2,669
import datetime import logging import time from .dist_util import get_dist_info, master_only The provided code snippet includes necessary dependencies for implementing the `get_env_info` function. Write a Python function `def get_env_info()` to solve the following problem: Get environment information. Currently, only ...
Get environment information. Currently, only log the software version.
2,670
import re The provided code snippet includes necessary dependencies for implementing the `read_data_from_tensorboard` function. Write a Python function `def read_data_from_tensorboard(log_path, tag)` to solve the following problem: Get raw data (steps and values) from tensorboard events. Args: log_path (str): Path to ...
Get raw data (steps and values) from tensorboard events. Args: log_path (str): Path to the tensorboard log. tag (str): tag to be read.
2,671
import re The provided code snippet includes necessary dependencies for implementing the `read_data_from_txt_2v` function. Write a Python function `def read_data_from_txt_2v(path, pattern, step_one=False)` to solve the following problem: Read data from txt with 2 returned values (usually [step, value]). Args: path (st...
Read data from txt with 2 returned values (usually [step, value]). Args: path (str): path to the txt file. pattern (str): re (regular expression) pattern. step_one (bool): add 1 to steps. Default: False.
2,672
import re The provided code snippet includes necessary dependencies for implementing the `read_data_from_txt_1v` function. Write a Python function `def read_data_from_txt_1v(path, pattern)` to solve the following problem: Read data from txt with 1 returned values. Args: path (str): path to the txt file. pattern (str):...
Read data from txt with 1 returned values. Args: path (str): path to the txt file. pattern (str): re (regular expression) pattern.
2,673
import re The provided code snippet includes necessary dependencies for implementing the `smooth_data` function. Write a Python function `def smooth_data(values, smooth_weight)` to solve the following problem: Smooth data using 1st-order IIR low-pass filter (what tensorflow does). Reference: https://github.com/tensorf...
Smooth data using 1st-order IIR low-pass filter (what tensorflow does). Reference: https://github.com/tensorflow/tensorboard/blob/f801ebf1f9fbfe2baee1ddd65714d0bccc640fb1/tensorboard/plugins/scalar/vz_line_chart/vz-line-chart.ts#L704 # noqa: E501 Args: values (list): A list of values to be smoothed. smooth_weight (floa...
2,674
import os import PIL import numpy as np import copy import torch from omegaconf import OmegaConf from PIL import Image from tqdm import trange from itertools import islice from einops import rearrange, repeat from torch import autocast from pytorch_lightning import seed_everything import torch.nn.functional as F from l...
null
2,675
import os import PIL import numpy as np import copy import torch from omegaconf import OmegaConf from PIL import Image from tqdm import trange from itertools import islice from einops import rearrange, repeat from torch import autocast from pytorch_lightning import seed_everything import torch.nn.functional as F from l...
null
2,676
import os import PIL import numpy as np import copy import torch from omegaconf import OmegaConf from PIL import Image from tqdm import trange from itertools import islice from einops import rearrange, repeat from torch import autocast from pytorch_lightning import seed_everything import torch.nn.functional as F from l...
null
2,677
import os import PIL import numpy as np import copy import torch from omegaconf import OmegaConf from PIL import Image from tqdm import trange from itertools import islice from einops import rearrange, repeat from torch import autocast from pytorch_lightning import seed_everything import torch.nn.functional as F from l...
null
2,678
import os import PIL import numpy as np import copy import torch from omegaconf import OmegaConf from PIL import Image from tqdm import trange from itertools import islice from einops import rearrange, repeat from torch import autocast from pytorch_lightning import seed_everything import torch.nn.functional as F from l...
null
2,679
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
Adaptive instance normalization. Adjust the reference features to have the similar color and illuminations as those in the degradate features. Args: content_feat (Tensor): The reference feature. style_feat (Tensor): The degradate features.
2,680
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
Create a list of timesteps to use from an original diffusion process, given the number of timesteps we want to take from equally-sized portions of the original process. For example, if there's 300 timesteps and the section counts are [10,15,20] then the first 100 timesteps are strided to be 10 timesteps, the second 100...
2,681
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,682
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,683
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,684
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
Adaptive instance normalization. Adjust the reference features to have the similar color and illuminations as those in the degradate features. Args: content_feat (Tensor): The reference feature. style_feat (Tensor): The degradate features.
2,685
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
Create a list of timesteps to use from an original diffusion process, given the number of timesteps we want to take from equally-sized portions of the original process. For example, if there's 300 timesteps and the section counts are [10,15,20] then the first 100 timesteps are strided to be 10 timesteps, the second 100...
2,686
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,687
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,688
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,689
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
Adaptive instance normalization. Adjust the reference features to have the similar color and illuminations as those in the degradate features. Args: content_feat (Tensor): The reference feature. style_feat (Tensor): The degradate features.
2,690
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
Create a list of timesteps to use from an original diffusion process, given the number of timesteps we want to take from equally-sized portions of the original process. For example, if there's 300 timesteps and the section counts are [10,15,20] then the first 100 timesteps are strided to be 10 timesteps, the second 100...
2,691
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,692
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,693
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,694
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,695
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
Adaptive instance normalization. Adjust the reference features to have the similar color and illuminations as those in the degradate features. Args: content_feat (Tensor): The reference feature. style_feat (Tensor): The degradate features.
2,696
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
Create a list of timesteps to use from an original diffusion process, given the number of timesteps we want to take from equally-sized portions of the original process. For example, if there's 300 timesteps and the section counts are [10,15,20] then the first 100 timesteps are strided to be 10 timesteps, the second 100...
2,697
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,698
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,699
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,700
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
Create a list of timesteps to use from an original diffusion process, given the number of timesteps we want to take from equally-sized portions of the original process. For example, if there's 300 timesteps and the section counts are [10,15,20] then the first 100 timesteps are strided to be 10 timesteps, the second 100...
2,701
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,702
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,703
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,704
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,705
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
Adaptive instance normalization. Adjust the reference features to have the similar color and illuminations as those in the degradate features. Args: content_feat (Tensor): The reference feature. style_feat (Tensor): The degradate features.
2,706
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
Create a list of timesteps to use from an original diffusion process, given the number of timesteps we want to take from equally-sized portions of the original process. For example, if there's 300 timesteps and the section counts are [10,15,20] then the first 100 timesteps are strided to be 10 timesteps, the second 100...
2,707
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,708
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,709
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,710
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
Create a list of timesteps to use from an original diffusion process, given the number of timesteps we want to take from equally-sized portions of the original process. For example, if there's 300 timesteps and the section counts are [10,15,20] then the first 100 timesteps are strided to be 10 timesteps, the second 100...
2,711
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,712
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,713
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,714
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
Create a list of timesteps to use from an original diffusion process, given the number of timesteps we want to take from equally-sized portions of the original process. For example, if there's 300 timesteps and the section counts are [10,15,20] then the first 100 timesteps are strided to be 10 timesteps, the second 100...
2,715
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,716
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,717
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,718
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
Adaptive instance normalization. Adjust the reference features to have the similar color and illuminations as those in the degradate features. Args: content_feat (Tensor): The reference feature. style_feat (Tensor): The degradate features.
2,719
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
Create a list of timesteps to use from an original diffusion process, given the number of timesteps we want to take from equally-sized portions of the original process. For example, if there's 300 timesteps and the section counts are [10,15,20] then the first 100 timesteps are strided to be 10 timesteps, the second 100...
2,720
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,721
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,722
import argparse, os, sys, glob import PIL import torch import numpy as np import torchvision from omegaconf import OmegaConf from PIL import Image from tqdm import tqdm, trange from itertools import islice from einops import rearrange, repeat from torchvision.utils import make_grid from torch import autocast from conte...
null
2,723
import torch from PIL import Image from torch import Tensor from torch.nn import functional as F from torchvision.transforms import ToTensor, ToPILImage def adaptive_instance_normalization(content_feat:Tensor, style_feat:Tensor): """Adaptive instance normalization. Adjust the reference features to have the simi...
null
2,724
import torch from PIL import Image from torch import Tensor from torch.nn import functional as F from torchvision.transforms import ToTensor, ToPILImage def wavelet_reconstruction(content_feat:Tensor, style_feat:Tensor): def wavelet_color_fix(target: Image, source: Image): # Convert images to tensors to_tensor...
null
2,725
import sys import cv2 import math import torch import random import numpy as np from scipy import fft from pathlib import Path from einops import rearrange from skimage import img_as_ubyte, img_as_float32 def calculate_psnr(im1, im2, border=0, ycbcr=False): def rgb2ycbcrTorch(im, only_y=True): def batch_PSNR(img, imcl...
null
2,726
import sys import cv2 import math import torch import random import numpy as np from scipy import fft from pathlib import Path from einops import rearrange from skimage import img_as_ubyte, img_as_float32 def calculate_ssim(im1, im2, border=0, ycbcr=False): ''' SSIM the same outputs as MATLAB's im1, im2: h ...
null
2,727
import sys import cv2 import math import torch import random import numpy as np from scipy import fft from pathlib import Path from einops import rearrange from skimage import img_as_ubyte, img_as_float32 The provided code snippet includes necessary dependencies for implementing the `normalize_np` function. Write a Py...
Input: im: h x w x c, numpy array Normalize: (im - mean) / std Reverse: im * std + mean
2,728
import sys import cv2 import math import torch import random import numpy as np from scipy import fft from pathlib import Path from einops import rearrange from skimage import img_as_ubyte, img_as_float32 The provided code snippet includes necessary dependencies for implementing the `normalize_th` function. Write a Py...
Input: im: b x c x h x w, torch tensor Normalize: (im - mean) / std Reverse: im * std + mean
2,729
import sys import cv2 import math import torch import random import numpy as np from scipy import fft from pathlib import Path from einops import rearrange from skimage import img_as_ubyte, img_as_float32 The provided code snippet includes necessary dependencies for implementing the `tensor2img` function. Write a Pyth...
Convert torch Tensors into image numpy arrays. After clamping to [min, max], values will be normalized to [0, 1]. Args: tensor (Tensor or list[Tensor]): Accept shapes: 1) 4D mini-batch Tensor of shape (B x 3/1 x H x W); 2) 3D Tensor of shape (3/1 x H x W); 3) 2D Tensor of shape (H x W). Tensor channel should be in RGB ...
2,730
import sys import cv2 import math import torch import random import numpy as np from scipy import fft from pathlib import Path from einops import rearrange from skimage import img_as_ubyte, img_as_float32 The provided code snippet includes necessary dependencies for implementing the `img2tensor` function. Write a Pyth...
Convert image numpy arrays into torch tensor. Args: imgs (Array or list[array]): Accept shapes: 3) list of numpy arrays 1) 3D numpy array of shape (H x W x 3/1); 2) 2D Tensor of shape (H x W). Tensor channel should be in RGB order. Returns: (array or list): 4D ndarray of shape (1 x C x H x W)
2,731
import sys import cv2 import math import torch import random import numpy as np from scipy import fft from pathlib import Path from einops import rearrange from skimage import img_as_ubyte, img_as_float32 def bgr2rgb(im): return cv2.cvtColor(im, cv2.COLOR_BGR2RGB) The provided code snippet includes necessary dependenc...
Read image. chn: 'rgb', 'bgr' or 'gray' out: im: h x w x c, numpy tensor
2,732
import sys import cv2 import math import torch import random import numpy as np from scipy import fft from pathlib import Path from einops import rearrange from skimage import img_as_ubyte, img_as_float32 def rgb2bgr(im): return cv2.cvtColor(im, cv2.COLOR_RGB2BGR) The provided code snippet includes necessary dependenc...
Save image. Input: im: h x w x c, numpy tensor path: the saving path chn: the channel order of the im,
2,733
import sys import cv2 import math import torch import random import numpy as np from scipy import fft from pathlib import Path from einops import rearrange from skimage import img_as_ubyte, img_as_float32 def bgr2rgb(im): return cv2.cvtColor(im, cv2.COLOR_BGR2RGB) def rgb2bgr(im): return cv2.cvtColor(im, cv2.COLOR_RGB2...
Input: im: h x w x 3 array qf: compress factor, (0, 100] chn_in: 'rgb' or 'bgr' Return: Compressed Image with channel order: chn_in
2,734
import sys import cv2 import math import torch import random import numpy as np from scipy import fft from pathlib import Path from einops import rearrange from skimage import img_as_ubyte, img_as_float32 The provided code snippet includes necessary dependencies for implementing the `data_aug_np` function. Write a Pyt...
Performs data augmentation of the input image Input: image: a cv2 (OpenCV) image mode: int. Choice of transformation to apply to the image 0 - no transformation 1 - flip up and down 2 - rotate counterwise 90 degree 3 - rotate 90 degree and flip up and down 4 - rotate 180 degree 5 - rotate 180 degree and flip 6 - rotate...
2,735
import sys import cv2 import math import torch import random import numpy as np from scipy import fft from pathlib import Path from einops import rearrange from skimage import img_as_ubyte, img_as_float32 The provided code snippet includes necessary dependencies for implementing the `inverse_data_aug_np` function. Wri...
Performs inverse data augmentation of the input image
2,736
import sys import cv2 import math import torch import random import numpy as np from scipy import fft from pathlib import Path from einops import rearrange from skimage import img_as_ubyte, img_as_float32 def imshow(x, title=None, cbar=False): import matplotlib.pyplot as plt plt.imshow(np.squeeze(x), interpola...
null
2,737
import sys import cv2 import math import torch import random import numpy as np from scipy import fft from pathlib import Path from einops import rearrange from skimage import img_as_ubyte, img_as_float32 The provided code snippet includes necessary dependencies for implementing the `imgrad` function. Write a Python f...
Calculate image gradient. Input: im: h x w x c numpy array
2,738
import sys import cv2 import math import torch import random import numpy as np from scipy import fft from pathlib import Path from einops import rearrange from skimage import img_as_ubyte, img_as_float32 def convfft(im, weight): ''' Convolution with FFT Input: im: h1 x w1 x c numpy array we...
Calculate image gradient. Input: im: h x w x c numpy array
2,739
import sys import cv2 import math import torch import random import numpy as np from scipy import fft from pathlib import Path from einops import rearrange from skimage import img_as_ubyte, img_as_float32 The provided code snippet includes necessary dependencies for implementing the `random_crop` function. Write a Pyt...
Randomly crop a patch from the give image.
2,740
import torch from torch import nn import torch.nn.functional as F from einops import repeat from taming.modules.discriminator.model import NLayerDiscriminator, weights_init from taming.modules.losses.lpips import LPIPS from taming.modules.losses.vqperceptual import hinge_d_loss, vanilla_d_loss def hinge_d_loss_with_ex...
null
2,741
import torch from torch import nn import torch.nn.functional as F from einops import repeat from taming.modules.discriminator.model import NLayerDiscriminator, weights_init from taming.modules.losses.lpips import LPIPS from taming.modules.losses.vqperceptual import hinge_d_loss, vanilla_d_loss def adopt_weight(weight,...
null
2,742
import torch from torch import nn import torch.nn.functional as F from einops import repeat from taming.modules.discriminator.model import NLayerDiscriminator, weights_init from taming.modules.losses.lpips import LPIPS from taming.modules.losses.vqperceptual import hinge_d_loss, vanilla_d_loss def measure_perplexity(p...
null
2,743
import torch from torch import nn import torch.nn.functional as F from einops import repeat from taming.modules.discriminator.model import NLayerDiscriminator, weights_init from taming.modules.losses.lpips import LPIPS from taming.modules.losses.vqperceptual import hinge_d_loss, vanilla_d_loss def l1(x, y): return...
null
2,744
import torch from torch import nn import torch.nn.functional as F from einops import repeat from taming.modules.discriminator.model import NLayerDiscriminator, weights_init from taming.modules.losses.lpips import LPIPS from taming.modules.losses.vqperceptual import hinge_d_loss, vanilla_d_loss def l2(x, y): return...
null
2,745
import math import torch import torch.nn as nn import torch.nn.functional as F import torch.utils.checkpoint as checkpoint from timm.models.layers import DropPath, to_2tuple, trunc_normal_ The provided code snippet includes necessary dependencies for implementing the `window_partition` function. Write a Python functio...
Args: x: (B, H, W, C) window_size (int): window size Returns: windows: (num_windows*B, window_size, window_size, C)
2,746
import math import torch import torch.nn as nn import torch.nn.functional as F import torch.utils.checkpoint as checkpoint from timm.models.layers import DropPath, to_2tuple, trunc_normal_ The provided code snippet includes necessary dependencies for implementing the `window_reverse` function. Write a Python function ...
Args: windows: (num_windows*B, window_size, window_size, C) window_size (int): Window size H (int): Height of image W (int): Width of image Returns: x: (B, H, W, C)
2,747
import re import torch import torch.nn as nn import torch.nn.functional as F import torch.nn.utils.spectral_norm as spectral_norm from ldm.modules.diffusionmodules.util import normalization def get_nonspade_norm_layer(opt, norm_type='instance'): # helper function to get # output channels of the previous layer ...
null
2,748
import os import math import torch import torch.nn as nn import numpy as np from einops import repeat from ldm.util import instantiate_from_config def make_beta_schedule(schedule, n_timestep, linear_start=1e-4, linear_end=2e-2, cosine_s=8e-3): if schedule == "linear": betas = ( torch.linspa...
null
2,749
import os import math import torch import torch.nn as nn import numpy as np from einops import repeat from ldm.util import instantiate_from_config def make_ddim_timesteps(ddim_discr_method, num_ddim_timesteps, num_ddpm_timesteps, verbose=True): if ddim_discr_method == 'uniform': c = num_ddpm_timesteps // n...
null
2,750
import os import math import torch import torch.nn as nn import numpy as np from einops import repeat from ldm.util import instantiate_from_config def make_ddim_sampling_parameters(alphacums, ddim_timesteps, eta, verbose=True): # select alphas for computing the variance schedule alphas = alphacums[ddim_timeste...
null
2,751
import os import math import torch import torch.nn as nn import numpy as np from einops import repeat from ldm.util import instantiate_from_config The provided code snippet includes necessary dependencies for implementing the `betas_for_alpha_bar` function. Write a Python function `def betas_for_alpha_bar(num_diffusio...
Create a beta schedule that discretizes the given alpha_t_bar function, which defines the cumulative product of (1-beta) over time from t = [0,1]. :param num_diffusion_timesteps: the number of betas to produce. :param alpha_bar: a lambda that takes an argument t from 0 to 1 and produces the cumulative product of (1-bet...
2,752
import os import math import torch import torch.nn as nn import numpy as np from einops import repeat from ldm.util import instantiate_from_config def extract_into_tensor(a, t, x_shape): b, *_ = t.shape out = a.gather(-1, t) return out.reshape(b, *((1,) * (len(x_shape) - 1)))
null
2,753
import os import math import torch import torch.nn as nn import numpy as np from einops import repeat from ldm.util import instantiate_from_config class CheckpointFunction(torch.autograd.Function): def forward(ctx, run_function, length, *args): ctx.run_function = run_function ctx.input_tensors = lis...
Evaluate a function without caching intermediate activations, allowing for reduced memory at the expense of extra compute in the backward pass. :param func: the function to evaluate. :param inputs: the argument sequence to pass to `func`. :param params: a sequence of parameters `func` depends on but does not explicitly...
2,754
import os import math import torch import torch.nn as nn import numpy as np from einops import repeat from ldm.util import instantiate_from_config The provided code snippet includes necessary dependencies for implementing the `timestep_embedding` function. Write a Python function `def timestep_embedding(timesteps, dim...
Create sinusoidal timestep embeddings. :param timesteps: a 1-D Tensor of N indices, one per batch element. These may be fractional. :param dim: the dimension of the output. :param max_period: controls the minimum frequency of the embeddings. :return: an [N x dim] Tensor of positional embeddings.
2,755
import os import math import torch import torch.nn as nn import numpy as np from einops import repeat from ldm.util import instantiate_from_config The provided code snippet includes necessary dependencies for implementing the `zero_module` function. Write a Python function `def zero_module(module)` to solve the follow...
Zero out the parameters of a module and return it.
2,756
import os import math import torch import torch.nn as nn import numpy as np from einops import repeat from ldm.util import instantiate_from_config The provided code snippet includes necessary dependencies for implementing the `scale_module` function. Write a Python function `def scale_module(module, scale)` to solve t...
Scale the parameters of a module and return it.
2,757
import os import math import torch import torch.nn as nn import numpy as np from einops import repeat from ldm.util import instantiate_from_config The provided code snippet includes necessary dependencies for implementing the `mean_flat` function. Write a Python function `def mean_flat(tensor)` to solve the following ...
Take the mean over all non-batch dimensions.
2,758
import os import math import torch import torch.nn as nn import numpy as np from einops import repeat from ldm.util import instantiate_from_config class GroupNorm32(nn.GroupNorm): def forward(self, x): return super().forward(x.float()).type(x.dtype) The provided code snippet includes necessary dependencies...
Make a standard normalization layer. :param channels: number of input channels. :return: an nn.Module for normalization.
2,759
import os import math import torch import torch.nn as nn import numpy as np from einops import repeat from ldm.util import instantiate_from_config The provided code snippet includes necessary dependencies for implementing the `conv_nd` function. Write a Python function `def conv_nd(dims, *args, **kwargs)` to solve the...
Create a 1D, 2D, or 3D convolution module.