| import os |
| import torch |
| import PIL.Image |
| import numpy as np |
| from torch import nn |
| import torch.distributed as dist |
| import timm.models as timm_hub |
|
|
| """Modified from https://github.com/CompVis/taming-transformers.git""" |
|
|
| import hashlib |
| import requests |
| from tqdm import tqdm |
| try: |
| import piq |
| except: |
| pass |
|
|
| _CONTEXT_PARALLEL_GROUP = None |
| _CONTEXT_PARALLEL_SIZE = None |
|
|
|
|
| def is_dist_avail_and_initialized(): |
| if not dist.is_available(): |
| return False |
| if not dist.is_initialized(): |
| return False |
| return True |
|
|
|
|
| def get_world_size(): |
| if not is_dist_avail_and_initialized(): |
| return 1 |
| return dist.get_world_size() |
|
|
|
|
| def get_rank(): |
| if not is_dist_avail_and_initialized(): |
| return 0 |
| return dist.get_rank() |
|
|
|
|
| def is_main_process(): |
| return get_rank() == 0 |
|
|
|
|
| def is_context_parallel_initialized(): |
| if _CONTEXT_PARALLEL_GROUP is None: |
| return False |
| else: |
| return True |
|
|
|
|
| def set_context_parallel_group(size, group): |
| global _CONTEXT_PARALLEL_GROUP |
| global _CONTEXT_PARALLEL_SIZE |
| _CONTEXT_PARALLEL_GROUP = group |
| _CONTEXT_PARALLEL_SIZE = size |
|
|
|
|
| def initialize_context_parallel(context_parallel_size): |
| global _CONTEXT_PARALLEL_GROUP |
| global _CONTEXT_PARALLEL_SIZE |
|
|
| assert _CONTEXT_PARALLEL_GROUP is None, "context parallel group is already initialized" |
| _CONTEXT_PARALLEL_SIZE = context_parallel_size |
|
|
| rank = torch.distributed.get_rank() |
| world_size = torch.distributed.get_world_size() |
|
|
| for i in range(0, world_size, context_parallel_size): |
| ranks = range(i, i + context_parallel_size) |
| group = torch.distributed.new_group(ranks) |
| if rank in ranks: |
| _CONTEXT_PARALLEL_GROUP = group |
| break |
|
|
|
|
| def get_context_parallel_group(): |
| assert _CONTEXT_PARALLEL_GROUP is not None, "context parallel group is not initialized" |
|
|
| return _CONTEXT_PARALLEL_GROUP |
|
|
|
|
| def get_context_parallel_world_size(): |
| assert _CONTEXT_PARALLEL_SIZE is not None, "context parallel size is not initialized" |
|
|
| return _CONTEXT_PARALLEL_SIZE |
|
|
|
|
| def get_context_parallel_rank(): |
| assert _CONTEXT_PARALLEL_SIZE is not None, "context parallel size is not initialized" |
|
|
| rank = get_rank() |
| cp_rank = rank % _CONTEXT_PARALLEL_SIZE |
| return cp_rank |
|
|
|
|
| def get_context_parallel_group_rank(): |
| assert _CONTEXT_PARALLEL_SIZE is not None, "context parallel size is not initialized" |
|
|
| rank = get_rank() |
| cp_group_rank = rank // _CONTEXT_PARALLEL_SIZE |
|
|
| return cp_group_rank |
|
|
|
|
| def download_cached_file(url, check_hash=True, progress=False): |
| """ |
| Download a file from a URL and cache it locally. If the file already exists, it is not downloaded again. |
| If distributed, only the main process downloads the file, and the other processes wait for the file to be downloaded. |
| """ |
|
|
| def get_cached_file_path(): |
| |
| parts = torch.hub.urlparse(url) |
| filename = os.path.basename(parts.path) |
| cached_file = os.path.join(timm_hub.get_cache_dir(), filename) |
|
|
| return cached_file |
|
|
| if is_main_process(): |
| timm_hub.download_cached_file(url, check_hash, progress) |
|
|
| if is_dist_avail_and_initialized(): |
| dist.barrier() |
|
|
| return get_cached_file_path() |
|
|
|
|
| def convert_weights_to_fp16(model: nn.Module): |
| """Convert applicable model parameters to fp16""" |
|
|
| def _convert_weights_to_fp16(l): |
| if isinstance(l, (nn.Conv1d, nn.Conv2d, nn.Conv3d, nn.Linear)): |
| l.weight.data = l.weight.data.to(torch.float16) |
| if l.bias is not None: |
| l.bias.data = l.bias.data.to(torch.float16) |
|
|
| model.apply(_convert_weights_to_fp16) |
|
|
|
|
| def convert_weights_to_bf16(model: nn.Module): |
| """Convert applicable model parameters to fp16""" |
|
|
| def _convert_weights_to_bf16(l): |
| if isinstance(l, (nn.Conv1d, nn.Conv2d, nn.Conv3d, nn.Linear)): |
| l.weight.data = l.weight.data.to(torch.bfloat16) |
| if l.bias is not None: |
| l.bias.data = l.bias.data.to(torch.bfloat16) |
|
|
| model.apply(_convert_weights_to_bf16) |
|
|
|
|
| def save_result(result, result_dir, filename, remove_duplicate="", save_format='json'): |
| import json |
| import jsonlines |
| print("Dump result") |
|
|
| |
| if not os.path.exists(result_dir): |
| if is_main_process(): |
| os.makedirs(result_dir) |
| if is_dist_avail_and_initialized(): |
| torch.distributed.barrier() |
|
|
| result_file = os.path.join( |
| result_dir, "%s_rank%d.json" % (filename, get_rank()) |
| ) |
| |
| final_result_file = os.path.join(result_dir, f"{filename}.{save_format}") |
|
|
| json.dump(result, open(result_file, "w")) |
|
|
| if is_dist_avail_and_initialized(): |
| torch.distributed.barrier() |
|
|
| if is_main_process(): |
| |
| |
| result = [] |
|
|
| for rank in range(get_world_size()): |
| result_file = os.path.join(result_dir, "%s_rank%d.json" % (filename, rank)) |
| res = json.load(open(result_file, "r")) |
| result += res |
|
|
| |
| if remove_duplicate: |
| result_new = [] |
| id_set = set() |
| for res in result: |
| if res[remove_duplicate] not in id_set: |
| id_set.add(res[remove_duplicate]) |
| result_new.append(res) |
| result = result_new |
|
|
| if save_format == 'json': |
| json.dump(result, open(final_result_file, "w")) |
| else: |
| assert save_format == 'jsonl', "Only support json adn jsonl format" |
| with jsonlines.open(final_result_file, "w") as writer: |
| writer.write_all(result) |
|
|
| |
|
|
| return final_result_file |
|
|
|
|
| |
| |
| def _resize_with_antialiasing(input, size, interpolation="bicubic", align_corners=True): |
| h, w = input.shape[-2:] |
| factors = (h / size[0], w / size[1]) |
|
|
| |
| |
| sigmas = ( |
| max((factors[0] - 1.0) / 2.0, 0.001), |
| max((factors[1] - 1.0) / 2.0, 0.001), |
| ) |
|
|
| |
| |
| |
| ks = int(max(2.0 * 2 * sigmas[0], 3)), int(max(2.0 * 2 * sigmas[1], 3)) |
|
|
| |
| if (ks[0] % 2) == 0: |
| ks = ks[0] + 1, ks[1] |
|
|
| if (ks[1] % 2) == 0: |
| ks = ks[0], ks[1] + 1 |
|
|
| input = _gaussian_blur2d(input, ks, sigmas) |
|
|
| output = torch.nn.functional.interpolate(input, size=size, mode=interpolation, align_corners=align_corners) |
| return output |
|
|
|
|
| def _compute_padding(kernel_size): |
| """Compute padding tuple.""" |
| |
| |
| if len(kernel_size) < 2: |
| raise AssertionError(kernel_size) |
| computed = [k - 1 for k in kernel_size] |
|
|
| |
| out_padding = 2 * len(kernel_size) * [0] |
|
|
| for i in range(len(kernel_size)): |
| computed_tmp = computed[-(i + 1)] |
|
|
| pad_front = computed_tmp // 2 |
| pad_rear = computed_tmp - pad_front |
|
|
| out_padding[2 * i + 0] = pad_front |
| out_padding[2 * i + 1] = pad_rear |
|
|
| return out_padding |
|
|
|
|
| def _filter2d(input, kernel): |
| |
| b, c, h, w = input.shape |
| tmp_kernel = kernel[:, None, ...].to(device=input.device, dtype=input.dtype) |
|
|
| tmp_kernel = tmp_kernel.expand(-1, c, -1, -1) |
|
|
| height, width = tmp_kernel.shape[-2:] |
|
|
| padding_shape: list[int] = _compute_padding([height, width]) |
| input = torch.nn.functional.pad(input, padding_shape, mode="reflect") |
|
|
| |
| tmp_kernel = tmp_kernel.reshape(-1, 1, height, width) |
| input = input.view(-1, tmp_kernel.size(0), input.size(-2), input.size(-1)) |
|
|
| |
| output = torch.nn.functional.conv2d(input, tmp_kernel, groups=tmp_kernel.size(0), padding=0, stride=1) |
|
|
| out = output.view(b, c, h, w) |
| return out |
|
|
|
|
| def _gaussian(window_size: int, sigma): |
| if isinstance(sigma, float): |
| sigma = torch.tensor([[sigma]]) |
|
|
| batch_size = sigma.shape[0] |
|
|
| x = (torch.arange(window_size, device=sigma.device, dtype=sigma.dtype) - window_size // 2).expand(batch_size, -1) |
|
|
| if window_size % 2 == 0: |
| x = x + 0.5 |
|
|
| gauss = torch.exp(-x.pow(2.0) / (2 * sigma.pow(2.0))) |
|
|
| return gauss / gauss.sum(-1, keepdim=True) |
|
|
|
|
| def _gaussian_blur2d(input, kernel_size, sigma): |
| if isinstance(sigma, tuple): |
| sigma = torch.tensor([sigma], dtype=input.dtype) |
| else: |
| sigma = sigma.to(dtype=input.dtype) |
|
|
| ky, kx = int(kernel_size[0]), int(kernel_size[1]) |
| bs = sigma.shape[0] |
| kernel_x = _gaussian(kx, sigma[:, 1].view(bs, 1)) |
| kernel_y = _gaussian(ky, sigma[:, 0].view(bs, 1)) |
| out_x = _filter2d(input, kernel_x[..., None, :]) |
| out = _filter2d(out_x, kernel_y[..., None]) |
|
|
| return out |
|
|
|
|
| URL_MAP = { |
| "vgg_lpips": "https://heibox.uni-heidelberg.de/f/607503859c864bc1b30b/?dl=1" |
| } |
|
|
| CKPT_MAP = { |
| "vgg_lpips": "vgg.pth" |
| } |
|
|
| MD5_MAP = { |
| "vgg_lpips": "d507d7349b931f0638a25a48a722f98a" |
| } |
|
|
|
|
| def download(url, local_path, chunk_size=1024): |
| os.makedirs(os.path.split(local_path)[0], exist_ok=True) |
| with requests.get(url, stream=True) as r: |
| total_size = int(r.headers.get("content-length", 0)) |
| with tqdm(total=total_size, unit="B", unit_scale=True) as pbar: |
| with open(local_path, "wb") as f: |
| for data in r.iter_content(chunk_size=chunk_size): |
| if data: |
| f.write(data) |
| pbar.update(chunk_size) |
|
|
|
|
| def md5_hash(path): |
| with open(path, "rb") as f: |
| content = f.read() |
| return hashlib.md5(content).hexdigest() |
|
|
|
|
| def get_ckpt_path(name, root, check=False): |
| assert name in URL_MAP |
| path = os.path.join(root, CKPT_MAP[name]) |
| print(md5_hash(path)) |
| if not os.path.exists(path) or (check and not md5_hash(path) == MD5_MAP[name]): |
| print("Downloading {} model from {} to {}".format(name, URL_MAP[name], path)) |
| download(URL_MAP[name], path) |
| md5 = md5_hash(path) |
| assert md5 == MD5_MAP[name], md5 |
| return path |
|
|
|
|
| class KeyNotFoundError(Exception): |
| def __init__(self, cause, keys=None, visited=None): |
| self.cause = cause |
| self.keys = keys |
| self.visited = visited |
| messages = list() |
| if keys is not None: |
| messages.append("Key not found: {}".format(keys)) |
| if visited is not None: |
| messages.append("Visited: {}".format(visited)) |
| messages.append("Cause:\n{}".format(cause)) |
| message = "\n".join(messages) |
| super().__init__(message) |
|
|
|
|
| def retrieve( |
| list_or_dict, key, splitval="/", default=None, expand=True, pass_success=False |
| ): |
| """Given a nested list or dict return the desired value at key expanding |
| callable nodes if necessary and :attr:`expand` is ``True``. The expansion |
| is done in-place. |
| |
| Parameters |
| ---------- |
| list_or_dict : list or dict |
| Possibly nested list or dictionary. |
| key : str |
| key/to/value, path like string describing all keys necessary to |
| consider to get to the desired value. List indices can also be |
| passed here. |
| splitval : str |
| String that defines the delimiter between keys of the |
| different depth levels in `key`. |
| default : obj |
| Value returned if :attr:`key` is not found. |
| expand : bool |
| Whether to expand callable nodes on the path or not. |
| |
| Returns |
| ------- |
| The desired value or if :attr:`default` is not ``None`` and the |
| :attr:`key` is not found returns ``default``. |
| |
| Raises |
| ------ |
| Exception if ``key`` not in ``list_or_dict`` and :attr:`default` is |
| ``None``. |
| """ |
|
|
| keys = key.split(splitval) |
|
|
| success = True |
| try: |
| visited = [] |
| parent = None |
| last_key = None |
| for key in keys: |
| if callable(list_or_dict): |
| if not expand: |
| raise KeyNotFoundError( |
| ValueError( |
| "Trying to get past callable node with expand=False." |
| ), |
| keys=keys, |
| visited=visited, |
| ) |
| list_or_dict = list_or_dict() |
| parent[last_key] = list_or_dict |
|
|
| last_key = key |
| parent = list_or_dict |
|
|
| try: |
| if isinstance(list_or_dict, dict): |
| list_or_dict = list_or_dict[key] |
| else: |
| list_or_dict = list_or_dict[int(key)] |
| except (KeyError, IndexError, ValueError) as e: |
| raise KeyNotFoundError(e, keys=keys, visited=visited) |
|
|
| visited += [key] |
| |
| if expand and callable(list_or_dict): |
| list_or_dict = list_or_dict() |
| parent[last_key] = list_or_dict |
| except KeyNotFoundError as e: |
| if default is None: |
| raise e |
| else: |
| list_or_dict = default |
| success = False |
|
|
| if not pass_success: |
| return list_or_dict |
| else: |
| return list_or_dict, success |