index int64 | repo_name string | branch_name string | path string | content string | import_graph string |
|---|---|---|---|---|---|
84,726 | Srikanth9142/pdfroom-restserver | refs/heads/master | /shelf/serializers.py | from rest_framework import serializers
from .models import Book,Analytic,Reader,ShelfBook,Comment
class BookSerializer(serializers.ModelSerializer):
class Meta:
model = Book
fields = '__all__'
class LikesViewSerializer(serializers.ModelSerializer):
class Meta:
model = Analytic
f... | {"/shelf/admin.py": ["/shelf/models.py"], "/shelf/serializers.py": ["/shelf/models.py"], "/shelf/views.py": ["/shelf/models.py", "/shelf/serializers.py", "/shelf/checkserver.py"]} |
84,727 | Srikanth9142/pdfroom-restserver | refs/heads/master | /shelf/checkserver.py | #import google_oauth
from google.oauth2 import id_token
from google.auth.transport import requests
# token = "eyJhbGciOiJSUzI1NiIsImtpZCI6Ijc2MmZhNjM3YWY5NTM1OTBkYjhiYjhhNjM2YmYxMWQ0MzYwYWJjOTgiLCJ0eXAiOiJKV1QifQ.eyJpc3MiOiJhY2NvdW50cy5nb29nbGUuY29tIiwiYXpwIjoiNDEzMDUxMjYwNzM3LWxuMGZhNGM2djJnOXQyZnMwZzd2Y2oxMDU1NTUwbG... | {"/shelf/admin.py": ["/shelf/models.py"], "/shelf/serializers.py": ["/shelf/models.py"], "/shelf/views.py": ["/shelf/models.py", "/shelf/serializers.py", "/shelf/checkserver.py"]} |
84,728 | Srikanth9142/pdfroom-restserver | refs/heads/master | /shelf/views.py | from django.shortcuts import render,redirect,reverse
from django.http import HttpResponse
from rest_framework import generics
from rest_framework.views import APIView
from rest_framework.response import Response
from rest_framework import authentication, permissions
from .models import Book,Analytic,Reader,ShelfBook,C... | {"/shelf/admin.py": ["/shelf/models.py"], "/shelf/serializers.py": ["/shelf/models.py"], "/shelf/views.py": ["/shelf/models.py", "/shelf/serializers.py", "/shelf/checkserver.py"]} |
84,729 | Srikanth9142/pdfroom-restserver | refs/heads/master | /shelf/models.py | from django.db import models
import uuid
# Create your models here.
class Book(models.Model):
bookid = models.IntegerField(primary_key=True)
name = models.CharField(max_length=100)
fileUrl = models.CharField(max_length=100)
coverphoto = models.ImageField(upload_to='media/')
author = models.CharFie... | {"/shelf/admin.py": ["/shelf/models.py"], "/shelf/serializers.py": ["/shelf/models.py"], "/shelf/views.py": ["/shelf/models.py", "/shelf/serializers.py", "/shelf/checkserver.py"]} |
84,730 | AlKravets/summer_practice_2019 | refs/heads/master | /distance_between_elements.py | import numpy as np
import math
import random
def euvlid_dist(a: list,b: list):
res = 0
for i in range(len(a)):
res += (a[i] - b[i])**2
return math.sqrt(res)
if __name__ == '__main__':
import Create_data
import divide_data
data = Create_data.create_first_data(40,3)
test_data, ... | {"/distance_between_elements.py": ["/Create_data.py", "/divide_data.py"], "/KNN_aloritm.py": ["/Create_data.py", "/divide_data.py", "/distance_between_elements.py"], "/divide_data.py": ["/Create_data.py"]} |
84,731 | AlKravets/summer_practice_2019 | refs/heads/master | /KNN_aloritm.py | import numpy as np
import math
import os
import time
import random
import Create_data
import divide_data
import distance_between_elements
random.seed(20)
def KNN(test_data: list, train_data: list, k: int, number_of_clases: int):
test_lables = []
for test_point in test_data:
#print(test_point)
... | {"/distance_between_elements.py": ["/Create_data.py", "/divide_data.py"], "/KNN_aloritm.py": ["/Create_data.py", "/divide_data.py", "/distance_between_elements.py"], "/divide_data.py": ["/Create_data.py"]} |
84,732 | AlKravets/summer_practice_2019 | refs/heads/master | /Create_data.py | import numpy as np
import math
import os
import pylab as pl
from matplotlib.colors import ListedColormap
import random
import urllib
import csv
def create_first_data(size_class,number_of_classes):
data = []
for classNum in range(number_of_classes):
#Choose random center of 2-dimensional gaussian
... | {"/distance_between_elements.py": ["/Create_data.py", "/divide_data.py"], "/KNN_aloritm.py": ["/Create_data.py", "/divide_data.py", "/distance_between_elements.py"], "/divide_data.py": ["/Create_data.py"]} |
84,733 | AlKravets/summer_practice_2019 | refs/heads/master | /Khachiyan_algorithm.py | import numpy as np
import os
import math
import matplotlib.pyplot as plt
def khaciyan_algorithm (P, toleranse):
dimension = P.shape[0]
N = P.shape[1]
count = 1
err = 1
Q = np.vstack((P,np.ones(N)))
print(Q)
u = np.ones(N)/N
print(u)
while err > toleranse:
U = n... | {"/distance_between_elements.py": ["/Create_data.py", "/divide_data.py"], "/KNN_aloritm.py": ["/Create_data.py", "/divide_data.py", "/distance_between_elements.py"], "/divide_data.py": ["/Create_data.py"]} |
84,734 | AlKravets/summer_practice_2019 | refs/heads/master | /gaussian_blur.py | import cv2
import numpy as np
import os
import math
# Расширим фото на размер половины окна. заполним новые ячейки значением крайних пикселей
def photo_extension(img: np.ndarray, ksize: list) -> np.ndarray:
height, width = img.shape[0], img.shape[1]
if len(img.shape) ==2:
new_img = np.zeros((height+... | {"/distance_between_elements.py": ["/Create_data.py", "/divide_data.py"], "/KNN_aloritm.py": ["/Create_data.py", "/divide_data.py", "/distance_between_elements.py"], "/divide_data.py": ["/Create_data.py"]} |
84,735 | AlKravets/summer_practice_2019 | refs/heads/master | /median_blur.py | import cv2
import numpy as np
import os
# Расширим фото на размер половины окна. заполним новые ячейки значением крайних пикселей
def photo_extension(img: np.ndarray, k: int) -> np.ndarray:
height, width = img.shape[0], img.shape[1]
if len(img.shape) ==2:
new_img = np.zeros((height+k-1, width+k-1), ... | {"/distance_between_elements.py": ["/Create_data.py", "/divide_data.py"], "/KNN_aloritm.py": ["/Create_data.py", "/divide_data.py", "/distance_between_elements.py"], "/divide_data.py": ["/Create_data.py"]} |
84,736 | AlKravets/summer_practice_2019 | refs/heads/master | /divide_data.py | import numpy as np
import math
import random
def divide_data(data, persent):
train_data = []
test_data = []
for row in data:
if random.random() < persent:
test_data.append(row)
else:
train_data.append(row)
return test_data, train_data
if __name__ == '__main__':
... | {"/distance_between_elements.py": ["/Create_data.py", "/divide_data.py"], "/KNN_aloritm.py": ["/Create_data.py", "/divide_data.py", "/distance_between_elements.py"], "/divide_data.py": ["/Create_data.py"]} |
84,737 | AlKravets/summer_practice_2019 | refs/heads/master | /New_gaussian_blur.py | import cv2
import numpy as np
import os
import math
# Расширим фото на размер половины окна. заполним новые ячейки значением крайних пикселей
def photo_extension(img: np.ndarray, ksize: list) -> np.ndarray:
height, width = img.shape[0], img.shape[1]
if len(img.shape) ==2:
new_img = np.zeros((height+k... | {"/distance_between_elements.py": ["/Create_data.py", "/divide_data.py"], "/KNN_aloritm.py": ["/Create_data.py", "/divide_data.py", "/distance_between_elements.py"], "/divide_data.py": ["/Create_data.py"]} |
84,754 | nagadomi/nunif | refs/heads/master | /waifu2x/gui.py | import nunif.pythonw_fix # noqa
import locale
import sys
import os
from os import path
import gc
import functools
from time import time
import threading
from pprint import pprint # noqa
import wx
from wx.lib.masked.numctrl import NumCtrl # noqa
from wx.lib.buttons import GenBitmapButton
from wx.lib.delayedresult impor... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,755 | nagadomi/nunif | refs/heads/master | /waifu2x/web/__main__.py | if __name__ == "__main__":
from os import path
from ..download_models import main as download_main
from .webgen.gen import main as webgen_main
model_dir = path.join(path.dirname(__file__), "..", "pretrained_models")
if not path.exists(model_dir):
download_main()
public_html_dir = path.... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,756 | nagadomi/nunif | refs/heads/master | /nunif/modules/res_block.py | import torch
import torch.nn as nn
from .norm import FRN2d, TLU2d
from .attention import SEBlock
from torch.nn.utils.parametrizations import spectral_norm
def parameterize_none(conv):
return conv
class ResBlock(nn.Module):
def __init__(
self,
in_channels, out_channels,
st... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,757 | nagadomi/nunif | refs/heads/master | /waifu2x/models/cunet.py | import torch
import torch.nn as nn
import torch.nn.functional as F
from nunif.models import I2IBaseModel, register_model
from nunif.modules import SEBlock
import copy
class UNetConv(nn.Module):
def __init__(self, in_channels, mid_channels, out_channels, se):
super(UNetConv, self).__init__()
self.c... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,758 | nagadomi/nunif | refs/heads/master | /nunif/utils/render.py | import math
import torch.nn.functional as F
from .. models import get_model_config, get_model_device
from .. device import autocast
from .seam_blending import SeamBlending
def tiled_render(x, model, tile_size=256, batch_size=4, enable_amp=False):
return SeamBlending.tiled_render(
x, model,
tile_si... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,759 | nagadomi/nunif | refs/heads/master | /iw3/locales/__init__.py | import os
from os import path
import yaml
def load_locales(locale_dir):
files = [path.join(locale_dir, f) for f in os.listdir(locale_dir) if f.endswith(".yml")]
locales = {}
for locale_file in files:
with open(locale_file, mode="r", encoding="utf-8") as f:
locale = yaml.load(f.read(), ... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,760 | nagadomi/nunif | refs/heads/master | /nunif/training/sampler.py | import torch
from torch.utils.data.sampler import WeightedRandomSampler
from collections import deque, defaultdict
from enum import Enum
from ..logger import logger
class MiningMethod(Enum):
LINEAR = 0 # linear scaling
TOP10 = 1 # Top 10% scaling
TOP20 = 2 # Top 20% scaling
class HardExampleSampler... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,761 | nagadomi/nunif | refs/heads/master | /nunif/utils/perlin2d.py | # Original work is perlin-numpy: https://github.com/pvigier/perlin-numpy)
# Pierre Vigier / MIT License
# Vadim Kantorov ported to pytorch: https://gist.github.com/vadimkantorov/ac1b097753f217c5c11bc2ff396e0a57
# some minor changes by nagdaomi
import torch
import math
def generate_perlin_noise_2d(shape, res, fade=la... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,762 | nagadomi/nunif | refs/heads/master | /iw3/training/sbs/stereoimage_generation.py | # file from https://github.com/thygate/stable-diffusion-webui-depthmap-script
# MIT License / Copyright (c) 2023 Bob Thiry
from numba import njit, prange
import numpy as np
from PIL import Image
def create_stereoimages(original_image, depthmap, divergence, modes=None, stereo_balance=0.0,
fill_t... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,763 | nagadomi/nunif | refs/heads/master | /nunif/initializer.py | import os
import torch
import random
import numpy as np
def disable_image_lib_threads():
# Disable OpenMP
# os.environ['OMP_NUM_THREADS'] = '1'
# os.environ['OMP_THREAD_LIMIT'] = '1'
# Disable ImageMagick's Threading
os.environ['MAGICK_THREAD_LIMIT'] = '1'
try:
from wand.resource impo... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,764 | nagadomi/nunif | refs/heads/master | /nunif/modules/auxiliary_loss.py | import torch
import torch.nn as nn
def auxiliary_loss(inputs, targets, modules, weights):
assert (len(inputs) == len(targets) and len(modules) == len(weights))
return sum([modules[i].forward(inputs[i], targets[i]) * weights[i] for i in range(len(inputs))])
class AuxiliaryLoss(nn.Module):
def __init__(se... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,765 | nagadomi/nunif | refs/heads/master | /iw3/models/__init__.py | from .row_flow import RowFlow
__all__ = ["RowFlow"]
| {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,766 | nagadomi/nunif | refs/heads/master | /playground/vae/train_mnist.py | # VAE
# python3 -m playground.vae.train_mnist --data-dir ./tmp/vae --model-dir ./tmp/vae
from torchvision.datasets import MNIST
from torchvision import transforms as T
from torchvision.transforms import functional as TF
from torchvision.utils import make_grid
import torch
from torch import nn
from os import path
from P... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,767 | nagadomi/nunif | refs/heads/master | /cliqa/filter_low_quality_jpeg.py | # Tool to filter low quality jpeg files
import os
from os import path
import argparse
import shutil
from tqdm import tqdm
from multiprocessing import cpu_count
from concurrent.futures import ThreadPoolExecutor as PoolExecutor
import torch
from nunif.models import load_model
from nunif.logger import logger
from .utils i... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,768 | nagadomi/nunif | refs/heads/master | /nunif/transforms/tta.py | import torch
def _hflip(x):
return torch.flip(x, (2,))
def _vflip(x):
return torch.flip(x, (1,))
def _tr_f(x):
return torch.rot90(x, 1, (1, 2))
def _itr_f(x):
return torch.rot90(x, -1, (1, 2))
def tta_split(x):
assert (isinstance(x, torch.Tensor) and x.dim() == 3)
x_hflip = _hflip(x)
... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,769 | nagadomi/nunif | refs/heads/master | /nunif/modules/charbonnier_loss.py | from torch import nn
import torch
def charbonnier_loss(input, target, reduction="mean", eps=1.0e-6):
loss = torch.sqrt(((input - target) ** 2) + eps ** 2)
if reduction is None or reduction == "none":
return loss
elif reduction == "mean":
return loss.mean()
else:
return loss.sum... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,770 | nagadomi/nunif | refs/heads/master | /font_resource/download_google_fonts.py | # Font downloader
# python3 -m font_resource.download_google_fonts
import requests
import os
from os import path
from urllib.parse import quote_plus as url_encode
import shutil
from nunif.utils.downloader import ArchiveDownloader
# Google fonts
GOOGLE_FONTS = [
"Noto Sans JP",
"Noto Serif JP",
"Zen Old ... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,771 | nagadomi/nunif | refs/heads/master | /iw3/utils.py | import os
from os import path
import torch
import torch.nn.functional as F
from torchvision.transforms import functional as TF, InterpolationMode
import argparse
from concurrent.futures import ThreadPoolExecutor as PoolExecutor
import math
from tqdm import tqdm
from PIL import Image, ImageDraw
from nunif.utils.image_lo... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,772 | nagadomi/nunif | refs/heads/master | /iw3/poc/divergence.py | import torch
import torchvision.transforms.functional as TF
import torch.nn.functional as F
from PIL import Image
def psnr(im1, im2):
mse = ((im1 - im2) ** 2).mean()
return 10 * torch.log10(1. / (mse + 1e-6))
def load_image(filename):
im = Image.open(filename)
im.load()
return im
def normalize... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,773 | nagadomi/nunif | refs/heads/master | /cliqa/models/grain_noise_level.py | import torch
import torch.nn as nn
from nunif.models import Model, register_model
from nunif.modules.res_block import ResBlockBNReLU
@register_model
class GrainNoiseLevel(Model):
name = "cliqa.grain_noise_level"
def __init__(self):
super().__init__({})
self.features = nn.Sequential(
... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,774 | nagadomi/nunif | refs/heads/master | /cliqa/training/grain_noise_trainer.py | import sys
from os import path
import argparse
import random
import math
import torch
import torch.nn as nn
import numpy as np
from torch.utils.data.dataset import Dataset
import torchvision.transforms as T
from torchvision.transforms import (
functional as TF,
)
from nunif.utils.image_loader import ImageLoader
fro... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,775 | nagadomi/nunif | refs/heads/master | /waifu2x/models/upconv_7.py | import torch
import torch.nn as nn
from nunif.models import I2IBaseModel, register_model
@register_model
class UpConv7(I2IBaseModel):
name = "waifu2x.upconv_7"
def __init__(self, in_channels=3, out_channels=3):
super(UpConv7, self).__init__(locals(), scale=2, offset=14, in_channels=in_channels)
... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,776 | nagadomi/nunif | refs/heads/master | /waifu2x/download_models.py | import shutil
from os import path
from nunif.utils.downloader import ArchiveDownloader
from nunif.logger import logger
VERSION = "20230504"
MODEL_DIR = path.join(path.dirname(__file__), "pretrained_models")
VERSION_FILE = path.join(MODEL_DIR, VERSION)
MODEL_URL = f"https://github.com/nagadomi/nunif/releases/download/... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,777 | nagadomi/nunif | refs/heads/master | /nunif/utils/downloader.py | import requests
import shutil
import os
from tqdm import tqdm
from tempfile import NamedTemporaryFile, mkdtemp
from abc import ABC, abstractmethod
from ..logger import logger
class Downloader(ABC):
def __init__(self, url=None, name=None, format=None, archive=False, **kwargs):
self.reset_param(url=url, nam... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,778 | nagadomi/nunif | refs/heads/master | /nunif/models/utils.py | from packaging import version as packaging_version
import sys
import torch
from datetime import datetime, timezone
from collections import OrderedDict
import torch.nn as nn
from . register import create_model
from . model import Model
from .. logger import logger
from .. device import create_device
PYTORCH2 = packagi... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,779 | nagadomi/nunif | refs/heads/master | /waifu2x/training/photo_noise.py | # Random noise for Photo, made at random.
import math
import random
from PIL import Image, ImageDraw
from torchvision import transforms as T
from torchvision.transforms import (
functional as TF,
InterpolationMode,
)
import torch
from nunif.utils.perlin2d import generate_perlin_noise_2d
from nunif.utils import ... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,780 | nagadomi/nunif | refs/heads/master | /waifu2x/ui_utils.py | import os
from os import path
import torch
from torchvision.transforms import (
functional as TF,
InterpolationMode)
from PIL import Image
import argparse
import csv
from tqdm import tqdm
from multiprocessing import cpu_count
from concurrent.futures import ThreadPoolExecutor as PoolExecutor
from nunif.logger im... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,781 | nagadomi/nunif | refs/heads/master | /playground/jpeg_qtable/make_extreme_jpeg.py | from PIL import Image
import os
from os import path
import random
ZIGZAG_SCAN_INDEX = [
0, 1, 8, 16, 9, 2, 3, 10,
17, 24, 32, 25, 18, 11, 4,
5, 12, 19, 26, 33, 40, 48,
41, 34, 27, 20, 13, 6, 7,
14, 21, 28, 35, 42, 49, 56,
57, 50, 43, 36, 29, 22, 15,
23, 30, 37, 44, 51, 58, 59,
52, 45, ... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,782 | nagadomi/nunif | refs/heads/master | /playground/depth/show.py | # Show ZoeDepth monocular depth estimation result as 3d point cloud data
# pip3 install opencv-python timm open3d
# python -m playground.depth.show -i playground/jpeg_qtable/images/donut_q100.jpg
import cv2
import torch
import torchvision.transforms.functional as TF
import numpy as np
import argparse
import open3d as o... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,783 | nagadomi/nunif | refs/heads/master | /nunif/utils/text.py | import re
import math
# (),(),『』,「」,〝〟,“”,"",'',``
# 512 is the threshold for skipping over a forgotten closing parenthesis
SPEECH_BLOCK_PATTERNS = (
r"([^()]{0,512})", r"\([^\(\)]{0,512}\)",
r"『[^『』]{0,512}』", r"「[^「」]{0,512}」",
r"〝[^〝〟]{0,512}〟", r"“[^“”]{0,512}”",
r'"[^"]{0,512}"', r"'[^']{0,512}'"... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,784 | nagadomi/nunif | refs/heads/master | /nunif/training/confusion_matrix.py | import torch
import sys
class SoftmaxConfusionMatrix():
def __init__(self, class_names, max_print_class=None):
self.class_names = class_names
self.num_classes = len(class_names)
self.max_print_class = max_print_class
self.confusion_matrix = torch.zeros(
(self.num_classe... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,785 | nagadomi/nunif | refs/heads/master | /waifu2x/training/trainer.py | from os import path
import sys
from time import time
import argparse
import torch
from . dataset import Waifu2xDataset
from .. models.discriminator import SelfSupervisedDiscriminator
from nunif.training.sampler import MiningMethod
from nunif.training.trainer import Trainer
from nunif.training.env import LuminancePSNREn... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,786 | nagadomi/nunif | refs/heads/master | /train.py | import argparse
from pprint import pprint
from nunif.addon import load_addons
from nunif.initializer import set_seed, disable_image_lib_threads
from nunif.training.trainer import create_trainer_default_parser
def main():
parser = argparse.ArgumentParser(formatter_class=argparse.ArgumentDefaultsHelpFormatter)
... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,787 | nagadomi/nunif | refs/heads/master | /nunif/modules/fourier_unit.py | import torch
import torch.nn as nn
from torch.nn.utils.parametrizations import spectral_norm as _spectral_norm
# NOTE: This module does not support export to ONNX (at 2023-04, rfftn and irfftn)
class FourierUnit(nn.Module):
""" From LaMa: Resolution-robust Large Mask Inpainting with Fourier Convolutions
... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,788 | nagadomi/nunif | refs/heads/master | /waifu2x/web/web_load_test.py | # test client for high server load
# Do not use this for remote web sites
# python3 -m waifu2x.web.web_load_test --image-dir /images --ntest 100
import requests
import argparse
import sys
from PIL import Image, UnidentifiedImageError
from concurrent.futures import ThreadPoolExecutor as PoolExecutor
import random
impor... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,789 | nagadomi/nunif | refs/heads/master | /nunif/modules/clamp_loss.py | import torch
from torch import nn
class ClampLoss(nn.Module):
""" Wrapper Module for `(clamp(input, 0, 1) - clamp(target, 0, 1))`
"""
def __init__(self, module, min_value=0, max_value=1, eta=0.001):
super().__init__()
self.module = module
self.min_value = min_value
self.max... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,790 | nagadomi/nunif | refs/heads/master | /nunif/optim/__init__.py | from . lion import Lion
__all__ = ["Lion"]
| {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,791 | nagadomi/nunif | refs/heads/master | /nunif/modules/norm.py | import torch
import torch.nn as nn
import torch.nn.functional as F
class TLU2d(nn.Module):
"""
from Filter Response Normalization Layer
"""
def __init__(self, num_features, eps=1e-06, channel_last=False):
super().__init__()
if channel_last:
self.tau = nn.Parameter(torch.zer... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,792 | nagadomi/nunif | refs/heads/master | /waifu2x/training/downscaling_test.py | from nunif.utils import pil_io
from nunif.transforms import image_magick as IM
from nunif.utils.image_loader import list_images
from torchvision.transforms import (
functional as TF,
InterpolationMode
)
import argparse
from os import path
import os
def modcrop(im):
w, h = im.size
w_pad = -(w % 4)
... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,793 | nagadomi/nunif | refs/heads/master | /waifu2x/unlimited_waifu2x/test_server.py | # python3 -m waifu2x.unlimited_waifu2x.test_server
# View at http://localhost:8812/
# Do not use this server in product environments.
import bottle
import argparse
from os import path
parser = argparse.ArgumentParser(formatter_class=argparse.ArgumentDefaultsHelpFormatter)
parser.add_argument("--cors", action="store_t... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,794 | nagadomi/nunif | refs/heads/master | /nunif/utils/alpha.py | import torch
import torch.nn as nn
class ChannelWiseSum(nn.Module):
def __init__(self, in_channels, kernel_size):
super().__init__()
self.conv = nn.Conv2d(in_channels, in_channels, kernel_size=kernel_size,
stride=1, padding=1, padding_mode="zeros", groups=in_channels,... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,795 | nagadomi/nunif | refs/heads/master | /waifu2x/training/replace_webp.py | # A tool to convert .png to .webp.
# WARNING: Note that the original .png will be deleted.
from PIL import Image
import argparse
from os import path
import os
from tqdm import tqdm
from multiprocessing import cpu_count
from torch.utils.data.dataset import Dataset
from torch.utils.data import DataLoader
from nunif.utils... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,796 | nagadomi/nunif | refs/heads/master | /nunif/utils/gui.py | import wx
from wx.lib.masked.timectrl import TimeCtrl as _TimeCtrl
import os
from os import path
import sys
import subprocess
myEVT_TQDM = wx.NewEventType()
EVT_TQDM = wx.PyEventBinder(myEVT_TQDM, 1)
class TQDMEvent(wx.PyCommandEvent):
def __init__(self, etype, eid, type=None, value=None, desc=None):
su... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,797 | nagadomi/nunif | refs/heads/master | /nunif/models/model.py | import torch
import torch.nn as nn
import copy
class Model(nn.Module):
name = "nunif.Model"
def __init__(self, kwargs):
super(Model, self).__init__()
self.kwargs = {}
self.updated_at = None
self.register_kwargs(kwargs)
def get_device(self):
return next(self.parame... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,798 | nagadomi/nunif | refs/heads/master | /nunif/utils/filename.py | import os
def set_image_ext(filename, format):
"""
Note that this function removes file extension(.*) first.
This may not work with filenames without extensions.
"""
format = format.lower()
filename = os.path.splitext(filename)[0]
if format == "png":
return filename + ".png"
el... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,799 | nagadomi/nunif | refs/heads/master | /create_training_data.py | import argparse
from pprint import pprint
from nunif.addon import load_addons
from nunif.initializer import set_seed, disable_image_lib_threads
def create_default_parser():
parser = argparse.ArgumentParser(
add_help=False,
formatter_class=argparse.ArgumentDefaultsHelpFormatter)
parser.add_argu... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,800 | nagadomi/nunif | refs/heads/master | /waifu2x/training/face_extractor.py | import cv2
import argparse
from tqdm import tqdm
import os
from os import path
def main():
parser = argparse.ArgumentParser(formatter_class=argparse.ArgumentDefaultsHelpFormatter)
parser.add_argument("--input-dir", "-i", type=str, required=True, help="input dir")
parser.add_argument("--output-dir", "-o", ... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,801 | nagadomi/nunif | refs/heads/master | /font_resource/list.py | from nunif.utils.font import FontInfo, FONT_NAME_ID
from glob import glob
from os import path
from pprint import pprint
import argparse
import html
import re
def escape(s):
if not isinstance(s, str):
return s
s = html.escape(s)
s = re.sub(r"([|`_*\[\]])", r"\\\1", s)
return s
def main():
... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,802 | nagadomi/nunif | refs/heads/master | /playground/gan/train_font_dcgan.py | # Generate JIS Level-1 Kanji Fonts with DCGAN
# python -m font_resource.download_google_fonts
# python -m playground.gan.train_font_dcgan --data-dir ./tmp/dcgan --model-dir ./tmp/dcgan
from os import path
from PIL import Image, ImageFont, ImageDraw
from collections import defaultdict
import torch
from torch import nn
f... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,803 | nagadomi/nunif | refs/heads/master | /iw3/training/sbs/trainer.py | from os import path
import argparse
import torch
import torch.nn as nn
from nunif.models import create_model, get_model_config
from nunif.training.env import RGBPSNREnv
from nunif.training.trainer import Trainer
from nunif.modules.lbp_loss import YLBP
from nunif.modules.clamp_loss import ClampLoss
from .dataset import ... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,804 | nagadomi/nunif | refs/heads/master | /nunif/utils/wand_io.py | import torch
from wand.image import Image, IMAGE_TYPES
from wand.api import library
from wand.color import Color
import io
from PIL import ImageCms
from ..transforms.functional import quantize256
sRGB_profile = ImageCms.core.profile_tobytes(ImageCms.createProfile("sRGB"))
GRAYSCALE_TYPES = {
"grayscale",
"g... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,805 | nagadomi/nunif | refs/heads/master | /waifu2x/json2pth.py | # convert old format json model files to pytorch model files
# DEBUG=1 python3 -m waifu2x.json2pth -i ./waifu2x_json_dir -o output_dir
import os
import argparse
from nunif.models import save_model, load_model, create_model
from .models import load_state_from_waifu2x_json
from nunif.logger import logger
def convert_vg... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,806 | nagadomi/nunif | refs/heads/master | /cliqa/models/jpeg_quality.py | import torch
import torch.nn as nn
from nunif.models import Model, register_model
from nunif.modules.res_block import ResBlockBNReLU
@register_model
class JPEGQuality(Model):
name = "cliqa.jpeg_quality"
def __init__(self):
super().__init__({})
self.features = nn.Sequential(
nn.Con... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,807 | nagadomi/nunif | refs/heads/master | /nunif/cli/i2i.py | # image to image
import os
from os import path
import torch
import torchvision.transforms.functional as TF
import argparse
from tqdm import tqdm
from concurrent.futures import ThreadPoolExecutor as PoolExecutor
from .. utils import tiled_render, simple_render, ImageLoader
from .. models import load_model, get_model_con... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,808 | nagadomi/nunif | refs/heads/master | /iw3/training/sbs/depthmap_utils.py | import torch
import torchvision.transforms.functional as TF
import numpy as np
from PIL import Image
from .stereoimage_generation import create_stereoimages
from ... import zoedepth_model as ZU
def normalize_depth(depth):
depth_min = depth.min()
depth_max = depth.max()
max_val = 0xffff
if depth_max - ... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,809 | nagadomi/nunif | refs/heads/master | /nunif/modules/functional.py | import torch
# for VAE
def gaussian_noise(mean, log_var):
# NOTE:
# math.exp(log_var * 0.5) == math.exp(log_var/math.log(math.e**2)) == math.sqrt(math.exp(log_var))
standard_deviation = torch.exp(log_var * 0.5)
noise = torch.randn(mean.shape, device=mean.get_device())
return mean + (noise * stand... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,810 | nagadomi/nunif | refs/heads/master | /font_resource/metadata.py | from os import path
import threading
import random
from fontTools.ttLib import TTFont
from PIL import Image, ImageFont
from .font_map import FONT_MAP
# TODO: make this list
# Use only Hiragana, Katakana, Alphabet and Kanji (No symbol characters due to rendering problem)
LV1_FONT_NAMES = ()
# LV1 + `ー`(Katakana-Hiragan... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,811 | nagadomi/nunif | refs/heads/master | /cliqa/models/__init__.py | from .jpeg_quality import JPEGQuality
from .grain_noise_level import GrainNoiseLevel
__all__ = ["JPEGQuality", "GrainNoiseLevel"]
| {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,812 | nagadomi/nunif | refs/heads/master | /iw3/models/row_flow.py | import torch
import torch.nn as nn
import torch.nn.functional as F
from nunif.models import I2IBaseModel, register_model
@register_model
class RowFlow(I2IBaseModel):
name = "sbs.row_flow"
def __init__(self):
# from diverdence==2.5, (0.5 * 2.5) / 100 * 2048 = 24, so offset must be > 24
super(R... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,813 | nagadomi/nunif | refs/heads/master | /cliqa/download_models.py | import shutil
from os import path
from nunif.utils.downloader import ArchiveDownloader
from nunif.logger import logger
class ModelDownloader(ArchiveDownloader):
def handle(self, src):
src = path.join(src, "pretrained_models")
dst = path.join(path.dirname(__file__), "pretrained_models")
log... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,814 | nagadomi/nunif | refs/heads/master | /waifu2x/models/__init__.py | from .cunet import CUNet, UpCUNet
from .swin_unet import SwinUNet, SwinUNet2x, SwinUNet4x, SwinUNetDownscaled
from .vgg_7 import VGG7
from .upconv_7 import UpConv7
from .json_model import load_state_from_waifu2x_json
from .discriminator import (
L3Discriminator, L3ConditionalDiscriminator,
L3V1Discriminator, L3... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,815 | nagadomi/nunif | refs/heads/master | /nunif/modules/lpips.py | import torch
from torch import nn
import lpips
from os import path
# MODEL_PATH = None
# MODEL_PATH = path.join(path.dirname(__file__), "_lpips_1.pth")
MODEL_PATH = path.join(path.dirname(__file__), "_lpips_2.pth")
class LPIPSWith(nn.Module):
def __init__(self, base_loss, weight=1.):
super().__init__()
... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,816 | nagadomi/nunif | refs/heads/master | /font_resource/utils.py | from nunif.logger import logger
from .metadata import FontInfo, VALIDATE_FONT_SIZE, FONT_MAP
from os import path
FONT_DIR = path.join(path.dirname(__file__), "fonts")
def normalize_font_name(font_name):
return font_name.replace("_", " ")
def native_path(posix_path):
if path.sep == "/":
return posi... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,817 | nagadomi/nunif | refs/heads/master | /nunif/transforms/image_magick.py | import random
from .. utils import wand_io
def to_wand_image(float_tensor):
return wand_io.to_image(float_tensor)
def to_tensor(im):
return wand_io.to_tensor(im)
# ref: https://github.com/tsurumeso/waifu2x-chainer/blob/master/lib/iproc.py
YUV420 = "2x2,1x1,1x1"
YUV444 = "1x1,1x1,1x1"
def jpeg_noise(x,... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,818 | nagadomi/nunif | refs/heads/master | /waifu2x/export_onnx.py | # convert pytorch models to onnx
# DEBUG=1 python3 -m waifu2x.export_onnx -i ./waifu2x/pretrained_models -o ./waifu2x/onnx_models
# NOTE: torchvision 0.14's SwinTransformer has bug in Dropout's training flag. Use 0.15 or later.
import os
from os import path
import argparse
from nunif.models import load_model
from nuni... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,819 | nagadomi/nunif | refs/heads/master | /waifu2x/utils.py | from os import path
from packaging import version as packaging_version
import torch
import torch.nn.functional as F
from nunif.transforms.tta import tta_merge, tta_split
from nunif.utils.render import tiled_render
from nunif.utils.alpha import AlphaBorderPadding
from nunif.models import (
load_model, get_model_conf... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,820 | nagadomi/nunif | refs/heads/master | /waifu2x/locales/__init__.py | import os
from os import path
import yaml
from .. web.webgen.gen import load_locales as load_webgen_locales
WEBGEN_TERMS = [
"artwork", "scan", "photo",
"noise_reduction", "nr_none", "nr_low", "nr_medium", "nr_high", "nr_highest",
"upscaling", "up_none",
]
def merge_en(webgen_locales, lang):
t = web... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,821 | nagadomi/nunif | refs/heads/master | /nunif/utils/video.py | import av
import os
from os import path
import math
from tqdm import tqdm
from PIL import Image
import mimetypes
import re
# Add video mimetypes that does not exist in mimetypes
mimetypes.add_type("video/x-ms-asf", ".asf")
mimetypes.add_type("video/x-ms-vob", ".vob")
mimetypes.add_type("video/divx", ".divx")
mimetype... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,822 | nagadomi/nunif | refs/heads/master | /waifu2x/web/server.py | import os
import sys
import math
from os import path
import posixpath
import torch
import argparse
import bottle
from bottle import request, HTTPResponse
import threading
import requests
import io
import json
from time import time
import hashlib
from configparser import ConfigParser
from diskcache import Cache
import p... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,823 | nagadomi/nunif | refs/heads/master | /nunif/device.py | import torch
def create_device_name(device_id):
if isinstance(device_id, (list, tuple)):
assert len(device_id) > 0
device_id = device_id[0]
if device_id < 0:
device_name = "cpu"
else:
if torch.cuda.is_available():
device_name = 'cuda:%d' % device_id
elif... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,824 | nagadomi/nunif | refs/heads/master | /iw3/zoedepth_model.py | import os
from os import path
import torch
import torch.nn.functional as F
from torchvision.transforms import functional as TF
from nunif.utils.ui import HiddenPrints, TorchHubDir
from nunif.device import create_device, autocast
HUB_MODEL_DIR = path.join(path.dirname(__file__), "pretrained_models", "hub")
def load_... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,825 | nagadomi/nunif | refs/heads/master | /nunif/logger.py | import os
import logging
def _setup():
logger = logging.getLogger("nunif")
if logger.hasHandlers():
logger.handlers.clear()
handler = logging.StreamHandler()
handler.setFormatter(logging.Formatter("%(asctime)s:%(name)s: [%(levelname)8s] %(message)s"))
debug = os.getenv("DEBUG")
if deb... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,826 | nagadomi/nunif | refs/heads/master | /nunif/modules/alex11_loss.py | import torch
from torch import nn
from .charbonnier_loss import CharbonnierLoss
from os import path
class Alex11Loss(nn.Module):
# AlexNet's first 11x11 linear filter
def __init__(self, in_channels):
assert in_channels in {1, 3}
super().__init__()
self.in_channels = in_channels
... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,827 | nagadomi/nunif | refs/heads/master | /nunif/models/onnx_helper_models.py | # helper models for onnxruntime-web
import torch
import torch.nn as nn
from torch.nn import functional as F
from torchvision.transforms import functional as TF
import onnx
import copy
from .model import I2IBaseModel
from ..utils.alpha import ChannelWiseSum
from ..logger import logger
class ONNXReflectionPadding(I2IBa... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,828 | nagadomi/nunif | refs/heads/master | /nunif/transforms/pair.py | import math
from PIL import Image
from torchvision import transforms as T
from torchvision.transforms import (
functional as TF,
InterpolationMode
)
import random
from .std import pad as safe_pad
def same_size(a, b):
if isinstance(a, Image.Image):
return a.size == b.size
else:
return a... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,829 | nagadomi/nunif | refs/heads/master | /waifu2x/training/text_image_generator.py | # random text image generator
# # DEBUG=1 python3 -m waifu2x.training.text_image_generator -n 100 -o ./text_test --bg-dir /bg/eval --seed 73
from PIL import Image, ImageDraw
import random
import math
import argparse
from tqdm import tqdm
import os
from os import path
from multiprocessing import cpu_count
import threadi... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,830 | nagadomi/nunif | refs/heads/master | /playground/jpeg_qtable/show.py | from PIL import Image
import argparse
import os
from os import path
def show(qtables):
for idx, dct_coef in qtables.items():
print(f"{idx}:")
for i in range(8):
row = [dct_coef[i * 8 + j] for j in range(8)]
print(row)
def main():
parser = argparse.ArgumentParser(forma... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,831 | nagadomi/nunif | refs/heads/master | /imagenet/training/trainer.py | import argparse
import torch
from torch import nn
from nunif.training.env import SoftmaxEnv
from nunif.training.trainer import Trainer
from nunif.models import create_model as nunif_create_model
from .dataset import ImageNetDataset
class ImageNetTrainer(Trainer):
def create_model(self):
kwargs = {}
... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,832 | nagadomi/nunif | refs/heads/master | /waifu2x/models/discriminator.py | import torch
import torch.nn as nn
import torch.nn.functional as F
from nunif.models import Model, register_model
from nunif.modules.attention import SEBlock, SNSEBlock
from nunif.modules.res_block import ResBlockGNLReLU, ResBlockSNLReLU
from nunif.modules.fourier_unit import FourierUnitSNLReLU
from torch.nn.utils.para... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,833 | nagadomi/nunif | refs/heads/master | /nunif/utils/pil_io.py | from PIL import Image, ImageCms, ImageMath, PngImagePlugin, UnidentifiedImageError
import io
import struct
import base64
import torchvision.transforms.functional as TF
from ..transforms.functional import quantize256
from ..logger import logger
sRGB_profile = ImageCms.createProfile("sRGB")
CIE_Gray_profile = ImageCms.... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,834 | nagadomi/nunif | refs/heads/master | /nunif/modules/multiscale_loss.py | import torch
import torch.nn as nn
import torch.nn.functional as F
class MultiscaleLoss(nn.Module):
""" Wrapper Module for `(loss(x, y) * w1 + loss(downscale(x), downscale(y)) * w2 ..`
"""
def __init__(self, module, scale_factors=(1, 2), weights=(0.8, 0.2), mode="bicubic"):
super().__init__()
... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,835 | nagadomi/nunif | refs/heads/master | /imagenet/models/torchvision_models.py | import torch
from torchvision.models.vgg import vgg11_bn, VGG11_BN_Weights
from torchvision.models.swin_transformer import swin_t, Swin_T_Weights
from nunif.models import SoftmaxBaseModel, register_model
import torch.nn.functional as F
from ..class_names import CLASS_NAMES
@register_model
class VGG11BN(SoftmaxBaseMod... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,836 | nagadomi/nunif | refs/heads/master | /text_resource/char/__init__.py | from .char import Char
__all__ = ["Char"]
| {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,837 | nagadomi/nunif | refs/heads/master | /font_resource/draw.py | from dataclasses import dataclass
from PIL import Image, ImageFont, ImageDraw, ImageFilter
from .metadata import ImageFonts, FontInfo
@dataclass
class CharBox():
label: str
x: int
y: int
width: int
height: int
has_letter_spacing: bool
class CharDraw():
def __init__(self, font_info, font_... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,838 | nagadomi/nunif | refs/heads/master | /waifu2x/benchmark.py | import os
from os import path
import math
import torch
import argparse
import csv
from torchvision.transforms import functional as TF
from nunif.transforms import functional as NF
import nunif.transforms.image_magick as IM
from nunif.logger import logger
from nunif.utils.image_loader import ImageLoader
from tqdm import... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,839 | nagadomi/nunif | refs/heads/master | /cliqa/training/jpeg_noise_trainer.py | import sys
from os import path
import argparse
import random
import torch
import torch.nn as nn
from torch.utils.data.dataset import Dataset
import torchvision.transforms as T
from torchvision.transforms import (
functional as TF,
)
from nunif.utils.image_loader import ImageLoader
from nunif.utils import pil_io
fro... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,840 | nagadomi/nunif | refs/heads/master | /text_resource/aozora/download.py | # Aozora bunko downloader
# python3 -m text_resource.aozora.download
import requests
import os
from os import path
from urllib.parse import quote_plus as url_encode
import shutil
from nunif.utils.downloader import ArchiveDownloader
AOZORA_TEXT_URL = "https://github.com/aozorahack/aozorabunko_text/archive/master.zip... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
84,841 | nagadomi/nunif | refs/heads/master | /waifu2x/nunif_addon.py | from nunif.addon import Addon
def addon_config():
return Waifu2xAddon()
class Waifu2xAddon(Addon):
def __init__(self):
super(Waifu2xAddon, self).__init__("waifu2x")
def register_create_training_data(self, subparsers, default_parser):
from .training.create_training_data import register
... | {"/waifu2x/gui.py": ["/nunif/pythonw_fix.py", "/waifu2x/ui_utils.py", "/nunif/utils/image_loader.py", "/nunif/utils/video.py", "/nunif/utils/gui.py", "/waifu2x/locales/__init__.py", "/waifu2x/__init__.py"], "/waifu2x/web/__main__.py": ["/waifu2x/download_models.py", "/waifu2x/web/webgen/gen.py", "/waifu2x/web/server.py... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.