| import os | |
| import sys | |
| import torch | |
| import dlib | |
| import cv2 | |
| import PIL | |
| import argparse | |
| from tqdm import tqdm | |
| import numpy as np | |
| import torch.nn.functional as F | |
| import torchvision | |
| from torchvision import transforms, utils | |
| from argparse import Namespace | |
| from torch import autograd, optim | |
| from utils.inference_utils import save_image | |
| from models.psp import pSp | |
| from models.stylegan2.model import Downsample | |
| import models.stylegan2.lpips as lpips | |
| def requires_grad(model, flag=True): | |
| for p in model.parameters(): | |
| p.requires_grad = flag | |
| class TrainOptions(): | |
| def __init__(self): | |
| self.parser = argparse.ArgumentParser(description="StyleGANEX Pretraining") | |
| self.parser.add_argument('--exp_dir', type=str, default='./logs/styleganex_pretrain/', help='Path to experiment output directory') | |
| self.parser.add_argument("--ckpt", type=str, default='./pretrained_models/psp_ffhq_encode.pt', help="path of the original psp model") | |
| self.parser.add_argument('--batch_size', default=8, type=int, help='Batch size for training') | |
| self.parser.add_argument('--learning_rate', default=0.0001, type=float, help='Optimizer learning rate') | |
| self.parser.add_argument('--max_steps', default=5000, type=int, help='Maximum number of training steps') | |
| self.parser.add_argument('--image_interval', default=100, type=int, help='Interval for logging train images during training') | |
| def parse(self): | |
| self.opt = self.parser.parse_args() | |
| args = vars(self.opt) | |
| print('Load options') | |
| for name, value in sorted(args.items()): | |
| print('%s: %s' % (str(name), str(value))) | |
| return self.opt | |
| if __name__ == "__main__": | |
| parser = TrainOptions() | |
| args = parser.parse() | |
| print('*'*98) | |
| device = "cuda" | |
| os.makedirs(args.exp_dir, exist_ok=True) | |
| checkpoint_path = os.path.join(args.exp_dir, 'checkpoints') | |
| log_path = os.path.join(args.exp_dir, 'logs') | |
| os.makedirs(checkpoint_path, exist_ok=True) | |
| os.makedirs(log_path, exist_ok=True) | |
| ckpt = torch.load(args.ckpt, map_location='cpu') | |
| opts = ckpt['opts'] | |
| opts['checkpoint_path'] = args.ckpt | |
| if 'output_size' not in opts: | |
| opts['output_size'] = 1024 | |
| if 'toonify_weights' not in opts: | |
| opts['toonify_weights'] = None | |
| opts = Namespace(**opts) | |
| pspex = pSp(opts).to(device).eval() | |
| pspex.latent_avg = pspex.latent_avg.to(device) | |
| requires_grad(pspex, False) | |
| requires_grad(pspex.encoder.featlayer, True) | |
| requires_grad(pspex.encoder.skiplayer, True) | |
| down = Downsample([1, 3, 3, 1], 2).to(device) | |
| requires_grad(down, False) | |
| percept = lpips.PerceptualLoss(model="net-lin", net="vgg", use_gpu=device.startswith("cuda")) | |
| requires_grad(percept.model.net, False) | |
| e_optim = optim.Adam( | |
| list(pspex.encoder.featlayer.parameters()) + list(pspex.encoder.skiplayer.parameters()), | |
| lr=args.learning_rate, | |
| betas=(0.9, 0.99), | |
| ) | |
| pbar = tqdm(range(args.max_steps), initial=0, dynamic_ncols=True, smoothing=0.01) | |
| recon_loss = torch.tensor(0.0, device=device) | |
| loss_dict = {} | |
| accum = 0.5 ** (32 / (10 * 1000)) | |
| for idx in pbar: | |
| with torch.no_grad(): | |
| noise_sample = torch.randn(args.batch_size, 512).cuda() | |
| img_gen, _ = pspex.decoder([noise_sample], input_is_latent=False, truncation=0.7, truncation_latent=0, randomize_noise=False) | |
| img_gen = torch.clamp(img_gen, -1, 1).detach() | |
| img_real = img_gen.clone() | |
| real_input = down(down(img_gen)).detach() | |
| style = pspex.encoder(real_input) + pspex.latent_avg.unsqueeze(0) | |
| if idx == 0: | |
| samplein = real_input.clone().detach() | |
| samplestyle = style.clone().detach() | |
| _, feat = pspex.encoder(real_input, return_feat=True) | |
| fake_img, _ = pspex.decoder([style], input_is_latent=True, randomize_noise=False, first_layer_feature=feat) | |
| recon_loss = F.mse_loss(fake_img, img_real) * 10 | |
| perct_loss = percept(down(fake_img), down(img_real).detach()).sum() | |
| e_loss = recon_loss + perct_loss | |
| loss_dict["er"] = recon_loss | |
| loss_dict["ef"] = perct_loss | |
| pspex.zero_grad() | |
| e_loss.backward() | |
| e_optim.step() | |
| er_loss_val = loss_dict["er"].mean().item() | |
| ef_loss_val = loss_dict["ef"].mean().item() | |
| pbar.set_description( | |
| ( | |
| f"iter: {idx:d}; er: {er_loss_val:.3f}; ef: {ef_loss_val:.3f}" | |
| ) | |
| ) | |
| if idx % args.image_interval == 0 or (idx+1) == args.max_steps: | |
| with torch.no_grad(): | |
| _, sample_feat = pspex.encoder(samplein, return_feat=True) | |
| sample, _ = pspex.decoder([samplestyle], input_is_latent=True, randomize_noise=False, first_layer_feature=sample_feat) | |
| sample = torch.cat((samplein, down(down(sample))), dim=0) | |
| save_image(torchvision.utils.make_grid(sample, args.batch_size, 1), f"%s/%05d.jpg"%(log_path, idx+1)) | |
| save_dict = { | |
| 'state_dict': pspex.state_dict(), | |
| 'opts': vars(pspex.opts) | |
| } | |
| if pspex.opts.start_from_latent_avg: | |
| save_dict['latent_avg'] = pspex.latent_avg | |
| torch.save( | |
| save_dict, | |
| f"%s/%05d.pt"%(checkpoint_path, idx+1), | |
| ) |