ameerazam08's picture
Upload folder using huggingface_hub
ba32b3e verified
import os
import sys
import torch
import dlib
import cv2
import PIL
import argparse
from tqdm import tqdm
import numpy as np
import torch.nn.functional as F
import torchvision
from torchvision import transforms, utils
from argparse import Namespace
from torch import autograd, optim
from utils.inference_utils import save_image
from models.psp import pSp
from models.stylegan2.model import Downsample
import models.stylegan2.lpips as lpips
def requires_grad(model, flag=True):
for p in model.parameters():
p.requires_grad = flag
class TrainOptions():
def __init__(self):
self.parser = argparse.ArgumentParser(description="StyleGANEX Pretraining")
self.parser.add_argument('--exp_dir', type=str, default='./logs/styleganex_pretrain/', help='Path to experiment output directory')
self.parser.add_argument("--ckpt", type=str, default='./pretrained_models/psp_ffhq_encode.pt', help="path of the original psp model")
self.parser.add_argument('--batch_size', default=8, type=int, help='Batch size for training')
self.parser.add_argument('--learning_rate', default=0.0001, type=float, help='Optimizer learning rate')
self.parser.add_argument('--max_steps', default=5000, type=int, help='Maximum number of training steps')
self.parser.add_argument('--image_interval', default=100, type=int, help='Interval for logging train images during training')
def parse(self):
self.opt = self.parser.parse_args()
args = vars(self.opt)
print('Load options')
for name, value in sorted(args.items()):
print('%s: %s' % (str(name), str(value)))
return self.opt
if __name__ == "__main__":
parser = TrainOptions()
args = parser.parse()
print('*'*98)
device = "cuda"
os.makedirs(args.exp_dir, exist_ok=True)
checkpoint_path = os.path.join(args.exp_dir, 'checkpoints')
log_path = os.path.join(args.exp_dir, 'logs')
os.makedirs(checkpoint_path, exist_ok=True)
os.makedirs(log_path, exist_ok=True)
ckpt = torch.load(args.ckpt, map_location='cpu')
opts = ckpt['opts']
opts['checkpoint_path'] = args.ckpt
if 'output_size' not in opts:
opts['output_size'] = 1024
if 'toonify_weights' not in opts:
opts['toonify_weights'] = None
opts = Namespace(**opts)
pspex = pSp(opts).to(device).eval()
pspex.latent_avg = pspex.latent_avg.to(device)
requires_grad(pspex, False)
requires_grad(pspex.encoder.featlayer, True)
requires_grad(pspex.encoder.skiplayer, True)
down = Downsample([1, 3, 3, 1], 2).to(device)
requires_grad(down, False)
percept = lpips.PerceptualLoss(model="net-lin", net="vgg", use_gpu=device.startswith("cuda"))
requires_grad(percept.model.net, False)
e_optim = optim.Adam(
list(pspex.encoder.featlayer.parameters()) + list(pspex.encoder.skiplayer.parameters()),
lr=args.learning_rate,
betas=(0.9, 0.99),
)
pbar = tqdm(range(args.max_steps), initial=0, dynamic_ncols=True, smoothing=0.01)
recon_loss = torch.tensor(0.0, device=device)
loss_dict = {}
accum = 0.5 ** (32 / (10 * 1000))
for idx in pbar:
with torch.no_grad():
noise_sample = torch.randn(args.batch_size, 512).cuda()
img_gen, _ = pspex.decoder([noise_sample], input_is_latent=False, truncation=0.7, truncation_latent=0, randomize_noise=False)
img_gen = torch.clamp(img_gen, -1, 1).detach()
img_real = img_gen.clone()
real_input = down(down(img_gen)).detach()
style = pspex.encoder(real_input) + pspex.latent_avg.unsqueeze(0)
if idx == 0:
samplein = real_input.clone().detach()
samplestyle = style.clone().detach()
_, feat = pspex.encoder(real_input, return_feat=True)
fake_img, _ = pspex.decoder([style], input_is_latent=True, randomize_noise=False, first_layer_feature=feat)
recon_loss = F.mse_loss(fake_img, img_real) * 10
perct_loss = percept(down(fake_img), down(img_real).detach()).sum()
e_loss = recon_loss + perct_loss
loss_dict["er"] = recon_loss
loss_dict["ef"] = perct_loss
pspex.zero_grad()
e_loss.backward()
e_optim.step()
er_loss_val = loss_dict["er"].mean().item()
ef_loss_val = loss_dict["ef"].mean().item()
pbar.set_description(
(
f"iter: {idx:d}; er: {er_loss_val:.3f}; ef: {ef_loss_val:.3f}"
)
)
if idx % args.image_interval == 0 or (idx+1) == args.max_steps:
with torch.no_grad():
_, sample_feat = pspex.encoder(samplein, return_feat=True)
sample, _ = pspex.decoder([samplestyle], input_is_latent=True, randomize_noise=False, first_layer_feature=sample_feat)
sample = torch.cat((samplein, down(down(sample))), dim=0)
save_image(torchvision.utils.make_grid(sample, args.batch_size, 1), f"%s/%05d.jpg"%(log_path, idx+1))
save_dict = {
'state_dict': pspex.state_dict(),
'opts': vars(pspex.opts)
}
if pspex.opts.start_from_latent_avg:
save_dict['latent_avg'] = pspex.latent_avg
torch.save(
save_dict,
f"%s/%05d.pt"%(checkpoint_path, idx+1),
)