| | import torch
|
| | import torch.nn as nn
|
| | import torch.nn.functional as F
|
| | from transformers import PreTrainedModel, PretrainedConfig
|
| |
|
| |
|
| | class BaseVAE(nn.Module):
|
| | def __init__(self, latent_dim=16):
|
| | super(BaseVAE, self).__init__()
|
| | self.latent_dim = latent_dim
|
| |
|
| | self.encoder = nn.Sequential(
|
| | nn.Conv2d(3, 32, 4, 2, 1),
|
| | nn.BatchNorm2d(32),
|
| | nn.ReLU(),
|
| | nn.Conv2d(32, 64, 4, 2, 1),
|
| | nn.BatchNorm2d(64),
|
| | nn.ReLU(),
|
| | nn.Conv2d(64, 128, 4, 2, 1),
|
| | nn.BatchNorm2d(128),
|
| | nn.ReLU(),
|
| | nn.Flatten()
|
| | )
|
| | self.fc_mu = nn.Linear(128 * 4 * 4, latent_dim)
|
| | self.fc_logvar = nn.Linear(128 * 4 * 4, latent_dim)
|
| |
|
| | self.decoder_input = nn.Linear(latent_dim, 128 * 4 * 4)
|
| | self.decoder = nn.Sequential(
|
| | nn.ConvTranspose2d(128, 64, 4, 2, 1),
|
| | nn.BatchNorm2d(64),
|
| | nn.ReLU(),
|
| | nn.ConvTranspose2d(64, 32, 4, 2, 1),
|
| | nn.BatchNorm2d(32),
|
| | nn.ReLU(),
|
| | nn.ConvTranspose2d(32, 3, 4, 2, 1),
|
| | nn.Sigmoid()
|
| | )
|
| |
|
| | def encode(self, x):
|
| | x = self.encoder(x)
|
| | mu = self.fc_mu(x)
|
| | logvar = self.fc_logvar(x)
|
| | return mu, logvar
|
| |
|
| | def reparameterize(self, mu, logvar):
|
| | std = torch.exp(0.5 * logvar)
|
| | eps = torch.randn_like(std)
|
| | return mu + eps * std
|
| |
|
| | def decode(self, z):
|
| | x = self.decoder_input(z)
|
| | x = x.view(-1, 128, 4, 4)
|
| | return self.decoder(x)
|
| |
|
| | def forward(self, x):
|
| | mu, logvar = self.encode(x)
|
| | z = self.reparameterize(mu, logvar)
|
| | recon = self.decode(z)
|
| | return recon, mu, logvar
|
| |
|
| | class VAEConfig(PretrainedConfig):
|
| | model_type = "vae"
|
| |
|
| | def __init__(self, latent_dim=16, **kwargs):
|
| | super().__init__(**kwargs)
|
| | self.latent_dim = latent_dim
|
| |
|
| | class VAEModel(PreTrainedModel):
|
| | config_class = VAEConfig
|
| |
|
| | def __init__(self, config):
|
| | super().__init__(config)
|
| | self.vae = BaseVAE(latent_dim=config.latent_dim)
|
| | self.post_init()
|
| |
|
| | def forward(self, x):
|
| | return self.vae(x)
|
| |
|
| | def encode(self, x):
|
| | return self.vae.encode(x)
|
| |
|
| | def decode(self, z):
|
| | return self.vae.decode(z)
|
| |
|
| | device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
| | model = VAEModel.from_pretrained("Tadas79/emoji-vae-init").to(device)
|
| | model.eval()
|
| |
|