|
|
|
|
|
import torch |
|
|
import torch.nn as nn |
|
|
|
|
|
|
|
|
LATENT_DIM = 32 |
|
|
COND_DIM = 4 |
|
|
|
|
|
class SimpleVAE(nn.Module): |
|
|
|
|
|
def __init__(self): |
|
|
super(SimpleVAE, self).__init__() |
|
|
|
|
|
|
|
|
self.encoder = nn.Sequential( |
|
|
nn.Conv2d(3, 16, 4, 2, 1), nn.ReLU(), |
|
|
nn.Conv2d(16, 32, 4, 2, 1), nn.ReLU(), |
|
|
nn.Flatten(), |
|
|
nn.Linear(32 * 16 * 16, 128), nn.ReLU(), |
|
|
nn.Linear(128, LATENT_DIM * 2) |
|
|
) |
|
|
|
|
|
|
|
|
self.decoder_input = nn.Linear(LATENT_DIM + COND_DIM, 32 * 16 * 16) |
|
|
self.decoder = nn.Sequential( |
|
|
nn.ConvTranspose2d(32, 16, 4, 2, 1), nn.ReLU(), |
|
|
nn.ConvTranspose2d(16, 3, 4, 2, 1), nn.Sigmoid() |
|
|
) |
|
|
|
|
|
def encode(self, x): |
|
|
result = self.encoder(x) |
|
|
mu, logvar = result.chunk(2, dim=1) |
|
|
return mu, logvar |
|
|
|
|
|
def reparameterize(self, mu, logvar): |
|
|
std = torch.exp(0.5 * logvar) |
|
|
eps = torch.randn_like(std) |
|
|
return eps * std + mu |
|
|
|
|
|
def decode(self, z, conditions): |
|
|
z_c = torch.cat([z, conditions], dim=1) |
|
|
h = self.decoder_input(z_c) |
|
|
h = h.view(-1, 32, 16, 16) |
|
|
return self.decoder(h) |
|
|
|
|
|
def forward(self, x, conditions): |
|
|
mu, logvar = self.encode(x) |
|
|
z = self.reparameterize(mu, logvar) |
|
|
recon_x = self.decode(z, conditions) |
|
|
return recon_x, mu, logvar |
|
|
|