|
|
| import torch.nn as nn |
| import torch.nn.functional as F |
|
|
| |
| class VAE(nn.Module): |
| def __init__(self, latent_dim=128): |
| super(VAE, self).__init__() |
| self.latent_dim = latent_dim |
| |
| |
| self.encoder = nn.Sequential( |
| nn.Conv2d(3, 32, kernel_size=4, stride=2, padding=1), nn.ReLU(), |
| nn.Conv2d(32, 64, kernel_size=4, stride=2, padding=1), nn.ReLU(), |
| nn.Conv2d(64, 128, kernel_size=4, stride=2, padding=1), nn.ReLU(), |
| nn.Conv2d(128, 256, kernel_size=4, stride=2, padding=1), nn.ReLU(), |
| nn.Flatten() |
| ) |
| self.fc_mu = nn.Linear(256 * 4 * 4, latent_dim) |
| self.fc_logvar = nn.Linear(256 * 4 * 4, latent_dim) |
|
|
| |
| self.decoder_input = nn.Linear(latent_dim, 256 * 4 * 4) |
| self.decoder = nn.Sequential( |
| nn.ConvTranspose2d(256, 128, kernel_size=4, stride=2, padding=1), nn.ReLU(), |
| nn.ConvTranspose2d(128, 64, kernel_size=4, stride=2, padding=1), nn.ReLU(), |
| nn.ConvTranspose2d(64, 32, kernel_size=4, stride=2, padding=1), nn.ReLU(), |
| nn.ConvTranspose2d(32, 3, kernel_size=4, stride=2, padding=1), |
| nn.Tanh() |
| ) |
|
|
| def decode(self, z): |
| h = self.decoder_input(z) |
| h = h.view(-1, 256, 4, 4) |
| return self.decoder(h) |
|
|