Commit ·
788a6e1
1
Parent(s): 1013ca2
Change any output errors
Browse files- config/celebahq.yaml +1 -1
- dataset/__pycache__/celeba.cpython-310.pyc +0 -0
- dataset/celeba.py +1 -1
- models/__pycache__/blocks.cpython-310.pyc +0 -0
- models/__pycache__/discriminator.cpython-310.pyc +0 -0
- models/__pycache__/lpips.cpython-310.pyc +0 -0
- models/__pycache__/vqvae.cpython-310.pyc +0 -0
- models/blocks.py +2 -2
- models/vqvae.py +1 -1
- scripts/train_vqvae.py → train_vqvae.py +3 -3
config/celebahq.yaml
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
dataset_config:
|
| 2 |
-
im_path: '
|
| 3 |
im_channels : 3
|
| 4 |
im_size : 256
|
| 5 |
name: 'celebhq'
|
|
|
|
| 1 |
dataset_config:
|
| 2 |
+
im_path: '/home/Latent-Diffusion-Conditional/dataset/dataset'
|
| 3 |
im_channels : 3
|
| 4 |
im_size : 256
|
| 5 |
name: 'celebhq'
|
dataset/__pycache__/celeba.cpython-310.pyc
ADDED
|
Binary file (2.01 kB). View file
|
|
|
dataset/celeba.py
CHANGED
|
@@ -18,7 +18,7 @@ class ParquetImageDataset(Dataset):
|
|
| 18 |
return len(self.data)
|
| 19 |
|
| 20 |
def __getitem__(self, idx):
|
| 21 |
-
image = Image.open(
|
| 22 |
caption = self.data.iloc[idx]['text']
|
| 23 |
im_tensor = transforms.Compose([
|
| 24 |
transforms.Resize(self.im_size),
|
|
|
|
| 18 |
return len(self.data)
|
| 19 |
|
| 20 |
def __getitem__(self, idx):
|
| 21 |
+
image = Image.open(self.data.iloc[idx]['image'])
|
| 22 |
caption = self.data.iloc[idx]['text']
|
| 23 |
im_tensor = transforms.Compose([
|
| 24 |
transforms.Resize(self.im_size),
|
models/__pycache__/blocks.cpython-310.pyc
ADDED
|
Binary file (12.5 kB). View file
|
|
|
models/__pycache__/discriminator.cpython-310.pyc
ADDED
|
Binary file (1.5 kB). View file
|
|
|
models/__pycache__/lpips.cpython-310.pyc
ADDED
|
Binary file (4.97 kB). View file
|
|
|
models/__pycache__/vqvae.cpython-310.pyc
ADDED
|
Binary file (3.83 kB). View file
|
|
|
models/blocks.py
CHANGED
|
@@ -101,8 +101,8 @@ class DownBlock(nn.Module):
|
|
| 101 |
]
|
| 102 |
)
|
| 103 |
|
| 104 |
-
self.resnet_down_conv = nn.Conv2d(
|
| 105 |
-
|
| 106 |
def forward(self, x, t_emb=None, context=None):
|
| 107 |
out = x
|
| 108 |
for i in range(self.num_layers):
|
|
|
|
| 101 |
]
|
| 102 |
)
|
| 103 |
|
| 104 |
+
self.resnet_down_conv = nn.Conv2d(out_channels, out_channels,
|
| 105 |
+
4, 2, 1) if self.down_sample else nn.Identity()
|
| 106 |
def forward(self, x, t_emb=None, context=None):
|
| 107 |
out = x
|
| 108 |
for i in range(self.num_layers):
|
models/vqvae.py
CHANGED
|
@@ -74,7 +74,7 @@ class VQVAE(nn.Module):
|
|
| 74 |
self.decoder_layers.append(UpBlock(self.down_channels[i], self.down_channels[i-1],
|
| 75 |
t_emb_dim=None, up_sample=self.down_sample[i-1], num_heads=self.num_heads,
|
| 76 |
num_layers=self.num_up_layers,
|
| 77 |
-
attn=self.
|
| 78 |
norm_channels=self.norm_channels))
|
| 79 |
|
| 80 |
self.decoder_norm_out = nn.GroupNorm(
|
|
|
|
| 74 |
self.decoder_layers.append(UpBlock(self.down_channels[i], self.down_channels[i-1],
|
| 75 |
t_emb_dim=None, up_sample=self.down_sample[i-1], num_heads=self.num_heads,
|
| 76 |
num_layers=self.num_up_layers,
|
| 77 |
+
attn=self.attns[i-1],
|
| 78 |
norm_channels=self.norm_channels))
|
| 79 |
|
| 80 |
self.decoder_norm_out = nn.GroupNorm(
|
scripts/train_vqvae.py → train_vqvae.py
RENAMED
|
@@ -27,7 +27,7 @@ def train(args):
|
|
| 27 |
|
| 28 |
autoencoder_config = config["autoencoder_params"]
|
| 29 |
train_config = config["train_config"]
|
| 30 |
-
dataset_config =
|
| 31 |
|
| 32 |
# Set seed for reproducability
|
| 33 |
seed = train_config["seed"]
|
|
@@ -37,7 +37,7 @@ def train(args):
|
|
| 37 |
model = VQVAE(
|
| 38 |
im_channels=dataset_config["im_channels"], model_config=autoencoder_config)
|
| 39 |
|
| 40 |
-
data_loader = create_dataloader(dataset_config[")
|
| 41 |
|
| 42 |
if not os.path.exists(train_config["task_name"]):
|
| 43 |
os.mkdir(train_config["task_name"])
|
|
@@ -169,6 +169,6 @@ if __name__ == "__main__":
|
|
| 169 |
parser = argparse.ArgumentParser(
|
| 170 |
description="Arguments for vq vae training")
|
| 171 |
parser.add_argument("--config_path", type=str,
|
| 172 |
-
dest="config_path", default="
|
| 173 |
args = parser.parse_args()
|
| 174 |
train(args)
|
|
|
|
| 27 |
|
| 28 |
autoencoder_config = config["autoencoder_params"]
|
| 29 |
train_config = config["train_config"]
|
| 30 |
+
dataset_config = config["dataset_config"]
|
| 31 |
|
| 32 |
# Set seed for reproducability
|
| 33 |
seed = train_config["seed"]
|
|
|
|
| 37 |
model = VQVAE(
|
| 38 |
im_channels=dataset_config["im_channels"], model_config=autoencoder_config)
|
| 39 |
|
| 40 |
+
data_loader = create_dataloader(dataset_config["im_path"])
|
| 41 |
|
| 42 |
if not os.path.exists(train_config["task_name"]):
|
| 43 |
os.mkdir(train_config["task_name"])
|
|
|
|
| 169 |
parser = argparse.ArgumentParser(
|
| 170 |
description="Arguments for vq vae training")
|
| 171 |
parser.add_argument("--config_path", type=str,
|
| 172 |
+
dest="config_path", default="config/celebahq.yaml")
|
| 173 |
args = parser.parse_args()
|
| 174 |
train(args)
|