mfliu commited on
Commit
7c82ff8
·
1 Parent(s): 14448bb

Add checkpoints for celeba(32/64/256)

Browse files
celeba256/celeba256_eps1.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:074ee91513ae2c57da811ef8ffd5661ee3b6679edc704877563e14225bc9842a
3
+ size 3182184173
celeba256/celeba256_eps1.yaml ADDED
@@ -0,0 +1,111 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ base_learning_rate: 1.0e-07
3
+ target: ldm.models.diffusion.ddpm.LatentDiffusion
4
+ params:
5
+ ckpt_path: models/ldm/cin256-v2/model.ckpt
6
+ ignore_keys:
7
+ - cond_stage_model
8
+ linear_start: 0.0015
9
+ linear_end: 0.0195
10
+ num_timesteps_cond: 1
11
+ log_every_t: 200
12
+ timesteps: 1000
13
+ first_stage_key: image
14
+ cond_stage_key: class_label
15
+ image_size: 64
16
+ channels: 3
17
+ train_condition_only: true
18
+ conditioning_key: crossattn
19
+ monitor: val/loss
20
+ use_ema: false
21
+ dp_config:
22
+ enabled: true
23
+ epsilon: 1
24
+ delta: 1.0e-05
25
+ max_grad_norm: 0.0005
26
+ poisson_sampling: true
27
+ max_batch_size: 15
28
+ unet_config:
29
+ target: ldm.modules.diffusionmodules.openaimodel.UNetModel
30
+ params:
31
+ image_size: 64
32
+ in_channels: 3
33
+ out_channels: 3
34
+ model_channels: 192
35
+ attention_resolutions:
36
+ - 8
37
+ - 4
38
+ - 2
39
+ num_res_blocks: 2
40
+ channel_mult:
41
+ - 1
42
+ - 2
43
+ - 3
44
+ - 5
45
+ num_heads: 1
46
+ use_spatial_transformer: true
47
+ transformer_depth: 1
48
+ context_dim: 512
49
+ first_stage_config:
50
+ target: ldm.models.autoencoder.VQModelInterface
51
+ params:
52
+ embed_dim: 3
53
+ n_embed: 8192
54
+ ddconfig:
55
+ double_z: false
56
+ z_channels: 3
57
+ resolution: 256
58
+ in_channels: 3
59
+ out_ch: 3
60
+ ch: 128
61
+ ch_mult:
62
+ - 1
63
+ - 2
64
+ - 4
65
+ num_res_blocks: 2
66
+ attn_resolutions: []
67
+ dropout: 0.0
68
+ lossconfig:
69
+ target: torch.nn.Identity
70
+ cond_stage_config:
71
+ target: ldm.modules.encoders.modules.ClassEmbedder
72
+ params:
73
+ n_classes: 2
74
+ embed_dim: 512
75
+ key: class_label
76
+ ablation_blocks: -1
77
+ data:
78
+ target: main.DataModuleFromConfig
79
+ params:
80
+ batch_size: 2048
81
+ num_workers: 6
82
+ wrap: false
83
+ train:
84
+ target: ldm.data.celebahq.CelebAHQ
85
+ params:
86
+ split: train
87
+ target_type: class
88
+ class_attrs:
89
+ - Male
90
+ size: 256
91
+ validation:
92
+ target: ldm.data.celebahq.CelebAHQ
93
+ params:
94
+ split: test
95
+ target_type: class
96
+ class_attrs:
97
+ - Male
98
+ size: 256
99
+ lightning:
100
+ callbacks:
101
+ image_logger:
102
+ target: main.ImageLogger
103
+ params:
104
+ batch_frequency: 25000
105
+ max_images: 8
106
+ increase_log_steps: false
107
+ trainer:
108
+ benchmark: true
109
+ max_epochs: 50
110
+ accelerator: gpu
111
+ gpus: 0,
celeba256/celeba256_eps10.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78ecd018a1856f019bec1b306ccddc59a3c4a231570e8639c259abd9062e3a59
3
+ size 3182184173
celeba256/celeba256_eps10.yaml ADDED
@@ -0,0 +1,111 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ base_learning_rate: 1.0e-07
3
+ target: ldm.models.diffusion.ddpm.LatentDiffusion
4
+ params:
5
+ ckpt_path: models/ldm/cin256-v2/model.ckpt
6
+ ignore_keys:
7
+ - cond_stage_model
8
+ linear_start: 0.0015
9
+ linear_end: 0.0195
10
+ num_timesteps_cond: 1
11
+ log_every_t: 200
12
+ timesteps: 1000
13
+ first_stage_key: image
14
+ cond_stage_key: class_label
15
+ image_size: 64
16
+ channels: 3
17
+ train_condition_only: true
18
+ conditioning_key: crossattn
19
+ monitor: val/loss
20
+ use_ema: false
21
+ dp_config:
22
+ enabled: true
23
+ epsilon: 10
24
+ delta: 1.0e-05
25
+ max_grad_norm: 0.0005
26
+ poisson_sampling: true
27
+ max_batch_size: 15
28
+ unet_config:
29
+ target: ldm.modules.diffusionmodules.openaimodel.UNetModel
30
+ params:
31
+ image_size: 64
32
+ in_channels: 3
33
+ out_channels: 3
34
+ model_channels: 192
35
+ attention_resolutions:
36
+ - 8
37
+ - 4
38
+ - 2
39
+ num_res_blocks: 2
40
+ channel_mult:
41
+ - 1
42
+ - 2
43
+ - 3
44
+ - 5
45
+ num_heads: 1
46
+ use_spatial_transformer: true
47
+ transformer_depth: 1
48
+ context_dim: 512
49
+ first_stage_config:
50
+ target: ldm.models.autoencoder.VQModelInterface
51
+ params:
52
+ embed_dim: 3
53
+ n_embed: 8192
54
+ ddconfig:
55
+ double_z: false
56
+ z_channels: 3
57
+ resolution: 256
58
+ in_channels: 3
59
+ out_ch: 3
60
+ ch: 128
61
+ ch_mult:
62
+ - 1
63
+ - 2
64
+ - 4
65
+ num_res_blocks: 2
66
+ attn_resolutions: []
67
+ dropout: 0.0
68
+ lossconfig:
69
+ target: torch.nn.Identity
70
+ cond_stage_config:
71
+ target: ldm.modules.encoders.modules.ClassEmbedder
72
+ params:
73
+ n_classes: 2
74
+ embed_dim: 512
75
+ key: class_label
76
+ ablation_blocks: -1
77
+ data:
78
+ target: main.DataModuleFromConfig
79
+ params:
80
+ batch_size: 2048
81
+ num_workers: 6
82
+ wrap: false
83
+ train:
84
+ target: ldm.data.celebahq.CelebAHQ
85
+ params:
86
+ split: train
87
+ target_type: class
88
+ class_attrs:
89
+ - Male
90
+ size: 256
91
+ validation:
92
+ target: ldm.data.celebahq.CelebAHQ
93
+ params:
94
+ split: test
95
+ target_type: class
96
+ class_attrs:
97
+ - Male
98
+ size: 256
99
+ lightning:
100
+ callbacks:
101
+ image_logger:
102
+ target: main.ImageLogger
103
+ params:
104
+ batch_frequency: 25000
105
+ max_images: 8
106
+ increase_log_steps: false
107
+ trainer:
108
+ benchmark: true
109
+ max_epochs: 50
110
+ accelerator: gpu
111
+ gpus: 0,
celeba256/celeba256_eps5.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea57477a701530fb09507b5295e3c291bba91678b3e45088a09ef8ca2accb09f
3
+ size 3182184173
celeba256/celeba256_eps5.yaml ADDED
@@ -0,0 +1,111 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ base_learning_rate: 1.0e-07
3
+ target: ldm.models.diffusion.ddpm.LatentDiffusion
4
+ params:
5
+ ckpt_path: models/ldm/cin256-v2/model.ckpt
6
+ ignore_keys:
7
+ - cond_stage_model
8
+ linear_start: 0.0015
9
+ linear_end: 0.0195
10
+ num_timesteps_cond: 1
11
+ log_every_t: 200
12
+ timesteps: 1000
13
+ first_stage_key: image
14
+ cond_stage_key: class_label
15
+ image_size: 64
16
+ channels: 3
17
+ train_condition_only: true
18
+ conditioning_key: crossattn
19
+ monitor: val/loss
20
+ use_ema: false
21
+ dp_config:
22
+ enabled: true
23
+ epsilon: 5
24
+ delta: 1.0e-05
25
+ max_grad_norm: 0.0005
26
+ poisson_sampling: true
27
+ max_batch_size: 15
28
+ unet_config:
29
+ target: ldm.modules.diffusionmodules.openaimodel.UNetModel
30
+ params:
31
+ image_size: 64
32
+ in_channels: 3
33
+ out_channels: 3
34
+ model_channels: 192
35
+ attention_resolutions:
36
+ - 8
37
+ - 4
38
+ - 2
39
+ num_res_blocks: 2
40
+ channel_mult:
41
+ - 1
42
+ - 2
43
+ - 3
44
+ - 5
45
+ num_heads: 1
46
+ use_spatial_transformer: true
47
+ transformer_depth: 1
48
+ context_dim: 512
49
+ first_stage_config:
50
+ target: ldm.models.autoencoder.VQModelInterface
51
+ params:
52
+ embed_dim: 3
53
+ n_embed: 8192
54
+ ddconfig:
55
+ double_z: false
56
+ z_channels: 3
57
+ resolution: 256
58
+ in_channels: 3
59
+ out_ch: 3
60
+ ch: 128
61
+ ch_mult:
62
+ - 1
63
+ - 2
64
+ - 4
65
+ num_res_blocks: 2
66
+ attn_resolutions: []
67
+ dropout: 0.0
68
+ lossconfig:
69
+ target: torch.nn.Identity
70
+ cond_stage_config:
71
+ target: ldm.modules.encoders.modules.ClassEmbedder
72
+ params:
73
+ n_classes: 2
74
+ embed_dim: 512
75
+ key: class_label
76
+ ablation_blocks: -1
77
+ data:
78
+ target: main.DataModuleFromConfig
79
+ params:
80
+ batch_size: 2048
81
+ num_workers: 6
82
+ wrap: false
83
+ train:
84
+ target: ldm.data.celebahq.CelebAHQ
85
+ params:
86
+ split: train
87
+ target_type: class
88
+ class_attrs:
89
+ - Male
90
+ size: 256
91
+ validation:
92
+ target: ldm.data.celebahq.CelebAHQ
93
+ params:
94
+ split: test
95
+ target_type: class
96
+ class_attrs:
97
+ - Male
98
+ size: 256
99
+ lightning:
100
+ callbacks:
101
+ image_logger:
102
+ target: main.ImageLogger
103
+ params:
104
+ batch_frequency: 25000
105
+ max_images: 8
106
+ increase_log_steps: false
107
+ trainer:
108
+ benchmark: true
109
+ max_epochs: 50
110
+ accelerator: gpu
111
+ gpus: 0,
celeba32/celeba32_eps1.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:158e5b0ca17e4a4ded118643f44bcc46b0830885b900aafb75ca0bad719876d4
3
+ size 1570608894
celeba32/celeba32_eps1.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ base_learning_rate: 5.0e-07
3
+ target: ldm.models.diffusion.ddpm.LatentDiffusion
4
+ params:
5
+ ckpt_path: ...
6
+ ignore_keys:
7
+ - cond_stage_model
8
+ linear_start: 0.0015
9
+ linear_end: 0.0195
10
+ timesteps: 1000
11
+ num_timesteps_cond: 1
12
+ first_stage_key: image
13
+ image_size: 16
14
+ channels: 3
15
+ cond_stage_trainable: false
16
+ concat_mode: false
17
+ scale_by_std: true
18
+ train_attention_only: true
19
+ attention_flag: self
20
+ monitor: val/loss_simple_ema
21
+ log_every_t: 200
22
+ dp_config:
23
+ enabled: true
24
+ epsilon: 1
25
+ delta: 1.0e-06
26
+ max_grad_norm: 0.0005
27
+ poisson_sampling: true
28
+ max_batch_size: 100
29
+ unet_config:
30
+ target: ldm.modules.diffusionmodules.openaimodel.UNetModel
31
+ params:
32
+ image_size: 16
33
+ in_channels: 3
34
+ out_channels: 3
35
+ model_channels: 192
36
+ attention_resolutions:
37
+ - 4
38
+ - 2
39
+ - 1
40
+ num_res_blocks: 2
41
+ channel_mult:
42
+ - 1
43
+ - 2
44
+ - 4
45
+ num_head_channels: 32
46
+ use_scale_shift_norm: true
47
+ resblock_updown: true
48
+ first_stage_config:
49
+ target: ldm.models.autoencoder.AutoencoderKL
50
+ params:
51
+ monitor: val/rec_loss
52
+ embed_dim: 3
53
+ lossconfig:
54
+ target: ldm.modules.losses.LPIPSWithDiscriminator
55
+ params:
56
+ disc_start: 50001
57
+ kl_weight: 1.0e-06
58
+ disc_weight: 0.5
59
+ ddconfig:
60
+ double_z: true
61
+ z_channels: 3
62
+ resolution: 32
63
+ in_channels: 3
64
+ out_ch: 3
65
+ ch: 128
66
+ ch_mult:
67
+ - 1
68
+ - 2
69
+ num_res_blocks: 2
70
+ attn_resolutions:
71
+ - 16
72
+ - 8
73
+ dropout: 0.0
74
+ cond_stage_config: __is_unconditional__
75
+ ablation_blocks: -1
76
+ data:
77
+ target: main.DataModuleFromConfig
78
+ params:
79
+ batch_size: 2048
80
+ num_workers: 4
81
+ wrap: false
82
+ train:
83
+ target: ldm.data.celeba.CelebATrain
84
+ params:
85
+ config:
86
+ size: 32
87
+ validation:
88
+ target: ldm.data.celeba.CelebAValidation
89
+ params:
90
+ config:
91
+ size: 32
92
+ lightning:
93
+ callbacks:
94
+ image_logger:
95
+ target: main.ImageLogger
96
+ params:
97
+ batch_frequency: 500
98
+ max_images: 8
99
+ increase_log_steps: false
100
+ trainer:
101
+ benchmark: true
102
+ max_epochs: 20
103
+ accelerator: gpu
104
+ gpus: 0,
celeba32/celeba32_eps10.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c4b0db5eb9e4a79f0e2fe4025e3714a93c2b7dfbcefc247bfd3992600113df33
3
+ size 1570608894
celeba32/celeba32_eps10.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ base_learning_rate: 5.0e-07
3
+ target: ldm.models.diffusion.ddpm.LatentDiffusion
4
+ params:
5
+ ckpt_path: ...
6
+ ignore_keys:
7
+ - cond_stage_model
8
+ linear_start: 0.0015
9
+ linear_end: 0.0195
10
+ timesteps: 1000
11
+ num_timesteps_cond: 1
12
+ first_stage_key: image
13
+ image_size: 16
14
+ channels: 3
15
+ cond_stage_trainable: false
16
+ concat_mode: false
17
+ scale_by_std: true
18
+ train_attention_only: true
19
+ attention_flag: self
20
+ monitor: val/loss_simple_ema
21
+ log_every_t: 200
22
+ dp_config:
23
+ enabled: true
24
+ epsilon: 10
25
+ delta: 1.0e-06
26
+ max_grad_norm: 0.0005
27
+ poisson_sampling: true
28
+ max_batch_size: 100
29
+ unet_config:
30
+ target: ldm.modules.diffusionmodules.openaimodel.UNetModel
31
+ params:
32
+ image_size: 16
33
+ in_channels: 3
34
+ out_channels: 3
35
+ model_channels: 192
36
+ attention_resolutions:
37
+ - 4
38
+ - 2
39
+ - 1
40
+ num_res_blocks: 2
41
+ channel_mult:
42
+ - 1
43
+ - 2
44
+ - 4
45
+ num_head_channels: 32
46
+ use_scale_shift_norm: true
47
+ resblock_updown: true
48
+ first_stage_config:
49
+ target: ldm.models.autoencoder.AutoencoderKL
50
+ params:
51
+ monitor: val/rec_loss
52
+ embed_dim: 3
53
+ lossconfig:
54
+ target: ldm.modules.losses.LPIPSWithDiscriminator
55
+ params:
56
+ disc_start: 50001
57
+ kl_weight: 1.0e-06
58
+ disc_weight: 0.5
59
+ ddconfig:
60
+ double_z: true
61
+ z_channels: 3
62
+ resolution: 32
63
+ in_channels: 3
64
+ out_ch: 3
65
+ ch: 128
66
+ ch_mult:
67
+ - 1
68
+ - 2
69
+ num_res_blocks: 2
70
+ attn_resolutions:
71
+ - 16
72
+ - 8
73
+ dropout: 0.0
74
+ cond_stage_config: __is_unconditional__
75
+ ablation_blocks: -1
76
+ data:
77
+ target: main.DataModuleFromConfig
78
+ params:
79
+ batch_size: 8192
80
+ num_workers: 4
81
+ wrap: false
82
+ train:
83
+ target: ldm.data.celeba.CelebATrain
84
+ params:
85
+ config:
86
+ size: 32
87
+ validation:
88
+ target: ldm.data.celeba.CelebAValidation
89
+ params:
90
+ config:
91
+ size: 32
92
+ lightning:
93
+ callbacks:
94
+ image_logger:
95
+ target: main.ImageLogger
96
+ params:
97
+ batch_frequency: 500
98
+ max_images: 8
99
+ increase_log_steps: false
100
+ trainer:
101
+ benchmark: true
102
+ max_epochs: 20
103
+ accelerator: gpu
104
+ gpus: 0,
celeba32/celeba32_eps5.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bfa5b3a832bd3412f2ec8cec73b43603a818b6be1a010ae49b86ec47094072af
3
+ size 1570608639
celeba32/celeba32_eps5.yaml ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ base_learning_rate: 5.0e-07
3
+ target: ldm.models.diffusion.ddpm.LatentDiffusion
4
+ params:
5
+ ckpt_path: ...
6
+ ignore_keys:
7
+ - cond_stage_model
8
+ linear_start: 0.0015
9
+ linear_end: 0.0195
10
+ timesteps: 1000
11
+ num_timesteps_cond: 1
12
+ first_stage_key: image
13
+ image_size: 16
14
+ channels: 3
15
+ cond_stage_trainable: false
16
+ concat_mode: false
17
+ scale_by_std: true
18
+ train_attention_only: true
19
+ attention_flag: self
20
+ monitor: val/loss_simple_ema
21
+ log_every_t: 200
22
+ dp_config:
23
+ enabled: true
24
+ epsilon: 5
25
+ delta: 1.0e-06
26
+ max_grad_norm: 0.0005
27
+ poisson_sampling: true
28
+ max_batch_size: 100
29
+ unet_config:
30
+ target: ldm.modules.diffusionmodules.openaimodel.UNetModel
31
+ params:
32
+ image_size: 16
33
+ in_channels: 3
34
+ out_channels: 3
35
+ model_channels: 192
36
+ attention_resolutions:
37
+ - 4
38
+ - 2
39
+ - 1
40
+ num_res_blocks: 2
41
+ channel_mult:
42
+ - 1
43
+ - 2
44
+ - 4
45
+ num_head_channels: 32
46
+ use_scale_shift_norm: true
47
+ resblock_updown: true
48
+ first_stage_config:
49
+ target: ldm.models.autoencoder.AutoencoderKL
50
+ params:
51
+ monitor: val/rec_loss
52
+ embed_dim: 3
53
+ lossconfig:
54
+ target: ldm.modules.losses.LPIPSWithDiscriminator
55
+ params:
56
+ disc_start: 50001
57
+ kl_weight: 1.0e-06
58
+ disc_weight: 0.5
59
+ ddconfig:
60
+ double_z: true
61
+ z_channels: 3
62
+ resolution: 32
63
+ in_channels: 3
64
+ out_ch: 3
65
+ ch: 128
66
+ ch_mult:
67
+ - 1
68
+ - 2
69
+ num_res_blocks: 2
70
+ attn_resolutions:
71
+ - 16
72
+ - 8
73
+ dropout: 0.0
74
+ cond_stage_config: __is_unconditional__
75
+ ablation_blocks: -1
76
+ data:
77
+ target: main.DataModuleFromConfig
78
+ params:
79
+ batch_size: 8192
80
+ num_workers: 4
81
+ wrap: false
82
+ train:
83
+ target: ldm.data.celeba.CelebATrain
84
+ params:
85
+ config:
86
+ size: 32
87
+ validation:
88
+ target: ldm.data.celeba.CelebAValidation
89
+ params:
90
+ config:
91
+ size: 32
92
+ lightning:
93
+ callbacks:
94
+ image_logger:
95
+ target: main.ImageLogger
96
+ params:
97
+ batch_frequency: 500
98
+ max_images: 8
99
+ increase_log_steps: false
100
+ trainer:
101
+ benchmark: true
102
+ max_epochs: 20
103
+ accelerator: gpu
104
+ gpus: 0,
celeba64/celeba64_eps1.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc76d82cd2afd88a0351c0d68cd92b68a0aa4eb712f5b138ecd98e9e9901ea92
3
+ size 542413849
celeba64/celeba64_eps1.yaml ADDED
@@ -0,0 +1,103 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ base_learning_rate: 1.0e-07
3
+ target: ldm.models.diffusion.ddpm.LatentDiffusion
4
+ params:
5
+ ckpt_path: ...
6
+ ignore_keys:
7
+ - cond_stage_model
8
+ - model_ema
9
+ linear_start: 0.0015
10
+ linear_end: 0.0195
11
+ timesteps: 1000
12
+ num_timesteps_cond: 1
13
+ first_stage_key: image
14
+ image_size: 32
15
+ channels: 3
16
+ cond_stage_trainable: false
17
+ train_attention_only: true
18
+ use_ema: false
19
+ monitor: val/loss
20
+ log_every_t: 200
21
+ dp_config:
22
+ enabled: true
23
+ epsilon: 1
24
+ delta: 1.0e-06
25
+ max_grad_norm: 0.0005
26
+ poisson_sampling: true
27
+ max_batch_size: 50
28
+ unet_config:
29
+ target: ldm.modules.diffusionmodules.openaimodel.UNetModel
30
+ params:
31
+ image_size: 32
32
+ in_channels: 3
33
+ out_channels: 3
34
+ model_channels: 128
35
+ attention_resolutions:
36
+ - 4
37
+ - 2
38
+ - 1
39
+ num_res_blocks: 2
40
+ channel_mult:
41
+ - 1
42
+ - 2
43
+ - 4
44
+ num_heads: 8
45
+ use_scale_shift_norm: true
46
+ resblock_updown: true
47
+ first_stage_config:
48
+ target: ldm.models.autoencoder.AutoencoderKL
49
+ params:
50
+ monitor: val/rec_loss
51
+ embed_dim: 3
52
+ lossconfig:
53
+ target: ldm.modules.losses.LPIPSWithDiscriminator
54
+ params:
55
+ disc_start: 50001
56
+ kl_weight: 1.0e-06
57
+ disc_weight: 0.5
58
+ ddconfig:
59
+ double_z: true
60
+ z_channels: 3
61
+ resolution: 64
62
+ in_channels: 3
63
+ out_ch: 3
64
+ ch: 192
65
+ ch_mult:
66
+ - 1
67
+ - 2
68
+ num_res_blocks: 2
69
+ attn_resolutions:
70
+ - 16
71
+ - 8
72
+ dropout: 0.0
73
+ cond_stage_config: __is_unconditional__
74
+ ablation_blocks: -1
75
+ data:
76
+ target: main.DataModuleFromConfig
77
+ params:
78
+ batch_size: 8192
79
+ num_workers: 6
80
+ wrap: false
81
+ train:
82
+ target: ldm.data.celeba.CelebATrain
83
+ params:
84
+ config:
85
+ size: 64
86
+ validation:
87
+ target: ldm.data.celeba.CelebAValidation
88
+ params:
89
+ config:
90
+ size: 64
91
+ lightning:
92
+ callbacks:
93
+ image_logger:
94
+ target: main.ImageLogger
95
+ params:
96
+ batch_frequency: 25000
97
+ max_images: 8
98
+ increase_log_steps: false
99
+ trainer:
100
+ benchmark: true
101
+ max_epochs: 70
102
+ accelerator: gpu
103
+ gpus: 0,
celeba64/celeba64_eps10.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4b6875c0421b5308da4aa58087a95e8cdf0b29ef3fd8d5cd50d5be37f7063a5
3
+ size 542413849
celeba64/celeba64_eps10.yaml ADDED
@@ -0,0 +1,103 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ base_learning_rate: 1.0e-07
3
+ target: ldm.models.diffusion.ddpm.LatentDiffusion
4
+ params:
5
+ ckpt_path: ...
6
+ ignore_keys:
7
+ - cond_stage_model
8
+ - model_ema
9
+ linear_start: 0.0015
10
+ linear_end: 0.0195
11
+ timesteps: 1000
12
+ num_timesteps_cond: 1
13
+ first_stage_key: image
14
+ image_size: 32
15
+ channels: 3
16
+ cond_stage_trainable: false
17
+ train_attention_only: true
18
+ use_ema: false
19
+ monitor: val/loss
20
+ log_every_t: 200
21
+ dp_config:
22
+ enabled: true
23
+ epsilon: 10
24
+ delta: 1.0e-06
25
+ max_grad_norm: 0.0005
26
+ poisson_sampling: true
27
+ max_batch_size: 50
28
+ unet_config:
29
+ target: ldm.modules.diffusionmodules.openaimodel.UNetModel
30
+ params:
31
+ image_size: 32
32
+ in_channels: 3
33
+ out_channels: 3
34
+ model_channels: 128
35
+ attention_resolutions:
36
+ - 4
37
+ - 2
38
+ - 1
39
+ num_res_blocks: 2
40
+ channel_mult:
41
+ - 1
42
+ - 2
43
+ - 4
44
+ num_heads: 8
45
+ use_scale_shift_norm: true
46
+ resblock_updown: true
47
+ first_stage_config:
48
+ target: ldm.models.autoencoder.AutoencoderKL
49
+ params:
50
+ monitor: val/rec_loss
51
+ embed_dim: 3
52
+ lossconfig:
53
+ target: ldm.modules.losses.LPIPSWithDiscriminator
54
+ params:
55
+ disc_start: 50001
56
+ kl_weight: 1.0e-06
57
+ disc_weight: 0.5
58
+ ddconfig:
59
+ double_z: true
60
+ z_channels: 3
61
+ resolution: 64
62
+ in_channels: 3
63
+ out_ch: 3
64
+ ch: 192
65
+ ch_mult:
66
+ - 1
67
+ - 2
68
+ num_res_blocks: 2
69
+ attn_resolutions:
70
+ - 16
71
+ - 8
72
+ dropout: 0.0
73
+ cond_stage_config: __is_unconditional__
74
+ ablation_blocks: -1
75
+ data:
76
+ target: main.DataModuleFromConfig
77
+ params:
78
+ batch_size: 8192
79
+ num_workers: 6
80
+ wrap: false
81
+ train:
82
+ target: ldm.data.celeba.CelebATrain
83
+ params:
84
+ config:
85
+ size: 64
86
+ validation:
87
+ target: ldm.data.celeba.CelebAValidation
88
+ params:
89
+ config:
90
+ size: 64
91
+ lightning:
92
+ callbacks:
93
+ image_logger:
94
+ target: main.ImageLogger
95
+ params:
96
+ batch_frequency: 25000
97
+ max_images: 8
98
+ increase_log_steps: false
99
+ trainer:
100
+ benchmark: true
101
+ max_epochs: 70
102
+ accelerator: gpu
103
+ gpus: 0,
celeba64/celeba64_eps5.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b13337d51f2602d8822b7cef7d6b043c1fa175c3ef38888b1ec8ce7d27e85d8
3
+ size 542413849
celeba64/celeba64_eps5.yaml ADDED
@@ -0,0 +1,103 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ base_learning_rate: 1.0e-07
3
+ target: ldm.models.diffusion.ddpm.LatentDiffusion
4
+ params:
5
+ ckpt_path: ...
6
+ ignore_keys:
7
+ - cond_stage_model
8
+ - model_ema
9
+ linear_start: 0.0015
10
+ linear_end: 0.0195
11
+ timesteps: 1000
12
+ num_timesteps_cond: 1
13
+ first_stage_key: image
14
+ image_size: 32
15
+ channels: 3
16
+ cond_stage_trainable: false
17
+ train_attention_only: true
18
+ use_ema: false
19
+ monitor: val/loss
20
+ log_every_t: 200
21
+ dp_config:
22
+ enabled: true
23
+ epsilon: 5
24
+ delta: 1.0e-06
25
+ max_grad_norm: 0.0005
26
+ poisson_sampling: true
27
+ max_batch_size: 50
28
+ unet_config:
29
+ target: ldm.modules.diffusionmodules.openaimodel.UNetModel
30
+ params:
31
+ image_size: 32
32
+ in_channels: 3
33
+ out_channels: 3
34
+ model_channels: 128
35
+ attention_resolutions:
36
+ - 4
37
+ - 2
38
+ - 1
39
+ num_res_blocks: 2
40
+ channel_mult:
41
+ - 1
42
+ - 2
43
+ - 4
44
+ num_heads: 8
45
+ use_scale_shift_norm: true
46
+ resblock_updown: true
47
+ first_stage_config:
48
+ target: ldm.models.autoencoder.AutoencoderKL
49
+ params:
50
+ monitor: val/rec_loss
51
+ embed_dim: 3
52
+ lossconfig:
53
+ target: ldm.modules.losses.LPIPSWithDiscriminator
54
+ params:
55
+ disc_start: 50001
56
+ kl_weight: 1.0e-06
57
+ disc_weight: 0.5
58
+ ddconfig:
59
+ double_z: true
60
+ z_channels: 3
61
+ resolution: 64
62
+ in_channels: 3
63
+ out_ch: 3
64
+ ch: 192
65
+ ch_mult:
66
+ - 1
67
+ - 2
68
+ num_res_blocks: 2
69
+ attn_resolutions:
70
+ - 16
71
+ - 8
72
+ dropout: 0.0
73
+ cond_stage_config: __is_unconditional__
74
+ ablation_blocks: -1
75
+ data:
76
+ target: main.DataModuleFromConfig
77
+ params:
78
+ batch_size: 8192
79
+ num_workers: 6
80
+ wrap: false
81
+ train:
82
+ target: ldm.data.celeba.CelebATrain
83
+ params:
84
+ config:
85
+ size: 64
86
+ validation:
87
+ target: ldm.data.celeba.CelebAValidation
88
+ params:
89
+ config:
90
+ size: 64
91
+ lightning:
92
+ callbacks:
93
+ image_logger:
94
+ target: main.ImageLogger
95
+ params:
96
+ batch_frequency: 25000
97
+ max_images: 8
98
+ increase_log_steps: false
99
+ trainer:
100
+ benchmark: true
101
+ max_epochs: 70
102
+ accelerator: gpu
103
+ gpus: 0,