AzhureRaven commited on
Commit
d045179
·
verified ·
1 Parent(s): e33d760

Upload 17 files

Browse files
model_index.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "StableDiffusionPipeline",
3
+ "_diffusers_version": "0.21.4",
4
+ "feature_extractor": [
5
+ null,
6
+ null
7
+ ],
8
+ "requires_safety_checker": null,
9
+ "safety_checker": [
10
+ null,
11
+ null
12
+ ],
13
+ "scheduler": [
14
+ "diffusers",
15
+ "DDPMScheduler"
16
+ ],
17
+ "text_encoder": [
18
+ "transformers",
19
+ "CLIPTextModel"
20
+ ],
21
+ "tokenizer": [
22
+ "transformers",
23
+ "CLIPTokenizer"
24
+ ],
25
+ "unet": [
26
+ "diffusers",
27
+ "UNet2DConditionModel"
28
+ ],
29
+ "vae": [
30
+ "diffusers",
31
+ "AutoencoderKL"
32
+ ]
33
+ }
optimizer_rico.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "doc": {
3
+ "base": "base optimizer configuration for unet and text encoder",
4
+ "text_encoder_overrides": "text encoder config overrides",
5
+ "text_encoder_lr_scale": "if LR not set on text encoder, sets the Lr to a multiple of the Base LR. for example, if base `lr` is 2e-6 and `text_encoder_lr_scale` is 0.5, the text encoder's LR will be set to `1e-6`.",
6
+ "-----------------": "-----------------",
7
+ "optimizer": "adamw, adamw8bit, lion",
8
+ "optimizer_desc": "'adamw' in standard 32bit, 'adamw8bit' is bitsandbytes, 'lion' is lucidrains",
9
+ "lr": "learning rate, if null will use CLI or main JSON config value",
10
+ "lr_scheduler": "'constant' or 'cosine'",
11
+ "lr_warmup_steps": "number of steps to warmup LR to target LR, if null will use CLI or default a value based on max epochs",
12
+ "lr_decay_steps": "number of steps to decay LR to zero for cosine, if null will use CLI or default a value based on max epochs",
13
+ "betas": "exponential decay rates for the moment estimates",
14
+ "epsilon": "value added to denominator for numerical stability, unused for lion",
15
+ "weight_decay": "weight decay (L2 penalty)",
16
+ "------------------": "-----------------",
17
+ "unfreeze_last_n_layers": "if not null, freeze all parameters in the text encoder except for the last n layers and the final layer norm"
18
+ },
19
+ "base": {
20
+ "optimizer": "adamw8bit",
21
+ "lr": 1e-6,
22
+ "lr_scheduler": "constant",
23
+ "lr_decay_steps": null,
24
+ "lr_warmup_steps": null,
25
+ "betas": [0.9, 0.999],
26
+ "epsilon": 1e-8,
27
+ "weight_decay": 0.010
28
+ },
29
+ "text_encoder_overrides": {
30
+ "optimizer": null,
31
+ "lr": 5e-7,
32
+ "lr_scheduler": "cosine",
33
+ "lr_decay_steps": null,
34
+ "lr_warmup_steps": null,
35
+ "betas": null,
36
+ "epsilon": null,
37
+ "weight_decay": null
38
+ },
39
+ "text_encoder_freezing": {
40
+ "unfreeze_last_n_layers": null
41
+ },
42
+ "apply_grad_scaler_step_tweaks": true
43
+ }
optimizer_rico_comp.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "doc": {
3
+ "base": "base optimizer configuration for unet and text encoder",
4
+ "text_encoder_overrides": "text encoder config overrides",
5
+ "text_encoder_lr_scale": "if LR not set on text encoder, sets the Lr to a multiple of the Base LR. for example, if base `lr` is 2e-6 and `text_encoder_lr_scale` is 0.5, the text encoder's LR will be set to `1e-6`.",
6
+ "-----------------": "-----------------",
7
+ "optimizer": "adamw, adamw8bit, lion",
8
+ "optimizer_desc": "'adamw' in standard 32bit, 'adamw8bit' is bitsandbytes, 'lion' is lucidrains",
9
+ "lr": "learning rate, if null will use CLI or main JSON config value",
10
+ "lr_scheduler": "'constant' or 'cosine'",
11
+ "lr_warmup_steps": "number of steps to warmup LR to target LR, if null will use CLI or default a value based on max epochs",
12
+ "lr_decay_steps": "number of steps to decay LR to zero for cosine, if null will use CLI or default a value based on max epochs",
13
+ "betas": "exponential decay rates for the moment estimates",
14
+ "epsilon": "value added to denominator for numerical stability, unused for lion",
15
+ "weight_decay": "weight decay (L2 penalty)",
16
+ "------------------": "-----------------",
17
+ "unfreeze_last_n_layers": "if not null, freeze all parameters in the text encoder except for the last n layers and the final layer norm"
18
+ },
19
+ "base": {
20
+ "optimizer": "adamw8bit",
21
+ "lr": 2.5e-6,
22
+ "lr_scheduler": "constant",
23
+ "lr_decay_steps": null,
24
+ "lr_warmup_steps": null,
25
+ "betas": [0.9, 0.999],
26
+ "epsilon": 1e-8,
27
+ "weight_decay": 0.010
28
+ },
29
+ "text_encoder_overrides": {
30
+ "optimizer": null,
31
+ "lr": 1.25e-6,
32
+ "lr_scheduler": "cosine",
33
+ "lr_decay_steps": null,
34
+ "lr_warmup_steps": null,
35
+ "betas": null,
36
+ "epsilon": null,
37
+ "weight_decay": null
38
+ },
39
+ "text_encoder_freezing": {
40
+ "unfreeze_last_n_layers": null
41
+ },
42
+ "apply_grad_scaler_step_tweaks": true
43
+ }
rico_diffusion-20250309-203322_main.json ADDED
@@ -0,0 +1,112 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": "rico_diffusion.json",
3
+ "attn_type": "sdp",
4
+ "batch_size": 7,
5
+ "ckpt_every_n_minutes": 1000000000.0,
6
+ "clip_grad_norm": null,
7
+ "clip_skip": 2,
8
+ "cond_dropout": 0.1,
9
+ "data_root": "/home/azhure/Tesis/save_data",
10
+ "disable_amp": false,
11
+ "disable_textenc_training": false,
12
+ "flip_p": 0.0,
13
+ "gpuid": 0,
14
+ "gradient_checkpointing": true,
15
+ "grad_accum": 1,
16
+ "logdir": "logs",
17
+ "log_step": 25,
18
+ "loss_type": "huber",
19
+ "max_epochs": 5,
20
+ "optimizer_config": "optimizer_rico.json",
21
+ "project_name": "rico_diffusion",
22
+ "resolution": 512,
23
+ "resume_ckpt": "logs/rico_diffusion_v2_comp_big-20250307-090306/ckpts/last-rico_diffusion_v2_comp_big-ep04-gs32932",
24
+ "sample_prompts": "sample_prompts_rico.json",
25
+ "sample_steps": 15000,
26
+ "save_ckpt_dir": "model_output",
27
+ "save_ckpts_from_n_epochs": 0,
28
+ "save_every_n_epochs": 5,
29
+ "save_optimizer": false,
30
+ "scale_lr": false,
31
+ "seed": 555,
32
+ "keep_tags": 3,
33
+ "shuffle_tags": false,
34
+ "validation_config": "validation_rico_off.json",
35
+ "zero_frequency_noise_ratio": 0.02,
36
+ "min_snr_gamma": 5.0,
37
+ "enable_zero_terminal_snr": true,
38
+ "amp": true,
39
+ "disable_unet_training": false,
40
+ "embedding_perturbation": 0.0,
41
+ "lr": null,
42
+ "lr_decay_steps": 124140,
43
+ "lr_scheduler": "constant",
44
+ "lr_warmup_steps": 2482,
45
+ "no_prepend_last": false,
46
+ "no_save_ckpt": false,
47
+ "plugins": null,
48
+ "run_name": null,
49
+ "save_full_precision": false,
50
+ "timestep_start": 0,
51
+ "timestep_end": 1000,
52
+ "train_sampler": "ddpm",
53
+ "wandb": false,
54
+ "write_schedule": false,
55
+ "rated_dataset": false,
56
+ "rated_dataset_target_dropout_percent": 50,
57
+ "load_settings_every_epoch": null,
58
+ "ema_decay_rate": null,
59
+ "ema_strength_target": null,
60
+ "ema_update_interval": 500,
61
+ "ema_device": "cpu",
62
+ "ema_sample_nonema_model": false,
63
+ "ema_sample_ema_model": false,
64
+ "ema_resume_model": null,
65
+ "pyramid_noise_discount": null,
66
+ "aspects": [
67
+ [
68
+ 512,
69
+ 512
70
+ ],
71
+ [
72
+ 576,
73
+ 448
74
+ ],
75
+ [
76
+ 448,
77
+ 576
78
+ ],
79
+ [
80
+ 640,
81
+ 384
82
+ ],
83
+ [
84
+ 384,
85
+ 640
86
+ ],
87
+ [
88
+ 768,
89
+ 320
90
+ ],
91
+ [
92
+ 320,
93
+ 768
94
+ ],
95
+ [
96
+ 896,
97
+ 256
98
+ ],
99
+ [
100
+ 256,
101
+ 896
102
+ ],
103
+ [
104
+ 1024,
105
+ 256
106
+ ],
107
+ [
108
+ 256,
109
+ 1024
110
+ ]
111
+ ]
112
+ }
rico_diffusion-20250309-203322_opt.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base": {
3
+ "optimizer": "adamw8bit",
4
+ "lr": 1e-06,
5
+ "lr_scheduler": "constant",
6
+ "lr_decay_steps": 124140,
7
+ "lr_warmup_steps": 2482,
8
+ "betas": [
9
+ 0.9,
10
+ 0.999
11
+ ],
12
+ "epsilon": 1e-08,
13
+ "weight_decay": 0.01
14
+ },
15
+ "text_encoder_overrides": {
16
+ "optimizer": "adamw8bit",
17
+ "lr": 5e-07,
18
+ "lr_scheduler": "cosine",
19
+ "lr_decay_steps": 124140,
20
+ "lr_warmup_steps": 2482,
21
+ "betas": [
22
+ 0.9,
23
+ 0.999
24
+ ],
25
+ "epsilon": 1e-08,
26
+ "weight_decay": 0.01
27
+ },
28
+ "text_encoder_freezing": {
29
+ "unfreeze_last_n_layers": null
30
+ },
31
+ "apply_grad_scaler_step_tweaks": true
32
+ }
rico_diffusion.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "attn_type": "sdp",
3
+ "batch_size": 7,
4
+ "ckpt_every_n_minutes": null,
5
+ "clip_grad_norm": null,
6
+ "clip_skip": 2,
7
+ "cond_dropout": 0.10,
8
+ "data_root": "/home/azhure/Tesis/save_data",
9
+ "disable_amp": false,
10
+ "disable_textenc_training": false,
11
+ "flip_p": 0.0,
12
+ "gpuid": 0,
13
+ "gradient_checkpointing": true,
14
+ "grad_accum": 1,
15
+ "logdir": "logs",
16
+ "log_step": 25,
17
+ "loss_type": "huber",
18
+ "max_epochs": 5,
19
+ "optimizer_config": "optimizer_rico.json",
20
+ "project_name": "rico_diffusion",
21
+ "resolution": 512,
22
+ "resume_ckpt": "logs/rico_diffusion_v2_comp_big-20250307-090306/ckpts/last-rico_diffusion_v2_comp_big-ep04-gs32932",
23
+ "sample_prompts": "sample_prompts_rico.json",
24
+ "sample_steps": 15000,
25
+ "save_ckpt_dir": "model_output",
26
+ "save_ckpts_from_n_epochs": 0,
27
+ "save_every_n_epochs": 5,
28
+ "save_optimizer": false,
29
+ "scale_lr": false,
30
+ "seed": 555,
31
+ "keep_tags": 3,
32
+ "shuffle_tags": false,
33
+ "validation_config": "validation_rico_off.json",
34
+ "zero_frequency_noise_ratio": 0.02,
35
+ "min_snr_gamma": 5.0,
36
+ "enable_zero_terminal_snr": true
37
+ }
rico_diffusion.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0318d1fc435edcbfde87f7df5c892dd5a347aa719323d75a222ec5d9a595aeb
3
+ size 2132625158
rico_diffusion_v2_comp.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "attn_type": "sdp",
3
+ "batch_size": 11,
4
+ "ckpt_every_n_minutes": null,
5
+ "clip_grad_norm": null,
6
+ "clip_skip": 2,
7
+ "cond_dropout": 0.00,
8
+ "data_root": "/home/azhure/Tesis/save_data_comp",
9
+ "disable_amp": false,
10
+ "disable_textenc_training": false,
11
+ "flip_p": 0.0,
12
+ "gpuid": 0,
13
+ "gradient_checkpointing": true,
14
+ "grad_accum": 1,
15
+ "logdir": "logs",
16
+ "log_step": 25,
17
+ "loss_type": "huber",
18
+ "max_epochs": 3,
19
+ "optimizer_config": "optimizer_rico_comp.json",
20
+ "project_name": "rico_diffusion_v2_comp",
21
+ "resolution": 384,
22
+ "resume_ckpt": "stable-diffusion-v1-5/stable-diffusion-v1-5",
23
+ "sample_prompts": "sample_prompts.txt",
24
+ "sample_steps": 999999999,
25
+ "save_ckpt_dir": "model_output",
26
+ "save_ckpts_from_n_epochs": 0,
27
+ "save_every_n_epochs": 10,
28
+ "save_optimizer": false,
29
+ "scale_lr": false,
30
+ "seed": 555,
31
+ "keep_tags": 3,
32
+ "shuffle_tags": false,
33
+ "validation_config": "validation_rico_off.json",
34
+ "zero_frequency_noise_ratio": 0.02,
35
+ "min_snr_gamma": 5.0,
36
+ "enable_zero_terminal_snr": true,
37
+ "no_save_ckpt": true
38
+ }
rico_diffusion_v2_comp_big.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "attn_type": "sdp",
3
+ "batch_size": 9,
4
+ "ckpt_every_n_minutes": null,
5
+ "clip_grad_norm": null,
6
+ "clip_skip": 2,
7
+ "cond_dropout": 0.00,
8
+ "data_root": "/home/azhure/Tesis/save_data_comp_big",
9
+ "disable_amp": false,
10
+ "disable_textenc_training": false,
11
+ "flip_p": 0.0,
12
+ "gpuid": 0,
13
+ "gradient_checkpointing": true,
14
+ "grad_accum": 1,
15
+ "logdir": "logs",
16
+ "log_step": 25,
17
+ "loss_type": "huber",
18
+ "max_epochs": 4,
19
+ "optimizer_config": "optimizer_rico_comp.json",
20
+ "project_name": "rico_diffusion_v2_comp_big",
21
+ "resolution": 448,
22
+ "resume_ckpt": "findlast",
23
+ "sample_prompts": "sample_prompts.txt",
24
+ "sample_steps": 999999999,
25
+ "save_ckpt_dir": "model_output",
26
+ "save_ckpts_from_n_epochs": 0,
27
+ "save_every_n_epochs": 10,
28
+ "save_optimizer": false,
29
+ "scale_lr": false,
30
+ "seed": 555,
31
+ "keep_tags": 3,
32
+ "shuffle_tags": false,
33
+ "validation_config": "validation_rico_off.json",
34
+ "zero_frequency_noise_ratio": 0.02,
35
+ "min_snr_gamma": 5.0,
36
+ "enable_zero_terminal_snr": true,
37
+ "no_save_ckpt": true
38
+ }
rico_diffusion_v2_comp_button.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "attn_type": "sdp",
3
+ "batch_size": 11,
4
+ "ckpt_every_n_minutes": null,
5
+ "clip_grad_norm": null,
6
+ "clip_skip": 2,
7
+ "cond_dropout": 0.00,
8
+ "data_root": "/home/azhure/Tesis/Button",
9
+ "disable_amp": false,
10
+ "disable_textenc_training": false,
11
+ "flip_p": 0.0,
12
+ "gpuid": 0,
13
+ "gradient_checkpointing": true,
14
+ "grad_accum": 1,
15
+ "logdir": "logs",
16
+ "log_step": 25,
17
+ "loss_type": "huber",
18
+ "max_epochs": 2,
19
+ "optimizer_config": "optimizer_rico_comp.json",
20
+ "project_name": "rico_diffusion_v2_comp_button",
21
+ "resolution": 384,
22
+ "resume_ckpt": "findlast",
23
+ "sample_prompts": "sample_prompts.txt",
24
+ "sample_steps": 999999999,
25
+ "save_ckpt_dir": "model_output",
26
+ "save_ckpts_from_n_epochs": 0,
27
+ "save_every_n_epochs": 10,
28
+ "save_optimizer": false,
29
+ "scale_lr": false,
30
+ "seed": 555,
31
+ "keep_tags": 3,
32
+ "shuffle_tags": false,
33
+ "validation_config": "validation_rico_off.json",
34
+ "zero_frequency_noise_ratio": 0.02,
35
+ "min_snr_gamma": 5.0,
36
+ "enable_zero_terminal_snr": true,
37
+ "no_save_ckpt": true
38
+ }
rico_diffusion_v2_comp_icon.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "attn_type": "sdp",
3
+ "batch_size": 11,
4
+ "ckpt_every_n_minutes": null,
5
+ "clip_grad_norm": null,
6
+ "clip_skip": 2,
7
+ "cond_dropout": 0.00,
8
+ "data_root": "/home/azhure/Tesis/Icon",
9
+ "disable_amp": false,
10
+ "disable_textenc_training": false,
11
+ "flip_p": 0.0,
12
+ "gpuid": 0,
13
+ "gradient_checkpointing": true,
14
+ "grad_accum": 1,
15
+ "logdir": "logs",
16
+ "log_step": 25,
17
+ "loss_type": "huber",
18
+ "max_epochs": 2,
19
+ "optimizer_config": "optimizer_rico_comp.json",
20
+ "project_name": "rico_diffusion_v2_comp_icon",
21
+ "resolution": 384,
22
+ "resume_ckpt": "findlast",
23
+ "sample_prompts": "sample_prompts.txt",
24
+ "sample_steps": 999999999,
25
+ "save_ckpt_dir": "model_output",
26
+ "save_ckpts_from_n_epochs": 0,
27
+ "save_every_n_epochs": 10,
28
+ "save_optimizer": false,
29
+ "scale_lr": false,
30
+ "seed": 555,
31
+ "keep_tags": 3,
32
+ "shuffle_tags": false,
33
+ "validation_config": "validation_rico_off.json",
34
+ "zero_frequency_noise_ratio": 0.02,
35
+ "min_snr_gamma": 5.0,
36
+ "enable_zero_terminal_snr": true,
37
+ "no_save_ckpt": true
38
+ }
rico_diffusion_v2_comp_image.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "attn_type": "sdp",
3
+ "batch_size": 11,
4
+ "ckpt_every_n_minutes": null,
5
+ "clip_grad_norm": null,
6
+ "clip_skip": 2,
7
+ "cond_dropout": 0.00,
8
+ "data_root": "/home/azhure/Tesis/Image",
9
+ "disable_amp": false,
10
+ "disable_textenc_training": false,
11
+ "flip_p": 0.0,
12
+ "gpuid": 0,
13
+ "gradient_checkpointing": true,
14
+ "grad_accum": 1,
15
+ "logdir": "logs",
16
+ "log_step": 25,
17
+ "loss_type": "huber",
18
+ "max_epochs": 2,
19
+ "optimizer_config": "optimizer_rico_comp.json",
20
+ "project_name": "rico_diffusion_v2_comp_image",
21
+ "resolution": 384,
22
+ "resume_ckpt": "findlast",
23
+ "sample_prompts": "sample_prompts.txt",
24
+ "sample_steps": 999999999,
25
+ "save_ckpt_dir": "model_output",
26
+ "save_ckpts_from_n_epochs": 0,
27
+ "save_every_n_epochs": 10,
28
+ "save_optimizer": false,
29
+ "scale_lr": false,
30
+ "seed": 555,
31
+ "keep_tags": 3,
32
+ "shuffle_tags": false,
33
+ "validation_config": "validation_rico_off.json",
34
+ "zero_frequency_noise_ratio": 0.02,
35
+ "min_snr_gamma": 5.0,
36
+ "enable_zero_terminal_snr": true,
37
+ "no_save_ckpt": true
38
+ }
rico_diffusion_v2_comp_list_item.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "attn_type": "sdp",
3
+ "batch_size": 11,
4
+ "ckpt_every_n_minutes": null,
5
+ "clip_grad_norm": null,
6
+ "clip_skip": 2,
7
+ "cond_dropout": 0.00,
8
+ "data_root": "/home/azhure/Tesis/List_Item",
9
+ "disable_amp": false,
10
+ "disable_textenc_training": false,
11
+ "flip_p": 0.0,
12
+ "gpuid": 0,
13
+ "gradient_checkpointing": true,
14
+ "grad_accum": 1,
15
+ "logdir": "logs",
16
+ "log_step": 25,
17
+ "loss_type": "huber",
18
+ "max_epochs": 2,
19
+ "optimizer_config": "optimizer_rico_comp.json",
20
+ "project_name": "rico_diffusion_v2_comp_list_item",
21
+ "resolution": 384,
22
+ "resume_ckpt": "findlast",
23
+ "sample_prompts": "sample_prompts.txt",
24
+ "sample_steps": 999999999,
25
+ "save_ckpt_dir": "model_output",
26
+ "save_ckpts_from_n_epochs": 0,
27
+ "save_every_n_epochs": 10,
28
+ "save_optimizer": false,
29
+ "scale_lr": false,
30
+ "seed": 555,
31
+ "keep_tags": 3,
32
+ "shuffle_tags": false,
33
+ "validation_config": "validation_rico_off.json",
34
+ "zero_frequency_noise_ratio": 0.02,
35
+ "min_snr_gamma": 5.0,
36
+ "enable_zero_terminal_snr": true,
37
+ "no_save_ckpt": true
38
+ }
rico_diffusion_v2_full.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "attn_type": "sdp",
3
+ "batch_size": 7,
4
+ "ckpt_every_n_minutes": null,
5
+ "clip_grad_norm": null,
6
+ "clip_skip": 2,
7
+ "cond_dropout": 0.10,
8
+ "data_root": "/home/azhure/Tesis/save_data",
9
+ "disable_amp": false,
10
+ "disable_textenc_training": false,
11
+ "flip_p": 0.0,
12
+ "gpuid": 0,
13
+ "gradient_checkpointing": true,
14
+ "grad_accum": 1,
15
+ "logdir": "logs",
16
+ "log_step": 25,
17
+ "loss_type": "huber",
18
+ "max_epochs": 5,
19
+ "optimizer_config": "optimizer_rico.json",
20
+ "project_name": "rico_diffusion_v2_full",
21
+ "resolution": 512,
22
+ "resume_ckpt": "findlast",
23
+ "sample_prompts": "sample_prompts_rico.json",
24
+ "sample_steps": 15000,
25
+ "save_ckpt_dir": "model_output",
26
+ "save_ckpts_from_n_epochs": 0,
27
+ "save_every_n_epochs": 5,
28
+ "save_optimizer": false,
29
+ "scale_lr": false,
30
+ "seed": 555,
31
+ "keep_tags": 3,
32
+ "shuffle_tags": false,
33
+ "validation_config": "validation_rico.json",
34
+ "zero_frequency_noise_ratio": 0.02,
35
+ "min_snr_gamma": 5.0,
36
+ "enable_zero_terminal_snr": true
37
+ }
sample_prompts_rico.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "batch_size": 1,
3
+ "seed": 555,
4
+ "cfgs": [7],
5
+ "scheduler": "ddim",
6
+ "num_inference_steps": 20,
7
+ "show_progress_bars": true,
8
+ "generate_pretrain_samples": true,
9
+ "samples": [
10
+ {
11
+ "prompt": "red Toolbar Upper Top containing Text Left Login inside and white Input Upper Top and white Input Lower Top and red Button login Upper Middle, Android UI, Medical, white background",
12
+ "size": [384, 640]
13
+ },
14
+ {
15
+ "prompt": "blue Toolbar Upper Top containing Text League Left inside and blue Multi Tab Top containing Text Button News Left Side Left and Text Button Schedule Right Side Left inside and white Card Middle containing Image soccer Cover Top and Text Soccer Match Bottom Left inside and white Card Bottom containing Image basketball Cover Top and Text Basketball Match Bottom Left inside, Android UI, Sports, white background",
16
+ "size": [384, 640]
17
+ }
18
+ ]
19
+ }
validation_rico.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "documentation": {
3
+ "validate_training": "If true, validate the training using a separate set of image/caption pairs, and log the results as `loss/val`. The curve will trend downwards as the model trains, then flatten and start to trend upwards as effective training finishes and the model begins to overfit the training data. Very useful for preventing overfitting, for checking if your learning rate is too low or too high, and for deciding when to stop training.",
4
+ "val_split_mode": "Either 'automatic' or 'manual', ignored if validate_training is false. 'automatic' val_split_mode picks a random subset of the training set (the number of items is controlled by auto_split_proportion) and removes them from training to use as a validation set. 'manual' val_split_mode lets you provide your own folder of validation items (images+captions), specified using 'manual_data_root'.",
5
+ "auto_split_proportion": "For 'automatic' val_split_mode, how much of the train dataset that should be removed to use for validation. Typical values are 0.15-0.2 (15-20% of the total dataset). Higher is more accurate but slower.",
6
+ "manual_data_root": "For 'manual' val_split_mode, the path to a folder containing validation items.",
7
+ "extra_manual_datasets": "Dictionary of 'name':'path' pairs defining additional validation datasets to load and log. eg { 'santa_suit': '/path/to/captioned_santa_suit_images', 'flamingo_suit': '/path/to/flamingo_suit_images' }",
8
+ "stabilize_training_loss": "If true, stabilize the train loss curves for `loss/epoch` and `loss/log step` by re-calculating training loss with a fixed random seed, and log the results as `loss/train-stabilized`. This more clearly shows the training progress, but it is not enough alone to tell you if you're overfitting.",
9
+ "stabilize_split_proportion": "For stabilize_training_loss, the proportion of the train dataset to overlap for stabilizing the train loss graph. Typical values are 0.15-0.2 (15-20% of the total dataset). Higher is more accurate but slower.",
10
+ "every_n_epochs": "How often to run validation (1=every epoch, 2=every second epoch; 0.5=twice per epoch, 0.33=three times per epoch, etc.).",
11
+ "seed": "The seed to use when running validation and stabilization passes.",
12
+ "use_relative_loss": "logs val/loss as negative relative to first pre-train val/loss value"
13
+ },
14
+ "validate_training": true,
15
+ "val_split_mode": "automatic",
16
+ "auto_split_proportion": 0.05,
17
+ "manual_data_root": null,
18
+ "extra_manual_datasets" : {},
19
+ "stabilize_training_loss": true,
20
+ "stabilize_split_proportion": 0.05,
21
+ "every_n_epochs": 1,
22
+ "seed": 555,
23
+ "use_relative_loss": false
24
+ }
validation_rico_off.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "documentation": {
3
+ "validate_training": "If true, validate the training using a separate set of image/caption pairs, and log the results as `loss/val`. The curve will trend downwards as the model trains, then flatten and start to trend upwards as effective training finishes and the model begins to overfit the training data. Very useful for preventing overfitting, for checking if your learning rate is too low or too high, and for deciding when to stop training.",
4
+ "val_split_mode": "Either 'automatic' or 'manual', ignored if validate_training is false. 'automatic' val_split_mode picks a random subset of the training set (the number of items is controlled by auto_split_proportion) and removes them from training to use as a validation set. 'manual' val_split_mode lets you provide your own folder of validation items (images+captions), specified using 'manual_data_root'.",
5
+ "auto_split_proportion": "For 'automatic' val_split_mode, how much of the train dataset that should be removed to use for validation. Typical values are 0.15-0.2 (15-20% of the total dataset). Higher is more accurate but slower.",
6
+ "manual_data_root": "For 'manual' val_split_mode, the path to a folder containing validation items.",
7
+ "extra_manual_datasets": "Dictionary of 'name':'path' pairs defining additional validation datasets to load and log. eg { 'santa_suit': '/path/to/captioned_santa_suit_images', 'flamingo_suit': '/path/to/flamingo_suit_images' }",
8
+ "stabilize_training_loss": "If true, stabilize the train loss curves for `loss/epoch` and `loss/log step` by re-calculating training loss with a fixed random seed, and log the results as `loss/train-stabilized`. This more clearly shows the training progress, but it is not enough alone to tell you if you're overfitting.",
9
+ "stabilize_split_proportion": "For stabilize_training_loss, the proportion of the train dataset to overlap for stabilizing the train loss graph. Typical values are 0.15-0.2 (15-20% of the total dataset). Higher is more accurate but slower.",
10
+ "every_n_epochs": "How often to run validation (1=every epoch, 2=every second epoch; 0.5=twice per epoch, 0.33=three times per epoch, etc.).",
11
+ "seed": "The seed to use when running validation and stabilization passes.",
12
+ "use_relative_loss": "logs val/loss as negative relative to first pre-train val/loss value"
13
+ },
14
+ "validate_training": false,
15
+ "val_split_mode": "automatic",
16
+ "auto_split_proportion": 0.05,
17
+ "manual_data_root": null,
18
+ "extra_manual_datasets" : {},
19
+ "stabilize_training_loss": false,
20
+ "stabilize_split_proportion": 0.05,
21
+ "every_n_epochs": 1,
22
+ "seed": 555,
23
+ "use_relative_loss": false
24
+ }