ehvris commited on
Commit
87b3abc
·
1 Parent(s): c9f60d2

Upload 160 files

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. train_shaco/log/20230820193555/network_train/events.out.tfevents.1692585442.DESKTOP-DK3AJDA.11420.0 +3 -0
  2. train_shaco/log/shaco_20230819-042813.json +109 -0
  3. train_shaco/log/shaco_20230819-043113.json +109 -0
  4. train_shaco/log/shaco_20230819-043426.json +109 -0
  5. train_shaco/model/FileName_20230820-183451.json +109 -0
  6. train_shaco/model/sample/prompt.txt +1 -0
  7. train_shaco/model/shaco_20230819-044008.json +109 -0
  8. train_shaco/model/shaco_20230819-044133.json +109 -0
  9. train_shaco/model/shaco_20230819-044207.json +109 -0
  10. train_shaco/model/shaco_20230819-044603.json +109 -0
  11. train_shaco/model/shaco_20230819-044629.json +109 -0
  12. train_shaco/model/shaco_20230819-044800.json +109 -0
  13. train_shaco/model/shaco_20230819-044821.json +109 -0
  14. train_shaco/model/shaco_20230819-044836.json +109 -0
  15. train_shaco/model/shaco_20230820-184523.json +109 -0
  16. train_shaco/regularization/1_person/04771-3283295968-person.png +3 -0
  17. train_shaco/regularization/1_person/04771-3283295968-person.txt +1 -0
  18. train_shaco/regularization/1_person/04772-3283295969-person.png +3 -0
  19. train_shaco/regularization/1_person/04772-3283295969-person.txt +1 -0
  20. train_shaco/regularization/1_person/04773-3283295970-person.png +3 -0
  21. train_shaco/regularization/1_person/04773-3283295970-person.txt +1 -0
  22. train_shaco/regularization/1_person/04774-2844224771-person.png +3 -0
  23. train_shaco/regularization/1_person/04774-2844224771-person.txt +1 -0
  24. train_shaco/regularization/1_person/04775-2844224772-person.png +3 -0
  25. train_shaco/regularization/1_person/04775-2844224772-person.txt +1 -0
  26. train_shaco/regularization/1_person/04776-2844224773-person.png +3 -0
  27. train_shaco/regularization/1_person/04776-2844224773-person.txt +1 -0
  28. train_shaco/regularization/1_person/04777-2844224774-person.png +3 -0
  29. train_shaco/regularization/1_person/04777-2844224774-person.txt +1 -0
  30. train_shaco/regularization/1_person/04778-2844224775-person.png +3 -0
  31. train_shaco/regularization/1_person/04778-2844224775-person.txt +1 -0
  32. train_shaco/regularization/1_person/04779-774901008-person.png +3 -0
  33. train_shaco/regularization/1_person/04779-774901008-person.txt +1 -0
  34. train_shaco/regularization/1_person/04780-774901009-person.png +3 -0
  35. train_shaco/regularization/1_person/04780-774901009-person.txt +1 -0
  36. train_shaco/regularization/1_person/04781-774901010-person.png +3 -0
  37. train_shaco/regularization/1_person/04781-774901010-person.txt +1 -0
  38. train_shaco/regularization/1_person/04782-774901011-person.png +3 -0
  39. train_shaco/regularization/1_person/04782-774901011-person.txt +1 -0
  40. train_shaco/regularization/1_person/04783-774901012-person.png +3 -0
  41. train_shaco/regularization/1_person/04783-774901012-person.txt +1 -0
  42. train_shaco/regularization/1_person/04784-4134988531-person.png +3 -0
  43. train_shaco/regularization/1_person/04784-4134988531-person.txt +1 -0
  44. train_shaco/regularization/1_person/04785-4134988532-person.png +3 -0
  45. train_shaco/regularization/1_person/04785-4134988532-person.txt +1 -0
  46. train_shaco/regularization/1_person/04786-4134988533-person.png +3 -0
  47. train_shaco/regularization/1_person/04786-4134988533-person.txt +1 -0
  48. train_shaco/regularization/1_person/04787-4134988534-person.png +3 -0
  49. train_shaco/regularization/1_person/04787-4134988534-person.txt +1 -0
  50. train_shaco/regularization/1_person/04788-4134988535-person.png +3 -0
train_shaco/log/20230820193555/network_train/events.out.tfevents.1692585442.DESKTOP-DK3AJDA.11420.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f274f72b7d615b5d295a2cf1f3a8e6ed0d164364b52510b52d847ee2cbddca65
3
+ size 321807
train_shaco/log/shaco_20230819-042813.json ADDED
@@ -0,0 +1,109 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "LoRA_type": "Standard",
3
+ "adaptive_noise_scale": 0,
4
+ "additional_parameters": "",
5
+ "block_alphas": "",
6
+ "block_dims": "",
7
+ "block_lr_zero_threshold": "",
8
+ "bucket_no_upscale": true,
9
+ "bucket_reso_steps": 1,
10
+ "cache_latents": true,
11
+ "cache_latents_to_disk": true,
12
+ "caption_dropout_every_n_epochs": 0.0,
13
+ "caption_dropout_rate": 0,
14
+ "caption_extension": ".none-use-foldername",
15
+ "clip_skip": "1",
16
+ "color_aug": false,
17
+ "conv_alpha": 64,
18
+ "conv_block_alphas": "",
19
+ "conv_block_dims": "",
20
+ "conv_dim": 64,
21
+ "decompose_both": false,
22
+ "dim_from_weights": false,
23
+ "down_lr_weight": "",
24
+ "enable_bucket": true,
25
+ "epoch": 4,
26
+ "factor": -1,
27
+ "flip_aug": false,
28
+ "full_bf16": false,
29
+ "full_fp16": false,
30
+ "gradient_accumulation_steps": 1.0,
31
+ "gradient_checkpointing": false,
32
+ "keep_tokens": "0",
33
+ "learning_rate": 4e-07,
34
+ "logging_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/model",
35
+ "lora_network_weights": "",
36
+ "lr_scheduler": "constant_with_warmup",
37
+ "lr_scheduler_num_cycles": "",
38
+ "lr_scheduler_power": "",
39
+ "lr_warmup": 8,
40
+ "max_bucket_reso": 2048,
41
+ "max_data_loader_n_workers": "0",
42
+ "max_resolution": "512,512",
43
+ "max_timestep": 1000,
44
+ "max_token_length": "75",
45
+ "max_train_epochs": "",
46
+ "mem_eff_attn": false,
47
+ "mid_lr_weight": "",
48
+ "min_bucket_reso": 256,
49
+ "min_snr_gamma": 10,
50
+ "min_timestep": 0,
51
+ "mixed_precision": "bf16",
52
+ "model_list": "custom",
53
+ "module_dropout": 0,
54
+ "multires_noise_discount": 0.2,
55
+ "multires_noise_iterations": 8,
56
+ "network_alpha": 64,
57
+ "network_dim": 64,
58
+ "network_dropout": 0,
59
+ "no_token_padding": false,
60
+ "noise_offset": 0.0357,
61
+ "noise_offset_type": "Original",
62
+ "num_cpu_threads_per_process": 2,
63
+ "optimizer": "Adafactor",
64
+ "optimizer_args": "scale_parameter=False relative_step=False warmup_init=False",
65
+ "output_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/log",
66
+ "output_name": "shaco",
67
+ "persistent_data_loader_workers": false,
68
+ "pretrained_model_name_or_path": "D:/StableDiffusion/stable-diffusion-webui-1.3.0/models/Stable-diffusion/Animay/mixProV3_v3.safetensors",
69
+ "prior_loss_weight": 1.0,
70
+ "random_crop": false,
71
+ "rank_dropout": 0,
72
+ "reg_data_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco",
73
+ "resume": "",
74
+ "sample_every_n_epochs": 0,
75
+ "sample_every_n_steps": 0,
76
+ "sample_prompts": "",
77
+ "sample_sampler": "euler_a",
78
+ "save_every_n_epochs": 1,
79
+ "save_every_n_steps": 0,
80
+ "save_last_n_steps": 0,
81
+ "save_last_n_steps_state": 0,
82
+ "save_model_as": "safetensors",
83
+ "save_precision": "fp16",
84
+ "save_state": false,
85
+ "scale_v_pred_loss_like_noise_pred": false,
86
+ "scale_weight_norms": 0,
87
+ "sdxl": true,
88
+ "sdxl_cache_text_encoder_outputs": false,
89
+ "sdxl_no_half_vae": true,
90
+ "seed": "",
91
+ "shuffle_caption": false,
92
+ "stop_text_encoder_training_pct": 0,
93
+ "text_encoder_lr": 0.0,
94
+ "train_batch_size": 1,
95
+ "train_data_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco",
96
+ "train_on_input": true,
97
+ "training_comment": "",
98
+ "unet_lr": 4e-07,
99
+ "unit": 1,
100
+ "up_lr_weight": "",
101
+ "use_cp": false,
102
+ "use_wandb": false,
103
+ "v2": false,
104
+ "v_parameterization": false,
105
+ "vae_batch_size": 0,
106
+ "wandb_api_key": "",
107
+ "weighted_captions": false,
108
+ "xformers": true
109
+ }
train_shaco/log/shaco_20230819-043113.json ADDED
@@ -0,0 +1,109 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "LoRA_type": "Standard",
3
+ "adaptive_noise_scale": 0,
4
+ "additional_parameters": "",
5
+ "block_alphas": "",
6
+ "block_dims": "",
7
+ "block_lr_zero_threshold": "",
8
+ "bucket_no_upscale": true,
9
+ "bucket_reso_steps": 1,
10
+ "cache_latents": true,
11
+ "cache_latents_to_disk": true,
12
+ "caption_dropout_every_n_epochs": 0.0,
13
+ "caption_dropout_rate": 0,
14
+ "caption_extension": ".none-use-foldername",
15
+ "clip_skip": "1",
16
+ "color_aug": false,
17
+ "conv_alpha": 64,
18
+ "conv_block_alphas": "",
19
+ "conv_block_dims": "",
20
+ "conv_dim": 64,
21
+ "decompose_both": false,
22
+ "dim_from_weights": false,
23
+ "down_lr_weight": "",
24
+ "enable_bucket": true,
25
+ "epoch": 4,
26
+ "factor": -1,
27
+ "flip_aug": false,
28
+ "full_bf16": false,
29
+ "full_fp16": false,
30
+ "gradient_accumulation_steps": 1.0,
31
+ "gradient_checkpointing": false,
32
+ "keep_tokens": "0",
33
+ "learning_rate": 4e-07,
34
+ "logging_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/model",
35
+ "lora_network_weights": "",
36
+ "lr_scheduler": "constant_with_warmup",
37
+ "lr_scheduler_num_cycles": "",
38
+ "lr_scheduler_power": "",
39
+ "lr_warmup": 8,
40
+ "max_bucket_reso": 2048,
41
+ "max_data_loader_n_workers": "0",
42
+ "max_resolution": "512,512",
43
+ "max_timestep": 1000,
44
+ "max_token_length": "75",
45
+ "max_train_epochs": "",
46
+ "mem_eff_attn": false,
47
+ "mid_lr_weight": "",
48
+ "min_bucket_reso": 256,
49
+ "min_snr_gamma": 10,
50
+ "min_timestep": 0,
51
+ "mixed_precision": "bf16",
52
+ "model_list": "custom",
53
+ "module_dropout": 0,
54
+ "multires_noise_discount": 0.2,
55
+ "multires_noise_iterations": 8,
56
+ "network_alpha": 64,
57
+ "network_dim": 64,
58
+ "network_dropout": 0,
59
+ "no_token_padding": false,
60
+ "noise_offset": 0.0357,
61
+ "noise_offset_type": "Original",
62
+ "num_cpu_threads_per_process": 2,
63
+ "optimizer": "Adafactor",
64
+ "optimizer_args": "scale_parameter=False relative_step=False warmup_init=False",
65
+ "output_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/log",
66
+ "output_name": "shaco",
67
+ "persistent_data_loader_workers": false,
68
+ "pretrained_model_name_or_path": "D:/StableDiffusion/stable-diffusion-webui-1.3.0/models/Stable-diffusion/Animay/mixProV3_v3.safetensors",
69
+ "prior_loss_weight": 1.0,
70
+ "random_crop": false,
71
+ "rank_dropout": 0,
72
+ "reg_data_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/1_person",
73
+ "resume": "",
74
+ "sample_every_n_epochs": 0,
75
+ "sample_every_n_steps": 0,
76
+ "sample_prompts": "",
77
+ "sample_sampler": "euler_a",
78
+ "save_every_n_epochs": 1,
79
+ "save_every_n_steps": 0,
80
+ "save_last_n_steps": 0,
81
+ "save_last_n_steps_state": 0,
82
+ "save_model_as": "safetensors",
83
+ "save_precision": "fp16",
84
+ "save_state": false,
85
+ "scale_v_pred_loss_like_noise_pred": false,
86
+ "scale_weight_norms": 0,
87
+ "sdxl": true,
88
+ "sdxl_cache_text_encoder_outputs": false,
89
+ "sdxl_no_half_vae": true,
90
+ "seed": "",
91
+ "shuffle_caption": false,
92
+ "stop_text_encoder_training_pct": 0,
93
+ "text_encoder_lr": 0.0,
94
+ "train_batch_size": 1,
95
+ "train_data_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/20_shaco person",
96
+ "train_on_input": true,
97
+ "training_comment": "",
98
+ "unet_lr": 4e-07,
99
+ "unit": 1,
100
+ "up_lr_weight": "",
101
+ "use_cp": false,
102
+ "use_wandb": false,
103
+ "v2": false,
104
+ "v_parameterization": false,
105
+ "vae_batch_size": 0,
106
+ "wandb_api_key": "",
107
+ "weighted_captions": false,
108
+ "xformers": true
109
+ }
train_shaco/log/shaco_20230819-043426.json ADDED
@@ -0,0 +1,109 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "LoRA_type": "Standard",
3
+ "adaptive_noise_scale": 0,
4
+ "additional_parameters": "",
5
+ "block_alphas": "",
6
+ "block_dims": "",
7
+ "block_lr_zero_threshold": "",
8
+ "bucket_no_upscale": true,
9
+ "bucket_reso_steps": 1,
10
+ "cache_latents": true,
11
+ "cache_latents_to_disk": true,
12
+ "caption_dropout_every_n_epochs": 0.0,
13
+ "caption_dropout_rate": 0,
14
+ "caption_extension": ".none-use-foldername",
15
+ "clip_skip": "1",
16
+ "color_aug": false,
17
+ "conv_alpha": 64,
18
+ "conv_block_alphas": "",
19
+ "conv_block_dims": "",
20
+ "conv_dim": 64,
21
+ "decompose_both": false,
22
+ "dim_from_weights": false,
23
+ "down_lr_weight": "",
24
+ "enable_bucket": true,
25
+ "epoch": 4,
26
+ "factor": -1,
27
+ "flip_aug": false,
28
+ "full_bf16": false,
29
+ "full_fp16": false,
30
+ "gradient_accumulation_steps": 1.0,
31
+ "gradient_checkpointing": false,
32
+ "keep_tokens": "0",
33
+ "learning_rate": 4e-07,
34
+ "logging_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/model",
35
+ "lora_network_weights": "",
36
+ "lr_scheduler": "constant_with_warmup",
37
+ "lr_scheduler_num_cycles": "",
38
+ "lr_scheduler_power": "",
39
+ "lr_warmup": 8,
40
+ "max_bucket_reso": 2048,
41
+ "max_data_loader_n_workers": "0",
42
+ "max_resolution": "512,512",
43
+ "max_timestep": 1000,
44
+ "max_token_length": "75",
45
+ "max_train_epochs": "",
46
+ "mem_eff_attn": false,
47
+ "mid_lr_weight": "",
48
+ "min_bucket_reso": 256,
49
+ "min_snr_gamma": 10,
50
+ "min_timestep": 0,
51
+ "mixed_precision": "bf16",
52
+ "model_list": "custom",
53
+ "module_dropout": 0,
54
+ "multires_noise_discount": 0.2,
55
+ "multires_noise_iterations": 8,
56
+ "network_alpha": 64,
57
+ "network_dim": 64,
58
+ "network_dropout": 0,
59
+ "no_token_padding": false,
60
+ "noise_offset": 0.0357,
61
+ "noise_offset_type": "Original",
62
+ "num_cpu_threads_per_process": 2,
63
+ "optimizer": "Adafactor",
64
+ "optimizer_args": "scale_parameter=False relative_step=False warmup_init=False",
65
+ "output_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/log",
66
+ "output_name": "shaco",
67
+ "persistent_data_loader_workers": false,
68
+ "pretrained_model_name_or_path": "D:/StableDiffusion/stable-diffusion-webui-1.3.0/models/Stable-diffusion/Animay/mixProV3_v3.safetensors",
69
+ "prior_loss_weight": 1.0,
70
+ "random_crop": false,
71
+ "rank_dropout": 0,
72
+ "reg_data_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/regularization",
73
+ "resume": "",
74
+ "sample_every_n_epochs": 0,
75
+ "sample_every_n_steps": 0,
76
+ "sample_prompts": "",
77
+ "sample_sampler": "euler_a",
78
+ "save_every_n_epochs": 1,
79
+ "save_every_n_steps": 0,
80
+ "save_last_n_steps": 0,
81
+ "save_last_n_steps_state": 0,
82
+ "save_model_as": "safetensors",
83
+ "save_precision": "fp16",
84
+ "save_state": false,
85
+ "scale_v_pred_loss_like_noise_pred": false,
86
+ "scale_weight_norms": 0,
87
+ "sdxl": true,
88
+ "sdxl_cache_text_encoder_outputs": false,
89
+ "sdxl_no_half_vae": true,
90
+ "seed": "",
91
+ "shuffle_caption": false,
92
+ "stop_text_encoder_training_pct": 0,
93
+ "text_encoder_lr": 0.0,
94
+ "train_batch_size": 1,
95
+ "train_data_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/train",
96
+ "train_on_input": true,
97
+ "training_comment": "",
98
+ "unet_lr": 4e-07,
99
+ "unit": 1,
100
+ "up_lr_weight": "",
101
+ "use_cp": false,
102
+ "use_wandb": false,
103
+ "v2": false,
104
+ "v_parameterization": false,
105
+ "vae_batch_size": 0,
106
+ "wandb_api_key": "",
107
+ "weighted_captions": false,
108
+ "xformers": true
109
+ }
train_shaco/model/FileName_20230820-183451.json ADDED
@@ -0,0 +1,109 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "LoRA_type": "Standard",
3
+ "adaptive_noise_scale": 0,
4
+ "additional_parameters": "",
5
+ "block_alphas": "",
6
+ "block_dims": "",
7
+ "block_lr_zero_threshold": "",
8
+ "bucket_no_upscale": true,
9
+ "bucket_reso_steps": 64,
10
+ "cache_latents": true,
11
+ "cache_latents_to_disk": false,
12
+ "caption_dropout_every_n_epochs": 0.0,
13
+ "caption_dropout_rate": 0,
14
+ "caption_extension": ".txt",
15
+ "clip_skip": 2,
16
+ "color_aug": false,
17
+ "conv_alpha": 1,
18
+ "conv_block_alphas": "",
19
+ "conv_block_dims": "",
20
+ "conv_dim": 1,
21
+ "decompose_both": false,
22
+ "dim_from_weights": false,
23
+ "down_lr_weight": "",
24
+ "enable_bucket": true,
25
+ "epoch": 1,
26
+ "factor": -1,
27
+ "flip_aug": false,
28
+ "full_bf16": false,
29
+ "full_fp16": false,
30
+ "gradient_accumulation_steps": 1.0,
31
+ "gradient_checkpointing": false,
32
+ "keep_tokens": "0",
33
+ "learning_rate": 0.0001,
34
+ "logging_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/log",
35
+ "lora_network_weights": "",
36
+ "lr_scheduler": "constant",
37
+ "lr_scheduler_num_cycles": "",
38
+ "lr_scheduler_power": "",
39
+ "lr_warmup": 0,
40
+ "max_bucket_reso": 2048,
41
+ "max_data_loader_n_workers": "1",
42
+ "max_resolution": "512,512",
43
+ "max_timestep": 1000,
44
+ "max_token_length": "75",
45
+ "max_train_epochs": "",
46
+ "mem_eff_attn": false,
47
+ "mid_lr_weight": "",
48
+ "min_bucket_reso": 256,
49
+ "min_snr_gamma": 0,
50
+ "min_timestep": 0,
51
+ "mixed_precision": "bf16",
52
+ "model_list": "custom",
53
+ "module_dropout": 0,
54
+ "multires_noise_discount": 0,
55
+ "multires_noise_iterations": 0,
56
+ "network_alpha": 128,
57
+ "network_dim": 128,
58
+ "network_dropout": 0,
59
+ "no_token_padding": false,
60
+ "noise_offset": 0,
61
+ "noise_offset_type": "Original",
62
+ "num_cpu_threads_per_process": 2,
63
+ "optimizer": "AdamW8bit",
64
+ "optimizer_args": "",
65
+ "output_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/model",
66
+ "output_name": "FileName",
67
+ "persistent_data_loader_workers": false,
68
+ "pretrained_model_name_or_path": "D:/StableDiffusion/stable-diffusion-webui-1.3.0/models/Stable-diffusion/Animay/mixProV3_v3.safetensors",
69
+ "prior_loss_weight": 1.0,
70
+ "random_crop": false,
71
+ "rank_dropout": 0,
72
+ "reg_data_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/regularization",
73
+ "resume": "",
74
+ "sample_every_n_epochs": 0,
75
+ "sample_every_n_steps": 0,
76
+ "sample_prompts": "",
77
+ "sample_sampler": "euler_a",
78
+ "save_every_n_epochs": 1,
79
+ "save_every_n_steps": 0,
80
+ "save_last_n_steps": 0,
81
+ "save_last_n_steps_state": 0,
82
+ "save_model_as": "safetensors",
83
+ "save_precision": "bf16",
84
+ "save_state": false,
85
+ "scale_v_pred_loss_like_noise_pred": false,
86
+ "scale_weight_norms": 0,
87
+ "sdxl": false,
88
+ "sdxl_cache_text_encoder_outputs": false,
89
+ "sdxl_no_half_vae": true,
90
+ "seed": "1234",
91
+ "shuffle_caption": false,
92
+ "stop_text_encoder_training_pct": 0,
93
+ "text_encoder_lr": 5e-05,
94
+ "train_batch_size": 2,
95
+ "train_data_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/train",
96
+ "train_on_input": true,
97
+ "training_comment": "",
98
+ "unet_lr": 0.0001,
99
+ "unit": 1,
100
+ "up_lr_weight": "",
101
+ "use_cp": false,
102
+ "use_wandb": "",
103
+ "v2": false,
104
+ "v_parameterization": false,
105
+ "vae_batch_size": 0,
106
+ "wandb_api_key": "False",
107
+ "weighted_captions": false,
108
+ "xformers": true
109
+ }
train_shaco/model/sample/prompt.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ potioncraft, woman, wearing white shirt, wearing necklace, medium breasts
train_shaco/model/shaco_20230819-044008.json ADDED
@@ -0,0 +1,109 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "LoRA_type": "Standard",
3
+ "adaptive_noise_scale": 0,
4
+ "additional_parameters": "",
5
+ "block_alphas": "",
6
+ "block_dims": "",
7
+ "block_lr_zero_threshold": "",
8
+ "bucket_no_upscale": true,
9
+ "bucket_reso_steps": 64,
10
+ "cache_latents": true,
11
+ "cache_latents_to_disk": false,
12
+ "caption_dropout_every_n_epochs": 0.0,
13
+ "caption_dropout_rate": 0,
14
+ "caption_extension": ".txt",
15
+ "clip_skip": 2,
16
+ "color_aug": false,
17
+ "conv_alpha": 1,
18
+ "conv_block_alphas": "",
19
+ "conv_block_dims": "",
20
+ "conv_dim": 1,
21
+ "decompose_both": false,
22
+ "dim_from_weights": false,
23
+ "down_lr_weight": "",
24
+ "enable_bucket": true,
25
+ "epoch": 1,
26
+ "factor": -1,
27
+ "flip_aug": false,
28
+ "full_bf16": false,
29
+ "full_fp16": false,
30
+ "gradient_accumulation_steps": 1.0,
31
+ "gradient_checkpointing": false,
32
+ "keep_tokens": "0",
33
+ "learning_rate": 0.0001,
34
+ "logging_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/log",
35
+ "lora_network_weights": "",
36
+ "lr_scheduler": "constant",
37
+ "lr_scheduler_num_cycles": "",
38
+ "lr_scheduler_power": "",
39
+ "lr_warmup": 0,
40
+ "max_bucket_reso": 2048,
41
+ "max_data_loader_n_workers": "1",
42
+ "max_resolution": "512,512",
43
+ "max_timestep": 1000,
44
+ "max_token_length": "75",
45
+ "max_train_epochs": "",
46
+ "mem_eff_attn": false,
47
+ "mid_lr_weight": "",
48
+ "min_bucket_reso": 64,
49
+ "min_snr_gamma": 0,
50
+ "min_timestep": 0,
51
+ "mixed_precision": "bf16",
52
+ "model_list": "custom",
53
+ "module_dropout": 0,
54
+ "multires_noise_discount": 0,
55
+ "multires_noise_iterations": 0,
56
+ "network_alpha": 128,
57
+ "network_dim": 128,
58
+ "network_dropout": 0,
59
+ "no_token_padding": false,
60
+ "noise_offset": 0,
61
+ "noise_offset_type": "Original",
62
+ "num_cpu_threads_per_process": 2,
63
+ "optimizer": "AdamW8bit",
64
+ "optimizer_args": "",
65
+ "output_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/model",
66
+ "output_name": "shaco",
67
+ "persistent_data_loader_workers": false,
68
+ "pretrained_model_name_or_path": "D:/StableDiffusion/stable-diffusion-webui-1.3.0/models/Stable-diffusion/Animay/mixProV3_v3.safetensors",
69
+ "prior_loss_weight": 1.0,
70
+ "random_crop": false,
71
+ "rank_dropout": 0,
72
+ "reg_data_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/regularization",
73
+ "resume": "",
74
+ "sample_every_n_epochs": 0,
75
+ "sample_every_n_steps": 0,
76
+ "sample_prompts": "",
77
+ "sample_sampler": "euler_a",
78
+ "save_every_n_epochs": 1,
79
+ "save_every_n_steps": 0,
80
+ "save_last_n_steps": 0,
81
+ "save_last_n_steps_state": 0,
82
+ "save_model_as": "safetensors",
83
+ "save_precision": "bf16",
84
+ "save_state": false,
85
+ "scale_v_pred_loss_like_noise_pred": false,
86
+ "scale_weight_norms": 0,
87
+ "sdxl": true,
88
+ "sdxl_cache_text_encoder_outputs": true,
89
+ "sdxl_no_half_vae": true,
90
+ "seed": "1234",
91
+ "shuffle_caption": false,
92
+ "stop_text_encoder_training_pct": 0,
93
+ "text_encoder_lr": 5e-05,
94
+ "train_batch_size": 2,
95
+ "train_data_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/train",
96
+ "train_on_input": false,
97
+ "training_comment": "",
98
+ "unet_lr": 0.0001,
99
+ "unit": 1,
100
+ "up_lr_weight": "",
101
+ "use_cp": false,
102
+ "use_wandb": "",
103
+ "v2": false,
104
+ "v_parameterization": false,
105
+ "vae_batch_size": 0,
106
+ "wandb_api_key": "False",
107
+ "weighted_captions": false,
108
+ "xformers": true
109
+ }
train_shaco/model/shaco_20230819-044133.json ADDED
@@ -0,0 +1,109 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "LoRA_type": "Standard",
3
+ "adaptive_noise_scale": 0,
4
+ "additional_parameters": "",
5
+ "block_alphas": "",
6
+ "block_dims": "",
7
+ "block_lr_zero_threshold": "",
8
+ "bucket_no_upscale": true,
9
+ "bucket_reso_steps": 64,
10
+ "cache_latents": true,
11
+ "cache_latents_to_disk": false,
12
+ "caption_dropout_every_n_epochs": 0.0,
13
+ "caption_dropout_rate": 0,
14
+ "caption_extension": ".txt",
15
+ "clip_skip": 2,
16
+ "color_aug": false,
17
+ "conv_alpha": 1,
18
+ "conv_block_alphas": "",
19
+ "conv_block_dims": "",
20
+ "conv_dim": 1,
21
+ "decompose_both": false,
22
+ "dim_from_weights": false,
23
+ "down_lr_weight": "",
24
+ "enable_bucket": true,
25
+ "epoch": 1,
26
+ "factor": -1,
27
+ "flip_aug": false,
28
+ "full_bf16": false,
29
+ "full_fp16": false,
30
+ "gradient_accumulation_steps": 1.0,
31
+ "gradient_checkpointing": false,
32
+ "keep_tokens": "0",
33
+ "learning_rate": 0.0001,
34
+ "logging_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/log",
35
+ "lora_network_weights": "",
36
+ "lr_scheduler": "constant",
37
+ "lr_scheduler_num_cycles": "",
38
+ "lr_scheduler_power": "",
39
+ "lr_warmup": 0,
40
+ "max_bucket_reso": 2048,
41
+ "max_data_loader_n_workers": "1",
42
+ "max_resolution": "512,512",
43
+ "max_timestep": 1000,
44
+ "max_token_length": "75",
45
+ "max_train_epochs": "",
46
+ "mem_eff_attn": false,
47
+ "mid_lr_weight": "",
48
+ "min_bucket_reso": 64,
49
+ "min_snr_gamma": 0,
50
+ "min_timestep": 0,
51
+ "mixed_precision": "bf16",
52
+ "model_list": "custom",
53
+ "module_dropout": 0,
54
+ "multires_noise_discount": 0,
55
+ "multires_noise_iterations": 0,
56
+ "network_alpha": 128,
57
+ "network_dim": 128,
58
+ "network_dropout": 0,
59
+ "no_token_padding": false,
60
+ "noise_offset": 0,
61
+ "noise_offset_type": "Original",
62
+ "num_cpu_threads_per_process": 2,
63
+ "optimizer": "AdamW8bit",
64
+ "optimizer_args": "",
65
+ "output_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/model",
66
+ "output_name": "shaco",
67
+ "persistent_data_loader_workers": false,
68
+ "pretrained_model_name_or_path": "D:/StableDiffusion/stable-diffusion-webui-1.3.0/models/Stable-diffusion/Animay/mixProV3_v3.safetensors",
69
+ "prior_loss_weight": 1.0,
70
+ "random_crop": false,
71
+ "rank_dropout": 0,
72
+ "reg_data_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/regularization",
73
+ "resume": "",
74
+ "sample_every_n_epochs": 1,
75
+ "sample_every_n_steps": 0,
76
+ "sample_prompts": "",
77
+ "sample_sampler": "euler_a",
78
+ "save_every_n_epochs": 1,
79
+ "save_every_n_steps": 0,
80
+ "save_last_n_steps": 0,
81
+ "save_last_n_steps_state": 0,
82
+ "save_model_as": "safetensors",
83
+ "save_precision": "bf16",
84
+ "save_state": false,
85
+ "scale_v_pred_loss_like_noise_pred": false,
86
+ "scale_weight_norms": 0,
87
+ "sdxl": true,
88
+ "sdxl_cache_text_encoder_outputs": true,
89
+ "sdxl_no_half_vae": true,
90
+ "seed": "1234",
91
+ "shuffle_caption": false,
92
+ "stop_text_encoder_training_pct": 0,
93
+ "text_encoder_lr": 5e-05,
94
+ "train_batch_size": 2,
95
+ "train_data_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/train",
96
+ "train_on_input": false,
97
+ "training_comment": "",
98
+ "unet_lr": 0.0001,
99
+ "unit": 1,
100
+ "up_lr_weight": "",
101
+ "use_cp": false,
102
+ "use_wandb": "",
103
+ "v2": false,
104
+ "v_parameterization": false,
105
+ "vae_batch_size": 0,
106
+ "wandb_api_key": "False",
107
+ "weighted_captions": false,
108
+ "xformers": true
109
+ }
train_shaco/model/shaco_20230819-044207.json ADDED
@@ -0,0 +1,109 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "LoRA_type": "Standard",
3
+ "adaptive_noise_scale": 0,
4
+ "additional_parameters": "",
5
+ "block_alphas": "",
6
+ "block_dims": "",
7
+ "block_lr_zero_threshold": "",
8
+ "bucket_no_upscale": true,
9
+ "bucket_reso_steps": 64,
10
+ "cache_latents": true,
11
+ "cache_latents_to_disk": false,
12
+ "caption_dropout_every_n_epochs": 0.0,
13
+ "caption_dropout_rate": 0,
14
+ "caption_extension": ".txt",
15
+ "clip_skip": 2,
16
+ "color_aug": false,
17
+ "conv_alpha": 1,
18
+ "conv_block_alphas": "",
19
+ "conv_block_dims": "",
20
+ "conv_dim": 1,
21
+ "decompose_both": false,
22
+ "dim_from_weights": false,
23
+ "down_lr_weight": "",
24
+ "enable_bucket": true,
25
+ "epoch": 1,
26
+ "factor": -1,
27
+ "flip_aug": false,
28
+ "full_bf16": true,
29
+ "full_fp16": false,
30
+ "gradient_accumulation_steps": 1.0,
31
+ "gradient_checkpointing": false,
32
+ "keep_tokens": "0",
33
+ "learning_rate": 0.0001,
34
+ "logging_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/log",
35
+ "lora_network_weights": "",
36
+ "lr_scheduler": "constant",
37
+ "lr_scheduler_num_cycles": "",
38
+ "lr_scheduler_power": "",
39
+ "lr_warmup": 0,
40
+ "max_bucket_reso": 2048,
41
+ "max_data_loader_n_workers": "1",
42
+ "max_resolution": "512,512",
43
+ "max_timestep": 1000,
44
+ "max_token_length": "75",
45
+ "max_train_epochs": "",
46
+ "mem_eff_attn": false,
47
+ "mid_lr_weight": "",
48
+ "min_bucket_reso": 64,
49
+ "min_snr_gamma": 0,
50
+ "min_timestep": 0,
51
+ "mixed_precision": "bf16",
52
+ "model_list": "custom",
53
+ "module_dropout": 0,
54
+ "multires_noise_discount": 0,
55
+ "multires_noise_iterations": 0,
56
+ "network_alpha": 128,
57
+ "network_dim": 128,
58
+ "network_dropout": 0,
59
+ "no_token_padding": false,
60
+ "noise_offset": 0,
61
+ "noise_offset_type": "Original",
62
+ "num_cpu_threads_per_process": 2,
63
+ "optimizer": "AdamW8bit",
64
+ "optimizer_args": "",
65
+ "output_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/model",
66
+ "output_name": "shaco",
67
+ "persistent_data_loader_workers": false,
68
+ "pretrained_model_name_or_path": "D:/StableDiffusion/stable-diffusion-webui-1.3.0/models/Stable-diffusion/Animay/mixProV3_v3.safetensors",
69
+ "prior_loss_weight": 1.0,
70
+ "random_crop": false,
71
+ "rank_dropout": 0,
72
+ "reg_data_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/regularization",
73
+ "resume": "",
74
+ "sample_every_n_epochs": 1,
75
+ "sample_every_n_steps": 0,
76
+ "sample_prompts": "",
77
+ "sample_sampler": "euler_a",
78
+ "save_every_n_epochs": 1,
79
+ "save_every_n_steps": 0,
80
+ "save_last_n_steps": 0,
81
+ "save_last_n_steps_state": 0,
82
+ "save_model_as": "safetensors",
83
+ "save_precision": "bf16",
84
+ "save_state": false,
85
+ "scale_v_pred_loss_like_noise_pred": false,
86
+ "scale_weight_norms": 0,
87
+ "sdxl": true,
88
+ "sdxl_cache_text_encoder_outputs": false,
89
+ "sdxl_no_half_vae": false,
90
+ "seed": "1234",
91
+ "shuffle_caption": false,
92
+ "stop_text_encoder_training_pct": 0,
93
+ "text_encoder_lr": 5e-05,
94
+ "train_batch_size": 2,
95
+ "train_data_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/train",
96
+ "train_on_input": false,
97
+ "training_comment": "",
98
+ "unet_lr": 0.0001,
99
+ "unit": 1,
100
+ "up_lr_weight": "",
101
+ "use_cp": false,
102
+ "use_wandb": "",
103
+ "v2": false,
104
+ "v_parameterization": false,
105
+ "vae_batch_size": 0,
106
+ "wandb_api_key": "False",
107
+ "weighted_captions": false,
108
+ "xformers": true
109
+ }
train_shaco/model/shaco_20230819-044603.json ADDED
@@ -0,0 +1,109 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "LoRA_type": "Standard",
3
+ "adaptive_noise_scale": 0,
4
+ "additional_parameters": "",
5
+ "block_alphas": "",
6
+ "block_dims": "",
7
+ "block_lr_zero_threshold": "",
8
+ "bucket_no_upscale": true,
9
+ "bucket_reso_steps": 64,
10
+ "cache_latents": true,
11
+ "cache_latents_to_disk": false,
12
+ "caption_dropout_every_n_epochs": 0.0,
13
+ "caption_dropout_rate": 0,
14
+ "caption_extension": ".txt",
15
+ "clip_skip": 2,
16
+ "color_aug": false,
17
+ "conv_alpha": 1,
18
+ "conv_block_alphas": "",
19
+ "conv_block_dims": "",
20
+ "conv_dim": 1,
21
+ "decompose_both": false,
22
+ "dim_from_weights": false,
23
+ "down_lr_weight": "",
24
+ "enable_bucket": true,
25
+ "epoch": 1,
26
+ "factor": -1,
27
+ "flip_aug": false,
28
+ "full_bf16": false,
29
+ "full_fp16": false,
30
+ "gradient_accumulation_steps": 1.0,
31
+ "gradient_checkpointing": false,
32
+ "keep_tokens": "0",
33
+ "learning_rate": 0.0001,
34
+ "logging_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/log",
35
+ "lora_network_weights": "",
36
+ "lr_scheduler": "constant",
37
+ "lr_scheduler_num_cycles": "",
38
+ "lr_scheduler_power": "",
39
+ "lr_warmup": 0,
40
+ "max_bucket_reso": 2048,
41
+ "max_data_loader_n_workers": "1",
42
+ "max_resolution": "512,512",
43
+ "max_timestep": 1000,
44
+ "max_token_length": "75",
45
+ "max_train_epochs": "",
46
+ "mem_eff_attn": false,
47
+ "mid_lr_weight": "",
48
+ "min_bucket_reso": 64,
49
+ "min_snr_gamma": 0,
50
+ "min_timestep": 0,
51
+ "mixed_precision": "bf16",
52
+ "model_list": "custom",
53
+ "module_dropout": 0,
54
+ "multires_noise_discount": 0,
55
+ "multires_noise_iterations": 0,
56
+ "network_alpha": 128,
57
+ "network_dim": 128,
58
+ "network_dropout": 0,
59
+ "no_token_padding": false,
60
+ "noise_offset": 0,
61
+ "noise_offset_type": "Original",
62
+ "num_cpu_threads_per_process": 2,
63
+ "optimizer": "AdamW8bit",
64
+ "optimizer_args": "",
65
+ "output_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/model",
66
+ "output_name": "shaco",
67
+ "persistent_data_loader_workers": false,
68
+ "pretrained_model_name_or_path": "D:/StableDiffusion/stable-diffusion-webui-1.3.0/models/Stable-diffusion/Animay/mixProV3_v3.safetensors",
69
+ "prior_loss_weight": 1.0,
70
+ "random_crop": false,
71
+ "rank_dropout": 0,
72
+ "reg_data_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/regularization",
73
+ "resume": "",
74
+ "sample_every_n_epochs": 1,
75
+ "sample_every_n_steps": 0,
76
+ "sample_prompts": "",
77
+ "sample_sampler": "euler_a",
78
+ "save_every_n_epochs": 1,
79
+ "save_every_n_steps": 0,
80
+ "save_last_n_steps": 0,
81
+ "save_last_n_steps_state": 0,
82
+ "save_model_as": "safetensors",
83
+ "save_precision": "bf16",
84
+ "save_state": false,
85
+ "scale_v_pred_loss_like_noise_pred": false,
86
+ "scale_weight_norms": 0,
87
+ "sdxl": true,
88
+ "sdxl_cache_text_encoder_outputs": true,
89
+ "sdxl_no_half_vae": true,
90
+ "seed": "1234",
91
+ "shuffle_caption": false,
92
+ "stop_text_encoder_training_pct": 0,
93
+ "text_encoder_lr": 5e-05,
94
+ "train_batch_size": 2,
95
+ "train_data_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/train",
96
+ "train_on_input": false,
97
+ "training_comment": "",
98
+ "unet_lr": 0.0001,
99
+ "unit": 1,
100
+ "up_lr_weight": "",
101
+ "use_cp": false,
102
+ "use_wandb": "",
103
+ "v2": false,
104
+ "v_parameterization": false,
105
+ "vae_batch_size": 0,
106
+ "wandb_api_key": "False",
107
+ "weighted_captions": false,
108
+ "xformers": true
109
+ }
train_shaco/model/shaco_20230819-044629.json ADDED
@@ -0,0 +1,109 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "LoRA_type": "Standard",
3
+ "adaptive_noise_scale": 0,
4
+ "additional_parameters": "",
5
+ "block_alphas": "",
6
+ "block_dims": "",
7
+ "block_lr_zero_threshold": "",
8
+ "bucket_no_upscale": true,
9
+ "bucket_reso_steps": 64,
10
+ "cache_latents": true,
11
+ "cache_latents_to_disk": false,
12
+ "caption_dropout_every_n_epochs": 0.0,
13
+ "caption_dropout_rate": 0,
14
+ "caption_extension": ".txt",
15
+ "clip_skip": 2,
16
+ "color_aug": false,
17
+ "conv_alpha": 1,
18
+ "conv_block_alphas": "",
19
+ "conv_block_dims": "",
20
+ "conv_dim": 1,
21
+ "decompose_both": false,
22
+ "dim_from_weights": false,
23
+ "down_lr_weight": "",
24
+ "enable_bucket": true,
25
+ "epoch": 1,
26
+ "factor": -1,
27
+ "flip_aug": false,
28
+ "full_bf16": false,
29
+ "full_fp16": false,
30
+ "gradient_accumulation_steps": 1.0,
31
+ "gradient_checkpointing": false,
32
+ "keep_tokens": "0",
33
+ "learning_rate": 0.0001,
34
+ "logging_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/log",
35
+ "lora_network_weights": "",
36
+ "lr_scheduler": "constant",
37
+ "lr_scheduler_num_cycles": "",
38
+ "lr_scheduler_power": "",
39
+ "lr_warmup": 0,
40
+ "max_bucket_reso": 2048,
41
+ "max_data_loader_n_workers": "1",
42
+ "max_resolution": "512,512",
43
+ "max_timestep": 1000,
44
+ "max_token_length": "75",
45
+ "max_train_epochs": "",
46
+ "mem_eff_attn": false,
47
+ "mid_lr_weight": "",
48
+ "min_bucket_reso": 64,
49
+ "min_snr_gamma": 0,
50
+ "min_timestep": 0,
51
+ "mixed_precision": "bf16",
52
+ "model_list": "custom",
53
+ "module_dropout": 0,
54
+ "multires_noise_discount": 0,
55
+ "multires_noise_iterations": 0,
56
+ "network_alpha": 128,
57
+ "network_dim": 128,
58
+ "network_dropout": 0,
59
+ "no_token_padding": false,
60
+ "noise_offset": 0,
61
+ "noise_offset_type": "Original",
62
+ "num_cpu_threads_per_process": 2,
63
+ "optimizer": "AdamW8bit",
64
+ "optimizer_args": "",
65
+ "output_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/model",
66
+ "output_name": "shaco",
67
+ "persistent_data_loader_workers": false,
68
+ "pretrained_model_name_or_path": "D:/StableDiffusion/stable-diffusion-webui-1.3.0/models/Stable-diffusion/Animay/mixProV3_v3.safetensors",
69
+ "prior_loss_weight": 1.0,
70
+ "random_crop": false,
71
+ "rank_dropout": 0,
72
+ "reg_data_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/regularization",
73
+ "resume": "",
74
+ "sample_every_n_epochs": 1,
75
+ "sample_every_n_steps": 0,
76
+ "sample_prompts": "",
77
+ "sample_sampler": "euler_a",
78
+ "save_every_n_epochs": 1,
79
+ "save_every_n_steps": 0,
80
+ "save_last_n_steps": 0,
81
+ "save_last_n_steps_state": 0,
82
+ "save_model_as": "safetensors",
83
+ "save_precision": "bf16",
84
+ "save_state": false,
85
+ "scale_v_pred_loss_like_noise_pred": false,
86
+ "scale_weight_norms": 0,
87
+ "sdxl": true,
88
+ "sdxl_cache_text_encoder_outputs": false,
89
+ "sdxl_no_half_vae": false,
90
+ "seed": "1234",
91
+ "shuffle_caption": false,
92
+ "stop_text_encoder_training_pct": 0,
93
+ "text_encoder_lr": 5e-05,
94
+ "train_batch_size": 2,
95
+ "train_data_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/train",
96
+ "train_on_input": false,
97
+ "training_comment": "",
98
+ "unet_lr": 0.0001,
99
+ "unit": 1,
100
+ "up_lr_weight": "",
101
+ "use_cp": false,
102
+ "use_wandb": "",
103
+ "v2": false,
104
+ "v_parameterization": false,
105
+ "vae_batch_size": 0,
106
+ "wandb_api_key": "False",
107
+ "weighted_captions": false,
108
+ "xformers": true
109
+ }
train_shaco/model/shaco_20230819-044800.json ADDED
@@ -0,0 +1,109 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "LoRA_type": "Standard",
3
+ "adaptive_noise_scale": 0,
4
+ "additional_parameters": "",
5
+ "block_alphas": "",
6
+ "block_dims": "",
7
+ "block_lr_zero_threshold": "",
8
+ "bucket_no_upscale": true,
9
+ "bucket_reso_steps": 64,
10
+ "cache_latents": true,
11
+ "cache_latents_to_disk": false,
12
+ "caption_dropout_every_n_epochs": 0.0,
13
+ "caption_dropout_rate": 0,
14
+ "caption_extension": ".txt",
15
+ "clip_skip": 2,
16
+ "color_aug": false,
17
+ "conv_alpha": 1,
18
+ "conv_block_alphas": "",
19
+ "conv_block_dims": "",
20
+ "conv_dim": 1,
21
+ "decompose_both": false,
22
+ "dim_from_weights": false,
23
+ "down_lr_weight": "",
24
+ "enable_bucket": true,
25
+ "epoch": 1,
26
+ "factor": -1,
27
+ "flip_aug": false,
28
+ "full_bf16": false,
29
+ "full_fp16": false,
30
+ "gradient_accumulation_steps": 1.0,
31
+ "gradient_checkpointing": false,
32
+ "keep_tokens": "0",
33
+ "learning_rate": 0.0001,
34
+ "logging_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/log",
35
+ "lora_network_weights": "",
36
+ "lr_scheduler": "constant",
37
+ "lr_scheduler_num_cycles": "",
38
+ "lr_scheduler_power": "",
39
+ "lr_warmup": 0,
40
+ "max_bucket_reso": 2048,
41
+ "max_data_loader_n_workers": "1",
42
+ "max_resolution": "512,512",
43
+ "max_timestep": 1000,
44
+ "max_token_length": "75",
45
+ "max_train_epochs": "",
46
+ "mem_eff_attn": false,
47
+ "mid_lr_weight": "",
48
+ "min_bucket_reso": 64,
49
+ "min_snr_gamma": 0,
50
+ "min_timestep": 0,
51
+ "mixed_precision": "bf16",
52
+ "model_list": "custom",
53
+ "module_dropout": 0,
54
+ "multires_noise_discount": 0,
55
+ "multires_noise_iterations": 0,
56
+ "network_alpha": 128,
57
+ "network_dim": 128,
58
+ "network_dropout": 0,
59
+ "no_token_padding": false,
60
+ "noise_offset": 0,
61
+ "noise_offset_type": "Original",
62
+ "num_cpu_threads_per_process": 2,
63
+ "optimizer": "AdamW8bit",
64
+ "optimizer_args": "",
65
+ "output_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/model",
66
+ "output_name": "shaco",
67
+ "persistent_data_loader_workers": false,
68
+ "pretrained_model_name_or_path": "D:/StableDiffusion/stable-diffusion-webui-1.3.0/models/Stable-diffusion/Animay/mixProV3_v3.safetensors",
69
+ "prior_loss_weight": 1.0,
70
+ "random_crop": false,
71
+ "rank_dropout": 0,
72
+ "reg_data_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/regularization",
73
+ "resume": "",
74
+ "sample_every_n_epochs": 1,
75
+ "sample_every_n_steps": 0,
76
+ "sample_prompts": "",
77
+ "sample_sampler": "euler_a",
78
+ "save_every_n_epochs": 1,
79
+ "save_every_n_steps": 0,
80
+ "save_last_n_steps": 0,
81
+ "save_last_n_steps_state": 0,
82
+ "save_model_as": "safetensors",
83
+ "save_precision": "bf16",
84
+ "save_state": false,
85
+ "scale_v_pred_loss_like_noise_pred": false,
86
+ "scale_weight_norms": 0,
87
+ "sdxl": true,
88
+ "sdxl_cache_text_encoder_outputs": false,
89
+ "sdxl_no_half_vae": false,
90
+ "seed": "1234",
91
+ "shuffle_caption": false,
92
+ "stop_text_encoder_training_pct": 0,
93
+ "text_encoder_lr": 5e-05,
94
+ "train_batch_size": 2,
95
+ "train_data_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/train",
96
+ "train_on_input": false,
97
+ "training_comment": "",
98
+ "unet_lr": 0.0001,
99
+ "unit": 1,
100
+ "up_lr_weight": "",
101
+ "use_cp": false,
102
+ "use_wandb": "",
103
+ "v2": true,
104
+ "v_parameterization": false,
105
+ "vae_batch_size": 0,
106
+ "wandb_api_key": "False",
107
+ "weighted_captions": false,
108
+ "xformers": true
109
+ }
train_shaco/model/shaco_20230819-044821.json ADDED
@@ -0,0 +1,109 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "LoRA_type": "Standard",
3
+ "adaptive_noise_scale": 0,
4
+ "additional_parameters": "",
5
+ "block_alphas": "",
6
+ "block_dims": "",
7
+ "block_lr_zero_threshold": "",
8
+ "bucket_no_upscale": true,
9
+ "bucket_reso_steps": 64,
10
+ "cache_latents": true,
11
+ "cache_latents_to_disk": false,
12
+ "caption_dropout_every_n_epochs": 0.0,
13
+ "caption_dropout_rate": 0,
14
+ "caption_extension": ".txt",
15
+ "clip_skip": 2,
16
+ "color_aug": false,
17
+ "conv_alpha": 1,
18
+ "conv_block_alphas": "",
19
+ "conv_block_dims": "",
20
+ "conv_dim": 1,
21
+ "decompose_both": false,
22
+ "dim_from_weights": false,
23
+ "down_lr_weight": "",
24
+ "enable_bucket": true,
25
+ "epoch": 1,
26
+ "factor": -1,
27
+ "flip_aug": false,
28
+ "full_bf16": false,
29
+ "full_fp16": false,
30
+ "gradient_accumulation_steps": 1.0,
31
+ "gradient_checkpointing": false,
32
+ "keep_tokens": "0",
33
+ "learning_rate": 0.0001,
34
+ "logging_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/log",
35
+ "lora_network_weights": "",
36
+ "lr_scheduler": "constant",
37
+ "lr_scheduler_num_cycles": "",
38
+ "lr_scheduler_power": "",
39
+ "lr_warmup": 0,
40
+ "max_bucket_reso": 2048,
41
+ "max_data_loader_n_workers": "1",
42
+ "max_resolution": "512,512",
43
+ "max_timestep": 1000,
44
+ "max_token_length": "75",
45
+ "max_train_epochs": "",
46
+ "mem_eff_attn": false,
47
+ "mid_lr_weight": "",
48
+ "min_bucket_reso": 64,
49
+ "min_snr_gamma": 0,
50
+ "min_timestep": 0,
51
+ "mixed_precision": "bf16",
52
+ "model_list": "custom",
53
+ "module_dropout": 0,
54
+ "multires_noise_discount": 0,
55
+ "multires_noise_iterations": 0,
56
+ "network_alpha": 128,
57
+ "network_dim": 128,
58
+ "network_dropout": 0,
59
+ "no_token_padding": false,
60
+ "noise_offset": 0,
61
+ "noise_offset_type": "Original",
62
+ "num_cpu_threads_per_process": 2,
63
+ "optimizer": "AdamW8bit",
64
+ "optimizer_args": "",
65
+ "output_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/model",
66
+ "output_name": "shaco",
67
+ "persistent_data_loader_workers": false,
68
+ "pretrained_model_name_or_path": "D:/StableDiffusion/stable-diffusion-webui-1.3.0/models/Stable-diffusion/Animay/mixProV3_v3.safetensors",
69
+ "prior_loss_weight": 1.0,
70
+ "random_crop": false,
71
+ "rank_dropout": 0,
72
+ "reg_data_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/regularization",
73
+ "resume": "",
74
+ "sample_every_n_epochs": 1,
75
+ "sample_every_n_steps": 0,
76
+ "sample_prompts": "",
77
+ "sample_sampler": "euler_a",
78
+ "save_every_n_epochs": 1,
79
+ "save_every_n_steps": 0,
80
+ "save_last_n_steps": 0,
81
+ "save_last_n_steps_state": 0,
82
+ "save_model_as": "safetensors",
83
+ "save_precision": "bf16",
84
+ "save_state": false,
85
+ "scale_v_pred_loss_like_noise_pred": false,
86
+ "scale_weight_norms": 0,
87
+ "sdxl": false,
88
+ "sdxl_cache_text_encoder_outputs": false,
89
+ "sdxl_no_half_vae": false,
90
+ "seed": "1234",
91
+ "shuffle_caption": false,
92
+ "stop_text_encoder_training_pct": 0,
93
+ "text_encoder_lr": 5e-05,
94
+ "train_batch_size": 2,
95
+ "train_data_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/train",
96
+ "train_on_input": false,
97
+ "training_comment": "",
98
+ "unet_lr": 0.0001,
99
+ "unit": 1,
100
+ "up_lr_weight": "",
101
+ "use_cp": false,
102
+ "use_wandb": "",
103
+ "v2": true,
104
+ "v_parameterization": false,
105
+ "vae_batch_size": 0,
106
+ "wandb_api_key": "False",
107
+ "weighted_captions": false,
108
+ "xformers": true
109
+ }
train_shaco/model/shaco_20230819-044836.json ADDED
@@ -0,0 +1,109 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "LoRA_type": "Standard",
3
+ "adaptive_noise_scale": 0,
4
+ "additional_parameters": "",
5
+ "block_alphas": "",
6
+ "block_dims": "",
7
+ "block_lr_zero_threshold": "",
8
+ "bucket_no_upscale": true,
9
+ "bucket_reso_steps": 64,
10
+ "cache_latents": true,
11
+ "cache_latents_to_disk": false,
12
+ "caption_dropout_every_n_epochs": 0.0,
13
+ "caption_dropout_rate": 0,
14
+ "caption_extension": ".txt",
15
+ "clip_skip": 2,
16
+ "color_aug": false,
17
+ "conv_alpha": 1,
18
+ "conv_block_alphas": "",
19
+ "conv_block_dims": "",
20
+ "conv_dim": 1,
21
+ "decompose_both": false,
22
+ "dim_from_weights": false,
23
+ "down_lr_weight": "",
24
+ "enable_bucket": true,
25
+ "epoch": 1,
26
+ "factor": -1,
27
+ "flip_aug": false,
28
+ "full_bf16": false,
29
+ "full_fp16": false,
30
+ "gradient_accumulation_steps": 1.0,
31
+ "gradient_checkpointing": false,
32
+ "keep_tokens": "0",
33
+ "learning_rate": 0.0001,
34
+ "logging_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/log",
35
+ "lora_network_weights": "",
36
+ "lr_scheduler": "constant",
37
+ "lr_scheduler_num_cycles": "",
38
+ "lr_scheduler_power": "",
39
+ "lr_warmup": 0,
40
+ "max_bucket_reso": 2048,
41
+ "max_data_loader_n_workers": "1",
42
+ "max_resolution": "512,512",
43
+ "max_timestep": 1000,
44
+ "max_token_length": "75",
45
+ "max_train_epochs": "",
46
+ "mem_eff_attn": false,
47
+ "mid_lr_weight": "",
48
+ "min_bucket_reso": 64,
49
+ "min_snr_gamma": 0,
50
+ "min_timestep": 0,
51
+ "mixed_precision": "bf16",
52
+ "model_list": "custom",
53
+ "module_dropout": 0,
54
+ "multires_noise_discount": 0,
55
+ "multires_noise_iterations": 0,
56
+ "network_alpha": 128,
57
+ "network_dim": 128,
58
+ "network_dropout": 0,
59
+ "no_token_padding": false,
60
+ "noise_offset": 0,
61
+ "noise_offset_type": "Original",
62
+ "num_cpu_threads_per_process": 2,
63
+ "optimizer": "AdamW8bit",
64
+ "optimizer_args": "",
65
+ "output_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/model",
66
+ "output_name": "shaco",
67
+ "persistent_data_loader_workers": false,
68
+ "pretrained_model_name_or_path": "D:/StableDiffusion/stable-diffusion-webui-1.3.0/models/Stable-diffusion/Animay/mixProV3_v3.safetensors",
69
+ "prior_loss_weight": 1.0,
70
+ "random_crop": false,
71
+ "rank_dropout": 0,
72
+ "reg_data_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/regularization",
73
+ "resume": "",
74
+ "sample_every_n_epochs": 1,
75
+ "sample_every_n_steps": 0,
76
+ "sample_prompts": "",
77
+ "sample_sampler": "euler_a",
78
+ "save_every_n_epochs": 1,
79
+ "save_every_n_steps": 0,
80
+ "save_last_n_steps": 0,
81
+ "save_last_n_steps_state": 0,
82
+ "save_model_as": "safetensors",
83
+ "save_precision": "bf16",
84
+ "save_state": false,
85
+ "scale_v_pred_loss_like_noise_pred": false,
86
+ "scale_weight_norms": 0,
87
+ "sdxl": false,
88
+ "sdxl_cache_text_encoder_outputs": false,
89
+ "sdxl_no_half_vae": false,
90
+ "seed": "1234",
91
+ "shuffle_caption": false,
92
+ "stop_text_encoder_training_pct": 0,
93
+ "text_encoder_lr": 5e-05,
94
+ "train_batch_size": 2,
95
+ "train_data_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/train",
96
+ "train_on_input": false,
97
+ "training_comment": "",
98
+ "unet_lr": 0.0001,
99
+ "unit": 1,
100
+ "up_lr_weight": "",
101
+ "use_cp": false,
102
+ "use_wandb": "",
103
+ "v2": true,
104
+ "v_parameterization": false,
105
+ "vae_batch_size": 0,
106
+ "wandb_api_key": "False",
107
+ "weighted_captions": false,
108
+ "xformers": true
109
+ }
train_shaco/model/shaco_20230820-184523.json ADDED
@@ -0,0 +1,109 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "LoRA_type": "Standard",
3
+ "adaptive_noise_scale": 0,
4
+ "additional_parameters": "",
5
+ "block_alphas": "",
6
+ "block_dims": "",
7
+ "block_lr_zero_threshold": "",
8
+ "bucket_no_upscale": true,
9
+ "bucket_reso_steps": 64,
10
+ "cache_latents": false,
11
+ "cache_latents_to_disk": false,
12
+ "caption_dropout_every_n_epochs": 0.0,
13
+ "caption_dropout_rate": 0,
14
+ "caption_extension": ".txt",
15
+ "clip_skip": "1",
16
+ "color_aug": false,
17
+ "conv_alpha": 4,
18
+ "conv_block_alphas": "",
19
+ "conv_block_dims": "",
20
+ "conv_dim": 4,
21
+ "decompose_both": false,
22
+ "dim_from_weights": false,
23
+ "down_lr_weight": "",
24
+ "enable_bucket": true,
25
+ "epoch": 1,
26
+ "factor": -1,
27
+ "flip_aug": false,
28
+ "full_bf16": false,
29
+ "full_fp16": false,
30
+ "gradient_accumulation_steps": 1.0,
31
+ "gradient_checkpointing": false,
32
+ "keep_tokens": "0",
33
+ "learning_rate": 0.0001,
34
+ "logging_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/log",
35
+ "lora_network_weights": "",
36
+ "lr_scheduler": "cosine",
37
+ "lr_scheduler_num_cycles": "",
38
+ "lr_scheduler_power": "",
39
+ "lr_warmup": 10,
40
+ "max_bucket_reso": 2048,
41
+ "max_data_loader_n_workers": "0",
42
+ "max_resolution": "512,512",
43
+ "max_timestep": 1000,
44
+ "max_token_length": "75",
45
+ "max_train_epochs": "",
46
+ "mem_eff_attn": false,
47
+ "mid_lr_weight": "",
48
+ "min_bucket_reso": 256,
49
+ "min_snr_gamma": 0,
50
+ "min_timestep": 0,
51
+ "mixed_precision": "fp16",
52
+ "model_list": "runwayml/stable-diffusion-v1-5",
53
+ "module_dropout": 0,
54
+ "multires_noise_discount": 0,
55
+ "multires_noise_iterations": 0,
56
+ "network_alpha": 128,
57
+ "network_dim": 128,
58
+ "network_dropout": 0,
59
+ "no_token_padding": false,
60
+ "noise_offset": 0,
61
+ "noise_offset_type": "Original",
62
+ "num_cpu_threads_per_process": 2,
63
+ "optimizer": "AdamW8bit",
64
+ "optimizer_args": "",
65
+ "output_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/model",
66
+ "output_name": "shaco",
67
+ "persistent_data_loader_workers": false,
68
+ "pretrained_model_name_or_path": "runwayml/stable-diffusion-v1-5",
69
+ "prior_loss_weight": 1.0,
70
+ "random_crop": false,
71
+ "rank_dropout": 0,
72
+ "reg_data_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/regularization",
73
+ "resume": "",
74
+ "sample_every_n_epochs": 0,
75
+ "sample_every_n_steps": 25,
76
+ "sample_prompts": "potioncraft, woman, wearing white shirt, wearing necklace, medium breasts",
77
+ "sample_sampler": "euler_a",
78
+ "save_every_n_epochs": 1,
79
+ "save_every_n_steps": 0,
80
+ "save_last_n_steps": 0,
81
+ "save_last_n_steps_state": 0,
82
+ "save_model_as": "safetensors",
83
+ "save_precision": "fp16",
84
+ "save_state": false,
85
+ "scale_v_pred_loss_like_noise_pred": false,
86
+ "scale_weight_norms": 1,
87
+ "sdxl": false,
88
+ "sdxl_cache_text_encoder_outputs": false,
89
+ "sdxl_no_half_vae": true,
90
+ "seed": "",
91
+ "shuffle_caption": false,
92
+ "stop_text_encoder_training_pct": 0,
93
+ "text_encoder_lr": 0.0,
94
+ "train_batch_size": 2,
95
+ "train_data_dir": "D:/StableDiffusion/kohya_ss/TrainingSets/train_shaco/train",
96
+ "train_on_input": false,
97
+ "training_comment": "",
98
+ "unet_lr": 0.0,
99
+ "unit": 1,
100
+ "up_lr_weight": "",
101
+ "use_cp": false,
102
+ "use_wandb": false,
103
+ "v2": false,
104
+ "v_parameterization": false,
105
+ "vae_batch_size": 0,
106
+ "wandb_api_key": "",
107
+ "weighted_captions": false,
108
+ "xformers": true
109
+ }
train_shaco/regularization/1_person/04771-3283295968-person.png ADDED

Git LFS Details

  • SHA256: 2120aa6e1e577b7bd0406d05a6aaf01b4bdd19390022013077d83a334e5f7e46
  • Pointer size: 131 Bytes
  • Size of remote file: 517 kB
train_shaco/regularization/1_person/04771-3283295968-person.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ person, a person standing in front of a wall of colorful items and a person wearing a person prison vest and a person with a backpack
train_shaco/regularization/1_person/04772-3283295969-person.png ADDED

Git LFS Details

  • SHA256: 90ea06bca9a57f9261c9c94953c0d6810d5a00c8aa4170a148592a2b1d5829a3
  • Pointer size: 131 Bytes
  • Size of remote file: 471 kB
train_shaco/regularization/1_person/04772-3283295969-person.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ person, a drawing of two women sitting at a table with food in their hands and a plate of food in their hand
train_shaco/regularization/1_person/04773-3283295970-person.png ADDED

Git LFS Details

  • SHA256: 9b3a9aebe9ba78736627a522fc82a673bc3af2d511f06c2c6f614bbebc385c59
  • Pointer size: 131 Bytes
  • Size of remote file: 523 kB
train_shaco/regularization/1_person/04773-3283295970-person.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ person, a painting of a man standing in front of a wall of colorful blocks and shapes, with a black body
train_shaco/regularization/1_person/04774-2844224771-person.png ADDED

Git LFS Details

  • SHA256: 490847ae3733d8519cdd1f2da2d8cd0485d4d3729bd544cf6cea6b9f672fdeee
  • Pointer size: 131 Bytes
  • Size of remote file: 506 kB
train_shaco/regularization/1_person/04774-2844224771-person.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ person, a man with a face painted with writing on his face and a tie around his neck and neckline,
train_shaco/regularization/1_person/04775-2844224772-person.png ADDED

Git LFS Details

  • SHA256: d377f4911831ea7f3567a4fc5d6bfa1e62cad5f9ae0a15f7462d3a63253ed054
  • Pointer size: 131 Bytes
  • Size of remote file: 419 kB
train_shaco/regularization/1_person/04775-2844224772-person.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ person, a painting of a woman with long hair and a crown on her head, in a dark room with a curtain
train_shaco/regularization/1_person/04776-2844224773-person.png ADDED

Git LFS Details

  • SHA256: 0c80e08b1c500ba1f6a289dca75d3e3f4b68affb2d1490aae257f3f97ac4158f
  • Pointer size: 131 Bytes
  • Size of remote file: 416 kB
train_shaco/regularization/1_person/04776-2844224773-person.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ person, a doll with a hat on sitting in a chair next to a planter and a window with a green plant
train_shaco/regularization/1_person/04777-2844224774-person.png ADDED

Git LFS Details

  • SHA256: 9b2f1fc404984e7bfe6398c7cf07edb57a371ca4f710af351f4ee39932166b6d
  • Pointer size: 131 Bytes
  • Size of remote file: 345 kB
train_shaco/regularization/1_person/04777-2844224774-person.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ person, a man in a red shirt holding a remote control in his hand and pointing at the camera with a surprised look on his face
train_shaco/regularization/1_person/04778-2844224775-person.png ADDED

Git LFS Details

  • SHA256: a5012646958b9f297f34d13598c81cca6b7e6aae3e90b160ee5b37076caabc76
  • Pointer size: 131 Bytes
  • Size of remote file: 418 kB
train_shaco/regularization/1_person/04778-2844224775-person.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ person, two men standing next to each other talking to each other while holding cell phones in their hands and looking at each other
train_shaco/regularization/1_person/04779-774901008-person.png ADDED

Git LFS Details

  • SHA256: ccd07b1c3f2f16faeaf4c1d889b2c88660847016ee622239d8db9f3b39475829
  • Pointer size: 131 Bytes
  • Size of remote file: 387 kB
train_shaco/regularization/1_person/04779-774901008-person.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ person, a drawing of a man with a mustache and a beard on his head with the words design ped - d'd
train_shaco/regularization/1_person/04780-774901009-person.png ADDED

Git LFS Details

  • SHA256: 99e55a2d5b3994309759b65d5f98cbf2d310298cd32dd2a7924c4921433309e3
  • Pointer size: 131 Bytes
  • Size of remote file: 344 kB
train_shaco/regularization/1_person/04780-774901009-person.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ person, a man with a bald head and a black shirt is staring at the camera with his eyes closed and his eyes closed
train_shaco/regularization/1_person/04781-774901010-person.png ADDED

Git LFS Details

  • SHA256: 72e8860308d39e1870a2bec4e5db7c686258dd6361be9a4b296d3414abc898a9
  • Pointer size: 131 Bytes
  • Size of remote file: 543 kB
train_shaco/regularization/1_person/04781-774901010-person.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ person, a man with glasses and a beard standing in front of a crowd of people with many different faces on it
train_shaco/regularization/1_person/04782-774901011-person.png ADDED

Git LFS Details

  • SHA256: 1c3a18076a231c9e551fd2f5f0b13c227eda309fcce2ea9df77b7e9bb7708f6e
  • Pointer size: 131 Bytes
  • Size of remote file: 472 kB
train_shaco/regularization/1_person/04782-774901011-person.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ person, a painting of a woman with a quote on it that says,'the name of the woman is the beautiful zoo, the young boy, the young boy, the boy, the boy, the boy, the boy, the boy, the boy, the boy, the boy, the boy, the boy, the boy, the boy, the boy,
train_shaco/regularization/1_person/04783-774901012-person.png ADDED

Git LFS Details

  • SHA256: e4befc58874690925a42bb0747472cf64151b2fbc58de10b2f237e14f4f1940a
  • Pointer size: 131 Bytes
  • Size of remote file: 383 kB
train_shaco/regularization/1_person/04783-774901012-person.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ person, a painting of a woman with glasses on a red background with a red background and a red background behind her
train_shaco/regularization/1_person/04784-4134988531-person.png ADDED

Git LFS Details

  • SHA256: afe2d303d2418f9cf6624d7fe5230dd9f52425f4fd3ebc3e407407b717d23ac1
  • Pointer size: 131 Bytes
  • Size of remote file: 376 kB
train_shaco/regularization/1_person/04784-4134988531-person.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ person, a painting of a man with a beard and a shirtless shirt on, with his arms crossed, in front of a green wall
train_shaco/regularization/1_person/04785-4134988532-person.png ADDED

Git LFS Details

  • SHA256: 2c049c717dac5c82e95c340084f20b0f9166eec2c488ae9faff757fec0071dd5
  • Pointer size: 131 Bytes
  • Size of remote file: 409 kB
train_shaco/regularization/1_person/04785-4134988532-person.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ person, a woman holding a banana in front of a wall with the word prison babe written on it and a hand holding a banana
train_shaco/regularization/1_person/04786-4134988533-person.png ADDED

Git LFS Details

  • SHA256: 15be3e3d2aa43f2212ab624092ad51731b2e6ea172d7bc357f08ac97ee78575c
  • Pointer size: 131 Bytes
  • Size of remote file: 451 kB
train_shaco/regularization/1_person/04786-4134988533-person.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ person, a painting of a man with a hat and glasses on a yellow background with a red, blue, green, and yellow background
train_shaco/regularization/1_person/04787-4134988534-person.png ADDED

Git LFS Details

  • SHA256: 06c3e443b6a021b289832ea5465377ef39089c430fbb616a9c1fa0f895ba16d3
  • Pointer size: 131 Bytes
  • Size of remote file: 469 kB
train_shaco/regularization/1_person/04787-4134988534-person.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ person, a man in a red shirt and white shorts is running in a marathon with a crowd of people behind him
train_shaco/regularization/1_person/04788-4134988535-person.png ADDED

Git LFS Details

  • SHA256: b67c2dedc787034e2dc98ab7efda5853b4bdc09de5ad4f6798328c7a37670338
  • Pointer size: 131 Bytes
  • Size of remote file: 447 kB