ixelszy commited on
Commit
23299fa
·
verified ·
1 Parent(s): 5274992

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +33 -0
  2. config_lora-20250602-145848.toml +48 -0
  3. config_lora-20250602-145929.toml +48 -0
  4. config_lora-20250602-150005.toml +48 -0
  5. config_lora-20250602-150102.toml +48 -0
  6. config_lora-20250602-150232.toml +49 -0
  7. config_lora-20250602-150344.toml +49 -0
  8. config_lora-20250602-150626.toml +49 -0
  9. config_lora-20250602-150707.toml +49 -0
  10. config_lora-20250602-151250.toml +49 -0
  11. config_lora-20250602-151500.toml +50 -0
  12. config_lora-20250602-151638.toml +50 -0
  13. config_lora-20250602-151841.toml +50 -0
  14. config_lora-20250602-152328.toml +51 -0
  15. config_lora-20250602-152659.toml +51 -0
  16. config_lora-20250602-152748.toml +51 -0
  17. config_lora-20250602-152812.toml +50 -0
  18. config_lora-20250602-152843.toml +51 -0
  19. config_lora-20250602-153053.toml +51 -0
  20. config_lora-20250602-153250.toml +49 -0
  21. config_lora-20250602-153310.toml +49 -0
  22. config_lora-20250602-153339.toml +49 -0
  23. config_lora-20250602-153522.toml +49 -0
  24. config_lora-20250602-155358.toml +51 -0
  25. img/100_Lena person/1.png +3 -0
  26. img/100_Lena person/1.txt +1 -0
  27. img/100_Lena person/10.png +3 -0
  28. img/100_Lena person/10.txt +1 -0
  29. img/100_Lena person/11.png +3 -0
  30. img/100_Lena person/11.txt +1 -0
  31. img/100_Lena person/12.png +3 -0
  32. img/100_Lena person/12.txt +1 -0
  33. img/100_Lena person/13.png +3 -0
  34. img/100_Lena person/13.txt +1 -0
  35. img/100_Lena person/14.png +3 -0
  36. img/100_Lena person/14.txt +1 -0
  37. img/100_Lena person/15.png +3 -0
  38. img/100_Lena person/15.txt +1 -0
  39. img/100_Lena person/2.png +3 -0
  40. img/100_Lena person/2.txt +1 -0
  41. img/100_Lena person/3.png +3 -0
  42. img/100_Lena person/3.txt +1 -0
  43. img/100_Lena person/4.png +3 -0
  44. img/100_Lena person/4.txt +1 -0
  45. img/100_Lena person/5.png +3 -0
  46. img/100_Lena person/5.txt +1 -0
  47. img/100_Lena person/6.png +3 -0
  48. img/100_Lena person/6.txt +1 -0
  49. img/100_Lena person/7.png +3 -0
  50. img/100_Lena person/7.txt +1 -0
.gitattributes CHANGED
@@ -33,3 +33,36 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ img/100_Lena[[:space:]]person/1.png filter=lfs diff=lfs merge=lfs -text
37
+ img/100_Lena[[:space:]]person/10.png filter=lfs diff=lfs merge=lfs -text
38
+ img/100_Lena[[:space:]]person/11.png filter=lfs diff=lfs merge=lfs -text
39
+ img/100_Lena[[:space:]]person/12.png filter=lfs diff=lfs merge=lfs -text
40
+ img/100_Lena[[:space:]]person/13.png filter=lfs diff=lfs merge=lfs -text
41
+ img/100_Lena[[:space:]]person/14.png filter=lfs diff=lfs merge=lfs -text
42
+ img/100_Lena[[:space:]]person/15.png filter=lfs diff=lfs merge=lfs -text
43
+ img/100_Lena[[:space:]]person/2.png filter=lfs diff=lfs merge=lfs -text
44
+ img/100_Lena[[:space:]]person/3.png filter=lfs diff=lfs merge=lfs -text
45
+ img/100_Lena[[:space:]]person/4.png filter=lfs diff=lfs merge=lfs -text
46
+ img/100_Lena[[:space:]]person/5.png filter=lfs diff=lfs merge=lfs -text
47
+ img/100_Lena[[:space:]]person/6.png filter=lfs diff=lfs merge=lfs -text
48
+ img/100_Lena[[:space:]]person/7.png filter=lfs diff=lfs merge=lfs -text
49
+ img/100_Lena[[:space:]]person/8.png filter=lfs diff=lfs merge=lfs -text
50
+ img/100_Lena[[:space:]]person/9.png filter=lfs diff=lfs merge=lfs -text
51
+ model/sample/last_e000001_00_20250602162805.png filter=lfs diff=lfs merge=lfs -text
52
+ model/sample/last_e000001_00_20250602164004.png filter=lfs diff=lfs merge=lfs -text
53
+ model/sample/last_e000001_01_20250602162809.png filter=lfs diff=lfs merge=lfs -text
54
+ model/sample/last_e000001_01_20250602164008.png filter=lfs diff=lfs merge=lfs -text
55
+ model/sample/last_e000001_02_20250602162813.png filter=lfs diff=lfs merge=lfs -text
56
+ model/sample/last_e000001_02_20250602164012.png filter=lfs diff=lfs merge=lfs -text
57
+ model/sample/last_e000002_00_20250602164323.png filter=lfs diff=lfs merge=lfs -text
58
+ model/sample/last_e000002_01_20250602164328.png filter=lfs diff=lfs merge=lfs -text
59
+ model/sample/last_e000002_02_20250602164332.png filter=lfs diff=lfs merge=lfs -text
60
+ model/sample/last_e000003_00_20250602164632.png filter=lfs diff=lfs merge=lfs -text
61
+ model/sample/last_e000003_01_20250602164636.png filter=lfs diff=lfs merge=lfs -text
62
+ model/sample/last_e000003_02_20250602164640.png filter=lfs diff=lfs merge=lfs -text
63
+ model/sample/last_e000004_00_20250602164941.png filter=lfs diff=lfs merge=lfs -text
64
+ model/sample/last_e000004_01_20250602164945.png filter=lfs diff=lfs merge=lfs -text
65
+ model/sample/last_e000004_02_20250602164949.png filter=lfs diff=lfs merge=lfs -text
66
+ model/sample/last_e000005_00_20250602165042.png filter=lfs diff=lfs merge=lfs -text
67
+ model/sample/last_e000005_01_20250602165046.png filter=lfs diff=lfs merge=lfs -text
68
+ model/sample/last_e000005_02_20250602165050.png filter=lfs diff=lfs merge=lfs -text
config_lora-20250602-145848.toml ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bucket_no_upscale = true
2
+ bucket_reso_steps = 64
3
+ cache_latents = true
4
+ caption_extension = ".txt"
5
+ clip_skip = 1
6
+ dynamo_backend = "no"
7
+ enable_bucket = true
8
+ epoch = 1
9
+ gradient_accumulation_steps = 1
10
+ huber_c = 0.1
11
+ huber_scale = 1
12
+ huber_schedule = "snr"
13
+ loss_type = "l2"
14
+ lr_scheduler = "constant"
15
+ lr_scheduler_args = []
16
+ lr_scheduler_num_cycles = 1
17
+ lr_scheduler_power = 1
18
+ max_bucket_reso = 2048
19
+ max_data_loader_n_workers = 0
20
+ max_grad_norm = 1
21
+ max_timestep = 1000
22
+ max_token_length = 75
23
+ max_train_steps = 1600
24
+ min_bucket_reso = 512
25
+ mixed_precision = "fp16"
26
+ network_alpha = 16
27
+ network_args = []
28
+ network_dim = 32
29
+ network_module = "networks.lora"
30
+ noise_offset_type = "Original"
31
+ optimizer_args = []
32
+ optimizer_type = "DAdaptAdam"
33
+ output_dir = "/workspace/kohya_ss/outputs"
34
+ output_name = "last"
35
+ pretrained_model_name_or_path = "coreml-community/coreml-RealismEngineSDXL-v10_SDXL_8bit"
36
+ prior_loss_weight = 1
37
+ resolution = "512,512"
38
+ sample_prompts = "/workspace/kohya_ss/outputs/sample/prompt.txt"
39
+ sample_sampler = "euler_a"
40
+ save_every_n_epochs = 1
41
+ save_model_as = "safetensors"
42
+ save_precision = "fp16"
43
+ text_encoder_lr = [ 4.5e-5, 4.5e-5,]
44
+ train_batch_size = 1
45
+ train_data_dir = "/workspace/kohya_ss/dataset/images"
46
+ unet_lr = 0.0001
47
+ wandb_run_name = "last"
48
+ xformers = true
config_lora-20250602-145929.toml ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bucket_no_upscale = true
2
+ bucket_reso_steps = 64
3
+ cache_latents = true
4
+ caption_extension = ".txt"
5
+ clip_skip = 1
6
+ dynamo_backend = "no"
7
+ enable_bucket = true
8
+ epoch = 1
9
+ gradient_accumulation_steps = 1
10
+ huber_c = 0.1
11
+ huber_scale = 1
12
+ huber_schedule = "snr"
13
+ loss_type = "l2"
14
+ lr_scheduler = "constant"
15
+ lr_scheduler_args = []
16
+ lr_scheduler_num_cycles = 1
17
+ lr_scheduler_power = 1
18
+ max_bucket_reso = 2048
19
+ max_data_loader_n_workers = 0
20
+ max_grad_norm = 1
21
+ max_timestep = 1000
22
+ max_token_length = 75
23
+ max_train_steps = 1600
24
+ min_bucket_reso = 512
25
+ mixed_precision = "fp16"
26
+ network_alpha = 16
27
+ network_args = []
28
+ network_dim = 32
29
+ network_module = "networks.lora"
30
+ noise_offset_type = "Original"
31
+ optimizer_args = []
32
+ optimizer_type = "DAdaptAdam"
33
+ output_dir = "/workspace/kohya_ss/outputs"
34
+ output_name = "last"
35
+ pretrained_model_name_or_path = "coreml-community/coreml-RealismEngineSDXL-v10_SDXL_8bit"
36
+ prior_loss_weight = 1
37
+ resolution = "512,512"
38
+ sample_prompts = "/workspace/kohya_ss/outputs/sample/prompt.txt"
39
+ sample_sampler = "euler_a"
40
+ save_every_n_epochs = 1
41
+ save_model_as = "safetensors"
42
+ save_precision = "fp16"
43
+ text_encoder_lr = [ 4.5e-5, 4.5e-5,]
44
+ train_batch_size = 1
45
+ train_data_dir = "/workspace/kohya_ss/img/40_Lena person"
46
+ unet_lr = 0.0001
47
+ wandb_run_name = "last"
48
+ xformers = true
config_lora-20250602-150005.toml ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bucket_no_upscale = true
2
+ bucket_reso_steps = 64
3
+ cache_latents = true
4
+ caption_extension = ".txt"
5
+ clip_skip = 1
6
+ dynamo_backend = "no"
7
+ enable_bucket = true
8
+ epoch = 1
9
+ gradient_accumulation_steps = 1
10
+ huber_c = 0.1
11
+ huber_scale = 1
12
+ huber_schedule = "snr"
13
+ loss_type = "l2"
14
+ lr_scheduler = "constant"
15
+ lr_scheduler_args = []
16
+ lr_scheduler_num_cycles = 1
17
+ lr_scheduler_power = 1
18
+ max_bucket_reso = 2048
19
+ max_data_loader_n_workers = 0
20
+ max_grad_norm = 1
21
+ max_timestep = 1000
22
+ max_token_length = 75
23
+ max_train_steps = 1600
24
+ min_bucket_reso = 512
25
+ mixed_precision = "fp16"
26
+ network_alpha = 16
27
+ network_args = []
28
+ network_dim = 32
29
+ network_module = "networks.lora"
30
+ noise_offset_type = "Original"
31
+ optimizer_args = []
32
+ optimizer_type = "DAdaptAdam"
33
+ output_dir = "/workspace/kohya_ss/outputs"
34
+ output_name = "last"
35
+ pretrained_model_name_or_path = "coreml-community/coreml-RealismEngineSDXL-v10_SDXL_8bit"
36
+ prior_loss_weight = 1
37
+ resolution = "512,512"
38
+ sample_prompts = "/workspace/kohya_ss/outputs/sample/prompt.txt"
39
+ sample_sampler = "euler_a"
40
+ save_every_n_epochs = 1
41
+ save_model_as = "safetensors"
42
+ save_precision = "fp16"
43
+ text_encoder_lr = [ 4.5e-5, 4.5e-5,]
44
+ train_batch_size = 1
45
+ train_data_dir = "/workspace/kohya_ss/img/40_Lena person"
46
+ unet_lr = 0.0001
47
+ wandb_run_name = "last"
48
+ xformers = true
config_lora-20250602-150102.toml ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bucket_no_upscale = true
2
+ bucket_reso_steps = 64
3
+ cache_latents = true
4
+ caption_extension = ".txt"
5
+ clip_skip = 1
6
+ dynamo_backend = "no"
7
+ enable_bucket = true
8
+ epoch = 1
9
+ gradient_accumulation_steps = 1
10
+ huber_c = 0.1
11
+ huber_scale = 1
12
+ huber_schedule = "snr"
13
+ loss_type = "l2"
14
+ lr_scheduler = "constant"
15
+ lr_scheduler_args = []
16
+ lr_scheduler_num_cycles = 1
17
+ lr_scheduler_power = 1
18
+ max_bucket_reso = 2048
19
+ max_data_loader_n_workers = 0
20
+ max_grad_norm = 1
21
+ max_timestep = 1000
22
+ max_token_length = 75
23
+ max_train_steps = 1600
24
+ min_bucket_reso = 512
25
+ mixed_precision = "fp16"
26
+ network_alpha = 16
27
+ network_args = []
28
+ network_dim = 32
29
+ network_module = "networks.lora"
30
+ noise_offset_type = "Original"
31
+ optimizer_args = []
32
+ optimizer_type = "AdamW8bit"
33
+ output_dir = "/workspace/kohya_ss/outputs"
34
+ output_name = "last"
35
+ pretrained_model_name_or_path = "coreml-community/coreml-RealismEngineSDXL-v10_SDXL_8bit"
36
+ prior_loss_weight = 1
37
+ resolution = "512,512"
38
+ sample_prompts = "/workspace/kohya_ss/outputs/sample/prompt.txt"
39
+ sample_sampler = "euler_a"
40
+ save_every_n_epochs = 1
41
+ save_model_as = "safetensors"
42
+ save_precision = "fp16"
43
+ text_encoder_lr = [ 4.5e-5, 4.5e-5,]
44
+ train_batch_size = 4
45
+ train_data_dir = "/workspace/kohya_ss/img/40_Lena person"
46
+ unet_lr = 0.0001
47
+ wandb_run_name = "last"
48
+ xformers = true
config_lora-20250602-150232.toml ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bucket_no_upscale = true
2
+ bucket_reso_steps = 64
3
+ cache_latents = true
4
+ caption_extension = ".txt"
5
+ clip_skip = 1
6
+ dynamo_backend = "no"
7
+ enable_bucket = true
8
+ epoch = 1
9
+ gradient_accumulation_steps = 1
10
+ huber_c = 0.1
11
+ huber_scale = 1
12
+ huber_schedule = "snr"
13
+ loss_type = "l2"
14
+ lr_scheduler = "constant"
15
+ lr_scheduler_args = []
16
+ lr_scheduler_num_cycles = 1
17
+ lr_scheduler_power = 1
18
+ max_bucket_reso = 2048
19
+ max_data_loader_n_workers = 0
20
+ max_grad_norm = 1
21
+ max_timestep = 1000
22
+ max_token_length = 75
23
+ max_train_steps = 1600
24
+ min_bucket_reso = 64
25
+ mixed_precision = "fp16"
26
+ network_alpha = 16
27
+ network_args = []
28
+ network_dim = 32
29
+ network_module = "networks.lora"
30
+ network_train_unet_only = true
31
+ noise_offset_type = "Original"
32
+ optimizer_args = []
33
+ optimizer_type = "AdamW8bit"
34
+ output_dir = "/workspace/kohya_ss/outputs"
35
+ output_name = "last"
36
+ pretrained_model_name_or_path = "coreml-community/coreml-RealismEngineSDXL-v10_SDXL_8bit"
37
+ prior_loss_weight = 1
38
+ resolution = "512,512"
39
+ sample_prompts = "/workspace/kohya_ss/outputs/sample/prompt.txt"
40
+ sample_sampler = "euler_a"
41
+ save_every_n_epochs = 1
42
+ save_model_as = "safetensors"
43
+ save_precision = "fp16"
44
+ text_encoder_lr = []
45
+ train_batch_size = 4
46
+ train_data_dir = "./img/40_Lena"
47
+ unet_lr = 0.0001
48
+ wandb_run_name = "last"
49
+ xformers = true
config_lora-20250602-150344.toml ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bucket_no_upscale = true
2
+ bucket_reso_steps = 64
3
+ cache_latents = true
4
+ caption_extension = ".txt"
5
+ clip_skip = 1
6
+ dynamo_backend = "no"
7
+ enable_bucket = true
8
+ epoch = 1
9
+ gradient_accumulation_steps = 1
10
+ huber_c = 0.1
11
+ huber_scale = 1
12
+ huber_schedule = "snr"
13
+ loss_type = "l2"
14
+ lr_scheduler = "constant"
15
+ lr_scheduler_args = []
16
+ lr_scheduler_num_cycles = 1
17
+ lr_scheduler_power = 1
18
+ max_bucket_reso = 2048
19
+ max_data_loader_n_workers = 0
20
+ max_grad_norm = 1
21
+ max_timestep = 1000
22
+ max_token_length = 75
23
+ max_train_steps = 1600
24
+ min_bucket_reso = 64
25
+ mixed_precision = "fp16"
26
+ network_alpha = 16
27
+ network_args = []
28
+ network_dim = 32
29
+ network_module = "networks.lora"
30
+ network_train_unet_only = true
31
+ noise_offset_type = "Original"
32
+ optimizer_args = []
33
+ optimizer_type = "AdamW8bit"
34
+ output_dir = "/workspace/kohya_ss/outputs"
35
+ output_name = "last"
36
+ pretrained_model_name_or_path = "coreml-community/coreml-RealismEngineSDXL-v10_SDXL_8bit"
37
+ prior_loss_weight = 1
38
+ resolution = "512,512"
39
+ sample_prompts = "/workspace/kohya_ss/outputs/sample/prompt.txt"
40
+ sample_sampler = "euler_a"
41
+ save_every_n_epochs = 1
42
+ save_model_as = "safetensors"
43
+ save_precision = "fp16"
44
+ text_encoder_lr = []
45
+ train_batch_size = 4
46
+ train_data_dir = "./img/40_Lena"
47
+ unet_lr = 0.0001
48
+ wandb_run_name = "last"
49
+ xformers = true
config_lora-20250602-150626.toml ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bucket_no_upscale = true
2
+ bucket_reso_steps = 64
3
+ cache_latents = true
4
+ caption_extension = ".txt"
5
+ clip_skip = 1
6
+ dynamo_backend = "no"
7
+ enable_bucket = true
8
+ epoch = 1
9
+ gradient_accumulation_steps = 1
10
+ huber_c = 0.1
11
+ huber_scale = 1
12
+ huber_schedule = "snr"
13
+ loss_type = "l2"
14
+ lr_scheduler = "constant"
15
+ lr_scheduler_args = []
16
+ lr_scheduler_num_cycles = 1
17
+ lr_scheduler_power = 1
18
+ max_bucket_reso = 2048
19
+ max_data_loader_n_workers = 0
20
+ max_grad_norm = 1
21
+ max_timestep = 1000
22
+ max_token_length = 75
23
+ max_train_steps = 1600
24
+ min_bucket_reso = 64
25
+ mixed_precision = "fp16"
26
+ network_alpha = 16
27
+ network_args = []
28
+ network_dim = 32
29
+ network_module = "networks.lora"
30
+ network_train_unet_only = true
31
+ noise_offset_type = "Original"
32
+ optimizer_args = []
33
+ optimizer_type = "AdamW8bit"
34
+ output_dir = "/workspace/kohya_ss/outputs"
35
+ output_name = "last"
36
+ pretrained_model_name_or_path = "coreml-community/coreml-RealismEngineSDXL-v10_SDXL_8bit"
37
+ prior_loss_weight = 1
38
+ resolution = "512,512"
39
+ sample_prompts = "/workspace/kohya_ss/outputs/sample/prompt.txt"
40
+ sample_sampler = "euler_a"
41
+ save_every_n_epochs = 1
42
+ save_model_as = "safetensors"
43
+ save_precision = "fp16"
44
+ text_encoder_lr = []
45
+ train_batch_size = 4
46
+ train_data_dir = "./img/40_Lena/"
47
+ unet_lr = 0.0001
48
+ wandb_run_name = "last"
49
+ xformers = true
config_lora-20250602-150707.toml ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bucket_no_upscale = true
2
+ bucket_reso_steps = 64
3
+ cache_latents = true
4
+ caption_extension = ".txt"
5
+ clip_skip = 1
6
+ dynamo_backend = "no"
7
+ enable_bucket = true
8
+ epoch = 1
9
+ gradient_accumulation_steps = 1
10
+ huber_c = 0.1
11
+ huber_scale = 1
12
+ huber_schedule = "snr"
13
+ loss_type = "l2"
14
+ lr_scheduler = "constant"
15
+ lr_scheduler_args = []
16
+ lr_scheduler_num_cycles = 1
17
+ lr_scheduler_power = 1
18
+ max_bucket_reso = 2048
19
+ max_data_loader_n_workers = 0
20
+ max_grad_norm = 1
21
+ max_timestep = 1000
22
+ max_token_length = 75
23
+ max_train_steps = 1600
24
+ min_bucket_reso = 64
25
+ mixed_precision = "fp16"
26
+ network_alpha = 16
27
+ network_args = []
28
+ network_dim = 32
29
+ network_module = "networks.lora"
30
+ network_train_unet_only = true
31
+ noise_offset_type = "Original"
32
+ optimizer_args = []
33
+ optimizer_type = "AdamW8bit"
34
+ output_dir = "/workspace/kohya_ss/outputs"
35
+ output_name = "last"
36
+ pretrained_model_name_or_path = "stabilityai/stable-diffusion-xl-base-1.0"
37
+ prior_loss_weight = 1
38
+ resolution = "512,512"
39
+ sample_prompts = "/workspace/kohya_ss/outputs/sample/prompt.txt"
40
+ sample_sampler = "euler_a"
41
+ save_every_n_epochs = 1
42
+ save_model_as = "safetensors"
43
+ save_precision = "fp16"
44
+ text_encoder_lr = []
45
+ train_batch_size = 4
46
+ train_data_dir = "./img/40_Lena"
47
+ unet_lr = 0.0001
48
+ wandb_run_name = "last"
49
+ xformers = true
config_lora-20250602-151250.toml ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bucket_no_upscale = true
2
+ bucket_reso_steps = 64
3
+ cache_latents = true
4
+ caption_extension = ".txt"
5
+ clip_skip = 1
6
+ dynamo_backend = "no"
7
+ enable_bucket = true
8
+ epoch = 1
9
+ gradient_accumulation_steps = 1
10
+ huber_c = 0.1
11
+ huber_scale = 1
12
+ huber_schedule = "snr"
13
+ loss_type = "l2"
14
+ lr_scheduler = "constant"
15
+ lr_scheduler_args = []
16
+ lr_scheduler_num_cycles = 1
17
+ lr_scheduler_power = 1
18
+ max_bucket_reso = 2048
19
+ max_data_loader_n_workers = 0
20
+ max_grad_norm = 1
21
+ max_timestep = 1000
22
+ max_token_length = 75
23
+ max_train_steps = 1600
24
+ min_bucket_reso = 64
25
+ mixed_precision = "fp16"
26
+ network_alpha = 16
27
+ network_args = []
28
+ network_dim = 32
29
+ network_module = "networks.lora"
30
+ network_train_unet_only = true
31
+ noise_offset_type = "Original"
32
+ optimizer_args = []
33
+ optimizer_type = "AdamW8bit"
34
+ output_dir = "/workspace/kohya_ss/outputs"
35
+ output_name = "last"
36
+ pretrained_model_name_or_path = "coreml-community/coreml-RealismEngineSDXL-v10_SDXL_8bit"
37
+ prior_loss_weight = 1
38
+ resolution = "512,512"
39
+ sample_prompts = "/workspace/kohya_ss/outputs/sample/prompt.txt"
40
+ sample_sampler = "euler_a"
41
+ save_every_n_epochs = 1
42
+ save_model_as = "safetensors"
43
+ save_precision = "fp16"
44
+ text_encoder_lr = []
45
+ train_batch_size = 4
46
+ train_data_dir = "./img/10_Lena person"
47
+ unet_lr = 0.0001
48
+ wandb_run_name = "last"
49
+ xformers = true
config_lora-20250602-151500.toml ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bucket_no_upscale = true
2
+ bucket_reso_steps = 64
3
+ cache_latents = true
4
+ caption_extension = ".txt"
5
+ clip_skip = 1
6
+ dynamo_backend = "no"
7
+ enable_bucket = true
8
+ epoch = 1
9
+ gradient_accumulation_steps = 1
10
+ huber_c = 0.1
11
+ huber_scale = 1
12
+ huber_schedule = "snr"
13
+ loss_type = "l2"
14
+ lr_scheduler = "constant"
15
+ lr_scheduler_args = []
16
+ lr_scheduler_num_cycles = 1
17
+ lr_scheduler_power = 1
18
+ max_bucket_reso = 2048
19
+ max_data_loader_n_workers = 0
20
+ max_grad_norm = 1
21
+ max_timestep = 1000
22
+ max_token_length = 75
23
+ max_train_epochs = 10
24
+ max_train_steps = 1600
25
+ min_bucket_reso = 64
26
+ mixed_precision = "fp16"
27
+ network_alpha = 16
28
+ network_args = []
29
+ network_dim = 32
30
+ network_module = "networks.lora"
31
+ network_train_unet_only = true
32
+ noise_offset_type = "Original"
33
+ optimizer_args = []
34
+ optimizer_type = "AdamW8bit"
35
+ output_dir = "/workspace/kohya_ss/outputs"
36
+ output_name = "last"
37
+ pretrained_model_name_or_path = "coreml-community/coreml-RealismEngineSDXL-v10_SDXL_8bit"
38
+ prior_loss_weight = 1
39
+ resolution = "512,512"
40
+ sample_prompts = "/workspace/kohya_ss/outputs/sample/prompt.txt"
41
+ sample_sampler = "euler_a"
42
+ save_every_n_epochs = 1
43
+ save_model_as = "safetensors"
44
+ save_precision = "fp16"
45
+ text_encoder_lr = []
46
+ train_batch_size = 4
47
+ train_data_dir = "/workspace/kohya_ss/img/10_Lena person"
48
+ unet_lr = 0.0001
49
+ wandb_run_name = "last"
50
+ xformers = true
config_lora-20250602-151638.toml ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bucket_no_upscale = true
2
+ bucket_reso_steps = 64
3
+ cache_latents = true
4
+ caption_extension = ".txt"
5
+ clip_skip = 1
6
+ dynamo_backend = "no"
7
+ enable_bucket = true
8
+ epoch = 1
9
+ gradient_accumulation_steps = 1
10
+ huber_c = 0.1
11
+ huber_scale = 1
12
+ huber_schedule = "snr"
13
+ loss_type = "l2"
14
+ lr_scheduler = "constant"
15
+ lr_scheduler_args = []
16
+ lr_scheduler_num_cycles = 1
17
+ lr_scheduler_power = 1
18
+ max_bucket_reso = 2048
19
+ max_data_loader_n_workers = 0
20
+ max_grad_norm = 1
21
+ max_timestep = 1000
22
+ max_token_length = 75
23
+ max_train_epochs = 10
24
+ max_train_steps = 1600
25
+ min_bucket_reso = 64
26
+ mixed_precision = "fp16"
27
+ network_alpha = 16
28
+ network_args = []
29
+ network_dim = 32
30
+ network_module = "networks.lora"
31
+ network_train_unet_only = true
32
+ noise_offset_type = "Original"
33
+ optimizer_args = []
34
+ optimizer_type = "AdamW8bit"
35
+ output_dir = "/workspace/kohya_ss/outputs"
36
+ output_name = "last"
37
+ pretrained_model_name_or_path = "coreml-community/coreml-RealismEngineSDXL-v10_SDXL_8bit"
38
+ prior_loss_weight = 1
39
+ resolution = "512,512"
40
+ sample_prompts = "/workspace/kohya_ss/outputs/sample/prompt.txt"
41
+ sample_sampler = "euler_a"
42
+ save_every_n_epochs = 1
43
+ save_model_as = "safetensors"
44
+ save_precision = "fp16"
45
+ text_encoder_lr = []
46
+ train_batch_size = 4
47
+ train_data_dir = "/workspace/kohya_ss/img/10_Lena person"
48
+ unet_lr = 0.0001
49
+ wandb_run_name = "last"
50
+ xformers = true
config_lora-20250602-151841.toml ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bucket_no_upscale = true
2
+ bucket_reso_steps = 64
3
+ cache_latents = true
4
+ caption_extension = ".txt"
5
+ clip_skip = 1
6
+ dynamo_backend = "no"
7
+ enable_bucket = true
8
+ epoch = 1
9
+ gradient_accumulation_steps = 1
10
+ huber_c = 0.1
11
+ huber_scale = 1
12
+ huber_schedule = "snr"
13
+ loss_type = "l2"
14
+ lr_scheduler = "constant"
15
+ lr_scheduler_args = []
16
+ lr_scheduler_num_cycles = 1
17
+ lr_scheduler_power = 1
18
+ max_bucket_reso = 2048
19
+ max_data_loader_n_workers = 0
20
+ max_grad_norm = 1
21
+ max_timestep = 1000
22
+ max_token_length = 75
23
+ max_train_epochs = 10
24
+ max_train_steps = 1600
25
+ min_bucket_reso = 64
26
+ mixed_precision = "fp16"
27
+ network_alpha = 16
28
+ network_args = []
29
+ network_dim = 32
30
+ network_module = "networks.lora"
31
+ network_train_unet_only = true
32
+ noise_offset_type = "Original"
33
+ optimizer_args = []
34
+ optimizer_type = "AdamW8bit"
35
+ output_dir = "/workspace/kohya_ss/outputs"
36
+ output_name = "last"
37
+ pretrained_model_name_or_path = "coreml-community/coreml-RealismEngineSDXL-v10_SDXL_8bit"
38
+ prior_loss_weight = 1
39
+ resolution = "512,512"
40
+ sample_prompts = "/workspace/kohya_ss/outputs/sample/prompt.txt"
41
+ sample_sampler = "euler_a"
42
+ save_every_n_epochs = 1
43
+ save_model_as = "safetensors"
44
+ save_precision = "fp16"
45
+ text_encoder_lr = []
46
+ train_batch_size = 4
47
+ train_data_dir = "/workspace/kohya_ss/img/10_Lena person"
48
+ unet_lr = 0.0001
49
+ wandb_run_name = "last"
50
+ xformers = true
config_lora-20250602-152328.toml ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bucket_no_upscale = true
2
+ bucket_reso_steps = 64
3
+ cache_latents = true
4
+ caption_extension = ".txt"
5
+ clip_skip = 1
6
+ dynamo_backend = "no"
7
+ enable_bucket = true
8
+ epoch = 1
9
+ gradient_accumulation_steps = 1
10
+ huber_c = 0.1
11
+ huber_scale = 1
12
+ huber_schedule = "snr"
13
+ loss_type = "l2"
14
+ lr_scheduler = "cosine"
15
+ lr_scheduler_args = []
16
+ lr_scheduler_num_cycles = 1
17
+ lr_scheduler_power = 1
18
+ lr_warmup_steps = 0.1
19
+ max_bucket_reso = 2048
20
+ max_data_loader_n_workers = 0
21
+ max_grad_norm = 1
22
+ max_timestep = 1000
23
+ max_token_length = 75
24
+ max_train_epochs = 10
25
+ max_train_steps = 1600
26
+ min_bucket_reso = 64
27
+ mixed_precision = "fp16"
28
+ network_alpha = 16
29
+ network_args = []
30
+ network_dim = 32
31
+ network_module = "networks.lora"
32
+ network_train_unet_only = true
33
+ noise_offset_type = "Original"
34
+ optimizer_args = []
35
+ optimizer_type = "AdamW8bit"
36
+ output_dir = "/workspace/kohya_ss/outputs"
37
+ output_name = "last"
38
+ pretrained_model_name_or_path = "coreml-community/coreml-RealismEngineSDXL-v10_SDXL_8bit"
39
+ prior_loss_weight = 1
40
+ resolution = "512,512"
41
+ sample_prompts = "/workspace/kohya_ss/outputs/sample/prompt.txt"
42
+ sample_sampler = "euler_a"
43
+ save_every_n_epochs = 1
44
+ save_model_as = "safetensors"
45
+ save_precision = "fp16"
46
+ text_encoder_lr = []
47
+ train_batch_size = 4
48
+ train_data_dir = "/workspace/kohya_ss/img/10_Lena person"
49
+ unet_lr = 0.0001
50
+ wandb_run_name = "last"
51
+ xformers = true
config_lora-20250602-152659.toml ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bucket_no_upscale = true
2
+ bucket_reso_steps = 64
3
+ cache_latents = true
4
+ caption_extension = ".txt"
5
+ clip_skip = 1
6
+ dynamo_backend = "no"
7
+ enable_bucket = true
8
+ epoch = 1
9
+ gradient_accumulation_steps = 1
10
+ huber_c = 0.1
11
+ huber_scale = 1
12
+ huber_schedule = "snr"
13
+ loss_type = "l2"
14
+ lr_scheduler = "cosine"
15
+ lr_scheduler_args = []
16
+ lr_scheduler_num_cycles = 1
17
+ lr_scheduler_power = 1
18
+ lr_warmup_steps = 0.1
19
+ max_bucket_reso = 2048
20
+ max_data_loader_n_workers = 0
21
+ max_grad_norm = 1
22
+ max_timestep = 1000
23
+ max_token_length = 75
24
+ max_train_epochs = 10
25
+ max_train_steps = 1600
26
+ min_bucket_reso = 64
27
+ mixed_precision = "fp16"
28
+ network_alpha = 16
29
+ network_args = []
30
+ network_dim = 32
31
+ network_module = "networks.lora"
32
+ network_train_unet_only = true
33
+ noise_offset_type = "Original"
34
+ optimizer_args = []
35
+ optimizer_type = "AdamW8bit"
36
+ output_dir = "/workspace/kohya_ss/outputs"
37
+ output_name = "last"
38
+ pretrained_model_name_or_path = "coreml-community/coreml-RealismEngineSDXL-v10_SDXL_8bit"
39
+ prior_loss_weight = 1
40
+ resolution = "512,512"
41
+ sample_prompts = "/workspace/kohya_ss/outputs/sample/prompt.txt"
42
+ sample_sampler = "euler_a"
43
+ save_every_n_epochs = 1
44
+ save_model_as = "safetensors"
45
+ save_precision = "fp16"
46
+ text_encoder_lr = []
47
+ train_batch_size = 4
48
+ train_data_dir = "/workspace/kohya_ss/img/10_Lena person"
49
+ unet_lr = 0.0001
50
+ wandb_run_name = "last"
51
+ xformers = true
config_lora-20250602-152748.toml ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bucket_no_upscale = true
2
+ bucket_reso_steps = 64
3
+ cache_latents = true
4
+ caption_extension = ".txt"
5
+ clip_skip = 1
6
+ dynamo_backend = "no"
7
+ enable_bucket = true
8
+ epoch = 1
9
+ gradient_accumulation_steps = 1
10
+ huber_c = 0.1
11
+ huber_scale = 1
12
+ huber_schedule = "snr"
13
+ loss_type = "l2"
14
+ lr_scheduler = "cosine"
15
+ lr_scheduler_args = []
16
+ lr_scheduler_num_cycles = 1
17
+ lr_scheduler_power = 1
18
+ lr_warmup_steps = 0.1
19
+ max_bucket_reso = 2048
20
+ max_data_loader_n_workers = 0
21
+ max_grad_norm = 1
22
+ max_timestep = 1000
23
+ max_token_length = 75
24
+ max_train_epochs = 10
25
+ max_train_steps = 1600
26
+ min_bucket_reso = 64
27
+ mixed_precision = "fp16"
28
+ network_alpha = 16
29
+ network_args = []
30
+ network_dim = 32
31
+ network_module = "networks.lora"
32
+ network_train_unet_only = true
33
+ noise_offset_type = "Original"
34
+ optimizer_args = []
35
+ optimizer_type = "AdamW8bit"
36
+ output_dir = "/workspace/kohya_ss/outputs"
37
+ output_name = "last"
38
+ pretrained_model_name_or_path = "coreml-community/coreml-RealismEngineSDXL-v10_SDXL_8bit"
39
+ prior_loss_weight = 1
40
+ resolution = "512,512"
41
+ sample_prompts = "/workspace/kohya_ss/outputs/sample/prompt.txt"
42
+ sample_sampler = "euler_a"
43
+ save_every_n_epochs = 1
44
+ save_model_as = "safetensors"
45
+ save_precision = "fp16"
46
+ text_encoder_lr = []
47
+ train_batch_size = 4
48
+ train_data_dir = "/workspace/kohya_ss/img/10_Lena person/images"
49
+ unet_lr = 0.0001
50
+ wandb_run_name = "last"
51
+ xformers = true
config_lora-20250602-152812.toml ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bucket_no_upscale = true
2
+ bucket_reso_steps = 64
3
+ cache_latents = true
4
+ caption_extension = ".txt"
5
+ clip_skip = 1
6
+ dynamo_backend = "no"
7
+ enable_bucket = true
8
+ epoch = 1
9
+ gradient_accumulation_steps = 1
10
+ huber_c = 0.1
11
+ huber_scale = 1
12
+ huber_schedule = "snr"
13
+ loss_type = "l2"
14
+ lr_scheduler = "constant"
15
+ lr_scheduler_args = []
16
+ lr_scheduler_num_cycles = 1
17
+ lr_scheduler_power = 1
18
+ max_bucket_reso = 2048
19
+ max_data_loader_n_workers = 0
20
+ max_grad_norm = 1
21
+ max_timestep = 1000
22
+ max_token_length = 75
23
+ max_train_epochs = 10
24
+ max_train_steps = 1600
25
+ min_bucket_reso = 64
26
+ mixed_precision = "fp16"
27
+ network_alpha = 16
28
+ network_args = []
29
+ network_dim = 32
30
+ network_module = "networks.lora"
31
+ network_train_unet_only = true
32
+ noise_offset_type = "Original"
33
+ optimizer_args = []
34
+ optimizer_type = "AdamW8bit"
35
+ output_dir = "/workspace/kohya_ss/outputs"
36
+ output_name = "last"
37
+ pretrained_model_name_or_path = "coreml-community/coreml-RealismEngineSDXL-v10_SDXL_8bit"
38
+ prior_loss_weight = 1
39
+ resolution = "512,512"
40
+ sample_prompts = "/workspace/kohya_ss/outputs/sample/prompt.txt"
41
+ sample_sampler = "euler_a"
42
+ save_every_n_epochs = 1
43
+ save_model_as = "safetensors"
44
+ save_precision = "fp16"
45
+ text_encoder_lr = []
46
+ train_batch_size = 4
47
+ train_data_dir = "/workspace/kohya_ss/img/10_Lena person/images"
48
+ unet_lr = 0.0001
49
+ wandb_run_name = "last"
50
+ xformers = true
config_lora-20250602-152843.toml ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bucket_no_upscale = true
2
+ bucket_reso_steps = 64
3
+ cache_latents = true
4
+ caption_extension = ".txt"
5
+ clip_skip = 1
6
+ dynamo_backend = "no"
7
+ enable_bucket = true
8
+ epoch = 1
9
+ gradient_accumulation_steps = 1
10
+ huber_c = 0.1
11
+ huber_scale = 1
12
+ huber_schedule = "snr"
13
+ loss_type = "l2"
14
+ lr_scheduler = "constant"
15
+ lr_scheduler_args = []
16
+ lr_scheduler_num_cycles = 1
17
+ lr_scheduler_power = 1
18
+ max_bucket_reso = 2048
19
+ max_data_loader_n_workers = 0
20
+ max_grad_norm = 1
21
+ max_timestep = 1000
22
+ max_token_length = 75
23
+ max_train_epochs = 10
24
+ max_train_steps = 1600
25
+ min_bucket_reso = 64
26
+ mixed_precision = "fp16"
27
+ network_alpha = 16
28
+ network_args = []
29
+ network_dim = 32
30
+ network_module = "networks.lora"
31
+ network_train_unet_only = true
32
+ noise_offset_type = "Original"
33
+ optimizer_args = []
34
+ optimizer_type = "AdamW8bit"
35
+ output_dir = "/workspace/kohya_ss/outputs"
36
+ output_name = "last"
37
+ pretrained_model_name_or_path = "stabilityai/stable-diffusion-2-base"
38
+ prior_loss_weight = 1
39
+ resolution = "512,512"
40
+ sample_prompts = "/workspace/kohya_ss/outputs/sample/prompt.txt"
41
+ sample_sampler = "euler_a"
42
+ save_every_n_epochs = 1
43
+ save_model_as = "safetensors"
44
+ save_precision = "fp16"
45
+ text_encoder_lr = []
46
+ train_batch_size = 4
47
+ train_data_dir = "/workspace/kohya_ss/img/10_Lena person/images"
48
+ unet_lr = 0.0001
49
+ v2 = true
50
+ wandb_run_name = "last"
51
+ xformers = true
config_lora-20250602-153053.toml ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bucket_no_upscale = true
2
+ bucket_reso_steps = 64
3
+ cache_latents = true
4
+ caption_extension = ".txt"
5
+ clip_skip = 1
6
+ dynamo_backend = "no"
7
+ enable_bucket = true
8
+ epoch = 1
9
+ gradient_accumulation_steps = 1
10
+ huber_c = 0.1
11
+ huber_scale = 1
12
+ huber_schedule = "snr"
13
+ loss_type = "l2"
14
+ lr_scheduler = "constant"
15
+ lr_scheduler_args = []
16
+ lr_scheduler_num_cycles = 1
17
+ lr_scheduler_power = 1
18
+ max_bucket_reso = 2048
19
+ max_data_loader_n_workers = 0
20
+ max_grad_norm = 1
21
+ max_timestep = 1000
22
+ max_token_length = 75
23
+ max_train_epochs = 10
24
+ max_train_steps = 1600
25
+ min_bucket_reso = 64
26
+ mixed_precision = "fp16"
27
+ network_alpha = 16
28
+ network_args = []
29
+ network_dim = 32
30
+ network_module = "networks.lora"
31
+ network_train_unet_only = true
32
+ noise_offset_type = "Original"
33
+ optimizer_args = []
34
+ optimizer_type = "AdamW8bit"
35
+ output_dir = "/workspace/kohya_ss/outputs"
36
+ output_name = "last"
37
+ pretrained_model_name_or_path = "stabilityai/stable-diffusion-2-base"
38
+ prior_loss_weight = 1
39
+ resolution = "512,512"
40
+ sample_prompts = "/workspace/kohya_ss/outputs/sample/prompt.txt"
41
+ sample_sampler = "euler_a"
42
+ save_every_n_epochs = 1
43
+ save_model_as = "safetensors"
44
+ save_precision = "fp16"
45
+ text_encoder_lr = []
46
+ train_batch_size = 4
47
+ train_data_dir = "/workspace/kohya_ss/img"
48
+ unet_lr = 0.0001
49
+ v2 = true
50
+ wandb_run_name = "last"
51
+ xformers = true
config_lora-20250602-153250.toml ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bucket_no_upscale = true
2
+ bucket_reso_steps = 64
3
+ cache_latents = true
4
+ caption_extension = ".txt"
5
+ clip_skip = 1
6
+ dynamo_backend = "no"
7
+ enable_bucket = true
8
+ epoch = 1
9
+ gradient_accumulation_steps = 1
10
+ huber_c = 0.1
11
+ huber_scale = 1
12
+ huber_schedule = "snr"
13
+ loss_type = "l2"
14
+ lr_scheduler = "constant"
15
+ lr_scheduler_args = []
16
+ lr_scheduler_num_cycles = 1
17
+ lr_scheduler_power = 1
18
+ max_bucket_reso = 2048
19
+ max_data_loader_n_workers = 0
20
+ max_grad_norm = 1
21
+ max_timestep = 1000
22
+ max_token_length = 75
23
+ max_train_steps = 1600
24
+ min_bucket_reso = 64
25
+ mixed_precision = "fp16"
26
+ network_alpha = 16
27
+ network_args = []
28
+ network_dim = 32
29
+ network_module = "networks.lora"
30
+ network_train_unet_only = true
31
+ noise_offset_type = "Original"
32
+ optimizer_args = []
33
+ optimizer_type = "AdamW8bit"
34
+ output_dir = "/workspace/kohya_ss/outputs"
35
+ output_name = "last"
36
+ pretrained_model_name_or_path = "coreml-community/coreml-RealismEngineSDXL-v10_SDXL_8bit"
37
+ prior_loss_weight = 1
38
+ resolution = "512,512"
39
+ sample_prompts = "/workspace/kohya_ss/outputs/sample/prompt.txt"
40
+ sample_sampler = "euler_a"
41
+ save_every_n_epochs = 1
42
+ save_model_as = "safetensors"
43
+ save_precision = "fp16"
44
+ text_encoder_lr = []
45
+ train_batch_size = 4
46
+ train_data_dir = "./img"
47
+ unet_lr = 0.0001
48
+ wandb_run_name = "last"
49
+ xformers = true
config_lora-20250602-153310.toml ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bucket_no_upscale = true
2
+ bucket_reso_steps = 64
3
+ cache_latents = true
4
+ caption_extension = ".txt"
5
+ clip_skip = 1
6
+ dynamo_backend = "no"
7
+ enable_bucket = true
8
+ epoch = 1
9
+ gradient_accumulation_steps = 1
10
+ huber_c = 0.1
11
+ huber_scale = 1
12
+ huber_schedule = "snr"
13
+ loss_type = "l2"
14
+ lr_scheduler = "constant"
15
+ lr_scheduler_args = []
16
+ lr_scheduler_num_cycles = 1
17
+ lr_scheduler_power = 1
18
+ max_bucket_reso = 2048
19
+ max_data_loader_n_workers = 0
20
+ max_grad_norm = 1
21
+ max_timestep = 1000
22
+ max_token_length = 75
23
+ max_train_steps = 1600
24
+ min_bucket_reso = 64
25
+ mixed_precision = "fp16"
26
+ network_alpha = 16
27
+ network_args = []
28
+ network_dim = 32
29
+ network_module = "networks.lora"
30
+ network_train_unet_only = true
31
+ noise_offset_type = "Original"
32
+ optimizer_args = []
33
+ optimizer_type = "AdamW8bit"
34
+ output_dir = "/workspace/kohya_ss/outputs"
35
+ output_name = "last"
36
+ pretrained_model_name_or_path = "coreml-community/coreml-RealismEngineSDXL-v10_SDXL_8bit"
37
+ prior_loss_weight = 1
38
+ resolution = "512,512"
39
+ sample_prompts = "/workspace/kohya_ss/outputs/sample/prompt.txt"
40
+ sample_sampler = "euler_a"
41
+ save_every_n_epochs = 1
42
+ save_model_as = "safetensors"
43
+ save_precision = "fp16"
44
+ text_encoder_lr = []
45
+ train_batch_size = 4
46
+ train_data_dir = "./img"
47
+ unet_lr = 0.0001
48
+ wandb_run_name = "last"
49
+ xformers = true
config_lora-20250602-153339.toml ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bucket_no_upscale = true
2
+ bucket_reso_steps = 64
3
+ cache_latents = true
4
+ caption_extension = ".txt"
5
+ clip_skip = 1
6
+ dynamo_backend = "no"
7
+ enable_bucket = true
8
+ epoch = 1
9
+ gradient_accumulation_steps = 1
10
+ huber_c = 0.1
11
+ huber_scale = 1
12
+ huber_schedule = "snr"
13
+ loss_type = "l2"
14
+ lr_scheduler = "constant"
15
+ lr_scheduler_args = []
16
+ lr_scheduler_num_cycles = 1
17
+ lr_scheduler_power = 1
18
+ max_bucket_reso = 2048
19
+ max_data_loader_n_workers = 0
20
+ max_grad_norm = 1
21
+ max_timestep = 1000
22
+ max_token_length = 75
23
+ max_train_steps = 1600
24
+ min_bucket_reso = 64
25
+ mixed_precision = "fp16"
26
+ network_alpha = 16
27
+ network_args = []
28
+ network_dim = 32
29
+ network_module = "networks.lora"
30
+ network_train_unet_only = true
31
+ noise_offset_type = "Original"
32
+ optimizer_args = []
33
+ optimizer_type = "AdamW8bit"
34
+ output_dir = "/workspace/kohya_ss/outputs"
35
+ output_name = "last"
36
+ pretrained_model_name_or_path = "coreml-community/coreml-RealismEngineSDXL-v10_SDXL_8bit"
37
+ prior_loss_weight = 1
38
+ resolution = "512,512"
39
+ sample_prompts = "/workspace/kohya_ss/outputs/sample/prompt.txt"
40
+ sample_sampler = "euler_a"
41
+ save_every_n_epochs = 1
42
+ save_model_as = "safetensors"
43
+ save_precision = "fp16"
44
+ text_encoder_lr = []
45
+ train_batch_size = 4
46
+ train_data_dir = "./img"
47
+ unet_lr = 0.0001
48
+ wandb_run_name = "last"
49
+ xformers = true
config_lora-20250602-153522.toml ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bucket_no_upscale = true
2
+ bucket_reso_steps = 64
3
+ cache_latents = true
4
+ caption_extension = ".txt"
5
+ clip_skip = 1
6
+ dynamo_backend = "no"
7
+ enable_bucket = true
8
+ epoch = 1
9
+ gradient_accumulation_steps = 1
10
+ huber_c = 0.1
11
+ huber_scale = 1
12
+ huber_schedule = "snr"
13
+ loss_type = "l2"
14
+ lr_scheduler = "constant"
15
+ lr_scheduler_args = []
16
+ lr_scheduler_num_cycles = 1
17
+ lr_scheduler_power = 1
18
+ max_bucket_reso = 2048
19
+ max_data_loader_n_workers = 0
20
+ max_grad_norm = 1
21
+ max_timestep = 1000
22
+ max_token_length = 75
23
+ max_train_steps = 1600
24
+ min_bucket_reso = 64
25
+ mixed_precision = "fp16"
26
+ network_alpha = 16
27
+ network_args = []
28
+ network_dim = 32
29
+ network_module = "networks.lora"
30
+ network_train_unet_only = true
31
+ noise_offset_type = "Original"
32
+ optimizer_args = []
33
+ optimizer_type = "AdamW8bit"
34
+ output_dir = "/workspace/kohya_ss/outputs"
35
+ output_name = "last"
36
+ pretrained_model_name_or_path = "stabilityai/stable-diffusion-xl-base-1.0"
37
+ prior_loss_weight = 1
38
+ resolution = "512,512"
39
+ sample_prompts = "/workspace/kohya_ss/outputs/sample/prompt.txt"
40
+ sample_sampler = "euler_a"
41
+ save_every_n_epochs = 1
42
+ save_model_as = "safetensors"
43
+ save_precision = "fp16"
44
+ text_encoder_lr = []
45
+ train_batch_size = 4
46
+ train_data_dir = "./img"
47
+ unet_lr = 0.0001
48
+ wandb_run_name = "last"
49
+ xformers = true
config_lora-20250602-155358.toml ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bucket_no_upscale = true
2
+ bucket_reso_steps = 64
3
+ cache_latents = true
4
+ caption_extension = ".txt"
5
+ clip_skip = 1
6
+ dynamo_backend = "no"
7
+ enable_bucket = true
8
+ epoch = 1
9
+ gradient_accumulation_steps = 1
10
+ huber_c = 0.1
11
+ huber_scale = 1
12
+ huber_schedule = "snr"
13
+ loss_type = "l2"
14
+ lr_scheduler = "constant"
15
+ lr_scheduler_args = []
16
+ lr_scheduler_num_cycles = 1
17
+ lr_scheduler_power = 1
18
+ max_bucket_reso = 2048
19
+ max_data_loader_n_workers = 0
20
+ max_grad_norm = 1
21
+ max_timestep = 1000
22
+ max_token_length = 75
23
+ max_train_steps = 1600
24
+ min_bucket_reso = 64
25
+ mixed_precision = "fp16"
26
+ network_alpha = 16
27
+ network_args = []
28
+ network_dim = 32
29
+ network_module = "networks.lora"
30
+ network_train_unet_only = true
31
+ noise_offset_type = "Original"
32
+ optimizer_args = []
33
+ optimizer_type = "DAdaptAdam"
34
+ output_dir = "/workspace/kohya_ss/outputs"
35
+ output_name = "last"
36
+ pretrained_model_name_or_path = "stabilityai/stable-diffusion-xl-base-1.0"
37
+ prior_loss_weight = 1
38
+ resolution = "512,512"
39
+ sample_every_n_epochs = 1
40
+ sample_every_n_steps = 5
41
+ sample_prompts = "/workspace/kohya_ss/outputs/sample/prompt.txt"
42
+ sample_sampler = "euler_a"
43
+ save_every_n_epochs = 1
44
+ save_model_as = "safetensors"
45
+ save_precision = "bf16"
46
+ text_encoder_lr = []
47
+ train_batch_size = 4
48
+ train_data_dir = "./img"
49
+ unet_lr = 0.0001
50
+ wandb_run_name = "last"
51
+ xformers = true
img/100_Lena person/1.png ADDED

Git LFS Details

  • SHA256: 3fda3993bbc558363bef06ac70bbbef2c8e76da046a92aefe91a52c37ffea870
  • Pointer size: 131 Bytes
  • Size of remote file: 500 kB
img/100_Lena person/1.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ 1girl, solo, long hair, breasts, looking at viewer, smile, brown hair, navel, holding, cleavage, brown eyes, medium breasts, sitting, closed mouth, underwear, collarbone, outdoors, day, water, bra, tree, lips, phone, cellphone, nature, smartphone, reflection, holding phone, sports bra, realistic, selfie, river, lake
img/100_Lena person/10.png ADDED

Git LFS Details

  • SHA256: 460a006467773674d99edfca774d49f4c27dcd6fec7577835ef1f9f38220d455
  • Pointer size: 131 Bytes
  • Size of remote file: 416 kB
img/100_Lena person/10.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ 1girl, solo, long hair, breasts, looking at viewer, brown hair, cleavage, brown eyes, medium breasts, underwear, collarbone, upper body, bra, mole, lips, freckles, mole on breast, realistic, pink bra
img/100_Lena person/11.png ADDED

Git LFS Details

  • SHA256: aeadf6ee31663c40017af0b46b633c9dae959f114d73aa1fdf11f8ff80d99130
  • Pointer size: 131 Bytes
  • Size of remote file: 385 kB
img/100_Lena person/11.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ 1girl, solo, long hair, breasts, looking at viewer, brown hair, navel, holding, medium breasts, sitting, closed mouth, nipples, green eyes, collarbone, nude, small breasts, indoors, mole, lips, fingernails, chair, phone, piercing, cellphone, tan, smartphone, freckles, tanlines, holding phone, realistic, selfie, office chair
img/100_Lena person/12.png ADDED

Git LFS Details

  • SHA256: f9ad5f2237ed4af16d1348607789ec4bfa54c21d0101188abfbbaa0ee180285b
  • Pointer size: 131 Bytes
  • Size of remote file: 479 kB
img/100_Lena person/12.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ 1girl, solo, long hair, breasts, looking at viewer, brown hair, black hair, navel, cleavage, brown eyes, medium breasts, collarbone, upper body, outdoors, sky, day, midriff, mole, covered nipples, tree, lips, crop top, tank top, ground vehicle, building, motor vehicle, freckles, mole on breast, city, realistic, car, road, street
img/100_Lena person/13.png ADDED

Git LFS Details

  • SHA256: e14433500ef150f9e74513c576ca616e8ba6243f5017da24bec5c2a09b32283d
  • Pointer size: 131 Bytes
  • Size of remote file: 504 kB
img/100_Lena person/13.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ long hair, breasts, looking at viewer, multiple girls, black hair, 1boy, brown eyes, jewelry, sitting, solo focus, day, midriff, indoors, necklace, mole, cup, lips, chair, table, realistic, coffee, cafe
img/100_Lena person/14.png ADDED

Git LFS Details

  • SHA256: 50fba3330f270bab0d656ee19a510c43e80316de032fa4c50ff45c6b64912a1d
  • Pointer size: 131 Bytes
  • Size of remote file: 366 kB
img/100_Lena person/14.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ 1girl, solo, long hair, breasts, looking at viewer, brown hair, navel, holding, brown eyes, medium breasts, sitting, closed mouth, nipples, small breasts, shorts, indoors, stomach, clothes lift, lips, short shorts, no bra, arm support, shirt lift, phone, cellphone, tank top, smartphone, holding phone, realistic, nose, selfie, camouflage, barcode
img/100_Lena person/15.png ADDED

Git LFS Details

  • SHA256: 163b41aa1037e7c92b9be2c27f418fb2d760ea94bb68244daa077bb254414985
  • Pointer size: 131 Bytes
  • Size of remote file: 430 kB
img/100_Lena person/15.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ 1girl, solo, long hair, looking at viewer, smile, simple background, brown hair, black hair, green eyes, parted lips, teeth, lips, eyelashes, portrait, close-up, freckles, realistic, nose
img/100_Lena person/2.png ADDED

Git LFS Details

  • SHA256: 3fda3993bbc558363bef06ac70bbbef2c8e76da046a92aefe91a52c37ffea870
  • Pointer size: 131 Bytes
  • Size of remote file: 500 kB
img/100_Lena person/2.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ 1girl, solo, long hair, breasts, looking at viewer, smile, brown hair, navel, holding, cleavage, brown eyes, medium breasts, sitting, closed mouth, underwear, collarbone, outdoors, day, water, bra, tree, lips, phone, cellphone, nature, smartphone, reflection, holding phone, sports bra, realistic, selfie, river, lake
img/100_Lena person/3.png ADDED

Git LFS Details

  • SHA256: 920e09a1518d7e97237b197d5828df479603df2c513243060355f51d333201ea
  • Pointer size: 131 Bytes
  • Size of remote file: 364 kB
img/100_Lena person/3.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ 1girl, solo, long hair, breasts, looking at viewer, smile, brown hair, navel, brown eyes, underwear, panties, indoors, bra, mole, flat chest, covered nipples, lips, pillow, bed, on bed, sports bra, realistic, dirty
img/100_Lena person/4.png ADDED

Git LFS Details

  • SHA256: c716800381264cc8794bedd27e0bd40f8a71c185810eb7eac652186b29a4fda7
  • Pointer size: 131 Bytes
  • Size of remote file: 397 kB
img/100_Lena person/4.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ 1girl, solo, long hair, breasts, looking at viewer, smile, brown hair, navel, holding, brown eyes, sitting, closed mouth, underwear, panties, small breasts, indoors, bra, lips, phone, cellphone, smartphone, holding phone, sports bra, realistic, nose, selfie, brand name imitation, grey panties
img/100_Lena person/5.png ADDED

Git LFS Details

  • SHA256: 6766424805f335461448fc73ea5a0e4d1ac0ce45fc7267de026ce4f89c631d58
  • Pointer size: 131 Bytes
  • Size of remote file: 375 kB
img/100_Lena person/5.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ 1girl, solo, long hair, breasts, looking at viewer, brown hair, holding, brown eyes, sitting, closed mouth, collarbone, small breasts, indoors, bra, mole, flat chest, covered nipples, lips, bed, phone, cellphone, smartphone, holding phone, sports bra, sportswear, realistic, nose, selfie, blue bra
img/100_Lena person/6.png ADDED

Git LFS Details

  • SHA256: 17ce2cad431912aab93db47098b08832dde7e2600063f94b9cb12c0aa0c27c19
  • Pointer size: 131 Bytes
  • Size of remote file: 364 kB
img/100_Lena person/6.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ 1girl, solo, long hair, breasts, looking at viewer, brown hair, black hair, navel, brown eyes, sitting, underwear, collarbone, panties, small breasts, indoors, bra, flat chest, covered nipples, lips, pillow, bed, piercing, thick eyebrows, blue panties, sports bra, realistic, nose, lamp, blue bra, navel piercing, aqua panties
img/100_Lena person/7.png ADDED

Git LFS Details

  • SHA256: 1283560d722f945e39882ecf31ea752a33a07f6c3f8dd1aec39098000495f9b8
  • Pointer size: 131 Bytes
  • Size of remote file: 380 kB
img/100_Lena person/7.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ 1girl, solo, long hair, breasts, looking at viewer, smile, brown hair, black hair, navel, holding, brown eyes, underwear, panties, small breasts, indoors, bra, white panties, covered nipples, lips, phone, underwear only, cellphone, smartphone, white bra, holding phone, sports bra, realistic, nose, selfie