bigbossmonster commited on
Commit
b30e135
·
verified ·
1 Parent(s): 0f9218e

Upload folder with Python script

Browse files
ai_yuzuki/ai_yuzuki_kk,-000001.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c385489a399fc4bd301f5bf6742d113cb34ada408a4deea30fa9d48547f1fce
3
+ size 33129496
ai_yuzuki/ai_yuzuki_kk,-000002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5f5603fc5177dee35467d16fb1476003ef73dee49917b04ea1c0237c970ca7b
3
+ size 33129496
ai_yuzuki/ai_yuzuki_kk,-000003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9d9d133bbef6afdff50db86b1122ab5467fd9c6cfc46249c23decfb0119426e
3
+ size 33129496
ai_yuzuki/ai_yuzuki_kk,-000004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d3a1c82ebc4a0ec5ba8f00bbd8731a7d2f6097de1179bebe006051de7100cee
3
+ size 33129496
ai_yuzuki/ai_yuzuki_kk,-000005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76625ce74f1aac3c8ce05cfc73b2aa2e82cba2f64b03ca0a10760df4a3ff3f87
3
+ size 33129496
ai_yuzuki/ai_yuzuki_kk,-000006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1210c57e6dfe9dfaefbe20361fe4edcf4daa438a60baea7717e5ea103a090ba7
3
+ size 33129496
ai_yuzuki/ai_yuzuki_kk,-000007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cf11bc66b6c1ddd99a2be6e23faac30fc553f3ab205d46750f6f544ca55c63f
3
+ size 33129496
ai_yuzuki/ai_yuzuki_kk,-000008.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9983d81790c6c601fda246a560dbaaf5f65fa3f967042aa80408b9f7e2e00d3e
3
+ size 33129496
ai_yuzuki/ai_yuzuki_kk,-000009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa53e7f7d1bad01cf4e6804ab2b43eaec201206b66ff82b361e96a5480fd38dc
3
+ size 33129496
ai_yuzuki/ai_yuzuki_kk,.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89f894356db7b90cd54aeea7acc6c7670da3ab250d399493f73a0350e5a39042
3
+ size 33129496
ai_yuzuki/ai_yuzuki_kk,_20241201-181301.json ADDED
@@ -0,0 +1,205 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "LoRA_type": "Standard",
3
+ "LyCORIS_preset": "full",
4
+ "adaptive_noise_scale": 0,
5
+ "additional_parameters": "",
6
+ "ae": "",
7
+ "apply_t5_attn_mask": false,
8
+ "async_upload": false,
9
+ "block_alphas": "",
10
+ "block_dims": "",
11
+ "block_lr_zero_threshold": "",
12
+ "blocks_to_swap": 0,
13
+ "bucket_no_upscale": true,
14
+ "bucket_reso_steps": 64,
15
+ "bypass_mode": false,
16
+ "cache_latents": true,
17
+ "cache_latents_to_disk": false,
18
+ "caption_dropout_every_n_epochs": 0,
19
+ "caption_dropout_rate": 0,
20
+ "caption_extension": ".txt",
21
+ "clip_l": "",
22
+ "clip_skip": 1,
23
+ "color_aug": false,
24
+ "constrain": 0,
25
+ "conv_alpha": 1,
26
+ "conv_block_alphas": "",
27
+ "conv_block_dims": "",
28
+ "conv_dim": 1,
29
+ "cpu_offload_checkpointing": false,
30
+ "dataset_config": "",
31
+ "debiased_estimation_loss": false,
32
+ "decompose_both": false,
33
+ "dim_from_weights": false,
34
+ "discrete_flow_shift": 3,
35
+ "dora_wd": false,
36
+ "double_blocks_to_swap": 0,
37
+ "down_lr_weight": "",
38
+ "dynamo_backend": "no",
39
+ "dynamo_mode": "default",
40
+ "dynamo_use_dynamic": false,
41
+ "dynamo_use_fullgraph": false,
42
+ "enable_all_linear": false,
43
+ "enable_bucket": true,
44
+ "epoch": 10,
45
+ "extra_accelerate_launch_args": "",
46
+ "factor": -1,
47
+ "flip_aug": false,
48
+ "flux1_cache_text_encoder_outputs": false,
49
+ "flux1_cache_text_encoder_outputs_to_disk": false,
50
+ "flux1_checkbox": false,
51
+ "fp8_base": false,
52
+ "fp8_base_unet": false,
53
+ "full_bf16": false,
54
+ "full_fp16": false,
55
+ "gpu_ids": "",
56
+ "gradient_accumulation_steps": 1,
57
+ "gradient_checkpointing": false,
58
+ "guidance_scale": 3.5,
59
+ "highvram": false,
60
+ "huber_c": 0.1,
61
+ "huber_schedule": "snr",
62
+ "huggingface_path_in_repo": "",
63
+ "huggingface_repo_id": "",
64
+ "huggingface_repo_type": "",
65
+ "huggingface_repo_visibility": "",
66
+ "huggingface_token": "",
67
+ "img_attn_dim": "",
68
+ "img_mlp_dim": "",
69
+ "img_mod_dim": "",
70
+ "in_dims": "",
71
+ "ip_noise_gamma": 0,
72
+ "ip_noise_gamma_random_strength": false,
73
+ "keep_tokens": 0,
74
+ "learning_rate": 0.0001,
75
+ "log_config": false,
76
+ "log_tracker_config": "",
77
+ "log_tracker_name": "",
78
+ "log_with": "",
79
+ "logging_dir": "",
80
+ "loraplus_lr_ratio": 0,
81
+ "loraplus_text_encoder_lr_ratio": 0,
82
+ "loraplus_unet_lr_ratio": 0,
83
+ "loss_type": "l2",
84
+ "lowvram": false,
85
+ "lr_scheduler": "cosine_with_restarts",
86
+ "lr_scheduler_args": "",
87
+ "lr_scheduler_num_cycles": 3,
88
+ "lr_scheduler_power": 1,
89
+ "lr_scheduler_type": "",
90
+ "lr_warmup": 10,
91
+ "lr_warmup_steps": 0,
92
+ "main_process_port": 0,
93
+ "masked_loss": false,
94
+ "max_bucket_reso": 2048,
95
+ "max_data_loader_n_workers": 0,
96
+ "max_grad_norm": 1,
97
+ "max_resolution": "768,768",
98
+ "max_timestep": 1000,
99
+ "max_token_length": 75,
100
+ "max_train_epochs": 0,
101
+ "max_train_steps": 0,
102
+ "mem_eff_attn": false,
103
+ "mem_eff_save": false,
104
+ "metadata_author": "",
105
+ "metadata_description": "",
106
+ "metadata_license": "",
107
+ "metadata_tags": "",
108
+ "metadata_title": "",
109
+ "mid_lr_weight": "",
110
+ "min_bucket_reso": 256,
111
+ "min_snr_gamma": 5,
112
+ "min_timestep": 0,
113
+ "mixed_precision": "fp16",
114
+ "model_list": "custom",
115
+ "model_prediction_type": "sigma_scaled",
116
+ "module_dropout": 0,
117
+ "multi_gpu": false,
118
+ "multires_noise_discount": 0.3,
119
+ "multires_noise_iterations": 0,
120
+ "network_alpha": 1,
121
+ "network_dim": 32,
122
+ "network_dropout": 0,
123
+ "network_weights": "",
124
+ "noise_offset": 0.1,
125
+ "noise_offset_random_strength": false,
126
+ "noise_offset_type": "Original",
127
+ "num_cpu_threads_per_process": 2,
128
+ "num_machines": 1,
129
+ "num_processes": 1,
130
+ "optimizer": "AdamW8bit",
131
+ "optimizer_args": "",
132
+ "output_dir": "/content/drive/MyDrive/lora_dec1",
133
+ "output_name": "ai_yuzuki_kk,",
134
+ "persistent_data_loader_workers": false,
135
+ "pretrained_model_name_or_path": "runwayml/stable-diffusion-v1-5",
136
+ "prior_loss_weight": 1,
137
+ "random_crop": false,
138
+ "rank_dropout": 0,
139
+ "rank_dropout_scale": false,
140
+ "reg_data_dir": "",
141
+ "rescaled": false,
142
+ "resume": "",
143
+ "resume_from_huggingface": "",
144
+ "sample_every_n_epochs": 0,
145
+ "sample_every_n_steps": 0,
146
+ "sample_prompts": "",
147
+ "sample_sampler": "euler_a",
148
+ "save_every_n_epochs": 1,
149
+ "save_every_n_steps": 0,
150
+ "save_last_n_epochs": 0,
151
+ "save_last_n_epochs_state": 0,
152
+ "save_last_n_steps": 0,
153
+ "save_last_n_steps_state": 0,
154
+ "save_model_as": "safetensors",
155
+ "save_precision": "fp16",
156
+ "save_state": false,
157
+ "save_state_on_train_end": false,
158
+ "save_state_to_huggingface": false,
159
+ "scale_v_pred_loss_like_noise_pred": false,
160
+ "scale_weight_norms": 0,
161
+ "sdxl": false,
162
+ "sdxl_cache_text_encoder_outputs": false,
163
+ "sdxl_no_half_vae": false,
164
+ "seed": 0,
165
+ "shuffle_caption": true,
166
+ "single_blocks_to_swap": 0,
167
+ "single_dim": "",
168
+ "single_mod_dim": "",
169
+ "skip_cache_check": false,
170
+ "split_mode": false,
171
+ "split_qkv": false,
172
+ "stop_text_encoder_training": 0,
173
+ "t5xxl": "",
174
+ "t5xxl_lr": 0,
175
+ "t5xxl_max_token_length": 512,
176
+ "text_encoder_lr": 5e-05,
177
+ "timestep_sampling": "sigma",
178
+ "train_batch_size": 4,
179
+ "train_blocks": "all",
180
+ "train_data_dir": "/content/img",
181
+ "train_double_block_indices": "all",
182
+ "train_norm": false,
183
+ "train_on_input": true,
184
+ "train_single_block_indices": "all",
185
+ "train_t5xxl": false,
186
+ "training_comment": "",
187
+ "txt_attn_dim": "",
188
+ "txt_mlp_dim": "",
189
+ "txt_mod_dim": "",
190
+ "unet_lr": 0.0005,
191
+ "unit": 1,
192
+ "up_lr_weight": "",
193
+ "use_cp": false,
194
+ "use_scalar": false,
195
+ "use_tucker": false,
196
+ "v2": false,
197
+ "v_parameterization": false,
198
+ "v_pred_like_loss": 0,
199
+ "vae": "",
200
+ "vae_batch_size": 0,
201
+ "wandb_api_key": "",
202
+ "wandb_run_name": "",
203
+ "weighted_captions": false,
204
+ "xformers": "xformers"
205
+ }
ai_yuzuki/config_lora-20241201-181301.toml ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bucket_no_upscale = true
2
+ bucket_reso_steps = 64
3
+ cache_latents = true
4
+ caption_extension = ".txt"
5
+ clip_skip = 1
6
+ dynamo_backend = "no"
7
+ enable_bucket = true
8
+ epoch = 10
9
+ gradient_accumulation_steps = 1
10
+ huber_c = 0.1
11
+ huber_schedule = "snr"
12
+ loss_type = "l2"
13
+ lr_scheduler = "cosine_with_restarts"
14
+ lr_scheduler_args = []
15
+ lr_scheduler_num_cycles = 3
16
+ lr_scheduler_power = 1
17
+ lr_warmup_steps = 0.1
18
+ max_bucket_reso = 2048
19
+ max_data_loader_n_workers = 0
20
+ max_grad_norm = 1
21
+ max_timestep = 1000
22
+ max_token_length = 75
23
+ max_train_steps = 1625
24
+ min_bucket_reso = 256
25
+ min_snr_gamma = 5
26
+ mixed_precision = "fp16"
27
+ network_alpha = 1
28
+ network_args = []
29
+ network_dim = 32
30
+ network_module = "networks.lora"
31
+ noise_offset = 0.1
32
+ noise_offset_type = "Original"
33
+ optimizer_args = []
34
+ optimizer_type = "AdamW8bit"
35
+ output_dir = "/content/drive/MyDrive/lora_dec1"
36
+ output_name = "ai_yuzuki_kk,"
37
+ pretrained_model_name_or_path = "runwayml/stable-diffusion-v1-5"
38
+ prior_loss_weight = 1
39
+ resolution = "768,768"
40
+ sample_prompts = "/content/drive/MyDrive/lora_dec1/sample/prompt.txt"
41
+ sample_sampler = "euler_a"
42
+ save_every_n_epochs = 1
43
+ save_model_as = "safetensors"
44
+ save_precision = "fp16"
45
+ shuffle_caption = true
46
+ text_encoder_lr = [ 5e-5, 5e-5,]
47
+ train_batch_size = 4
48
+ train_data_dir = "/content/img"
49
+ unet_lr = 0.0005
50
+ wandb_run_name = "ai_yuzuki_kk,"
51
+ xformers = true
ai_yuzuki/sample/prompt.txt ADDED
File without changes
amin_niina/amin_niina_kk-000001.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b87d3b157c9fb09763487ae05793f75a45d49f123da252ac368ef0f222ad59c0
3
+ size 37866880
amin_niina/amin_niina_kk-000002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc6ed2890566365a760427fda84f70e4d6a39f9fd385a565f42f30904dd78cb0
3
+ size 37866880
amin_niina/amin_niina_kk-000003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:869cc416c96dd42ed9c14917aa54c0079f97912a7fe03723f0fb7c9f613fc525
3
+ size 37866880
amin_niina/amin_niina_kk-000004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f397b5db8f8562f167de964d49f6479a9ca1d5b2a29c365b849c01d490a6a03f
3
+ size 37866880
amin_niina/amin_niina_kk-000005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69128cfd4f03b05b9f5809796edaf06f769a75fb187f61a9204f14ce5eae1409
3
+ size 37866880
amin_niina/amin_niina_kk-000006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:373ad7bbb2d976d24a72169483878f1675b256b22e87f8f35f2dfb04b28adc28
3
+ size 37866880
amin_niina/amin_niina_kk-000007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0dfb3d45a8fe168792e3c5686f2f1630ca54bdee58ff33cd30e0a7c363400480
3
+ size 37866880
amin_niina/amin_niina_kk-000008.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b92091792840a6a928ad26007a7784da994c7475707a826255bf682cd05633b
3
+ size 37866880
amin_niina/amin_niina_kk-000009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2132941c182d843c399c3fc7f8a9545581478c2130b7f668d1b84355d04dd43f
3
+ size 37866880
amin_niina/amin_niina_kk.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72b74369e8b6249098d68afb8eb1a45c604cdea32f6cf06d4ee0afd34bdaa331
3
+ size 37866880
amin_niina/amin_niina_kk_20241014-194038.json ADDED
@@ -0,0 +1,159 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "LoRA_type": "Standard",
3
+ "LyCORIS_preset": "full",
4
+ "adaptive_noise_scale": 0,
5
+ "additional_parameters": "",
6
+ "async_upload": false,
7
+ "block_alphas": "",
8
+ "block_dims": "",
9
+ "block_lr_zero_threshold": "",
10
+ "bucket_no_upscale": true,
11
+ "bucket_reso_steps": 64,
12
+ "bypass_mode": false,
13
+ "cache_latents": true,
14
+ "cache_latents_to_disk": false,
15
+ "caption_dropout_every_n_epochs": 0,
16
+ "caption_dropout_rate": 0,
17
+ "caption_extension": ".txt",
18
+ "clip_skip": 1,
19
+ "color_aug": false,
20
+ "constrain": 0,
21
+ "conv_alpha": 1,
22
+ "conv_block_alphas": "",
23
+ "conv_block_dims": "",
24
+ "conv_dim": 1,
25
+ "dataset_config": "",
26
+ "debiased_estimation_loss": false,
27
+ "decompose_both": false,
28
+ "dim_from_weights": false,
29
+ "dora_wd": false,
30
+ "down_lr_weight": "",
31
+ "dynamo_backend": "no",
32
+ "dynamo_mode": "default",
33
+ "dynamo_use_dynamic": false,
34
+ "dynamo_use_fullgraph": false,
35
+ "enable_bucket": true,
36
+ "epoch": 10,
37
+ "extra_accelerate_launch_args": "",
38
+ "factor": -1,
39
+ "flip_aug": false,
40
+ "fp8_base": false,
41
+ "full_bf16": false,
42
+ "full_fp16": false,
43
+ "gpu_ids": "",
44
+ "gradient_accumulation_steps": 1,
45
+ "gradient_checkpointing": false,
46
+ "huber_c": 0.1,
47
+ "huber_schedule": "snr",
48
+ "huggingface_path_in_repo": "",
49
+ "huggingface_repo_id": "",
50
+ "huggingface_repo_type": "",
51
+ "huggingface_repo_visibility": "",
52
+ "huggingface_token": "",
53
+ "ip_noise_gamma": 0,
54
+ "ip_noise_gamma_random_strength": false,
55
+ "keep_tokens": 0,
56
+ "learning_rate": 0.0001,
57
+ "log_tracker_config": "",
58
+ "log_tracker_name": "",
59
+ "log_with": "",
60
+ "logging_dir": "",
61
+ "loss_type": "l2",
62
+ "lr_scheduler": "cosine_with_restarts",
63
+ "lr_scheduler_args": "",
64
+ "lr_scheduler_num_cycles": 3,
65
+ "lr_scheduler_power": 1,
66
+ "lr_warmup": 10,
67
+ "main_process_port": 0,
68
+ "masked_loss": false,
69
+ "max_bucket_reso": 2048,
70
+ "max_data_loader_n_workers": 0,
71
+ "max_grad_norm": 1,
72
+ "max_resolution": "768,768",
73
+ "max_timestep": 1000,
74
+ "max_token_length": 75,
75
+ "max_train_epochs": 0,
76
+ "max_train_steps": 0,
77
+ "mem_eff_attn": false,
78
+ "metadata_author": "",
79
+ "metadata_description": "",
80
+ "metadata_license": "",
81
+ "metadata_tags": "",
82
+ "metadata_title": "",
83
+ "mid_lr_weight": "",
84
+ "min_bucket_reso": 256,
85
+ "min_snr_gamma": 5,
86
+ "min_timestep": 0,
87
+ "mixed_precision": "fp16",
88
+ "model_list": "custom",
89
+ "module_dropout": 0,
90
+ "multi_gpu": false,
91
+ "multires_noise_discount": 0.3,
92
+ "multires_noise_iterations": 0,
93
+ "network_alpha": 1,
94
+ "network_dim": 32,
95
+ "network_dropout": 0,
96
+ "network_weights": "",
97
+ "noise_offset": 0.1,
98
+ "noise_offset_random_strength": false,
99
+ "noise_offset_type": "Original",
100
+ "num_cpu_threads_per_process": 2,
101
+ "num_machines": 1,
102
+ "num_processes": 1,
103
+ "optimizer": "AdamW8bit",
104
+ "optimizer_args": "",
105
+ "output_dir": "/content/drive/MyDrive/lora_nina",
106
+ "output_name": "amin_niina_kk",
107
+ "persistent_data_loader_workers": false,
108
+ "pretrained_model_name_or_path": "runwayml/stable-diffusion-v1-5",
109
+ "prior_loss_weight": 1,
110
+ "random_crop": false,
111
+ "rank_dropout": 0,
112
+ "rank_dropout_scale": false,
113
+ "reg_data_dir": "",
114
+ "rescaled": false,
115
+ "resume": "",
116
+ "resume_from_huggingface": "",
117
+ "sample_every_n_epochs": 0,
118
+ "sample_every_n_steps": 0,
119
+ "sample_prompts": "",
120
+ "sample_sampler": "euler_a",
121
+ "save_every_n_epochs": 1,
122
+ "save_every_n_steps": 0,
123
+ "save_last_n_steps": 0,
124
+ "save_last_n_steps_state": 0,
125
+ "save_model_as": "safetensors",
126
+ "save_precision": "fp16",
127
+ "save_state": false,
128
+ "save_state_on_train_end": false,
129
+ "save_state_to_huggingface": false,
130
+ "scale_v_pred_loss_like_noise_pred": false,
131
+ "scale_weight_norms": 0,
132
+ "sdxl": false,
133
+ "sdxl_cache_text_encoder_outputs": false,
134
+ "sdxl_no_half_vae": false,
135
+ "seed": 0,
136
+ "shuffle_caption": true,
137
+ "stop_text_encoder_training_pct": 0,
138
+ "text_encoder_lr": 5e-05,
139
+ "train_batch_size": 4,
140
+ "train_data_dir": "/content/img",
141
+ "train_norm": false,
142
+ "train_on_input": true,
143
+ "training_comment": "",
144
+ "unet_lr": 0.0005,
145
+ "unit": 1,
146
+ "up_lr_weight": "",
147
+ "use_cp": false,
148
+ "use_scalar": false,
149
+ "use_tucker": false,
150
+ "v2": false,
151
+ "v_parameterization": false,
152
+ "v_pred_like_loss": 0,
153
+ "vae": "",
154
+ "vae_batch_size": 0,
155
+ "wandb_api_key": "",
156
+ "wandb_run_name": "",
157
+ "weighted_captions": false,
158
+ "xformers": "xformers"
159
+ }
amin_niina/config_lora-20241014-194038.toml ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bucket_no_upscale = true
2
+ bucket_reso_steps = 64
3
+ cache_latents = true
4
+ caption_extension = ".txt"
5
+ clip_skip = 1
6
+ dynamo_backend = "no"
7
+ enable_bucket = true
8
+ epoch = 10
9
+ gradient_accumulation_steps = 1
10
+ huber_c = 0.1
11
+ huber_schedule = "snr"
12
+ learning_rate = 0.0001
13
+ loss_type = "l2"
14
+ lr_scheduler = "cosine_with_restarts"
15
+ lr_scheduler_args = []
16
+ lr_scheduler_num_cycles = 3
17
+ lr_scheduler_power = 1
18
+ lr_warmup_steps = 68
19
+ max_bucket_reso = 2048
20
+ max_data_loader_n_workers = 0
21
+ max_grad_norm = 1
22
+ max_timestep = 1000
23
+ max_token_length = 75
24
+ max_train_steps = 675
25
+ min_bucket_reso = 256
26
+ min_snr_gamma = 5
27
+ mixed_precision = "fp16"
28
+ multires_noise_discount = 0.3
29
+ network_alpha = 1
30
+ network_args = []
31
+ network_dim = 32
32
+ network_module = "networks.lora"
33
+ noise_offset = 0.1
34
+ noise_offset_type = "Original"
35
+ optimizer_args = []
36
+ optimizer_type = "AdamW8bit"
37
+ output_dir = "/content/drive/MyDrive/lora_nina"
38
+ output_name = "amin_niina_kk"
39
+ pretrained_model_name_or_path = "runwayml/stable-diffusion-v1-5"
40
+ prior_loss_weight = 1
41
+ resolution = "768,768"
42
+ sample_prompts = "/content/drive/MyDrive/lora_nina/prompt.txt"
43
+ sample_sampler = "euler_a"
44
+ save_every_n_epochs = 1
45
+ save_model_as = "safetensors"
46
+ save_precision = "fp16"
47
+ shuffle_caption = true
48
+ text_encoder_lr = 5e-5
49
+ train_batch_size = 4
50
+ train_data_dir = "/content/img"
51
+ unet_lr = 0.0005
52
+ xformers = true
amin_niina/prompt.txt ADDED
File without changes