feat: upload matsumoto-sarina lora model
Browse files
matsumoto-sarina_config/config_file.toml
CHANGED
|
@@ -24,7 +24,7 @@ lr_warmup_steps = 3
|
|
| 24 |
debug_dataset = false
|
| 25 |
in_json = "/content/LoRA/meta_lat.json"
|
| 26 |
train_data_dir = "/content/LoRA/train_data"
|
| 27 |
-
dataset_repeats =
|
| 28 |
shuffle_caption = true
|
| 29 |
keep_tokens = 0
|
| 30 |
resolution = "512,512"
|
|
@@ -39,12 +39,12 @@ token_warmup_step = 0
|
|
| 39 |
output_dir = "/content/LoRA/output"
|
| 40 |
output_name = "matsumoto-sarina"
|
| 41 |
save_precision = "fp16"
|
| 42 |
-
save_every_n_epochs =
|
| 43 |
-
train_batch_size =
|
| 44 |
max_token_length = 225
|
| 45 |
mem_eff_attn = false
|
| 46 |
xformers = true
|
| 47 |
-
max_train_epochs =
|
| 48 |
max_data_loader_n_workers = 8
|
| 49 |
persistent_data_loader_workers = true
|
| 50 |
gradient_checkpointing = false
|
|
|
|
| 24 |
debug_dataset = false
|
| 25 |
in_json = "/content/LoRA/meta_lat.json"
|
| 26 |
train_data_dir = "/content/LoRA/train_data"
|
| 27 |
+
dataset_repeats = 27
|
| 28 |
shuffle_caption = true
|
| 29 |
keep_tokens = 0
|
| 30 |
resolution = "512,512"
|
|
|
|
| 39 |
output_dir = "/content/LoRA/output"
|
| 40 |
output_name = "matsumoto-sarina"
|
| 41 |
save_precision = "fp16"
|
| 42 |
+
save_every_n_epochs = 4
|
| 43 |
+
train_batch_size = 3
|
| 44 |
max_token_length = 225
|
| 45 |
mem_eff_attn = false
|
| 46 |
xformers = true
|
| 47 |
+
max_train_epochs = 4
|
| 48 |
max_data_loader_n_workers = 8
|
| 49 |
persistent_data_loader_workers = true
|
| 50 |
gradient_checkpointing = false
|