licyk commited on
Commit
a00372c
·
verified ·
1 Parent(s): 419a665

Upload config.yaml

Browse files
logs/20260206040621/wandb/run-20260206_040726-b1ypuzjm/files/config.yaml ADDED
@@ -0,0 +1,158 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.22.3
4
+ e:
5
+ lbu2nuaa3yreho3amdlfrp0vmcbm1k89:
6
+ args:
7
+ - --pretrained_model_name_or_path=/kaggle/sd-models/Illustrious-XL-v0.1.safetensors
8
+ - --vae=/kaggle/sd-models/sdxl_fp16_fix_vae.safetensors
9
+ - --train_data_dir=/kaggle/dataset/na_kyo_1
10
+ - --output_name=ill-xl-01-na_kyo_2
11
+ - --output_dir=/kaggle/working/model/ill-xl-01-na_kyo_2
12
+ - --wandb_run_name=ill-xl-01-na_kyo_2
13
+ - --log_tracker_name=Illustrious-XL-v0.1_LoRA
14
+ - --prior_loss_weight=1
15
+ - --resolution=1024,1024
16
+ - --enable_bucket
17
+ - --min_bucket_reso=256
18
+ - --max_bucket_reso=4096
19
+ - --bucket_reso_steps=64
20
+ - --save_model_as=safetensors
21
+ - --save_precision=fp16
22
+ - --save_every_n_epochs=1
23
+ - --max_train_epochs=2
24
+ - --train_batch_size=6
25
+ - --gradient_checkpointing
26
+ - --network_train_unet_only
27
+ - --learning_rate=0.0001
28
+ - --unet_lr=0.0003
29
+ - --text_encoder_lr=0.00001
30
+ - --lr_scheduler=constant_with_warmup
31
+ - --lr_warmup_steps=100
32
+ - --optimizer_type=Lion8bit
33
+ - --network_module=lycoris.kohya
34
+ - --network_dim=100000
35
+ - --network_alpha=100000
36
+ - --network_args
37
+ - conv_dim=100000
38
+ - conv_alpha=100000
39
+ - algo=lokr
40
+ - dropout=0
41
+ - factor=8
42
+ - train_norm=True
43
+ - preset=full
44
+ - --optimizer_args
45
+ - weight_decay=0.05
46
+ - betas=0.9,0.95
47
+ - --log_with=all
48
+ - --logging_dir=/kaggle/working/model/logs
49
+ - --caption_extension=.txt
50
+ - --shuffle_caption
51
+ - --keep_tokens=0
52
+ - --max_token_length=225
53
+ - --seed=1337
54
+ - --mixed_precision=fp16
55
+ - --xformers
56
+ - --cache_latents
57
+ - --cache_latents_to_disk
58
+ - --persistent_data_loader_workers
59
+ - --debiased_estimation_loss
60
+ - --vae_batch_size=4
61
+ - --full_fp16
62
+ codePath: sdxl_train_network.py
63
+ codePathLocal: sdxl_train_network.py
64
+ cpu_count: 2
65
+ cpu_count_logical: 4
66
+ cudaVersion: "13.0"
67
+ disk:
68
+ /:
69
+ total: "8656922775552"
70
+ used: "7138780241920"
71
+ email: 15177915035@163.com
72
+ executable: /usr/bin/python3
73
+ git:
74
+ commit: b996440c5fb89b91079dbf0a6f1ddea42efdc018
75
+ remote: https://github.com/kohya-ss/sd-scripts
76
+ gpu: Tesla T4
77
+ gpu_count: 2
78
+ gpu_nvidia:
79
+ - architecture: Turing
80
+ cudaCores: 2560
81
+ memoryTotal: "16106127360"
82
+ name: Tesla T4
83
+ uuid: GPU-b7d94609-8e99-b2cb-88e8-66a4aed3a642
84
+ - architecture: Turing
85
+ cudaCores: 2560
86
+ memoryTotal: "16106127360"
87
+ name: Tesla T4
88
+ uuid: GPU-cd4a0348-a775-0039-349d-2b37a640c5d9
89
+ host: 6cf858f42711
90
+ memory:
91
+ total: "33662472192"
92
+ os: Linux-6.6.113+-x86_64-with-glibc2.35
93
+ program: /kaggle/sd-scripts/sdxl_train_network.py
94
+ python: CPython 3.12.12
95
+ root: /kaggle/working/model/logs/20260206040621
96
+ startedAt: "2026-02-06T04:07:26.724885Z"
97
+ writerId: lbu2nuaa3yreho3amdlfrp0vmcbm1k89
98
+ m:
99
+ - "1": global_step
100
+ "6":
101
+ - 3
102
+ - 2
103
+ "7": []
104
+ - "1": epoch
105
+ "6":
106
+ - 3
107
+ - 2
108
+ "7": []
109
+ - "1": val_step
110
+ "6":
111
+ - 3
112
+ - 2
113
+ "7": []
114
+ python_version: 3.12.12
115
+ t:
116
+ "1":
117
+ - 1
118
+ - 2
119
+ - 3
120
+ - 5
121
+ - 11
122
+ - 12
123
+ - 41
124
+ - 45
125
+ - 49
126
+ - 53
127
+ - 71
128
+ - 83
129
+ - 98
130
+ - 105
131
+ "2":
132
+ - 1
133
+ - 2
134
+ - 3
135
+ - 5
136
+ - 11
137
+ - 12
138
+ - 41
139
+ - 45
140
+ - 49
141
+ - 53
142
+ - 71
143
+ - 83
144
+ - 98
145
+ - 105
146
+ "3":
147
+ - 2
148
+ - 7
149
+ - 13
150
+ - 61
151
+ - 65
152
+ "4": 3.12.12
153
+ "5": 0.22.3
154
+ "6": 4.54.1
155
+ "8":
156
+ - 2
157
+ "12": 0.22.3
158
+ "13": linux-x86_64