ohcaidek commited on
Commit
f9f13d2
·
verified ·
1 Parent(s): 7c68174

Upload folder using huggingface_hub

Browse files
.job_config.json ADDED
@@ -0,0 +1,122 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "job": "extension",
3
+ "config": {
4
+ "name": "anw0ll",
5
+ "process": [
6
+ {
7
+ "type": "diffusion_trainer",
8
+ "training_folder": "/workspace/ai-toolkit/output",
9
+ "sqlite_db_path": "/workspace/ai-toolkit/aitk_db.db",
10
+ "device": "cuda",
11
+ "trigger_word": null,
12
+ "performance_log_every": 10,
13
+ "network": {
14
+ "type": "lora",
15
+ "linear": 16,
16
+ "linear_alpha": 16,
17
+ "conv": 16,
18
+ "conv_alpha": 16,
19
+ "lokr_full_rank": true,
20
+ "lokr_factor": -1,
21
+ "network_kwargs": {
22
+ "ignore_if_contains": []
23
+ }
24
+ },
25
+ "save": {
26
+ "dtype": "bf16",
27
+ "save_every": 250,
28
+ "max_step_saves_to_keep": 8,
29
+ "save_format": "diffusers",
30
+ "push_to_hub": false
31
+ },
32
+ "datasets": [
33
+ {
34
+ "folder_path": "/workspace/ai-toolkit/datasets/anw0ll",
35
+ "mask_path": null,
36
+ "mask_min_value": 0.1,
37
+ "default_caption": "",
38
+ "caption_ext": "txt",
39
+ "caption_dropout_rate": 0.05,
40
+ "cache_latents_to_disk": false,
41
+ "is_reg": false,
42
+ "network_weight": 1,
43
+ "resolution": [
44
+ 512,
45
+ 768,
46
+ 1024
47
+ ],
48
+ "controls": [],
49
+ "shrink_video_to_frames": true,
50
+ "num_frames": 1,
51
+ "do_i2v": true,
52
+ "flip_x": false,
53
+ "flip_y": false
54
+ }
55
+ ],
56
+ "train": {
57
+ "batch_size": 1,
58
+ "bypass_guidance_embedding": false,
59
+ "steps": 3000,
60
+ "gradient_accumulation": 1,
61
+ "train_unet": true,
62
+ "train_text_encoder": false,
63
+ "gradient_checkpointing": true,
64
+ "noise_scheduler": "flowmatch",
65
+ "optimizer": "adamw8bit",
66
+ "timestep_type": "sigmoid",
67
+ "content_or_style": "balanced",
68
+ "optimizer_params": {
69
+ "weight_decay": 0.0001
70
+ },
71
+ "unload_text_encoder": false,
72
+ "cache_text_embeddings": true,
73
+ "lr": 0.0002,
74
+ "ema_config": {
75
+ "use_ema": false,
76
+ "ema_decay": 0.99
77
+ },
78
+ "skip_first_sample": false,
79
+ "force_first_sample": false,
80
+ "disable_sampling": false,
81
+ "dtype": "bf16",
82
+ "diff_output_preservation": false,
83
+ "diff_output_preservation_multiplier": 1,
84
+ "diff_output_preservation_class": "person",
85
+ "switch_boundary_every": 1,
86
+ "loss_type": "mse"
87
+ },
88
+ "model": {
89
+ "name_or_path": "Qwen/Qwen-Image",
90
+ "quantize": true,
91
+ "qtype": "uint6",
92
+ "quantize_te": true,
93
+ "qtype_te": "qfloat8",
94
+ "arch": "qwen_image",
95
+ "low_vram": true,
96
+ "model_kwargs": {},
97
+ "layer_offloading": false,
98
+ "layer_offloading_text_encoder_percent": 1,
99
+ "layer_offloading_transformer_percent": 1
100
+ },
101
+ "sample": {
102
+ "sampler": "flowmatch",
103
+ "sample_every": 250,
104
+ "width": 1024,
105
+ "height": 1024,
106
+ "samples": [],
107
+ "neg": "",
108
+ "seed": 42,
109
+ "walk_seed": true,
110
+ "guidance_scale": 4,
111
+ "sample_steps": 25,
112
+ "num_frames": 1,
113
+ "fps": 1
114
+ }
115
+ }
116
+ ]
117
+ },
118
+ "meta": {
119
+ "name": "[name]",
120
+ "version": "1.0"
121
+ }
122
+ }
anw0ll.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b602bb6dd95f679946b1e0c0f978e5f45c3ca78b438e531dc590ea69499b221
3
+ size 295146160
anw0ll_000002000.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d22562e241b243e80ea6c9b88aeb08f6654ae9cfbe562715ce097da41e5d6f8
3
+ size 295146160
anw0ll_000002250.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ec61190a0372eb3439200f522d9c649b47a33ac10c7d335b875504ff01b8a5d
3
+ size 295146160
anw0ll_000002500.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:03e0f62ec55e18082d7d578b83205d79508e69cdf3ab973cc72e0c03c1ca36c3
3
+ size 295146160
anw0ll_000002750.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa9a0f94de14a28dafb3771f3d74ec4e5dae36b1899e6cddabd79ac33bfa2905
3
+ size 295146160
config.yaml ADDED
@@ -0,0 +1,103 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ job: extension
2
+ config:
3
+ name: anw0ll
4
+ process:
5
+ - type: diffusion_trainer
6
+ training_folder: /workspace/ai-toolkit/output
7
+ sqlite_db_path: /workspace/ai-toolkit/aitk_db.db
8
+ device: cuda
9
+ trigger_word: null
10
+ performance_log_every: 10
11
+ network:
12
+ type: lora
13
+ linear: 16
14
+ linear_alpha: 16
15
+ conv: 16
16
+ conv_alpha: 16
17
+ lokr_full_rank: true
18
+ lokr_factor: -1
19
+ network_kwargs:
20
+ ignore_if_contains: []
21
+ save:
22
+ dtype: bf16
23
+ save_every: 250
24
+ max_step_saves_to_keep: 8
25
+ save_format: diffusers
26
+ push_to_hub: false
27
+ datasets:
28
+ - folder_path: /workspace/ai-toolkit/datasets/anw0ll
29
+ mask_path: null
30
+ mask_min_value: 0.1
31
+ default_caption: ''
32
+ caption_ext: txt
33
+ caption_dropout_rate: 0.05
34
+ cache_latents_to_disk: false
35
+ is_reg: false
36
+ network_weight: 1
37
+ resolution:
38
+ - 512
39
+ - 768
40
+ - 1024
41
+ controls: []
42
+ shrink_video_to_frames: true
43
+ num_frames: 1
44
+ do_i2v: true
45
+ flip_x: false
46
+ flip_y: false
47
+ train:
48
+ batch_size: 1
49
+ bypass_guidance_embedding: false
50
+ steps: 3000
51
+ gradient_accumulation: 1
52
+ train_unet: true
53
+ train_text_encoder: false
54
+ gradient_checkpointing: true
55
+ noise_scheduler: flowmatch
56
+ optimizer: adamw8bit
57
+ timestep_type: sigmoid
58
+ content_or_style: balanced
59
+ optimizer_params:
60
+ weight_decay: 0.0001
61
+ unload_text_encoder: false
62
+ cache_text_embeddings: true
63
+ lr: 0.0002
64
+ ema_config:
65
+ use_ema: false
66
+ ema_decay: 0.99
67
+ skip_first_sample: false
68
+ force_first_sample: false
69
+ disable_sampling: false
70
+ dtype: bf16
71
+ diff_output_preservation: false
72
+ diff_output_preservation_multiplier: 1
73
+ diff_output_preservation_class: person
74
+ switch_boundary_every: 1
75
+ loss_type: mse
76
+ model:
77
+ name_or_path: Qwen/Qwen-Image
78
+ quantize: true
79
+ qtype: uint6
80
+ quantize_te: true
81
+ qtype_te: qfloat8
82
+ arch: qwen_image
83
+ low_vram: true
84
+ model_kwargs: {}
85
+ layer_offloading: false
86
+ layer_offloading_text_encoder_percent: 1
87
+ layer_offloading_transformer_percent: 1
88
+ sample:
89
+ sampler: flowmatch
90
+ sample_every: 250
91
+ width: 1024
92
+ height: 1024
93
+ samples: []
94
+ neg: ''
95
+ seed: 42
96
+ walk_seed: true
97
+ guidance_scale: 4
98
+ sample_steps: 25
99
+ num_frames: 1
100
+ fps: 1
101
+ meta:
102
+ name: anw0ll
103
+ version: '1.0'
log.txt ADDED
The diff for this file is too large to render. See raw diff
 
logs/0_log.txt ADDED
@@ -0,0 +1,206 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0
  0%| | 0/60 [00:00<?, ?it/s]
1
  0%| | 0/60 [00:00<?, ?it/s]
2
  2%|1 | 1/60 [00:00<00:48, 1.22it/s]
3
  2%|1 | 1/60 [00:00<00:48, 1.22it/s]
4
  3%|3 | 2/60 [00:01<00:34, 1.70it/s]
5
  3%|3 | 2/60 [00:01<00:34, 1.70it/s]
6
  5%|5 | 3/60 [00:01<00:29, 1.96it/s]
7
  5%|5 | 3/60 [00:01<00:29, 1.96it/s]
8
  7%|6 | 4/60 [00:02<00:26, 2.15it/s]
9
  7%|6 | 4/60 [00:02<00:26, 2.15it/s]
10
  8%|8 | 5/60 [00:02<00:22, 2.47it/s]
11
  8%|8 | 5/60 [00:02<00:22, 2.47it/s]
12
  10%|# | 6/60 [00:02<00:19, 2.73it/s]
13
  10%|# | 6/60 [00:02<00:19, 2.73it/s]
14
  12%|#1 | 7/60 [00:02<00:18, 2.92it/s]
15
  12%|#1 | 7/60 [00:02<00:18, 2.92it/s]
16
  13%|#3 | 8/60 [00:03<00:17, 3.05it/s]
17
  13%|#3 | 8/60 [00:03<00:17, 3.05it/s]
18
  15%|#5 | 9/60 [00:03<00:16, 3.15it/s]
19
  15%|#5 | 9/60 [00:03<00:16, 3.15it/s]
20
  17%|#6 | 10/60 [00:03<00:15, 3.24it/s]
21
  17%|#6 | 10/60 [00:03<00:15, 3.24it/s]
22
  18%|#8 | 11/60 [00:04<00:14, 3.30it/s]
23
  18%|#8 | 11/60 [00:04<00:14, 3.30it/s]
24
  20%|## | 12/60 [00:04<00:15, 3.00it/s]
25
  20%|## | 12/60 [00:04<00:15, 3.00it/s]
26
  22%|##1 | 13/60 [00:04<00:16, 2.87it/s]
27
  22%|##1 | 13/60 [00:04<00:16, 2.87it/s]
28
  23%|##3 | 14/60 [00:05<00:16, 2.82it/s]
29
  23%|##3 | 14/60 [00:05<00:16, 2.82it/s]
30
  25%|##5 | 15/60 [00:06<00:21, 2.08it/s]
31
  25%|##5 | 15/60 [00:06<00:21, 2.08it/s]
32
  27%|##6 | 16/60 [00:06<00:19, 2.22it/s]
33
  27%|##6 | 16/60 [00:06<00:19, 2.22it/s]
34
  28%|##8 | 17/60 [00:06<00:18, 2.30it/s]
35
  28%|##8 | 17/60 [00:06<00:18, 2.30it/s]
36
  30%|### | 18/60 [00:07<00:17, 2.34it/s]
37
  30%|### | 18/60 [00:07<00:17, 2.34it/s]
38
  32%|###1 | 19/60 [00:07<00:17, 2.38it/s]
39
  32%|###1 | 19/60 [00:07<00:17, 2.38it/s]
40
  33%|###3 | 20/60 [00:08<00:16, 2.40it/s]
41
  33%|###3 | 20/60 [00:08<00:16, 2.40it/s]
42
  35%|###5 | 21/60 [00:08<00:16, 2.43it/s]
43
  35%|###5 | 21/60 [00:08<00:16, 2.43it/s]
44
  37%|###6 | 22/60 [00:09<00:20, 1.87it/s]
45
  37%|###6 | 22/60 [00:09<00:20, 1.87it/s]
46
  38%|###8 | 23/60 [00:09<00:18, 1.96it/s]
47
  38%|###8 | 23/60 [00:09<00:18, 1.96it/s]
48
  40%|#### | 24/60 [00:10<00:16, 2.12it/s]
49
  40%|#### | 24/60 [00:10<00:16, 2.12it/s]
50
  42%|####1 | 25/60 [00:10<00:15, 2.26it/s]
51
  42%|####1 | 25/60 [00:10<00:15, 2.26it/s]
52
  43%|####3 | 26/60 [00:10<00:14, 2.35it/s]
53
  43%|####3 | 26/60 [00:10<00:14, 2.35it/s]
54
  45%|####5 | 27/60 [00:11<00:13, 2.39it/s]
55
  45%|####5 | 27/60 [00:11<00:13, 2.39it/s]
56
  47%|####6 | 28/60 [00:11<00:13, 2.38it/s]
57
  47%|####6 | 28/60 [00:11<00:13, 2.38it/s]
58
  48%|####8 | 29/60 [00:12<00:13, 2.36it/s]
59
  48%|####8 | 29/60 [00:12<00:13, 2.36it/s]
60
  50%|##### | 30/60 [00:12<00:15, 1.97it/s]
61
  50%|##### | 30/60 [00:12<00:15, 1.97it/s]
62
  52%|#####1 | 31/60 [00:13<00:13, 2.22it/s]
63
  52%|#####1 | 31/60 [00:13<00:13, 2.22it/s]
64
  53%|#####3 | 32/60 [00:13<00:11, 2.41it/s]
65
  53%|#####3 | 32/60 [00:13<00:11, 2.41it/s]
66
  55%|#####5 | 33/60 [00:13<00:10, 2.58it/s]
67
  55%|#####5 | 33/60 [00:13<00:10, 2.58it/s]
68
  57%|#####6 | 34/60 [00:14<00:09, 2.69it/s]
69
  57%|#####6 | 34/60 [00:14<00:09, 2.69it/s]
70
  58%|#####8 | 35/60 [00:14<00:08, 2.84it/s]
71
  58%|#####8 | 35/60 [00:14<00:08, 2.84it/s]
72
  60%|###### | 36/60 [00:14<00:08, 2.73it/s]
73
  60%|###### | 36/60 [00:14<00:08, 2.73it/s]
74
  62%|######1 | 37/60 [00:15<00:11, 2.01it/s]
75
  62%|######1 | 37/60 [00:15<00:11, 2.01it/s]
76
  63%|######3 | 38/60 [00:16<00:10, 2.10it/s]
77
  63%|######3 | 38/60 [00:16<00:10, 2.10it/s]
78
  65%|######5 | 39/60 [00:16<00:09, 2.14it/s]
79
  65%|######5 | 39/60 [00:16<00:09, 2.14it/s]
80
  67%|######6 | 40/60 [00:16<00:09, 2.21it/s]
81
  67%|######6 | 40/60 [00:16<00:09, 2.21it/s]
82
  68%|######8 | 41/60 [00:17<00:08, 2.30it/s]
83
  68%|######8 | 41/60 [00:17<00:08, 2.30it/s]
84
  70%|####### | 42/60 [00:17<00:07, 2.33it/s]
85
  70%|####### | 42/60 [00:17<00:07, 2.33it/s]
86
  72%|#######1 | 43/60 [00:18<00:07, 2.38it/s]
87
  72%|#######1 | 43/60 [00:18<00:07, 2.38it/s]
88
  73%|#######3 | 44/60 [00:18<00:08, 1.85it/s]
89
  73%|#######3 | 44/60 [00:18<00:08, 1.85it/s]
90
  75%|#######5 | 45/60 [00:19<00:07, 2.00it/s]
91
  75%|#######5 | 45/60 [00:19<00:07, 2.00it/s]
92
  77%|#######6 | 46/60 [00:19<00:06, 2.08it/s]
93
  77%|#######6 | 46/60 [00:19<00:06, 2.08it/s]
94
  78%|#######8 | 47/60 [00:20<00:06, 2.15it/s]
95
  78%|#######8 | 47/60 [00:20<00:06, 2.15it/s]
96
  80%|######## | 48/60 [00:20<00:05, 2.24it/s]
97
  80%|######## | 48/60 [00:20<00:05, 2.24it/s]
98
  82%|########1 | 49/60 [00:21<00:04, 2.32it/s]
99
  82%|########1 | 49/60 [00:21<00:04, 2.32it/s]
100
  83%|########3 | 50/60 [00:21<00:04, 2.37it/s]
101
  83%|########3 | 50/60 [00:21<00:04, 2.37it/s]
102
  85%|########5 | 51/60 [00:22<00:04, 1.84it/s]
103
  85%|########5 | 51/60 [00:22<00:04, 1.84it/s]
104
  87%|########6 | 52/60 [00:22<00:04, 1.96it/s]
105
  87%|########6 | 52/60 [00:22<00:04, 1.96it/s]
106
  88%|########8 | 53/60 [00:23<00:03, 2.08it/s]
107
  88%|########8 | 53/60 [00:23<00:03, 2.08it/s]
108
  90%|######### | 54/60 [00:23<00:02, 2.31it/s]
109
  90%|######### | 54/60 [00:23<00:02, 2.31it/s]
110
  92%|#########1| 55/60 [00:23<00:01, 2.53it/s]
111
  92%|#########1| 55/60 [00:23<00:01, 2.53it/s]
112
  93%|#########3| 56/60 [00:24<00:01, 2.71it/s]
113
  93%|#########3| 56/60 [00:24<00:01, 2.71it/s]
114
  95%|#########5| 57/60 [00:24<00:01, 2.85it/s]
115
  95%|#########5| 57/60 [00:24<00:01, 2.85it/s]
116
  97%|#########6| 58/60 [00:24<00:00, 2.95it/s]
117
  97%|#########6| 58/60 [00:24<00:00, 2.95it/s]
118
  98%|#########8| 59/60 [00:25<00:00, 2.38it/s]
119
  98%|#########8| 59/60 [00:25<00:00, 2.38it/s]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Running 1 job
2
+ {
3
+ "type": "diffusion_trainer",
4
+ "training_folder": "/workspace/ai-toolkit/output",
5
+ "sqlite_db_path": "/workspace/ai-toolkit/aitk_db.db",
6
+ "device": "cuda",
7
+ "trigger_word": null,
8
+ "performance_log_every": 10,
9
+ "network": {
10
+ "type": "lora",
11
+ "linear": 16,
12
+ "linear_alpha": 16,
13
+ "conv": 16,
14
+ "conv_alpha": 16,
15
+ "lokr_full_rank": true,
16
+ "lokr_factor": -1,
17
+ "network_kwargs": {
18
+ "ignore_if_contains": []
19
+ }
20
+ },
21
+ "save": {
22
+ "dtype": "bf16",
23
+ "save_every": 250,
24
+ "max_step_saves_to_keep": 8,
25
+ "save_format": "diffusers",
26
+ "push_to_hub": false
27
+ },
28
+ "datasets": [
29
+ {
30
+ "folder_path": "/workspace/ai-toolkit/datasets/anw0ll",
31
+ "mask_path": null,
32
+ "mask_min_value": 0.1,
33
+ "default_caption": "",
34
+ "caption_ext": "txt",
35
+ "caption_dropout_rate": 0.05,
36
+ "cache_latents_to_disk": false,
37
+ "is_reg": false,
38
+ "network_weight": 1,
39
+ "resolution": [
40
+ 512,
41
+ 768,
42
+ 1024
43
+ ],
44
+ "controls": [],
45
+ "shrink_video_to_frames": true,
46
+ "num_frames": 1,
47
+ "do_i2v": true,
48
+ "flip_x": false,
49
+ "flip_y": false
50
+ }
51
+ ],
52
+ "train": {
53
+ "batch_size": 1,
54
+ "bypass_guidance_embedding": false,
55
+ "steps": 3000,
56
+ "gradient_accumulation": 1,
57
+ "train_unet": true,
58
+ "train_text_encoder": false,
59
+ "gradient_checkpointing": true,
60
+ "noise_scheduler": "flowmatch",
61
+ "optimizer": "adamw8bit",
62
+ "timestep_type": "sigmoid",
63
+ "content_or_style": "balanced",
64
+ "optimizer_params": {
65
+ "weight_decay": 0.0001
66
+ },
67
+ "unload_text_encoder": false,
68
+ "cache_text_embeddings": true,
69
+ "lr": 0.0002,
70
+ "ema_config": {
71
+ "use_ema": false,
72
+ "ema_decay": 0.99
73
+ },
74
+ "skip_first_sample": false,
75
+ "force_first_sample": false,
76
+ "disable_sampling": false,
77
+ "dtype": "bf16",
78
+ "diff_output_preservation": false,
79
+ "diff_output_preservation_multiplier": 1,
80
+ "diff_output_preservation_class": "person",
81
+ "switch_boundary_every": 1,
82
+ "loss_type": "mse"
83
+ },
84
+ "model": {
85
+ "name_or_path": "Qwen/Qwen-Image",
86
+ "quantize": true,
87
+ "qtype": "uint6",
88
+ "quantize_te": true,
89
+ "qtype_te": "qfloat8",
90
+ "arch": "qwen_image",
91
+ "low_vram": true,
92
+ "model_kwargs": {},
93
+ "layer_offloading": false,
94
+ "layer_offloading_text_encoder_percent": 1,
95
+ "layer_offloading_transformer_percent": 1
96
+ },
97
+ "sample": {
98
+ "sampler": "flowmatch",
99
+ "sample_every": 250,
100
+ "width": 1024,
101
+ "height": 1024,
102
+ "samples": [],
103
+ "neg": "",
104
+ "seed": 42,
105
+ "walk_seed": true,
106
+ "guidance_scale": 4,
107
+ "sample_steps": 25,
108
+ "num_frames": 1,
109
+ "fps": 1
110
+ }
111
+ }
112
+ Using SQLite database at /workspace/ai-toolkit/aitk_db.db
113
+ Job ID: "c0204165-6ad4-405b-954b-58c03d00ba04"
114
+
115
+ #############################################
116
+ # Running job: anw0ll
117
+ #############################################
118
+
119
+
120
+ Running 1 process
121
+ Loading Qwen Image model
122
+ Loading transformer
123
+
124
+
125
+ Quantizing Transformer
126
+ - quantizing 60 transformer blocks
127
+
128
  0%| | 0/60 [00:00<?, ?it/s]
129
  0%| | 0/60 [00:00<?, ?it/s]
130
  2%|1 | 1/60 [00:00<00:48, 1.22it/s]
131
  2%|1 | 1/60 [00:00<00:48, 1.22it/s]
132
  3%|3 | 2/60 [00:01<00:34, 1.70it/s]
133
  3%|3 | 2/60 [00:01<00:34, 1.70it/s]
134
  5%|5 | 3/60 [00:01<00:29, 1.96it/s]
135
  5%|5 | 3/60 [00:01<00:29, 1.96it/s]
136
  7%|6 | 4/60 [00:02<00:26, 2.15it/s]
137
  7%|6 | 4/60 [00:02<00:26, 2.15it/s]
138
  8%|8 | 5/60 [00:02<00:22, 2.47it/s]
139
  8%|8 | 5/60 [00:02<00:22, 2.47it/s]
140
  10%|# | 6/60 [00:02<00:19, 2.73it/s]
141
  10%|# | 6/60 [00:02<00:19, 2.73it/s]
142
  12%|#1 | 7/60 [00:02<00:18, 2.92it/s]
143
  12%|#1 | 7/60 [00:02<00:18, 2.92it/s]
144
  13%|#3 | 8/60 [00:03<00:17, 3.05it/s]
145
  13%|#3 | 8/60 [00:03<00:17, 3.05it/s]
146
  15%|#5 | 9/60 [00:03<00:16, 3.15it/s]
147
  15%|#5 | 9/60 [00:03<00:16, 3.15it/s]
148
  17%|#6 | 10/60 [00:03<00:15, 3.24it/s]
149
  17%|#6 | 10/60 [00:03<00:15, 3.24it/s]
150
  18%|#8 | 11/60 [00:04<00:14, 3.30it/s]
151
  18%|#8 | 11/60 [00:04<00:14, 3.30it/s]
152
  20%|## | 12/60 [00:04<00:15, 3.00it/s]
153
  20%|## | 12/60 [00:04<00:15, 3.00it/s]
154
  22%|##1 | 13/60 [00:04<00:16, 2.87it/s]
155
  22%|##1 | 13/60 [00:04<00:16, 2.87it/s]
156
  23%|##3 | 14/60 [00:05<00:16, 2.82it/s]
157
  23%|##3 | 14/60 [00:05<00:16, 2.82it/s]
158
  25%|##5 | 15/60 [00:06<00:21, 2.08it/s]
159
  25%|##5 | 15/60 [00:06<00:21, 2.08it/s]
160
  27%|##6 | 16/60 [00:06<00:19, 2.22it/s]
161
  27%|##6 | 16/60 [00:06<00:19, 2.22it/s]
162
  28%|##8 | 17/60 [00:06<00:18, 2.30it/s]
163
  28%|##8 | 17/60 [00:06<00:18, 2.30it/s]
164
  30%|### | 18/60 [00:07<00:17, 2.34it/s]
165
  30%|### | 18/60 [00:07<00:17, 2.34it/s]
166
  32%|###1 | 19/60 [00:07<00:17, 2.38it/s]
167
  32%|###1 | 19/60 [00:07<00:17, 2.38it/s]
168
  33%|###3 | 20/60 [00:08<00:16, 2.40it/s]
169
  33%|###3 | 20/60 [00:08<00:16, 2.40it/s]
170
  35%|###5 | 21/60 [00:08<00:16, 2.43it/s]
171
  35%|###5 | 21/60 [00:08<00:16, 2.43it/s]
172
  37%|###6 | 22/60 [00:09<00:20, 1.87it/s]
173
  37%|###6 | 22/60 [00:09<00:20, 1.87it/s]
174
  38%|###8 | 23/60 [00:09<00:18, 1.96it/s]
175
  38%|###8 | 23/60 [00:09<00:18, 1.96it/s]
176
  40%|#### | 24/60 [00:10<00:16, 2.12it/s]
177
  40%|#### | 24/60 [00:10<00:16, 2.12it/s]
178
  42%|####1 | 25/60 [00:10<00:15, 2.26it/s]
179
  42%|####1 | 25/60 [00:10<00:15, 2.26it/s]
180
  43%|####3 | 26/60 [00:10<00:14, 2.35it/s]
181
  43%|####3 | 26/60 [00:10<00:14, 2.35it/s]
182
  45%|####5 | 27/60 [00:11<00:13, 2.39it/s]
183
  45%|####5 | 27/60 [00:11<00:13, 2.39it/s]
184
  47%|####6 | 28/60 [00:11<00:13, 2.38it/s]
185
  47%|####6 | 28/60 [00:11<00:13, 2.38it/s]
186
  48%|####8 | 29/60 [00:12<00:13, 2.36it/s]
187
  48%|####8 | 29/60 [00:12<00:13, 2.36it/s]
188
  50%|##### | 30/60 [00:12<00:15, 1.97it/s]
189
  50%|##### | 30/60 [00:12<00:15, 1.97it/s]
190
  52%|#####1 | 31/60 [00:13<00:13, 2.22it/s]
191
  52%|#####1 | 31/60 [00:13<00:13, 2.22it/s]
192
  53%|#####3 | 32/60 [00:13<00:11, 2.41it/s]
193
  53%|#####3 | 32/60 [00:13<00:11, 2.41it/s]
194
  55%|#####5 | 33/60 [00:13<00:10, 2.58it/s]
195
  55%|#####5 | 33/60 [00:13<00:10, 2.58it/s]
196
  57%|#####6 | 34/60 [00:14<00:09, 2.69it/s]
197
  57%|#####6 | 34/60 [00:14<00:09, 2.69it/s]
198
  58%|#####8 | 35/60 [00:14<00:08, 2.84it/s]
199
  58%|#####8 | 35/60 [00:14<00:08, 2.84it/s]
200
  60%|###### | 36/60 [00:14<00:08, 2.73it/s]
201
  60%|###### | 36/60 [00:14<00:08, 2.73it/s]
202
  62%|######1 | 37/60 [00:15<00:11, 2.01it/s]
203
  62%|######1 | 37/60 [00:15<00:11, 2.01it/s]
204
  63%|######3 | 38/60 [00:16<00:10, 2.10it/s]
205
  63%|######3 | 38/60 [00:16<00:10, 2.10it/s]
206
  65%|######5 | 39/60 [00:16<00:09, 2.14it/s]
207
  65%|######5 | 39/60 [00:16<00:09, 2.14it/s]
208
  67%|######6 | 40/60 [00:16<00:09, 2.21it/s]
209
  67%|######6 | 40/60 [00:16<00:09, 2.21it/s]
210
  68%|######8 | 41/60 [00:17<00:08, 2.30it/s]
211
  68%|######8 | 41/60 [00:17<00:08, 2.30it/s]
212
  70%|####### | 42/60 [00:17<00:07, 2.33it/s]
213
  70%|####### | 42/60 [00:17<00:07, 2.33it/s]
214
  72%|#######1 | 43/60 [00:18<00:07, 2.38it/s]
215
  72%|#######1 | 43/60 [00:18<00:07, 2.38it/s]
216
  73%|#######3 | 44/60 [00:18<00:08, 1.85it/s]
217
  73%|#######3 | 44/60 [00:18<00:08, 1.85it/s]
218
  75%|#######5 | 45/60 [00:19<00:07, 2.00it/s]
219
  75%|#######5 | 45/60 [00:19<00:07, 2.00it/s]
220
  77%|#######6 | 46/60 [00:19<00:06, 2.08it/s]
221
  77%|#######6 | 46/60 [00:19<00:06, 2.08it/s]
222
  78%|#######8 | 47/60 [00:20<00:06, 2.15it/s]
223
  78%|#######8 | 47/60 [00:20<00:06, 2.15it/s]
224
  80%|######## | 48/60 [00:20<00:05, 2.24it/s]
225
  80%|######## | 48/60 [00:20<00:05, 2.24it/s]
226
  82%|########1 | 49/60 [00:21<00:04, 2.32it/s]
227
  82%|########1 | 49/60 [00:21<00:04, 2.32it/s]
228
  83%|########3 | 50/60 [00:21<00:04, 2.37it/s]
229
  83%|########3 | 50/60 [00:21<00:04, 2.37it/s]
230
  85%|########5 | 51/60 [00:22<00:04, 1.84it/s]
231
  85%|########5 | 51/60 [00:22<00:04, 1.84it/s]
232
  87%|########6 | 52/60 [00:22<00:04, 1.96it/s]
233
  87%|########6 | 52/60 [00:22<00:04, 1.96it/s]
234
  88%|########8 | 53/60 [00:23<00:03, 2.08it/s]
235
  88%|########8 | 53/60 [00:23<00:03, 2.08it/s]
236
  90%|######### | 54/60 [00:23<00:02, 2.31it/s]
237
  90%|######### | 54/60 [00:23<00:02, 2.31it/s]
238
  92%|#########1| 55/60 [00:23<00:01, 2.53it/s]
239
  92%|#########1| 55/60 [00:23<00:01, 2.53it/s]
240
  93%|#########3| 56/60 [00:24<00:01, 2.71it/s]
241
  93%|#########3| 56/60 [00:24<00:01, 2.71it/s]
242
  95%|#########5| 57/60 [00:24<00:01, 2.85it/s]
243
  95%|#########5| 57/60 [00:24<00:01, 2.85it/s]
244
  97%|#########6| 58/60 [00:24<00:00, 2.95it/s]
245
  97%|#########6| 58/60 [00:24<00:00, 2.95it/s]
246
  98%|#########8| 59/60 [00:25<00:00, 2.38it/s]
247
  98%|#########8| 59/60 [00:25<00:00, 2.38it/s]
248
+
249
+ - quantizing extras
250
+ Moving transformer to CPU
251
+ Text Encoder
252
+
253
+
254
+ Error running job: CUDA out of memory. Tried to allocate 130.00 MiB. GPU 0 has a total capacity of 31.36 GiB of which 21.69 MiB is free. Process 21085 has 29.54 GiB memory in use. Including non-PyTorch memory, this process has 1.78 GiB memory in use. Of the allocated memory 1.20 GiB is allocated by PyTorch, and 17.99 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
255
+
256
+ ========================================
257
+ Result:
258
+ - 0 completed jobs
259
+ - 1 failure
260
+ ========================================
261
+ Traceback (most recent call last):
262
+ Traceback (most recent call last):
263
+ File "/workspace/ai-toolkit/run.py", line 120, in <module>
264
+ File "/workspace/ai-toolkit/run.py", line 120, in <module>
265
+ main()main()
266
+
267
+ File "/workspace/ai-toolkit/run.py", line 108, in main
268
+ File "/workspace/ai-toolkit/run.py", line 108, in main
269
+ raise eraise e
270
+
271
+ File "/workspace/ai-toolkit/run.py", line 96, in main
272
+ File "/workspace/ai-toolkit/run.py", line 96, in main
273
+ job.run()job.run()
274
+
275
+ File "/workspace/ai-toolkit/jobs/ExtensionJob.py", line 22, in run
276
+ File "/workspace/ai-toolkit/jobs/ExtensionJob.py", line 22, in run
277
+ process.run()process.run()
278
+
279
+ File "/workspace/ai-toolkit/jobs/process/BaseSDTrainProcess.py", line 1564, in run
280
+ File "/workspace/ai-toolkit/jobs/process/BaseSDTrainProcess.py", line 1564, in run
281
+ self.sd.load_model()self.sd.load_model()
282
+
283
+ File "/workspace/ai-toolkit/extensions_built_in/diffusion_models/qwen_image/qwen_image.py", line 161, in load_model
284
+ File "/workspace/ai-toolkit/extensions_built_in/diffusion_models/qwen_image/qwen_image.py", line 161, in load_model
285
+ text_encoder.to(self.device_torch, dtype=dtype)text_encoder.to(self.device_torch, dtype=dtype)
286
+
287
+ File "/venv/main/lib/python3.12/site-packages/transformers/modeling_utils.py", line 3851, in to
288
+ File "/venv/main/lib/python3.12/site-packages/transformers/modeling_utils.py", line 3851, in to
289
+ return super().to(*args, **kwargs)return super().to(*args, **kwargs)
290
+
291
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
292
+
293
+ File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1355, in to
294
+ File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1355, in to
295
+ return self._apply(convert)return self._apply(convert)
296
+
297
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
298
+
299
+ File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 915, in _apply
300
+ File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 915, in _apply
301
+ module._apply(fn)module._apply(fn)
302
+
303
+ File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 915, in _apply
304
+ File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 915, in _apply
305
+ module._apply(fn)module._apply(fn)
306
+
307
+ File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 915, in _apply
308
+ File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 915, in _apply
309
+ module._apply(fn)module._apply(fn)
310
+
311
+ [Previous line repeated 3 more times]
312
+ [Previous line repeated 3 more times]
313
+ File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 942, in _apply
314
+ File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 942, in _apply
315
+ param_applied = fn(param)param_applied = fn(param)
316
+
317
+ ^^^^^^^^^^^^^^^^^^
318
+
319
+ File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1341, in convert
320
+ File "/venv/main/lib/python3.12/site-packages/torch/nn/modules/module.py", line 1341, in convert
321
+ return t.to(return t.to(
322
+
323
+ ^^^^^^^^^^
324
+
325
+ torchtorch..OutOfMemoryErrorOutOfMemoryError: : CUDA out of memory. Tried to allocate 130.00 MiB. GPU 0 has a total capacity of 31.36 GiB of which 21.69 MiB is free. Process 21085 has 29.54 GiB memory in use. Including non-PyTorch memory, this process has 1.78 GiB memory in use. Of the allocated memory 1.20 GiB is allocated by PyTorch, and 17.99 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)CUDA out of memory. Tried to allocate 130.00 MiB. GPU 0 has a total capacity of 31.36 GiB of which 21.69 MiB is free. Process 21085 has 29.54 GiB memory in use. Including non-PyTorch memory, this process has 1.78 GiB memory in use. Of the allocated memory 1.20 GiB is allocated by PyTorch, and 17.99 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
326
+
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7833cea23a1ef1384912a33c17d4ccd01a9d1d29b4677d30d5b98feae47c933e
3
+ size 300280909
pid.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ 4288