Ekanari commited on
Commit
7922fcd
·
verified ·
1 Parent(s): bd4f872

Upload 183 files

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. qwen2509_object_removal_512/.job_config.json +144 -0
  2. qwen2509_object_removal_512/config.yaml +117 -0
  3. qwen2509_object_removal_512/log.txt +0 -0
  4. qwen2509_object_removal_512/logs/0_log.txt +317 -0
  5. qwen2509_object_removal_512/logs/1_log.txt +0 -0
  6. qwen2509_object_removal_512/logs/2_log.txt +181 -0
  7. qwen2509_object_removal_512/logs/3_log.txt +761 -0
  8. qwen2509_object_removal_512/logs/4_log.txt +0 -0
  9. qwen2509_object_removal_512/optimizer.pt +3 -0
  10. qwen2509_object_removal_512/pid.txt +1 -0
  11. qwen2509_object_removal_512/qwen2509_object_removal_512.safetensors +3 -0
  12. qwen2509_object_removal_512/qwen2509_object_removal_512_000012750.safetensors +3 -0
  13. qwen2509_object_removal_512/qwen2509_object_removal_512_000013000.safetensors +3 -0
  14. qwen2509_object_removal_512/qwen2509_object_removal_512_000013250.safetensors +3 -0
  15. qwen2509_object_removal_512/qwen2509_object_removal_512_000013500.safetensors +3 -0
  16. qwen2509_object_removal_512/samples/1763725336846__000000000_0.jpg +0 -0
  17. qwen2509_object_removal_512/samples/1763725779399__000000000_1.jpg +0 -0
  18. qwen2509_object_removal_512/samples/1763726223923__000000000_2.jpg +0 -0
  19. qwen2509_object_removal_512/samples/1763730040026__000000250_0.jpg +0 -0
  20. qwen2509_object_removal_512/samples/1763730486421__000000250_1.jpg +0 -0
  21. qwen2509_object_removal_512/samples/1763730932813__000000250_2.jpg +0 -0
  22. qwen2509_object_removal_512/samples/1763734729969__000000500_0.jpg +0 -0
  23. qwen2509_object_removal_512/samples/1763735177000__000000500_1.jpg +0 -0
  24. qwen2509_object_removal_512/samples/1763735623411__000000500_2.jpg +0 -0
  25. qwen2509_object_removal_512/samples/1763739419862__000000750_0.jpg +0 -0
  26. qwen2509_object_removal_512/samples/1763739866343__000000750_1.jpg +0 -0
  27. qwen2509_object_removal_512/samples/1763740312288__000000750_2.jpg +0 -0
  28. qwen2509_object_removal_512/samples/1763750974794__000001000_0.jpg +0 -0
  29. qwen2509_object_removal_512/samples/1763751440727__000001000_1.jpg +0 -0
  30. qwen2509_object_removal_512/samples/1763751914145__000001000_2.jpg +0 -0
  31. qwen2509_object_removal_512/samples/1763755708153__000001250_0.jpg +0 -0
  32. qwen2509_object_removal_512/samples/1763756150890__000001250_1.jpg +0 -0
  33. qwen2509_object_removal_512/samples/1763756594297__000001250_2.jpg +0 -0
  34. qwen2509_object_removal_512/samples/1763760376880__000001500_0.jpg +0 -0
  35. qwen2509_object_removal_512/samples/1763760820202__000001500_1.jpg +0 -0
  36. qwen2509_object_removal_512/samples/1763761263792__000001500_2.jpg +0 -0
  37. qwen2509_object_removal_512/samples/1763765063596__000001750_0.jpg +0 -0
  38. qwen2509_object_removal_512/samples/1763765510827__000001750_1.jpg +0 -0
  39. qwen2509_object_removal_512/samples/1763765958366__000001750_2.jpg +0 -0
  40. qwen2509_object_removal_512/samples/1763769758775__000002000_0.jpg +0 -0
  41. qwen2509_object_removal_512/samples/1763770203788__000002000_1.jpg +0 -0
  42. qwen2509_object_removal_512/samples/1763770648127__000002000_2.jpg +0 -0
  43. qwen2509_object_removal_512/samples/1763774448325__000002250_0.jpg +0 -0
  44. qwen2509_object_removal_512/samples/1763774890888__000002250_1.jpg +0 -0
  45. qwen2509_object_removal_512/samples/1763775332817__000002250_2.jpg +0 -0
  46. qwen2509_object_removal_512/samples/1763779130716__000002500_0.jpg +0 -0
  47. qwen2509_object_removal_512/samples/1763779574800__000002500_1.jpg +0 -0
  48. qwen2509_object_removal_512/samples/1763780018302__000002500_2.jpg +0 -0
  49. qwen2509_object_removal_512/samples/1763783822810__000002750_0.jpg +0 -0
  50. qwen2509_object_removal_512/samples/1763784268071__000002750_1.jpg +0 -0
qwen2509_object_removal_512/.job_config.json ADDED
@@ -0,0 +1,144 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "job": "extension",
3
+ "config": {
4
+ "name": "qwen2509_object_removal_512",
5
+ "process": [
6
+ {
7
+ "type": "diffusion_trainer",
8
+ "training_folder": "/mnt/d/Github/ai-toolkit/output/",
9
+ "sqlite_db_path": "/mnt/d/Github/ai-toolkit/aitk_db.db",
10
+ "device": "cuda",
11
+ "trigger_word": null,
12
+ "performance_log_every": 10,
13
+ "network": {
14
+ "type": "lora",
15
+ "linear": 16,
16
+ "linear_alpha": 16,
17
+ "conv": 16,
18
+ "conv_alpha": 16,
19
+ "lokr_full_rank": true,
20
+ "lokr_factor": -1,
21
+ "network_kwargs": {
22
+ "ignore_if_contains": []
23
+ }
24
+ },
25
+ "save": {
26
+ "dtype": "bf16",
27
+ "save_every": 250,
28
+ "max_step_saves_to_keep": 4,
29
+ "save_format": "diffusers",
30
+ "push_to_hub": false
31
+ },
32
+ "datasets": [
33
+ {
34
+ "folder_path": "/mnt/d/Github/ai-toolkit/datasets/qwen_or/train/target",
35
+ "mask_path": null,
36
+ "mask_min_value": 0.1,
37
+ "default_caption": "Remove Object",
38
+ "caption_ext": "txt",
39
+ "caption_dropout_rate": 0.05,
40
+ "cache_latents_to_disk": true,
41
+ "is_reg": false,
42
+ "network_weight": 1,
43
+ "resolution": [
44
+ 512
45
+ ],
46
+ "controls": [],
47
+ "shrink_video_to_frames": true,
48
+ "num_frames": 1,
49
+ "do_i2v": true,
50
+ "flip_x": false,
51
+ "flip_y": false,
52
+ "control_path_1": "/mnt/d/Github/ai-toolkit/datasets/qwen_or/train/with_object",
53
+ "control_path_2": "/mnt/d/Github/ai-toolkit/datasets/qwen_or/train/mask",
54
+ "control_path_3": "/mnt/d/Github/ai-toolkit/datasets/qwen_or/train/mae_output"
55
+ }
56
+ ],
57
+ "train": {
58
+ "batch_size": 1,
59
+ "bypass_guidance_embedding": false,
60
+ "steps": 13582,
61
+ "gradient_accumulation": 1,
62
+ "train_unet": true,
63
+ "train_text_encoder": false,
64
+ "gradient_checkpointing": true,
65
+ "noise_scheduler": "flowmatch",
66
+ "optimizer": "adamw8bit",
67
+ "timestep_type": "weighted",
68
+ "content_or_style": "balanced",
69
+ "optimizer_params": {
70
+ "weight_decay": 0.0001
71
+ },
72
+ "unload_text_encoder": false,
73
+ "cache_text_embeddings": true,
74
+ "lr": 0.0001,
75
+ "ema_config": {
76
+ "use_ema": false,
77
+ "ema_decay": 0.99
78
+ },
79
+ "skip_first_sample": false,
80
+ "force_first_sample": false,
81
+ "disable_sampling": false,
82
+ "dtype": "bf16",
83
+ "diff_output_preservation": false,
84
+ "diff_output_preservation_multiplier": 1,
85
+ "diff_output_preservation_class": "person",
86
+ "switch_boundary_every": 1,
87
+ "loss_type": "mse"
88
+ },
89
+ "model": {
90
+ "name_or_path": "Qwen/Qwen-Image-Edit-2509",
91
+ "quantize": true,
92
+ "qtype": "uint3|ostris/accuracy_recovery_adapters/qwen_image_edit_2509_torchao_uint3.safetensors",
93
+ "quantize_te": true,
94
+ "qtype_te": "uint4",
95
+ "arch": "qwen_image_edit_plus",
96
+ "low_vram": true,
97
+ "model_kwargs": {
98
+ "match_target_res": true
99
+ },
100
+ "layer_offloading": true,
101
+ "layer_offloading_text_encoder_percent": 0,
102
+ "layer_offloading_transformer_percent": 1
103
+ },
104
+ "sample": {
105
+ "sampler": "flowmatch",
106
+ "sample_every": 250,
107
+ "width": 512,
108
+ "height": 512,
109
+ "samples": [
110
+ {
111
+ "prompt": "Remove Object",
112
+ "ctrl_img_1": "/mnt/d/Github/ai-toolkit/data/images/8184cef5-9303-484b-bc7c-fd3d2e865e24.jpg",
113
+ "ctrl_img_2": "/mnt/d/Github/ai-toolkit/data/images/74de8410-0630-41b7-bf80-747bdc03df50.jpg",
114
+ "ctrl_img_3": "/mnt/d/Github/ai-toolkit/data/images/4b3a523f-8bd2-4db2-b734-ebe0b629f17a.jpg"
115
+ },
116
+ {
117
+ "prompt": "Remove Object",
118
+ "ctrl_img_1": "/mnt/d/Github/ai-toolkit/data/images/66a6d4bc-7c71-40f3-a8b7-55719a0d0596.jpg",
119
+ "ctrl_img_2": "/mnt/d/Github/ai-toolkit/data/images/3f7e34c1-54d0-4095-b4ed-ec7f8f3ed475.jpg",
120
+ "ctrl_img_3": "/mnt/d/Github/ai-toolkit/data/images/17521be8-a7fe-4240-9340-da1657874313.jpg"
121
+ },
122
+ {
123
+ "prompt": "Remove Object",
124
+ "ctrl_img_1": "/mnt/d/Github/ai-toolkit/data/images/dba76474-f415-4e8b-a410-caa17c1ea158.jpg",
125
+ "ctrl_img_2": "/mnt/d/Github/ai-toolkit/data/images/47b8f626-147d-41da-a8c0-6c21382ae206.jpg",
126
+ "ctrl_img_3": "/mnt/d/Github/ai-toolkit/data/images/7dfc0f6d-6c7f-4f09-b855-248f6f39614a.jpg"
127
+ }
128
+ ],
129
+ "neg": "",
130
+ "seed": 42,
131
+ "walk_seed": true,
132
+ "guidance_scale": 4,
133
+ "sample_steps": 25,
134
+ "num_frames": 1,
135
+ "fps": 1
136
+ }
137
+ }
138
+ ]
139
+ },
140
+ "meta": {
141
+ "name": "[name]",
142
+ "version": "1.0"
143
+ }
144
+ }
qwen2509_object_removal_512/config.yaml ADDED
@@ -0,0 +1,117 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ job: extension
2
+ config:
3
+ name: qwen2509_object_removal_512
4
+ process:
5
+ - type: diffusion_trainer
6
+ training_folder: /mnt/d/Github/ai-toolkit/output/
7
+ sqlite_db_path: /mnt/d/Github/ai-toolkit/aitk_db.db
8
+ device: cuda
9
+ trigger_word: null
10
+ performance_log_every: 10
11
+ network:
12
+ type: lora
13
+ linear: 16
14
+ linear_alpha: 16
15
+ conv: 16
16
+ conv_alpha: 16
17
+ lokr_full_rank: true
18
+ lokr_factor: -1
19
+ network_kwargs:
20
+ ignore_if_contains: []
21
+ save:
22
+ dtype: bf16
23
+ save_every: 250
24
+ max_step_saves_to_keep: 4
25
+ save_format: diffusers
26
+ push_to_hub: false
27
+ datasets:
28
+ - folder_path: /mnt/d/Github/ai-toolkit/datasets/qwen_or/train/target
29
+ mask_path: null
30
+ mask_min_value: 0.1
31
+ default_caption: Remove Object
32
+ caption_ext: txt
33
+ caption_dropout_rate: 0.05
34
+ cache_latents_to_disk: true
35
+ is_reg: false
36
+ network_weight: 1
37
+ resolution:
38
+ - 512
39
+ controls: []
40
+ shrink_video_to_frames: true
41
+ num_frames: 1
42
+ do_i2v: true
43
+ flip_x: false
44
+ flip_y: false
45
+ control_path_1: /mnt/d/Github/ai-toolkit/datasets/qwen_or/train/with_object
46
+ control_path_2: /mnt/d/Github/ai-toolkit/datasets/qwen_or/train/mask
47
+ control_path_3: /mnt/d/Github/ai-toolkit/datasets/qwen_or/train/mae_output
48
+ train:
49
+ batch_size: 1
50
+ bypass_guidance_embedding: false
51
+ steps: 13582
52
+ gradient_accumulation: 1
53
+ train_unet: true
54
+ train_text_encoder: false
55
+ gradient_checkpointing: true
56
+ noise_scheduler: flowmatch
57
+ optimizer: adamw8bit
58
+ timestep_type: weighted
59
+ content_or_style: balanced
60
+ optimizer_params:
61
+ weight_decay: 0.0001
62
+ unload_text_encoder: false
63
+ cache_text_embeddings: true
64
+ lr: 0.0001
65
+ ema_config:
66
+ use_ema: false
67
+ ema_decay: 0.99
68
+ skip_first_sample: false
69
+ force_first_sample: false
70
+ disable_sampling: false
71
+ dtype: bf16
72
+ diff_output_preservation: false
73
+ diff_output_preservation_multiplier: 1
74
+ diff_output_preservation_class: person
75
+ switch_boundary_every: 1
76
+ loss_type: mse
77
+ model:
78
+ name_or_path: Qwen/Qwen-Image-Edit-2509
79
+ quantize: true
80
+ qtype: uint3|ostris/accuracy_recovery_adapters/qwen_image_edit_2509_torchao_uint3.safetensors
81
+ quantize_te: true
82
+ qtype_te: uint4
83
+ arch: qwen_image_edit_plus
84
+ low_vram: true
85
+ model_kwargs:
86
+ match_target_res: true
87
+ layer_offloading: true
88
+ layer_offloading_text_encoder_percent: 0
89
+ layer_offloading_transformer_percent: 1
90
+ sample:
91
+ sampler: flowmatch
92
+ sample_every: 250
93
+ width: 512
94
+ height: 512
95
+ samples:
96
+ - prompt: Remove Object
97
+ ctrl_img_1: /mnt/d/Github/ai-toolkit/data/images/8184cef5-9303-484b-bc7c-fd3d2e865e24.jpg
98
+ ctrl_img_2: /mnt/d/Github/ai-toolkit/data/images/74de8410-0630-41b7-bf80-747bdc03df50.jpg
99
+ ctrl_img_3: /mnt/d/Github/ai-toolkit/data/images/4b3a523f-8bd2-4db2-b734-ebe0b629f17a.jpg
100
+ - prompt: Remove Object
101
+ ctrl_img_1: /mnt/d/Github/ai-toolkit/data/images/66a6d4bc-7c71-40f3-a8b7-55719a0d0596.jpg
102
+ ctrl_img_2: /mnt/d/Github/ai-toolkit/data/images/3f7e34c1-54d0-4095-b4ed-ec7f8f3ed475.jpg
103
+ ctrl_img_3: /mnt/d/Github/ai-toolkit/data/images/17521be8-a7fe-4240-9340-da1657874313.jpg
104
+ - prompt: Remove Object
105
+ ctrl_img_1: /mnt/d/Github/ai-toolkit/data/images/dba76474-f415-4e8b-a410-caa17c1ea158.jpg
106
+ ctrl_img_2: /mnt/d/Github/ai-toolkit/data/images/47b8f626-147d-41da-a8c0-6c21382ae206.jpg
107
+ ctrl_img_3: /mnt/d/Github/ai-toolkit/data/images/7dfc0f6d-6c7f-4f09-b855-248f6f39614a.jpg
108
+ neg: ''
109
+ seed: 42
110
+ walk_seed: true
111
+ guidance_scale: 4
112
+ sample_steps: 25
113
+ num_frames: 1
114
+ fps: 1
115
+ meta:
116
+ name: qwen2509_object_removal_512
117
+ version: '1.0'
qwen2509_object_removal_512/log.txt ADDED
The diff for this file is too large to render. See raw diff
 
qwen2509_object_removal_512/logs/0_log.txt ADDED
@@ -0,0 +1,317 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Running 1 job
2
+ {
3
+ "type": "diffusion_trainer",
4
+ "training_folder": "D:\\Github\\ai-toolkit\\output",
5
+ "sqlite_db_path": "/mnt/d/Github/ai-toolkit/aitk_db.db",
6
+ "device": "cuda",
7
+ "trigger_word": null,
8
+ "performance_log_every": 10,
9
+ "network": {
10
+ "type": "lora",
11
+ "linear": 16,
12
+ "linear_alpha": 16,
13
+ "conv": 16,
14
+ "conv_alpha": 16,
15
+ "lokr_full_rank": true,
16
+ "lokr_factor": -1,
17
+ "network_kwargs": {
18
+ "ignore_if_contains": []
19
+ }
20
+ },
21
+ "save": {
22
+ "dtype": "bf16",
23
+ "save_every": 250,
24
+ "max_step_saves_to_keep": 4,
25
+ "save_format": "diffusers",
26
+ "push_to_hub": false
27
+ },
28
+ "datasets": [
29
+ {
30
+ "folder_path": "D:\\Github\\ai-toolkit\\datasets\\qwen_or\\train/mae_output",
31
+ "mask_path": null,
32
+ "mask_min_value": 0.1,
33
+ "default_caption": "Remove Object",
34
+ "caption_ext": "txt",
35
+ "caption_dropout_rate": 0.05,
36
+ "cache_latents_to_disk": true,
37
+ "is_reg": false,
38
+ "network_weight": 1,
39
+ "resolution": [
40
+ 512
41
+ ],
42
+ "controls": [],
43
+ "shrink_video_to_frames": true,
44
+ "num_frames": 1,
45
+ "do_i2v": true,
46
+ "flip_x": false,
47
+ "flip_y": false,
48
+ "control_path_1": "D:\\Github\\ai-toolkit\\datasets\\qwen_or\\train/with_object",
49
+ "control_path_2": "D:\\Github\\ai-toolkit\\datasets\\qwen_or\\train/mask",
50
+ "control_path_3": "D:\\Github\\ai-toolkit\\datasets\\qwen_or\\train/mae_output"
51
+ }
52
+ ],
53
+ "train": {
54
+ "batch_size": 1,
55
+ "bypass_guidance_embedding": false,
56
+ "steps": 13582,
57
+ "gradient_accumulation": 1,
58
+ "train_unet": true,
59
+ "train_text_encoder": false,
60
+ "gradient_checkpointing": true,
61
+ "noise_scheduler": "flowmatch",
62
+ "optimizer": "adamw8bit",
63
+ "timestep_type": "weighted",
64
+ "content_or_style": "balanced",
65
+ "optimizer_params": {
66
+ "weight_decay": 0.0001
67
+ },
68
+ "unload_text_encoder": false,
69
+ "cache_text_embeddings": true,
70
+ "lr": 0.0001,
71
+ "ema_config": {
72
+ "use_ema": false,
73
+ "ema_decay": 0.99
74
+ },
75
+ "skip_first_sample": false,
76
+ "force_first_sample": false,
77
+ "disable_sampling": false,
78
+ "dtype": "bf16",
79
+ "diff_output_preservation": false,
80
+ "diff_output_preservation_multiplier": 1,
81
+ "diff_output_preservation_class": "person",
82
+ "switch_boundary_every": 1,
83
+ "loss_type": "mse"
84
+ },
85
+ "model": {
86
+ "name_or_path": "Qwen/Qwen-Image-Edit-2509",
87
+ "quantize": true,
88
+ "qtype": "uint3|ostris/accuracy_recovery_adapters/qwen_image_edit_2509_torchao_uint3.safetensors",
89
+ "quantize_te": true,
90
+ "qtype_te": "uint4",
91
+ "arch": "qwen_image_edit_plus",
92
+ "low_vram": true,
93
+ "model_kwargs": {
94
+ "match_target_res": true
95
+ },
96
+ "layer_offloading": true,
97
+ "layer_offloading_text_encoder_percent": 0,
98
+ "layer_offloading_transformer_percent": 1
99
+ },
100
+ "sample": {
101
+ "sampler": "flowmatch",
102
+ "sample_every": 250,
103
+ "width": 512,
104
+ "height": 512,
105
+ "samples": [
106
+ {
107
+ "prompt": "Remove Object",
108
+ "ctrl_img_1": "D:\\Github\\ai-toolkit\\data\\images\\7dd1fedd-6445-465d-a53b-4e351946420a.jpg",
109
+ "ctrl_img_2": "D:\\Github\\ai-toolkit\\data\\images\\1e1341a1-6079-44bb-9eb3-aac41fa2c07c.jpg",
110
+ "ctrl_img_3": "D:\\Github\\ai-toolkit\\data\\images\\f50ebe43-94e4-4b5a-a243-3ffec9f98542.jpg"
111
+ },
112
+ {
113
+ "prompt": "Remove Object",
114
+ "ctrl_img_1": "D:\\Github\\ai-toolkit\\data\\images\\a772832e-8670-40d8-ae54-3acdc19f5491.jpg",
115
+ "ctrl_img_2": "D:\\Github\\ai-toolkit\\data\\images\\77783712-d7fb-4ed4-b958-f12b26ce2fb0.jpg",
116
+ "ctrl_img_3": "D:\\Github\\ai-toolkit\\data\\images\\dfb20c07-2aef-4325-a855-573468ed0a68.jpg"
117
+ },
118
+ {
119
+ "prompt": "Remove Object",
120
+ "ctrl_img_1": "D:\\Github\\ai-toolkit\\data\\images\\54207063-d910-4070-9033-019359d11116.jpg",
121
+ "ctrl_img_2": "D:\\Github\\ai-toolkit\\data\\images\\4a2705bb-f1f1-4576-bb5b-3573b48913a9.jpg",
122
+ "ctrl_img_3": "D:\\Github\\ai-toolkit\\data\\images\\11bc3d9c-a786-4719-aa57-015677eb8cde.jpg"
123
+ }
124
+ ],
125
+ "neg": "",
126
+ "seed": 42,
127
+ "walk_seed": true,
128
+ "guidance_scale": 4,
129
+ "sample_steps": 25,
130
+ "num_frames": 1,
131
+ "fps": 1
132
+ }
133
+ }
134
+ Using SQLite database at /mnt/d/Github/ai-toolkit/aitk_db.db
135
+ Job ID: "ab192e4e-2fbc-4fe0-8067-c4de78a005b2"
136
+
137
+ #############################################
138
+ # Running job: qwen2509_object_removal_512
139
+ #############################################
140
+
141
+
142
+ Running 1 process
143
+ Loading Qwen Image model
144
+ Loading transformer
145
+
146
+
147
+ Quantizing Transformer
148
+ Grabbing lora from the hub: ostris/accuracy_recovery_adapters/qwen_image_edit_2509_torchao_uint3.safetensors
149
+ create LoRA network. base dim (rank): 16, alpha: 16
150
+ neuron dropout: p=None, rank dropout: p=None, module dropout: p=None
151
+ create LoRA for Text Encoder: 0 modules.
152
+ create LoRA for U-Net: 846 modules.
153
+ enable LoRA for U-Net
154
+ Missing keys: []
155
+
156
+
157
+ - quantizing additional layers
158
+ Moving transformer to CPU
159
+ Text Encoder
160
+
161
+
162
+ Quantizing Text Encoder
163
+ Failed to quantize : Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
164
+ Failed to quantize model: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
165
+ Failed to quantize model.visual: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
166
+ Failed to quantize model.visual.blocks: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
167
+ Failed to quantize model.visual.blocks.0: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
168
+ Failed to quantize model.visual.blocks.0.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
169
+ Failed to quantize model.visual.blocks.0.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
170
+ Failed to quantize model.visual.blocks.1: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
171
+ Failed to quantize model.visual.blocks.1.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
172
+ Failed to quantize model.visual.blocks.1.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
173
+ Failed to quantize model.visual.blocks.2: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
174
+ Failed to quantize model.visual.blocks.2.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
175
+ Failed to quantize model.visual.blocks.2.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
176
+ Failed to quantize model.visual.blocks.3: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
177
+ Failed to quantize model.visual.blocks.3.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
178
+ Failed to quantize model.visual.blocks.3.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
179
+ Failed to quantize model.visual.blocks.4: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
180
+ Failed to quantize model.visual.blocks.4.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
181
+ Failed to quantize model.visual.blocks.4.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
182
+ Failed to quantize model.visual.blocks.5: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
183
+ Failed to quantize model.visual.blocks.5.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
184
+ Failed to quantize model.visual.blocks.5.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
185
+ Failed to quantize model.visual.blocks.6: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
186
+ Failed to quantize model.visual.blocks.6.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
187
+ Failed to quantize model.visual.blocks.6.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
188
+ Failed to quantize model.visual.blocks.7: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
189
+ Failed to quantize model.visual.blocks.7.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
190
+ Failed to quantize model.visual.blocks.7.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
191
+ Failed to quantize model.visual.blocks.8: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
192
+ Failed to quantize model.visual.blocks.8.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
193
+ Failed to quantize model.visual.blocks.8.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
194
+ Failed to quantize model.visual.blocks.9: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
195
+ Failed to quantize model.visual.blocks.9.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
196
+ Failed to quantize model.visual.blocks.9.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
197
+ Failed to quantize model.visual.blocks.10: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
198
+ Failed to quantize model.visual.blocks.10.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
199
+ Failed to quantize model.visual.blocks.10.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
200
+ Failed to quantize model.visual.blocks.11: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
201
+ Failed to quantize model.visual.blocks.11.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
202
+ Failed to quantize model.visual.blocks.11.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
203
+ Failed to quantize model.visual.blocks.12: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
204
+ Failed to quantize model.visual.blocks.12.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
205
+ Failed to quantize model.visual.blocks.12.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
206
+ Failed to quantize model.visual.blocks.13: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
207
+ Failed to quantize model.visual.blocks.13.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
208
+ Failed to quantize model.visual.blocks.13.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
209
+ Failed to quantize model.visual.blocks.14: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
210
+ Failed to quantize model.visual.blocks.14.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
211
+ Failed to quantize model.visual.blocks.14.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
212
+ Failed to quantize model.visual.blocks.15: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
213
+ Failed to quantize model.visual.blocks.15.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
214
+ Failed to quantize model.visual.blocks.15.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
215
+ Failed to quantize model.visual.blocks.16: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
216
+ Failed to quantize model.visual.blocks.16.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
217
+ Failed to quantize model.visual.blocks.16.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
218
+ Failed to quantize model.visual.blocks.17: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
219
+ Failed to quantize model.visual.blocks.17.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
220
+ Failed to quantize model.visual.blocks.17.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
221
+ Failed to quantize model.visual.blocks.18: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
222
+ Failed to quantize model.visual.blocks.18.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
223
+ Failed to quantize model.visual.blocks.18.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
224
+ Failed to quantize model.visual.blocks.19: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
225
+ Failed to quantize model.visual.blocks.19.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
226
+ Failed to quantize model.visual.blocks.19.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
227
+ Failed to quantize model.visual.blocks.20: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
228
+ Failed to quantize model.visual.blocks.20.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
229
+ Failed to quantize model.visual.blocks.20.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
230
+ Failed to quantize model.visual.blocks.21: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
231
+ Failed to quantize model.visual.blocks.21.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
232
+ Failed to quantize model.visual.blocks.21.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
233
+ Failed to quantize model.visual.blocks.22: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
234
+ Failed to quantize model.visual.blocks.22.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
235
+ Failed to quantize model.visual.blocks.22.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
236
+ Failed to quantize model.visual.blocks.23: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
237
+ Failed to quantize model.visual.blocks.23.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
238
+ Failed to quantize model.visual.blocks.23.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
239
+ Failed to quantize model.visual.blocks.24: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
240
+ Failed to quantize model.visual.blocks.24.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
241
+ Failed to quantize model.visual.blocks.24.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
242
+ Failed to quantize model.visual.blocks.25: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
243
+ Failed to quantize model.visual.blocks.25.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
244
+ Failed to quantize model.visual.blocks.25.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
245
+ Failed to quantize model.visual.blocks.26: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
246
+ Failed to quantize model.visual.blocks.26.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
247
+ Failed to quantize model.visual.blocks.26.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
248
+ Failed to quantize model.visual.blocks.27: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
249
+ Failed to quantize model.visual.blocks.27.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
250
+ Failed to quantize model.visual.blocks.27.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
251
+ Failed to quantize model.visual.blocks.28: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
252
+ Failed to quantize model.visual.blocks.28.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
253
+ Failed to quantize model.visual.blocks.28.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
254
+ Failed to quantize model.visual.blocks.29: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
255
+ Failed to quantize model.visual.blocks.29.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
256
+ Failed to quantize model.visual.blocks.29.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
257
+ Failed to quantize model.visual.blocks.30: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
258
+ Failed to quantize model.visual.blocks.30.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
259
+ Failed to quantize model.visual.blocks.30.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
260
+ Failed to quantize model.visual.blocks.31: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
261
+ Failed to quantize model.visual.blocks.31.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
262
+ Failed to quantize model.visual.blocks.31.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
263
+ Loading VAE
264
+ Making pipe
265
+ Preparing Model
266
+ Model Loaded
267
+ create LoRA network. base dim (rank): 16, alpha: 16
268
+ neuron dropout: p=None, rank dropout: p=None, module dropout: p=None
269
+ apply LoRA to Conv2d with kernel size (3,3). dim (rank): 16, alpha: 16
270
+ create LoRA for Text Encoder: 0 modules.
271
+ create LoRA for U-Net: 840 modules.
272
+ enable LoRA for U-Net
273
+ Error running job: [Errno 2] No such file or directory: 'D:\\Github\\ai-toolkit\\datasets\\qwen_or\\train/mae_output'
274
+
275
+ ========================================
276
+ Result:
277
+ - 0 completed jobs
278
+ - 1 failure
279
+ ========================================
280
+ Traceback (most recent call last):
281
+ Traceback (most recent call last):
282
+ File "/mnt/d/Github/ai-toolkit/run.py", line 120, in <module>
283
+ File "/mnt/d/Github/ai-toolkit/run.py", line 120, in <module>
284
+ main()main()
285
+
286
+ File "/mnt/d/Github/ai-toolkit/run.py", line 108, in main
287
+ File "/mnt/d/Github/ai-toolkit/run.py", line 108, in main
288
+ raise eraise e
289
+
290
+ File "/mnt/d/Github/ai-toolkit/run.py", line 96, in main
291
+ File "/mnt/d/Github/ai-toolkit/run.py", line 96, in main
292
+ job.run()job.run()
293
+
294
+ File "/mnt/d/Github/ai-toolkit/jobs/ExtensionJob.py", line 22, in run
295
+ File "/mnt/d/Github/ai-toolkit/jobs/ExtensionJob.py", line 22, in run
296
+ process.run()process.run()
297
+
298
+ File "/mnt/d/Github/ai-toolkit/jobs/process/BaseSDTrainProcess.py", line 1995, in run
299
+ File "/mnt/d/Github/ai-toolkit/jobs/process/BaseSDTrainProcess.py", line 1995, in run
300
+ self.data_loader = get_dataloader_from_datasets(self.datasets, self.train_config.batch_size, self.sd)self.data_loader = get_dataloader_from_datasets(self.datasets, self.train_config.batch_size, self.sd)
301
+
302
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
303
+
304
+ File "/mnt/d/Github/ai-toolkit/toolkit/data_loader.py", line 626, in get_dataloader_from_datasets
305
+ File "/mnt/d/Github/ai-toolkit/toolkit/data_loader.py", line 626, in get_dataloader_from_datasets
306
+ dataset = AiToolkitDataset(config, batch_size=batch_size, sd=sd)dataset = AiToolkitDataset(config, batch_size=batch_size, sd=sd)
307
+
308
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
309
+
310
+ File "/mnt/d/Github/ai-toolkit/toolkit/data_loader.py", line 431, in __init__
311
+ File "/mnt/d/Github/ai-toolkit/toolkit/data_loader.py", line 431, in __init__
312
+ with open(self.dataset_path, 'r') as f:with open(self.dataset_path, 'r') as f:
313
+
314
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
315
+
316
+ FileNotFoundErrorFileNotFoundError: : [Errno 2] No such file or directory: 'D:\\Github\\ai-toolkit\\datasets\\qwen_or\\train/mae_output'[Errno 2] No such file or directory: 'D:\\Github\\ai-toolkit\\datasets\\qwen_or\\train/mae_output'
317
+
qwen2509_object_removal_512/logs/1_log.txt ADDED
The diff for this file is too large to render. See raw diff
 
qwen2509_object_removal_512/logs/2_log.txt ADDED
@@ -0,0 +1,181 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Running 1 job
2
+ {
3
+ "type": "diffusion_trainer",
4
+ "training_folder": "D:\\Github\\ai-toolkit\\output",
5
+ "sqlite_db_path": "/mnt/d/Github/ai-toolkit/aitk_db.db",
6
+ "device": "cuda",
7
+ "trigger_word": null,
8
+ "performance_log_every": 10,
9
+ "network": {
10
+ "type": "lora",
11
+ "linear": 16,
12
+ "linear_alpha": 16,
13
+ "conv": 16,
14
+ "conv_alpha": 16,
15
+ "lokr_full_rank": true,
16
+ "lokr_factor": -1,
17
+ "network_kwargs": {
18
+ "ignore_if_contains": []
19
+ }
20
+ },
21
+ "save": {
22
+ "dtype": "bf16",
23
+ "save_every": 250,
24
+ "max_step_saves_to_keep": 4,
25
+ "save_format": "diffusers",
26
+ "push_to_hub": false
27
+ },
28
+ "datasets": [
29
+ {
30
+ "folder_path": "/mnt/d/Github/ai-toolkit/datasets/qwen_or/train/target",
31
+ "mask_path": null,
32
+ "mask_min_value": 0.1,
33
+ "default_caption": "Remove Object",
34
+ "caption_ext": "txt",
35
+ "caption_dropout_rate": 0.05,
36
+ "cache_latents_to_disk": true,
37
+ "is_reg": false,
38
+ "network_weight": 1,
39
+ "resolution": [
40
+ 512
41
+ ],
42
+ "controls": [],
43
+ "shrink_video_to_frames": true,
44
+ "num_frames": 1,
45
+ "do_i2v": true,
46
+ "flip_x": false,
47
+ "flip_y": false,
48
+ "control_path_1": "/mnt/d/Github/ai-toolkit/datasets/qwen_or/train/with_object",
49
+ "control_path_2": "/mnt/d/Github/ai-toolkit/datasets/qwen_or/train/mask",
50
+ "control_path_3": "/mnt/d/Github/ai-toolkit/datasets/qwen_or/train/mae_output"
51
+ }
52
+ ],
53
+ "train": {
54
+ "batch_size": 1,
55
+ "bypass_guidance_embedding": false,
56
+ "steps": 13582,
57
+ "gradient_accumulation": 1,
58
+ "train_unet": true,
59
+ "train_text_encoder": false,
60
+ "gradient_checkpointing": true,
61
+ "noise_scheduler": "flowmatch",
62
+ "optimizer": "adamw8bit",
63
+ "timestep_type": "weighted",
64
+ "content_or_style": "balanced",
65
+ "optimizer_params": {
66
+ "weight_decay": 0.0001
67
+ },
68
+ "unload_text_encoder": false,
69
+ "cache_text_embeddings": true,
70
+ "lr": 0.0001,
71
+ "ema_config": {
72
+ "use_ema": false,
73
+ "ema_decay": 0.99
74
+ },
75
+ "skip_first_sample": false,
76
+ "force_first_sample": false,
77
+ "disable_sampling": false,
78
+ "dtype": "bf16",
79
+ "diff_output_preservation": false,
80
+ "diff_output_preservation_multiplier": 1,
81
+ "diff_output_preservation_class": "person",
82
+ "switch_boundary_every": 1,
83
+ "loss_type": "mse"
84
+ },
85
+ "model": {
86
+ "name_or_path": "Qwen/Qwen-Image-Edit-2509",
87
+ "quantize": true,
88
+ "qtype": "uint3|ostris/accuracy_recovery_adapters/qwen_image_edit_2509_torchao_uint3.safetensors",
89
+ "quantize_te": true,
90
+ "qtype_te": "uint4",
91
+ "arch": "qwen_image_edit_plus",
92
+ "low_vram": true,
93
+ "model_kwargs": {
94
+ "match_target_res": true
95
+ },
96
+ "layer_offloading": true,
97
+ "layer_offloading_text_encoder_percent": 0,
98
+ "layer_offloading_transformer_percent": 1
99
+ },
100
+ "sample": {
101
+ "sampler": "flowmatch",
102
+ "sample_every": 250,
103
+ "width": 512,
104
+ "height": 512,
105
+ "samples": [
106
+ {
107
+ "prompt": "Remove Object",
108
+ "ctrl_img_1": "/mnt/d/Github/ai-toolkit/data/images/8184cef5-9303-484b-bc7c-fd3d2e865e24.jpg",
109
+ "ctrl_img_2": "/mnt/d/Github/ai-toolkit/data/images/74de8410-0630-41b7-bf80-747bdc03df50.jpg",
110
+ "ctrl_img_3": "/mnt/d/Github/ai-toolkit/data/images/4b3a523f-8bd2-4db2-b734-ebe0b629f17a.jpg"
111
+ },
112
+ {
113
+ "prompt": "Remove Object",
114
+ "ctrl_img_1": "/mnt/d/Github/ai-toolkit/data/images/66a6d4bc-7c71-40f3-a8b7-55719a0d0596.jpg",
115
+ "ctrl_img_2": "/mnt/d/Github/ai-toolkit/data/images/3f7e34c1-54d0-4095-b4ed-ec7f8f3ed475.jpg",
116
+ "ctrl_img_3": "/mnt/d/Github/ai-toolkit/data/images/17521be8-a7fe-4240-9340-da1657874313.jpg"
117
+ },
118
+ {
119
+ "prompt": "Remove Object",
120
+ "ctrl_img_1": "/mnt/d/Github/ai-toolkit/data/images/dba76474-f415-4e8b-a410-caa17c1ea158.jpg",
121
+ "ctrl_img_2": "/mnt/d/Github/ai-toolkit/data/images/47b8f626-147d-41da-a8c0-6c21382ae206.jpg",
122
+ "ctrl_img_3": "/mnt/d/Github/ai-toolkit/data/images/7dfc0f6d-6c7f-4f09-b855-248f6f39614a.jpg"
123
+ }
124
+ ],
125
+ "neg": "",
126
+ "seed": 42,
127
+ "walk_seed": true,
128
+ "guidance_scale": 4,
129
+ "sample_steps": 25,
130
+ "num_frames": 1,
131
+ "fps": 1
132
+ }
133
+ }
134
+ Using SQLite database at /mnt/d/Github/ai-toolkit/aitk_db.db
135
+ Job ID: "ab192e4e-2fbc-4fe0-8067-c4de78a005b2"
136
+
137
+ #############################################
138
+ # Running job: qwen2509_object_removal_512
139
+ #############################################
140
+
141
+
142
+ Running 1 process
143
+ Error running job: Job stopped
144
+
145
+ ========================================
146
+ Result:
147
+ - 0 completed jobs
148
+ - 1 failure
149
+ ========================================
150
+ Traceback (most recent call last):
151
+ Traceback (most recent call last):
152
+ File "/mnt/d/Github/ai-toolkit/run.py", line 120, in <module>
153
+ File "/mnt/d/Github/ai-toolkit/run.py", line 120, in <module>
154
+ main()main()
155
+
156
+ File "/mnt/d/Github/ai-toolkit/run.py", line 108, in main
157
+ File "/mnt/d/Github/ai-toolkit/run.py", line 108, in main
158
+ raise eraise e
159
+
160
+ File "/mnt/d/Github/ai-toolkit/run.py", line 96, in main
161
+ File "/mnt/d/Github/ai-toolkit/run.py", line 96, in main
162
+ job.run()job.run()
163
+
164
+ File "/mnt/d/Github/ai-toolkit/jobs/ExtensionJob.py", line 22, in run
165
+ File "/mnt/d/Github/ai-toolkit/jobs/ExtensionJob.py", line 22, in run
166
+ process.run()process.run()
167
+
168
+ File "/mnt/d/Github/ai-toolkit/jobs/process/BaseSDTrainProcess.py", line 1513, in run
169
+ File "/mnt/d/Github/ai-toolkit/jobs/process/BaseSDTrainProcess.py", line 1513, in run
170
+ self.hook_before_model_load()self.hook_before_model_load()
171
+
172
+ File "/mnt/d/Github/ai-toolkit/extensions_built_in/sd_trainer/DiffusionTrainer.py", line 269, in hook_before_model_load
173
+ File "/mnt/d/Github/ai-toolkit/extensions_built_in/sd_trainer/DiffusionTrainer.py", line 269, in hook_before_model_load
174
+ self.maybe_stop()self.maybe_stop()
175
+
176
+ File "/mnt/d/Github/ai-toolkit/extensions_built_in/sd_trainer/DiffusionTrainer.py", line 147, in maybe_stop
177
+ File "/mnt/d/Github/ai-toolkit/extensions_built_in/sd_trainer/DiffusionTrainer.py", line 147, in maybe_stop
178
+ raise Exception("Job stopped")raise Exception("Job stopped")
179
+
180
+ ExceptionException: : Job stoppedJob stopped
181
+
qwen2509_object_removal_512/logs/3_log.txt ADDED
@@ -0,0 +1,761 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Running 1 job
2
+ {
3
+ "type": "diffusion_trainer",
4
+ "training_folder": "D:\\Github\\ai-toolkit\\output",
5
+ "sqlite_db_path": "/mnt/d/Github/ai-toolkit/aitk_db.db",
6
+ "device": "cuda",
7
+ "trigger_word": null,
8
+ "performance_log_every": 10,
9
+ "network": {
10
+ "type": "lora",
11
+ "linear": 16,
12
+ "linear_alpha": 16,
13
+ "conv": 16,
14
+ "conv_alpha": 16,ten
15
+ "lokr_full_rank": true,
16
+ "lokr_factor": -1,
17
+ "network_kwargs": {
18
+ "ignore_if_contains": []
19
+ }
20
+ },
21
+ "save": {
22
+ "dtype": "bf16",
23
+ "save_every": 250,
24
+ "max_step_saves_to_keep": 4,
25
+ "save_format": "diffusers",
26
+ "push_to_hub": false
27
+ },
28
+ "datasets": [
29
+ {
30
+ "folder_path": "/mnt/d/Github/ai-toolkit/datasets/qwen_or/train/target",
31
+ "mask_path": null,
32
+ "mask_min_value": 0.1,
33
+ "default_caption": "Remove Object",
34
+ "caption_ext": "txt",
35
+ "caption_dropout_rate": 0.05,
36
+ "cache_latents_to_disk": true,
37
+ "is_reg": false,
38
+ "network_weight": 1,
39
+ "resolution": [
40
+ 512
41
+ ],
42
+ "controls": [],
43
+ "shrink_video_to_frames": true,
44
+ "num_frames": 1,
45
+ "do_i2v": true,
46
+ "flip_x": false,
47
+ "flip_y": false,
48
+ "control_path_1": "/mnt/d/Github/ai-toolkit/datasets/qwen_or/train/with_object",
49
+ "control_path_2": "/mnt/d/Github/ai-toolkit/datasets/qwen_or/train/mask",
50
+ "control_path_3": "/mnt/d/Github/ai-toolkit/datasets/qwen_or/train/mae_output"
51
+ }
52
+ ],
53
+ "train": {
54
+ "batch_size": 1,
55
+ "bypass_guidance_embedding": false,
56
+ "steps": 13582,
57
+ "gradient_accumulation": 1,
58
+ "train_unet": true,
59
+ "train_text_encoder": false,
60
+ "gradient_checkpointing": true,
61
+ "noise_scheduler": "flowmatch",
62
+ "optimizer": "adamw8bit",
63
+ "timestep_type": "weighted",
64
+ "content_or_style": "balanced",
65
+ "optimizer_params": {
66
+ "weight_decay": 0.0001
67
+ },
68
+ "unload_text_encoder": false,
69
+ "cache_text_embeddings": true,
70
+ "lr": 0.0001,
71
+ "ema_config": {
72
+ "use_ema": false,
73
+ "ema_decay": 0.99
74
+ },
75
+ "skip_first_sample": false,
76
+ "force_first_sample": false,
77
+ "disable_sampling": false,
78
+ "dtype": "bf16",
79
+ "diff_output_preservation": false,
80
+ "diff_output_preservation_multiplier": 1,
81
+ "diff_output_preservation_class": "person",
82
+ "switch_boundary_every": 1,
83
+ "loss_type": "mse"
84
+ },
85
+ "model": {
86
+ "name_or_path": "Qwen/Qwen-Image-Edit-2509",
87
+ "quantize": true,
88
+ "qtype": "uint3|ostris/accuracy_recovery_adapters/qwen_image_edit_2509_torchao_uint3.safetensors",
89
+ "quantize_te": true,
90
+ "qtype_te": "uint4",
91
+ "arch": "qwen_image_edit_plus",
92
+ "low_vram": true,
93
+ "model_kwargs": {
94
+ "match_target_res": true
95
+ },
96
+ "layer_offloading": true,
97
+ "layer_offloading_text_encoder_percent": 0,
98
+ "layer_offloading_transformer_percent": 1
99
+ },
100
+ "sample": {
101
+ "sampler": "flowmatch",
102
+ "sample_every": 250,
103
+ "width": 512,
104
+ "height": 512,
105
+ "samples": [
106
+ {
107
+ "prompt": "Remove Object",
108
+ "ctrl_img_1": "/mnt/d/Github/ai-toolkit/data/images/8184cef5-9303-484b-bc7c-fd3d2e865e24.jpg",
109
+ "ctrl_img_2": "/mnt/d/Github/ai-toolkit/data/images/74de8410-0630-41b7-bf80-747bdc03df50.jpg",
110
+ "ctrl_img_3": "/mnt/d/Github/ai-toolkit/data/images/4b3a523f-8bd2-4db2-b734-ebe0b629f17a.jpg"
111
+ },
112
+ {
113
+ "prompt": "Remove Object",
114
+ "ctrl_img_1": "/mnt/d/Github/ai-toolkit/data/images/66a6d4bc-7c71-40f3-a8b7-55719a0d0596.jpg",
115
+ "ctrl_img_2": "/mnt/d/Github/ai-toolkit/data/images/3f7e34c1-54d0-4095-b4ed-ec7f8f3ed475.jpg",
116
+ "ctrl_img_3": "/mnt/d/Github/ai-toolkit/data/images/17521be8-a7fe-4240-9340-da1657874313.jpg"
117
+ },
118
+ {
119
+ "prompt": "Remove Object",
120
+ "ctrl_img_1": "/mnt/d/Github/ai-toolkit/data/images/dba76474-f415-4e8b-a410-caa17c1ea158.jpg",
121
+ "ctrl_img_2": "/mnt/d/Github/ai-toolkit/data/images/47b8f626-147d-41da-a8c0-6c21382ae206.jpg",
122
+ "ctrl_img_3": "/mnt/d/Github/ai-toolkit/data/images/7dfc0f6d-6c7f-4f09-b855-248f6f39614a.jpg"
123
+ }
124
+ ],
125
+ "neg": "",
126
+ "seed": 42,
127
+ "walk_seed": true,
128
+ "guidance_scale": 4,
129
+ "sample_steps": 25,
130
+ "num_frames": 1,
131
+ "fps": 1
132
+ }
133
+ }
134
+ Using SQLite database at /mnt/d/Github/ai-toolkit/aitk_db.db
135
+ Job ID: "ab192e4e-2fbc-4fe0-8067-c4de78a005b2"
136
+
137
+ #############################################
138
+ # Running job: qwen2509_object_removal_512
139
+ #############################################
140
+
141
+
142
+ Running 1 process
143
+ Loading Qwen Image model
144
+ Loading transformer
145
+
146
+ Loading checkpoint shards: 0%| | 0/5 [00:00<?, ?it/s]
147
+ Loading checkpoint shards: 0%| | 0/5 [00:00<?, ?it/s]
148
+ Loading checkpoint shards: 20%|## | 1/5 [00:00<00:00, 9.22it/s]
149
+ Loading checkpoint shards: 20%|## | 1/5 [00:00<00:00, 9.22it/s]
150
+ Loading checkpoint shards: 40%|#### | 2/5 [00:00<00:00, 9.41it/s]
151
+ Loading checkpoint shards: 40%|#### | 2/5 [00:00<00:00, 9.41it/s]
152
+ Loading checkpoint shards: 60%|###### | 3/5 [00:00<00:00, 9.45it/s]
153
+ Loading checkpoint shards: 60%|###### | 3/5 [00:00<00:00, 9.45it/s]
154
+ Loading checkpoint shards: 80%|######## | 4/5 [00:00<00:00, 9.50it/s]
155
+ Loading checkpoint shards: 80%|######## | 4/5 [00:00<00:00, 9.50it/s]
156
+ Loading checkpoint shards: 100%|##########| 5/5 [00:00<00:00, 11.45it/s]
157
+ Loading checkpoint shards: 100%|##########| 5/5 [00:00<00:00, 11.45it/s]
158
+
159
+ Quantizing Transformer
160
+ Grabbing lora from the hub: ostris/accuracy_recovery_adapters/qwen_image_edit_2509_torchao_uint3.safetensors
161
+ create LoRA network. base dim (rank): 16, alpha: 16
162
+ neuron dropout: p=None, rank dropout: p=None, module dropout: p=None
163
+ create LoRA for Text Encoder: 0 modules.
164
+ create LoRA for U-Net: 846 modules.
165
+ enable LoRA for U-Net
166
+ Missing keys: []
167
+
168
+ Attaching quantization: 0%| | 0/846 [00:00<?, ?it/s]
169
+ Attaching quantization: 0%| | 0/846 [00:00<?, ?it/s]
170
+ Attaching quantization: 1%| | 5/846 [00:00<00:31, 26.53it/s]
171
+ Attaching quantization: 1%| | 5/846 [00:00<00:31, 26.53it/s]
172
+ Attaching quantization: 1%|1 | 12/846 [00:00<00:19, 43.12it/s]
173
+ Attaching quantization: 1%|1 | 12/846 [00:00<00:19, 43.12it/s]
174
+ Attaching quantization: 2%|2 | 17/846 [00:00<00:35, 23.11it/s]
175
+ Attaching quantization: 2%|2 | 17/846 [00:00<00:35, 23.11it/s]
176
+ Attaching quantization: 2%|2 | 21/846 [00:00<00:38, 21.70it/s]
177
+ Attaching quantization: 2%|2 | 21/846 [00:00<00:38, 21.70it/s]
178
+ Attaching quantization: 3%|3 | 28/846 [00:01<00:30, 26.80it/s]
179
+ Attaching quantization: 3%|3 | 28/846 [00:01<00:30, 26.80it/s]
180
+ Attaching quantization: 4%|3 | 32/846 [00:01<00:39, 20.42it/s]
181
+ Attaching quantization: 4%|3 | 32/846 [00:01<00:39, 20.42it/s]
182
+ Attaching quantization: 4%|4 | 35/846 [00:01<00:39, 20.64it/s]
183
+ Attaching quantization: 4%|4 | 35/846 [00:01<00:39, 20.64it/s]
184
+ Attaching quantization: 5%|4 | 42/846 [00:01<00:30, 26.19it/s]
185
+ Attaching quantization: 5%|4 | 42/846 [00:01<00:30, 26.19it/s]
186
+ Attaching quantization: 5%|5 | 45/846 [00:01<00:38, 21.06it/s]
187
+ Attaching quantization: 5%|5 | 45/846 [00:01<00:38, 21.06it/s]
188
+ Attaching quantization: 6%|5 | 48/846 [00:02<00:41, 19.46it/s]
189
+ Attaching quantization: 6%|5 | 48/846 [00:02<00:41, 19.46it/s]
190
+ Attaching quantization: 7%|6 | 55/846 [00:02<00:28, 27.95it/s]
191
+ Attaching quantization: 7%|6 | 55/846 [00:02<00:28, 27.95it/s]
192
+ Attaching quantization: 7%|6 | 59/846 [00:02<00:37, 21.04it/s]
193
+ Attaching quantization: 7%|6 | 59/846 [00:02<00:37, 21.04it/s]
194
+ Attaching quantization: 7%|7 | 62/846 [00:02<00:40, 19.60it/s]
195
+ Attaching quantization: 7%|7 | 62/846 [00:02<00:40, 19.60it/s]
196
+ Attaching quantization: 8%|8 | 69/846 [00:02<00:28, 27.65it/s]
197
+ Attaching quantization: 8%|8 | 69/846 [00:02<00:28, 27.65it/s]
198
+ Attaching quantization: 9%|8 | 73/846 [00:03<00:37, 20.58it/s]
199
+ Attaching quantization: 9%|8 | 73/846 [00:03<00:37, 20.58it/s]
200
+ Attaching quantization: 9%|8 | 76/846 [00:03<00:40, 18.98it/s]
201
+ Attaching quantization: 9%|8 | 76/846 [00:03<00:40, 18.98it/s]
202
+ Attaching quantization: 10%|9 | 83/846 [00:03<00:28, 26.78it/s]
203
+ Attaching quantization: 10%|9 | 83/846 [00:03<00:28, 26.78it/s]
204
+ Attaching quantization: 10%|# | 87/846 [00:03<00:36, 20.62it/s]
205
+ Attaching quantization: 10%|# | 87/846 [00:03<00:36, 20.62it/s]
206
+ Attaching quantization: 11%|# | 90/846 [00:04<00:39, 19.23it/s]
207
+ Attaching quantization: 11%|# | 90/846 [00:04<00:39, 19.23it/s]
208
+ Attaching quantization: 11%|#1 | 97/846 [00:04<00:27, 26.87it/s]
209
+ Attaching quantization: 11%|#1 | 97/846 [00:04<00:27, 26.87it/s]
210
+ Attaching quantization: 12%|#1 | 101/846 [00:04<00:36, 20.31it/s]
211
+ Attaching quantization: 12%|#1 | 101/846 [00:04<00:36, 20.31it/s]
212
+ Attaching quantization: 12%|#2 | 104/846 [00:04<00:39, 18.90it/s]
213
+ Attaching quantization: 12%|#2 | 104/846 [00:04<00:39, 18.90it/s]
214
+ Attaching quantization: 13%|#3 | 111/846 [00:04<00:27, 26.57it/s]
215
+ Attaching quantization: 13%|#3 | 111/846 [00:04<00:27, 26.57it/s]
216
+ Attaching quantization: 14%|#3 | 115/846 [00:05<00:36, 19.83it/s]
217
+ Attaching quantization: 14%|#3 | 115/846 [00:05<00:36, 19.83it/s]
218
+ Attaching quantization: 14%|#3 | 118/846 [00:05<00:39, 18.56it/s]
219
+ Attaching quantization: 14%|#3 | 118/846 [00:05<00:39, 18.56it/s]
220
+ Attaching quantization: 15%|#4 | 123/846 [00:05<00:31, 23.32it/s]
221
+ Attaching quantization: 15%|#4 | 123/846 [00:05<00:31, 23.32it/s]
222
+ Attaching quantization: 15%|#5 | 127/846 [00:05<00:32, 22.46it/s]
223
+ Attaching quantization: 15%|#5 | 127/846 [00:05<00:32, 22.46it/s]
224
+ Attaching quantization: 15%|#5 | 130/846 [00:05<00:39, 18.30it/s]
225
+ Attaching quantization: 15%|#5 | 130/846 [00:05<00:39, 18.30it/s]
226
+ Attaching quantization: 16%|#5 | 133/846 [00:06<00:38, 18.58it/s]
227
+ Attaching quantization: 16%|#5 | 133/846 [00:06<00:38, 18.58it/s]
228
+ Attaching quantization: 16%|#6 | 139/846 [00:06<00:27, 25.91it/s]
229
+ Attaching quantization: 16%|#6 | 139/846 [00:06<00:27, 25.91it/s]
230
+ Attaching quantization: 17%|#6 | 143/846 [00:06<00:36, 19.13it/s]
231
+ Attaching quantization: 17%|#6 | 143/846 [00:06<00:36, 19.13it/s]
232
+ Attaching quantization: 17%|#7 | 146/846 [00:06<00:39, 17.94it/s]
233
+ Attaching quantization: 17%|#7 | 146/846 [00:06<00:39, 17.94it/s]
234
+ Attaching quantization: 18%|#8 | 153/846 [00:06<00:26, 26.17it/s]
235
+ Attaching quantization: 18%|#8 | 153/846 [00:06<00:26, 26.17it/s]
236
+ Attaching quantization: 19%|#8 | 157/846 [00:07<00:35, 19.57it/s]
237
+ Attaching quantization: 19%|#8 | 157/846 [00:07<00:35, 19.57it/s]
238
+ Attaching quantization: 19%|#8 | 160/846 [00:07<00:39, 17.39it/s]
239
+ Attaching quantization: 19%|#8 | 160/846 [00:07<00:39, 17.39it/s]
240
+ Attaching quantization: 20%|#9 | 166/846 [00:07<00:29, 23.38it/s]
241
+ Attaching quantization: 20%|#9 | 166/846 [00:07<00:29, 23.38it/s]
242
+ Attaching quantization: 20%|## | 170/846 [00:07<00:33, 19.95it/s]
243
+ Attaching quantization: 20%|## | 170/846 [00:07<00:33, 19.95it/s]
244
+ Attaching quantization: 20%|## | 173/846 [00:08<00:38, 17.43it/s]
245
+ Attaching quantization: 20%|## | 173/846 [00:08<00:38, 17.43it/s]
246
+ Attaching quantization: 21%|##1 | 180/846 [00:08<00:26, 25.22it/s]
247
+ Attaching quantization: 21%|##1 | 180/846 [00:08<00:26, 25.22it/s]
248
+ Attaching quantization: 22%|##1 | 184/846 [00:08<00:31, 21.18it/s]
249
+ Attaching quantization: 22%|##1 | 184/846 [00:08<00:31, 21.18it/s]
250
+ Attaching quantization: 22%|##2 | 187/846 [00:08<00:37, 17.53it/s]
251
+ Attaching quantization: 22%|##2 | 187/846 [00:08<00:37, 17.53it/s]
252
+ Attaching quantization: 23%|##2 | 194/846 [00:08<00:26, 25.05it/s]
253
+ Attaching quantization: 23%|##2 | 194/846 [00:08<00:26, 25.05it/s]
254
+ Attaching quantization: 23%|##3 | 198/846 [00:09<00:30, 21.28it/s]
255
+ Attaching quantization: 23%|##3 | 198/846 [00:09<00:30, 21.28it/s]
256
+ Attaching quantization: 24%|##3 | 201/846 [00:09<00:35, 18.27it/s]
257
+ Attaching quantization: 24%|##3 | 201/846 [00:09<00:35, 18.27it/s]
258
+ Attaching quantization: 25%|##4 | 208/846 [00:09<00:25, 25.45it/s]
259
+ Attaching quantization: 25%|##4 | 208/846 [00:09<00:25, 25.45it/s]
260
+ Attaching quantization: 25%|##5 | 212/846 [00:09<00:31, 20.17it/s]
261
+ Attaching quantization: 25%|##5 | 212/846 [00:09<00:31, 20.17it/s]
262
+ Attaching quantization: 25%|##5 | 215/846 [00:10<00:37, 17.03it/s]
263
+ Attaching quantization: 25%|##5 | 215/846 [00:10<00:37, 17.03it/s]
264
+ Attaching quantization: 26%|##6 | 222/846 [00:10<00:25, 24.60it/s]
265
+ Attaching quantization: 26%|##6 | 222/846 [00:10<00:25, 24.60it/s]
266
+ Attaching quantization: 27%|##6 | 226/846 [00:10<00:31, 19.76it/s]
267
+ Attaching quantization: 27%|##6 | 226/846 [00:10<00:31, 19.76it/s]
268
+ Attaching quantization: 27%|##7 | 229/846 [00:10<00:36, 16.73it/s]
269
+ Attaching quantization: 27%|##7 | 229/846 [00:10<00:36, 16.73it/s]
270
+ Attaching quantization: 28%|##7 | 236/846 [00:10<00:25, 24.13it/s]
271
+ Attaching quantization: 28%|##7 | 236/846 [00:10<00:25, 24.13it/s]
272
+ Attaching quantization: 28%|##8 | 240/846 [00:11<00:29, 20.20it/s]
273
+ Attaching quantization: 28%|##8 | 240/846 [00:11<00:29, 20.20it/s]
274
+ Attaching quantization: 29%|##8 | 243/846 [00:11<00:35, 16.99it/s]
275
+ Attaching quantization: 29%|##8 | 243/846 [00:11<00:35, 16.99it/s]
276
+ Attaching quantization: 29%|##9 | 248/846 [00:11<00:27, 21.66it/s]
277
+ Attaching quantization: 29%|##9 | 248/846 [00:11<00:27, 21.66it/s]
278
+ Attaching quantization: 30%|##9 | 252/846 [00:11<00:25, 23.16it/s]
279
+ Attaching quantization: 30%|##9 | 252/846 [00:11<00:25, 23.16it/s]
280
+ Attaching quantization: 30%|### | 255/846 [00:12<00:33, 17.60it/s]
281
+ Attaching quantization: 30%|### | 255/846 [00:12<00:33, 17.60it/s]
282
+ Attaching quantization: 30%|### | 258/846 [00:12<00:35, 16.34it/s]
283
+ Attaching quantization: 30%|### | 258/846 [00:12<00:35, 16.34it/s]
284
+ Attaching quantization: 31%|###1 | 265/846 [00:12<00:23, 24.55it/s]
285
+ Attaching quantization: 31%|###1 | 265/846 [00:12<00:23, 24.55it/s]
286
+ Attaching quantization: 32%|###1 | 269/846 [00:12<00:31, 18.24it/s]
287
+ Attaching quantization: 32%|###1 | 269/846 [00:12<00:31, 18.24it/s]
288
+ Attaching quantization: 32%|###2 | 272/846 [00:12<00:33, 17.03it/s]
289
+ Attaching quantization: 32%|###2 | 272/846 [00:12<00:33, 17.03it/s]
290
+ Attaching quantization: 33%|###2 | 279/846 [00:13<00:22, 24.78it/s]
291
+ Attaching quantization: 33%|###2 | 279/846 [00:13<00:22, 24.78it/s]
292
+ Attaching quantization: 33%|###3 | 283/846 [00:13<00:30, 18.48it/s]
293
+ Attaching quantization: 33%|###3 | 283/846 [00:13<00:30, 18.48it/s]
294
+ Attaching quantization: 34%|###3 | 286/846 [00:13<00:34, 16.17it/s]
295
+ Attaching quantization: 34%|###3 | 286/846 [00:13<00:34, 16.17it/s]
296
+ Attaching quantization: 35%|###4 | 292/846 [00:13<00:24, 22.37it/s]
297
+ Attaching quantization: 35%|###4 | 292/846 [00:13<00:24, 22.37it/s]
298
+ Attaching quantization: 35%|###4 | 296/846 [00:14<00:33, 16.38it/s]
299
+ Attaching quantization: 35%|###4 | 296/846 [00:14<00:33, 16.38it/s]
300
+ Attaching quantization: 35%|###5 | 299/846 [00:14<00:42, 12.91it/s]
301
+ Attaching quantization: 35%|###5 | 299/846 [00:14<00:42, 12.91it/s]
302
+ Attaching quantization: 36%|###6 | 305/846 [00:14<00:29, 18.43it/s]
303
+ Attaching quantization: 36%|###6 | 305/846 [00:14<00:29, 18.43it/s]
304
+ Attaching quantization: 37%|###6 | 309/846 [00:14<00:29, 18.28it/s]
305
+ Attaching quantization: 37%|###6 | 309/846 [00:14<00:29, 18.28it/s]
306
+ Attaching quantization: 37%|###6 | 312/846 [00:15<00:37, 14.36it/s]
307
+ Attaching quantization: 37%|###6 | 312/846 [00:15<00:37, 14.36it/s]
308
+ Attaching quantization: 37%|###7 | 315/846 [00:15<00:37, 14.21it/s]
309
+ Attaching quantization: 37%|###7 | 315/846 [00:15<00:37, 14.21it/s]
310
+ Attaching quantization: 38%|###7 | 318/846 [00:15<00:32, 16.36it/s]
311
+ Attaching quantization: 38%|###7 | 318/846 [00:15<00:32, 16.36it/s]
312
+ Attaching quantization: 38%|###8 | 322/846 [00:15<00:30, 17.08it/s]
313
+ Attaching quantization: 38%|###8 | 322/846 [00:15<00:30, 17.08it/s]
314
+ Attaching quantization: 38%|###8 | 325/846 [00:16<00:39, 13.09it/s]
315
+ Attaching quantization: 38%|###8 | 325/846 [00:16<00:39, 13.09it/s]
316
+ Attaching quantization: 39%|###8 | 327/846 [00:16<00:51, 9.99it/s]
317
+ Attaching quantization: 39%|###8 | 327/846 [00:16<00:51, 9.99it/s]
318
+ Attaching quantization: 39%|###9 | 333/846 [00:16<00:32, 15.95it/s]
319
+ Attaching quantization: 39%|###9 | 333/846 [00:16<00:32, 15.95it/s]
320
+ Attaching quantization: 40%|###9 | 336/846 [00:16<00:33, 15.32it/s]
321
+ Attaching quantization: 40%|###9 | 336/846 [00:16<00:33, 15.32it/s]
322
+ Attaching quantization: 40%|#### | 339/846 [00:18<01:16, 6.65it/s]
323
+ Attaching quantization: 40%|#### | 339/846 [00:18<01:16, 6.65it/s]
324
+ Attaching quantization: 40%|#### | 341/846 [00:18<01:36, 5.24it/s]
325
+ Attaching quantization: 40%|#### | 341/846 [00:18<01:36, 5.24it/s]
326
+ Attaching quantization: 41%|####1 | 347/846 [00:18<00:55, 8.94it/s]
327
+ Attaching quantization: 41%|####1 | 347/846 [00:18<00:55, 8.94it/s]
328
+ Attaching quantization: 41%|####1 | 350/846 [00:19<00:51, 9.64it/s]
329
+ Attaching quantization: 41%|####1 | 350/846 [00:19<00:51, 9.64it/s]
330
+ Attaching quantization: 42%|####1 | 353/846 [00:19<01:06, 7.36it/s]
331
+ Attaching quantization: 42%|####1 | 353/846 [00:19<01:06, 7.36it/s]
332
+ Attaching quantization: 42%|####1 | 355/846 [00:21<01:53, 4.32it/s]
333
+ Attaching quantization: 42%|####1 | 355/846 [00:21<01:53, 4.32it/s]
334
+ Attaching quantization: 43%|####2 | 360/846 [00:21<01:09, 6.95it/s]
335
+ Attaching quantization: 43%|####2 | 360/846 [00:21<01:09, 6.95it/s]
336
+ Attaching quantization: 43%|####3 | 364/846 [00:23<01:58, 4.06it/s]
337
+ Attaching quantization: 43%|####3 | 364/846 [00:23<01:58, 4.06it/s]
338
+ Attaching quantization: 43%|####3 | 366/846 [00:23<02:03, 3.90it/s]
339
+ Attaching quantization: 43%|####3 | 366/846 [00:23<02:03, 3.90it/s]
340
+ Attaching quantization: 43%|####3 | 368/846 [00:23<01:49, 4.38it/s]
341
+ Attaching quantization: 43%|####3 | 368/846 [00:23<01:49, 4.38it/s]
342
+ Attaching quantization: 44%|####3 | 370/846 [00:24<01:42, 4.67it/s]
343
+ Attaching quantization: 44%|####3 | 370/846 [00:24<01:42, 4.67it/s]
344
+ Attaching quantization: 44%|####4 | 376/846 [00:24<00:55, 8.51it/s]
345
+ Attaching quantization: 44%|####4 | 376/846 [00:24<00:55, 8.51it/s]
346
+ Attaching quantization: 45%|####4 | 379/846 [00:24<01:00, 7.66it/s]
347
+ Attaching quantization: 45%|####4 | 379/846 [00:24<01:00, 7.66it/s]
348
+ Attaching quantization: 45%|####5 | 381/846 [00:25<01:19, 5.82it/s]
349
+ Attaching quantization: 45%|####5 | 381/846 [00:25<01:19, 5.82it/s]
350
+ Attaching quantization: 45%|####5 | 383/846 [00:26<01:39, 4.65it/s]
351
+ Attaching quantization: 45%|####5 | 383/846 [00:26<01:39, 4.65it/s]
352
+ Attaching quantization: 46%|####5 | 388/846 [00:26<00:59, 7.75it/s]
353
+ Attaching quantization: 46%|####5 | 388/846 [00:26<00:59, 7.75it/s]
354
+ Attaching quantization: 46%|####6 | 392/846 [00:26<00:51, 8.84it/s]
355
+ Attaching quantization: 46%|####6 | 392/846 [00:26<00:51, 8.84it/s]
356
+ Attaching quantization: 47%|####6 | 394/846 [00:27<01:00, 7.50it/s]
357
+ Attaching quantization: 47%|####6 | 394/846 [00:27<01:00, 7.50it/s]
358
+ Attaching quantization: 47%|####6 | 396/846 [00:27<01:00, 7.40it/s]
359
+ Attaching quantization: 47%|####6 | 396/846 [00:27<01:00, 7.40it/s]
360
+ Attaching quantization: 47%|####7 | 398/846 [00:27<01:06, 6.76it/s]
361
+ Attaching quantization: 47%|####7 | 398/846 [00:27<01:06, 6.76it/s]
362
+ Attaching quantization: 48%|####7 | 402/846 [00:27<00:43, 10.13it/s]
363
+ Attaching quantization: 48%|####7 | 402/846 [00:27<00:43, 10.13it/s]
364
+ Attaching quantization: 48%|####7 | 406/846 [00:28<00:39, 11.07it/s]
365
+ Attaching quantization: 48%|####7 | 406/846 [00:28<00:39, 11.07it/s]
366
+ Attaching quantization: 48%|####8 | 408/846 [00:28<00:52, 8.31it/s]
367
+ Attaching quantization: 48%|####8 | 408/846 [00:28<00:52, 8.31it/s]
368
+ Attaching quantization: 48%|####8 | 410/846 [00:29<01:08, 6.40it/s]
369
+ Attaching quantization: 48%|####8 | 410/846 [00:29<01:08, 6.40it/s]
370
+ Attaching quantization: 49%|####8 | 412/846 [00:29<01:08, 6.38it/s]
371
+ Attaching quantization: 49%|####8 | 412/846 [00:29<01:08, 6.38it/s]
372
+ Attaching quantization: 49%|####9 | 417/846 [00:29<00:40, 10.71it/s]
373
+ Attaching quantization: 49%|####9 | 417/846 [00:29<00:40, 10.71it/s]
374
+ Attaching quantization: 50%|####9 | 420/846 [00:29<00:40, 10.41it/s]
375
+ Attaching quantization: 50%|####9 | 420/846 [00:29<00:40, 10.41it/s]
376
+ Attaching quantization: 50%|####9 | 422/846 [00:30<00:53, 7.90it/s]
377
+ Attaching quantization: 50%|####9 | 422/846 [00:30<00:53, 7.90it/s]
378
+ Attaching quantization: 50%|##### | 424/846 [00:30<00:56, 7.50it/s]
379
+ Attaching quantization: 50%|##### | 424/846 [00:30<00:56, 7.50it/s]
380
+ Attaching quantization: 50%|##### | 426/846 [00:31<00:58, 7.14it/s]
381
+ Attaching quantization: 50%|##### | 426/846 [00:31<00:58, 7.14it/s]
382
+ Attaching quantization: 51%|#####1 | 432/846 [00:31<00:32, 12.92it/s]
383
+ Attaching quantization: 51%|#####1 | 432/846 [00:31<00:32, 12.92it/s]
384
+ Attaching quantization: 51%|#####1 | 435/846 [00:31<00:39, 10.35it/s]
385
+ Attaching quantization: 51%|#####1 | 435/846 [00:31<00:39, 10.35it/s]
386
+ Attaching quantization: 52%|#####1 | 437/846 [00:31<00:46, 8.84it/s]
387
+ Attaching quantization: 52%|#####1 | 437/846 [00:31<00:46, 8.84it/s]
388
+ Attaching quantization: 52%|#####1 | 439/846 [00:32<00:54, 7.43it/s]
389
+ Attaching quantization: 52%|#####1 | 439/846 [00:32<00:54, 7.43it/s]
390
+ Attaching quantization: 53%|#####2 | 445/846 [00:32<00:30, 12.95it/s]
391
+ Attaching quantization: 53%|#####2 | 445/846 [00:32<00:30, 12.95it/s]
392
+ Attaching quantization: 53%|#####2 | 448/846 [00:32<00:31, 12.63it/s]
393
+ Attaching quantization: 53%|#####2 | 448/846 [00:32<00:31, 12.63it/s]
394
+ Attaching quantization: 53%|#####3 | 451/846 [00:33<00:38, 10.28it/s]
395
+ Attaching quantization: 53%|#####3 | 451/846 [00:33<00:38, 10.28it/s]
396
+ Attaching quantization: 54%|#####3 | 453/846 [00:33<00:47, 8.30it/s]
397
+ Attaching quantization: 54%|#####3 | 453/846 [00:33<00:47, 8.30it/s]
398
+ Attaching quantization: 54%|#####4 | 459/846 [00:33<00:27, 13.83it/s]
399
+ Attaching quantization: 54%|#####4 | 459/846 [00:33<00:27, 13.83it/s]
400
+ Attaching quantization: 55%|#####4 | 462/846 [00:34<00:29, 13.09it/s]
401
+ Attaching quantization: 55%|#####4 | 462/846 [00:34<00:29, 13.09it/s]
402
+ Attaching quantization: 55%|#####4 | 465/846 [00:34<00:36, 10.48it/s]
403
+ Attaching quantization: 55%|#####4 | 465/846 [00:34<00:36, 10.48it/s]
404
+ Attaching quantization: 55%|#####5 | 467/846 [00:34<00:41, 9.10it/s]
405
+ Attaching quantization: 55%|#####5 | 467/846 [00:34<00:41, 9.10it/s]
406
+ Attaching quantization: 56%|#####5 | 473/846 [00:34<00:25, 14.73it/s]
407
+ Attaching quantization: 56%|#####5 | 473/846 [00:34<00:25, 14.73it/s]
408
+ Attaching quantization: 56%|#####6 | 476/846 [00:35<00:25, 14.48it/s]
409
+ Attaching quantization: 56%|#####6 | 476/846 [00:35<00:25, 14.48it/s]
410
+ Attaching quantization: 57%|#####6 | 479/846 [00:35<00:33, 10.95it/s]
411
+ Attaching quantization: 57%|#####6 | 479/846 [00:35<00:33, 10.95it/s]
412
+ Attaching quantization: 57%|#####6 | 481/846 [00:35<00:38, 9.41it/s]
413
+ Attaching quantization: 57%|#####6 | 481/846 [00:35<00:38, 9.41it/s]
414
+ Attaching quantization: 57%|#####7 | 486/846 [00:36<00:25, 14.10it/s]
415
+ Attaching quantization: 57%|#####7 | 486/846 [00:36<00:25, 14.10it/s]
416
+ Attaching quantization: 58%|#####7 | 490/846 [00:36<00:24, 14.78it/s]
417
+ Attaching quantization: 58%|#####7 | 490/846 [00:36<00:24, 14.78it/s]
418
+ Attaching quantization: 58%|#####8 | 493/846 [00:36<00:30, 11.53it/s]
419
+ Attaching quantization: 58%|#####8 | 493/846 [00:36<00:30, 11.53it/s]
420
+ Attaching quantization: 59%|#####8 | 495/846 [00:37<00:35, 9.82it/s]
421
+ Attaching quantization: 59%|#####8 | 495/846 [00:37<00:35, 9.82it/s]
422
+ Attaching quantization: 59%|#####9 | 501/846 [00:37<00:22, 15.68it/s]
423
+ Attaching quantization: 59%|#####9 | 501/846 [00:37<00:22, 15.68it/s]
424
+ Attaching quantization: 60%|#####9 | 504/846 [00:37<00:22, 15.21it/s]
425
+ Attaching quantization: 60%|#####9 | 504/846 [00:37<00:22, 15.21it/s]
426
+ Attaching quantization: 60%|#####9 | 507/846 [00:37<00:29, 11.32it/s]
427
+ Attaching quantization: 60%|#####9 | 507/846 [00:37<00:29, 11.32it/s]
428
+ Attaching quantization: 60%|###### | 509/846 [00:38<00:35, 9.55it/s]
429
+ Attaching quantization: 60%|###### | 509/846 [00:38<00:35, 9.55it/s]
430
+ Attaching quantization: 61%|###### | 515/846 [00:38<00:21, 15.45it/s]
431
+ Attaching quantization: 61%|###### | 515/846 [00:38<00:21, 15.45it/s]
432
+ Attaching quantization: 61%|######1 | 518/846 [00:38<00:21, 15.17it/s]
433
+ Attaching quantization: 61%|######1 | 518/846 [00:38<00:21, 15.17it/s]
434
+ Attaching quantization: 62%|######1 | 521/846 [00:38<00:26, 12.43it/s]
435
+ Attaching quantization: 62%|######1 | 521/846 [00:38<00:26, 12.43it/s]
436
+ Attaching quantization: 62%|######1 | 523/846 [00:39<00:30, 10.68it/s]
437
+ Attaching quantization: 62%|######1 | 523/846 [00:39<00:30, 10.68it/s]
438
+ Attaching quantization: 63%|######2 | 529/846 [00:39<00:18, 17.06it/s]
439
+ Attaching quantization: 63%|######2 | 529/846 [00:39<00:18, 17.06it/s]
440
+ Attaching quantization: 63%|######2 | 532/846 [00:39<00:18, 17.08it/s]
441
+ Attaching quantization: 63%|######2 | 532/846 [00:39<00:18, 17.08it/s]
442
+ Attaching quantization: 63%|######3 | 535/846 [00:39<00:23, 13.34it/s]
443
+ Attaching quantization: 63%|######3 | 535/846 [00:39<00:23, 13.34it/s]
444
+ Attaching quantization: 63%|######3 | 537/846 [00:40<00:27, 11.27it/s]
445
+ Attaching quantization: 63%|######3 | 537/846 [00:40<00:27, 11.27it/s]
446
+ Attaching quantization: 64%|######4 | 543/846 [00:40<00:16, 17.83it/s]
447
+ Attaching quantization: 64%|######4 | 543/846 [00:40<00:16, 17.83it/s]
448
+ Attaching quantization: 65%|######4 | 546/846 [00:40<00:17, 17.16it/s]
449
+ Attaching quantization: 65%|######4 | 546/846 [00:40<00:17, 17.16it/s]
450
+ Attaching quantization: 65%|######4 | 549/846 [00:40<00:22, 13.24it/s]
451
+ Attaching quantization: 65%|######4 | 549/846 [00:40<00:22, 13.24it/s]
452
+ Attaching quantization: 65%|######5 | 551/846 [00:41<00:26, 11.21it/s]
453
+ Attaching quantization: 65%|######5 | 551/846 [00:41<00:26, 11.21it/s]
454
+ Attaching quantization: 66%|######5 | 557/846 [00:41<00:16, 17.56it/s]
455
+ Attaching quantization: 66%|######5 | 557/846 [00:41<00:16, 17.56it/s]
456
+ Attaching quantization: 66%|######6 | 560/846 [00:41<00:16, 17.58it/s]
457
+ Attaching quantization: 66%|######6 | 560/846 [00:41<00:16, 17.58it/s]
458
+ Attaching quantization: 67%|######6 | 563/846 [00:41<00:20, 13.72it/s]
459
+ Attaching quantization: 67%|######6 | 563/846 [00:41<00:20, 13.72it/s]
460
+ Attaching quantization: 67%|######6 | 565/846 [00:41<00:25, 11.12it/s]
461
+ Attaching quantization: 67%|######6 | 565/846 [00:41<00:25, 11.12it/s]
462
+ Attaching quantization: 67%|######7 | 570/846 [00:42<00:17, 16.12it/s]
463
+ Attaching quantization: 67%|######7 | 570/846 [00:42<00:17, 16.12it/s]
464
+ Attaching quantization: 68%|######7 | 574/846 [00:42<00:16, 16.29it/s]
465
+ Attaching quantization: 68%|######7 | 574/846 [00:42<00:16, 16.29it/s]
466
+ Attaching quantization: 68%|######8 | 577/846 [00:42<00:20, 12.89it/s]
467
+ Attaching quantization: 68%|######8 | 577/846 [00:42<00:20, 12.89it/s]
468
+ Attaching quantization: 68%|######8 | 579/846 [00:42<00:23, 11.23it/s]
469
+ Attaching quantization: 68%|######8 | 579/846 [00:42<00:23, 11.23it/s]
470
+ Attaching quantization: 69%|######9 | 584/846 [00:43<00:16, 16.32it/s]
471
+ Attaching quantization: 69%|######9 | 584/846 [00:43<00:16, 16.32it/s]
472
+ Attaching quantization: 70%|######9 | 588/846 [00:43<00:15, 16.71it/s]
473
+ Attaching quantization: 70%|######9 | 588/846 [00:43<00:15, 16.71it/s]
474
+ Attaching quantization: 70%|######9 | 591/846 [00:43<00:20, 12.62it/s]
475
+ Attaching quantization: 70%|######9 | 591/846 [00:43<00:20, 12.62it/s]
476
+ Attaching quantization: 70%|####### | 593/846 [00:44<00:23, 10.76it/s]
477
+ Attaching quantization: 70%|####### | 593/846 [00:44<00:23, 10.76it/s]
478
+ Attaching quantization: 71%|####### | 598/846 [00:44<00:15, 15.80it/s]
479
+ Attaching quantization: 71%|####### | 598/846 [00:44<00:15, 15.80it/s]
480
+ Attaching quantization: 71%|#######1 | 602/846 [00:44<00:14, 17.22it/s]
481
+ Attaching quantization: 71%|#######1 | 602/846 [00:44<00:14, 17.22it/s]
482
+ Attaching quantization: 72%|#######1 | 605/846 [00:44<00:18, 13.22it/s]
483
+ Attaching quantization: 72%|#######1 | 605/846 [00:44<00:18, 13.22it/s]
484
+ Attaching quantization: 72%|#######1 | 607/846 [00:44<00:21, 11.30it/s]
485
+ Attaching quantization: 72%|#######1 | 607/846 [00:44<00:21, 11.30it/s]
486
+ Attaching quantization: 72%|#######2 | 613/846 [00:45<00:13, 17.87it/s]
487
+ Attaching quantization: 72%|#######2 | 613/846 [00:45<00:13, 17.87it/s]
488
+ Attaching quantization: 73%|#######2 | 616/846 [00:45<00:12, 18.11it/s]
489
+ Attaching quantization: 73%|#######2 | 616/846 [00:45<00:12, 18.11it/s]
490
+ Attaching quantization: 73%|#######3 | 619/846 [00:45<00:16, 13.54it/s]
491
+ Attaching quantization: 73%|#######3 | 619/846 [00:45<00:16, 13.54it/s]
492
+ Attaching quantization: 74%|#######3 | 622/846 [00:45<00:18, 12.40it/s]
493
+ Attaching quantization: 74%|#######3 | 622/846 [00:45<00:18, 12.40it/s]
494
+ Attaching quantization: 74%|#######4 | 628/846 [00:46<00:11, 18.62it/s]
495
+ Attaching quantization: 74%|#######4 | 628/846 [00:46<00:11, 18.62it/s]
496
+ Attaching quantization: 75%|#######4 | 631/846 [00:46<00:13, 16.13it/s]
497
+ Attaching quantization: 75%|#######4 | 631/846 [00:46<00:13, 16.13it/s]
498
+ Attaching quantization: 75%|#######4 | 634/846 [00:46<00:16, 12.90it/s]
499
+ Attaching quantization: 75%|#######4 | 634/846 [00:46<00:16, 12.90it/s]
500
+ Attaching quantization: 75%|#######5 | 636/846 [00:46<00:16, 12.43it/s]
501
+ Attaching quantization: 75%|#######5 | 636/846 [00:46<00:16, 12.43it/s]
502
+ Attaching quantization: 76%|#######6 | 644/846 [00:47<00:10, 18.61it/s]
503
+ Attaching quantization: 76%|#######6 | 644/846 [00:47<00:10, 18.61it/s]
504
+ Attaching quantization: 76%|#######6 | 647/846 [00:47<00:13, 14.90it/s]
505
+ Attaching quantization: 76%|#######6 | 647/846 [00:47<00:13, 14.90it/s]
506
+ Attaching quantization: 77%|#######6 | 649/846 [00:47<00:16, 11.90it/s]
507
+ Attaching quantization: 77%|#######6 | 649/846 [00:47<00:16, 11.90it/s]
508
+ Attaching quantization: 78%|#######7 | 657/846 [00:47<00:09, 20.19it/s]
509
+ Attaching quantization: 78%|#######7 | 657/846 [00:47<00:09, 20.19it/s]
510
+ Attaching quantization: 78%|#######8 | 661/846 [00:48<00:19, 9.35it/s]
511
+ Attaching quantization: 78%|#######8 | 661/846 [00:48<00:19, 9.35it/s]
512
+ Attaching quantization: 78%|#######8 | 664/846 [00:49<00:18, 9.58it/s]
513
+ Attaching quantization: 78%|#######8 | 664/846 [00:49<00:18, 9.58it/s]
514
+ Attaching quantization: 79%|#######9 | 672/846 [00:49<00:12, 14.36it/s]
515
+ Attaching quantization: 79%|#######9 | 672/846 [00:49<00:12, 14.36it/s]
516
+ Attaching quantization: 80%|#######9 | 675/846 [00:49<00:13, 12.92it/s]
517
+ Attaching quantization: 80%|#######9 | 675/846 [00:49<00:13, 12.92it/s]
518
+ Attaching quantization: 80%|######## | 677/846 [00:50<00:14, 11.82it/s]
519
+ Attaching quantization: 80%|######## | 677/846 [00:50<00:14, 11.82it/s]
520
+ Attaching quantization: 81%|######## | 684/846 [00:50<00:08, 18.50it/s]
521
+ Attaching quantization: 81%|######## | 684/846 [00:50<00:08, 18.50it/s]
522
+ Attaching quantization: 81%|########1 | 688/846 [00:50<00:09, 16.36it/s]
523
+ Attaching quantization: 81%|########1 | 688/846 [00:50<00:09, 16.36it/s]
524
+ Attaching quantization: 82%|########1 | 691/846 [00:50<00:10, 14.66it/s]
525
+ Attaching quantization: 82%|########1 | 691/846 [00:50<00:10, 14.66it/s]
526
+ Attaching quantization: 83%|########2 | 700/846 [00:50<00:06, 22.45it/s]
527
+ Attaching quantization: 83%|########2 | 700/846 [00:50<00:06, 22.45it/s]
528
+ Attaching quantization: 83%|########3 | 704/846 [00:51<00:07, 17.98it/s]
529
+ Attaching quantization: 83%|########3 | 704/846 [00:51<00:07, 17.98it/s]
530
+ Attaching quantization: 84%|########3 | 707/846 [00:51<00:07, 18.42it/s]
531
+ Attaching quantization: 84%|########3 | 707/846 [00:51<00:07, 18.42it/s]
532
+ Attaching quantization: 84%|########4 | 714/846 [00:51<00:05, 24.22it/s]
533
+ Attaching quantization: 84%|########4 | 714/846 [00:51<00:05, 24.22it/s]
534
+ Attaching quantization: 85%|########4 | 717/846 [00:51<00:06, 19.63it/s]
535
+ Attaching quantization: 85%|########4 | 717/846 [00:51<00:06, 19.63it/s]
536
+ Attaching quantization: 85%|########5 | 720/846 [00:52<00:06, 18.26it/s]
537
+ Attaching quantization: 85%|########5 | 720/846 [00:52<00:06, 18.26it/s]
538
+ Attaching quantization: 86%|########6 | 728/846 [00:52<00:04, 25.30it/s]
539
+ Attaching quantization: 86%|########6 | 728/846 [00:52<00:04, 25.30it/s]
540
+ Attaching quantization: 86%|########6 | 731/846 [00:52<00:05, 20.12it/s]
541
+ Attaching quantization: 86%|########6 | 731/846 [00:52<00:05, 20.12it/s]
542
+ Attaching quantization: 87%|########6 | 734/846 [00:52<00:06, 18.49it/s]
543
+ Attaching quantization: 87%|########6 | 734/846 [00:52<00:06, 18.49it/s]
544
+ Attaching quantization: 88%|########7 | 742/846 [00:52<00:04, 25.14it/s]
545
+ Attaching quantization: 88%|########7 | 742/846 [00:52<00:04, 25.14it/s]
546
+ Attaching quantization: 88%|########8 | 745/846 [00:53<00:05, 20.19it/s]
547
+ Attaching quantization: 88%|########8 | 745/846 [00:53<00:05, 20.19it/s]
548
+ Attaching quantization: 88%|########8 | 748/846 [00:53<00:05, 18.40it/s]
549
+ Attaching quantization: 88%|########8 | 748/846 [00:53<00:05, 18.40it/s]
550
+ Attaching quantization: 89%|########9 | 756/846 [00:53<00:03, 25.60it/s]
551
+ Attaching quantization: 89%|########9 | 756/846 [00:53<00:03, 25.60it/s]
552
+ Attaching quantization: 90%|########9 | 759/846 [00:53<00:04, 20.21it/s]
553
+ Attaching quantization: 90%|########9 | 759/846 [00:53<00:04, 20.21it/s]
554
+ Attaching quantization: 90%|######### | 762/846 [00:54<00:04, 18.54it/s]
555
+ Attaching quantization: 90%|######### | 762/846 [00:54<00:04, 18.54it/s]
556
+ Attaching quantization: 91%|#########1| 770/846 [00:54<00:03, 25.33it/s]
557
+ Attaching quantization: 91%|#########1| 770/846 [00:54<00:03, 25.33it/s]
558
+ Attaching quantization: 91%|#########1| 773/846 [00:54<00:03, 20.17it/s]
559
+ Attaching quantization: 91%|#########1| 773/846 [00:54<00:03, 20.17it/s]
560
+ Attaching quantization: 92%|#########1| 776/846 [00:54<00:03, 18.28it/s]
561
+ Attaching quantization: 92%|#########1| 776/846 [00:54<00:03, 18.28it/s]
562
+ Attaching quantization: 93%|#########2| 784/846 [00:54<00:02, 24.91it/s]
563
+ Attaching quantization: 93%|#########2| 784/846 [00:54<00:02, 24.91it/s]
564
+ Attaching quantization: 93%|#########3| 787/846 [00:55<00:02, 19.93it/s]
565
+ Attaching quantization: 93%|#########3| 787/846 [00:55<00:02, 19.93it/s]
566
+ Attaching quantization: 93%|#########3| 790/846 [00:55<00:03, 18.37it/s]
567
+ Attaching quantization: 93%|#########3| 790/846 [00:55<00:03, 18.37it/s]
568
+ Attaching quantization: 94%|#########4| 798/846 [00:55<00:01, 25.35it/s]
569
+ Attaching quantization: 94%|#########4| 798/846 [00:55<00:01, 25.35it/s]
570
+ Attaching quantization: 95%|#########4| 801/846 [00:55<00:02, 20.28it/s]
571
+ Attaching quantization: 95%|#########4| 801/846 [00:55<00:02, 20.28it/s]
572
+ Attaching quantization: 95%|#########5| 804/846 [00:56<00:02, 18.52it/s]
573
+ Attaching quantization: 95%|#########5| 804/846 [00:56<00:02, 18.52it/s]
574
+ Attaching quantization: 96%|#########5| 812/846 [00:56<00:01, 25.59it/s]
575
+ Attaching quantization: 96%|#########5| 812/846 [00:56<00:01, 25.59it/s]
576
+ Attaching quantization: 96%|#########6| 815/846 [00:56<00:01, 20.46it/s]
577
+ Attaching quantization: 96%|#########6| 815/846 [00:56<00:01, 20.46it/s]
578
+ Attaching quantization: 97%|#########6| 818/846 [00:56<00:01, 18.82it/s]
579
+ Attaching quantization: 97%|#########6| 818/846 [00:56<00:01, 18.82it/s]
580
+ Attaching quantization: 98%|#########7| 826/846 [00:56<00:00, 26.21it/s]
581
+ Attaching quantization: 98%|#########7| 826/846 [00:56<00:00, 26.21it/s]
582
+ Attaching quantization: 98%|#########7| 829/846 [00:57<00:00, 20.73it/s]
583
+ Attaching quantization: 98%|#########7| 829/846 [00:57<00:00, 20.73it/s]
584
+ Attaching quantization: 98%|#########8| 832/846 [00:57<00:00, 19.06it/s]
585
+ Attaching quantization: 98%|#########8| 832/846 [00:57<00:00, 19.06it/s]
586
+ Attaching quantization: 99%|#########9| 840/846 [00:57<00:00, 26.15it/s]
587
+ Attaching quantization: 99%|#########9| 840/846 [00:57<00:00, 26.15it/s]
588
+ Attaching quantization: 100%|#########9| 843/846 [00:57<00:00, 20.53it/s]
589
+ Attaching quantization: 100%|#########9| 843/846 [00:57<00:00, 20.53it/s]
590
+ Attaching quantization: 100%|##########| 846/846 [00:57<00:00, 21.00it/s]
591
+ Attaching quantization: 100%|##########| 846/846 [00:57<00:00, 21.00it/s]
592
+ Attaching quantization: 100%|##########| 846/846 [00:57<00:00, 14.62it/s]
593
+ Attaching quantization: 100%|##########| 846/846 [00:57<00:00, 14.62it/s]
594
+
595
+ - quantizing additional layers
596
+ Moving transformer to CPU
597
+ Text Encoder
598
+
599
+ Loading checkpoint shards: 0%| | 0/4 [00:00<?, ?it/s]
600
+ Loading checkpoint shards: 0%| | 0/4 [00:00<?, ?it/s]
601
+ Loading checkpoint shards: 50%|##### | 2/4 [00:00<00:00, 15.73it/s]
602
+ Loading checkpoint shards: 50%|##### | 2/4 [00:00<00:00, 15.73it/s]
603
+ Loading checkpoint shards: 100%|##########| 4/4 [00:00<00:00, 24.39it/s]
604
+ Loading checkpoint shards: 100%|##########| 4/4 [00:00<00:00, 24.39it/s]
605
+
606
+ Quantizing Text Encoder
607
+ Failed to quantize : Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
608
+ Failed to quantize model: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
609
+ Failed to quantize model.visual: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
610
+ Failed to quantize model.visual.blocks: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
611
+ Failed to quantize model.visual.blocks.0: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
612
+ Failed to quantize model.visual.blocks.0.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
613
+ Failed to quantize model.visual.blocks.0.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
614
+ Failed to quantize model.visual.blocks.1: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
615
+ Failed to quantize model.visual.blocks.1.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
616
+ Failed to quantize model.visual.blocks.1.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
617
+ Failed to quantize model.visual.blocks.2: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
618
+ Failed to quantize model.visual.blocks.2.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
619
+ Failed to quantize model.visual.blocks.2.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
620
+ Failed to quantize model.visual.blocks.3: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
621
+ Failed to quantize model.visual.blocks.3.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
622
+ Failed to quantize model.visual.blocks.3.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
623
+ Failed to quantize model.visual.blocks.4: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
624
+ Failed to quantize model.visual.blocks.4.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
625
+ Failed to quantize model.visual.blocks.4.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
626
+ Failed to quantize model.visual.blocks.5: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
627
+ Failed to quantize model.visual.blocks.5.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
628
+ Failed to quantize model.visual.blocks.5.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
629
+ Failed to quantize model.visual.blocks.6: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
630
+ Failed to quantize model.visual.blocks.6.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
631
+ Failed to quantize model.visual.blocks.6.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
632
+ Failed to quantize model.visual.blocks.7: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
633
+ Failed to quantize model.visual.blocks.7.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
634
+ Failed to quantize model.visual.blocks.7.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
635
+ Failed to quantize model.visual.blocks.8: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
636
+ Failed to quantize model.visual.blocks.8.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
637
+ Failed to quantize model.visual.blocks.8.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
638
+ Failed to quantize model.visual.blocks.9: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
639
+ Failed to quantize model.visual.blocks.9.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
640
+ Failed to quantize model.visual.blocks.9.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
641
+ Failed to quantize model.visual.blocks.10: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
642
+ Failed to quantize model.visual.blocks.10.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
643
+ Failed to quantize model.visual.blocks.10.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
644
+ Failed to quantize model.visual.blocks.11: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
645
+ Failed to quantize model.visual.blocks.11.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
646
+ Failed to quantize model.visual.blocks.11.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
647
+ Failed to quantize model.visual.blocks.12: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
648
+ Failed to quantize model.visual.blocks.12.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
649
+ Failed to quantize model.visual.blocks.12.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
650
+ Failed to quantize model.visual.blocks.13: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
651
+ Failed to quantize model.visual.blocks.13.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
652
+ Failed to quantize model.visual.blocks.13.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
653
+ Failed to quantize model.visual.blocks.14: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
654
+ Failed to quantize model.visual.blocks.14.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
655
+ Failed to quantize model.visual.blocks.14.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
656
+ Failed to quantize model.visual.blocks.15: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
657
+ Failed to quantize model.visual.blocks.15.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
658
+ Failed to quantize model.visual.blocks.15.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
659
+ Failed to quantize model.visual.blocks.16: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
660
+ Failed to quantize model.visual.blocks.16.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
661
+ Failed to quantize model.visual.blocks.16.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
662
+ Failed to quantize model.visual.blocks.17: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
663
+ Failed to quantize model.visual.blocks.17.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
664
+ Failed to quantize model.visual.blocks.17.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
665
+ Failed to quantize model.visual.blocks.18: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
666
+ Failed to quantize model.visual.blocks.18.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
667
+ Failed to quantize model.visual.blocks.18.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
668
+ Failed to quantize model.visual.blocks.19: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
669
+ Failed to quantize model.visual.blocks.19.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
670
+ Failed to quantize model.visual.blocks.19.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
671
+ Failed to quantize model.visual.blocks.20: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
672
+ Failed to quantize model.visual.blocks.20.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
673
+ Failed to quantize model.visual.blocks.20.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
674
+ Failed to quantize model.visual.blocks.21: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
675
+ Failed to quantize model.visual.blocks.21.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
676
+ Failed to quantize model.visual.blocks.21.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
677
+ Failed to quantize model.visual.blocks.22: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
678
+ Failed to quantize model.visual.blocks.22.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
679
+ Failed to quantize model.visual.blocks.22.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
680
+ Failed to quantize model.visual.blocks.23: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
681
+ Failed to quantize model.visual.blocks.23.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
682
+ Failed to quantize model.visual.blocks.23.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
683
+ Failed to quantize model.visual.blocks.24: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
684
+ Failed to quantize model.visual.blocks.24.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
685
+ Failed to quantize model.visual.blocks.24.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
686
+ Failed to quantize model.visual.blocks.25: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
687
+ Failed to quantize model.visual.blocks.25.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
688
+ Failed to quantize model.visual.blocks.25.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
689
+ Failed to quantize model.visual.blocks.26: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
690
+ Failed to quantize model.visual.blocks.26.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
691
+ Failed to quantize model.visual.blocks.26.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
692
+ Failed to quantize model.visual.blocks.27: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
693
+ Failed to quantize model.visual.blocks.27.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
694
+ Failed to quantize model.visual.blocks.27.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
695
+ Failed to quantize model.visual.blocks.28: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
696
+ Failed to quantize model.visual.blocks.28.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
697
+ Failed to quantize model.visual.blocks.28.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
698
+ Failed to quantize model.visual.blocks.29: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
699
+ Failed to quantize model.visual.blocks.29.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
700
+ Failed to quantize model.visual.blocks.29.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
701
+ Failed to quantize model.visual.blocks.30: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
702
+ Failed to quantize model.visual.blocks.30.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
703
+ Failed to quantize model.visual.blocks.30.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
704
+ Failed to quantize model.visual.blocks.31: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
705
+ Failed to quantize model.visual.blocks.31.mlp: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
706
+ Failed to quantize model.visual.blocks.31.mlp.down_proj: Expecting input size at 1 dimension: 3420 to be divisible by block_size at 1 dimension: 64
707
+ Loading VAE
708
+ Making pipe
709
+ Preparing Model
710
+ Model Loaded
711
+ create LoRA network. base dim (rank): 16, alpha: 16
712
+ neuron dropout: p=None, rank dropout: p=None, module dropout: p=None
713
+ apply LoRA to Conv2d with kernel size (3,3). dim (rank): 16, alpha: 16
714
+ create LoRA for Text Encoder: 0 modules.
715
+ create LoRA for U-Net: 840 modules.
716
+ enable LoRA for U-Net
717
+ #### IMPORTANT RESUMING FROM D:\Github\ai-toolkit\output/qwen2509_object_removal_512/qwen2509_object_removal_512_000000750.safetensors ####
718
+ Loading from D:\Github\ai-toolkit\output/qwen2509_object_removal_512/qwen2509_object_removal_512_000000750.safetensors
719
+ Missing keys: []
720
+ Found step 750 in metadata, starting from there
721
+ Loading optimizer state from D:\Github\ai-toolkit\output/qwen2509_object_removal_512/optimizer.pt
722
+ Updating optimizer LR from params
723
+ Error running job: Job stopped
724
+
725
+ ========================================
726
+ Result:
727
+ - 0 completed jobs
728
+ - 1 failure
729
+ ========================================
730
+ Traceback (most recent call last):
731
+ Traceback (most recent call last):
732
+ File "/mnt/d/Github/ai-toolkit/run.py", line 120, in <module>
733
+ File "/mnt/d/Github/ai-toolkit/run.py", line 120, in <module>
734
+ main()main()
735
+
736
+ File "/mnt/d/Github/ai-toolkit/run.py", line 108, in main
737
+ File "/mnt/d/Github/ai-toolkit/run.py", line 108, in main
738
+ raise eraise e
739
+
740
+ File "/mnt/d/Github/ai-toolkit/run.py", line 96, in main
741
+ File "/mnt/d/Github/ai-toolkit/run.py", line 96, in main
742
+ job.run()job.run()
743
+
744
+ File "/mnt/d/Github/ai-toolkit/jobs/ExtensionJob.py", line 22, in run
745
+ File "/mnt/d/Github/ai-toolkit/jobs/ExtensionJob.py", line 22, in run
746
+ process.run()process.run()
747
+
748
+ File "/mnt/d/Github/ai-toolkit/jobs/process/BaseSDTrainProcess.py", line 1992, in run
749
+ File "/mnt/d/Github/ai-toolkit/jobs/process/BaseSDTrainProcess.py", line 1992, in run
750
+ self.before_dataset_load()self.before_dataset_load()
751
+
752
+ File "/mnt/d/Github/ai-toolkit/extensions_built_in/sd_trainer/DiffusionTrainer.py", line 275, in before_dataset_load
753
+ File "/mnt/d/Github/ai-toolkit/extensions_built_in/sd_trainer/DiffusionTrainer.py", line 275, in before_dataset_load
754
+ self.maybe_stop()self.maybe_stop()
755
+
756
+ File "/mnt/d/Github/ai-toolkit/extensions_built_in/sd_trainer/DiffusionTrainer.py", line 147, in maybe_stop
757
+ File "/mnt/d/Github/ai-toolkit/extensions_built_in/sd_trainer/DiffusionTrainer.py", line 147, in maybe_stop
758
+ raise Exception("Job stopped")raise Exception("Job stopped")
759
+
760
+ ExceptionException: : Job stoppedJob stopped
761
+
qwen2509_object_removal_512/logs/4_log.txt ADDED
The diff for this file is too large to render. See raw diff
 
qwen2509_object_removal_512/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66d380506dabbab2fab6dffacf79f38c03c7557b5156a74035504fc15bf31f4b
3
+ size 300280909
qwen2509_object_removal_512/pid.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ 3351
qwen2509_object_removal_512/qwen2509_object_removal_512.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:973d76f618896db942e1e1c49785741df67eaceec9e20d7f79f0a56c8ab47ee0
3
+ size 295146200
qwen2509_object_removal_512/qwen2509_object_removal_512_000012750.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f19528b6c3c5fe30f2bff99b8589734b965efc981b89e09291fc702f61c2a97
3
+ size 295146200
qwen2509_object_removal_512/qwen2509_object_removal_512_000013000.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59838fd0c9f737f6ed33991f351cd2401b5ac37658182f6166481a4b8456ca7d
3
+ size 295146200
qwen2509_object_removal_512/qwen2509_object_removal_512_000013250.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f958e9778697f488be284f531a94498dfbcf38a9fc47d751c098f858e66bba1
3
+ size 295146200
qwen2509_object_removal_512/qwen2509_object_removal_512_000013500.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:768f71dab2d75343b4877b5a0d202186e4bb94ae4eb6076ed377e44d576053c3
3
+ size 295146200
qwen2509_object_removal_512/samples/1763725336846__000000000_0.jpg ADDED
qwen2509_object_removal_512/samples/1763725779399__000000000_1.jpg ADDED
qwen2509_object_removal_512/samples/1763726223923__000000000_2.jpg ADDED
qwen2509_object_removal_512/samples/1763730040026__000000250_0.jpg ADDED
qwen2509_object_removal_512/samples/1763730486421__000000250_1.jpg ADDED
qwen2509_object_removal_512/samples/1763730932813__000000250_2.jpg ADDED
qwen2509_object_removal_512/samples/1763734729969__000000500_0.jpg ADDED
qwen2509_object_removal_512/samples/1763735177000__000000500_1.jpg ADDED
qwen2509_object_removal_512/samples/1763735623411__000000500_2.jpg ADDED
qwen2509_object_removal_512/samples/1763739419862__000000750_0.jpg ADDED
qwen2509_object_removal_512/samples/1763739866343__000000750_1.jpg ADDED
qwen2509_object_removal_512/samples/1763740312288__000000750_2.jpg ADDED
qwen2509_object_removal_512/samples/1763750974794__000001000_0.jpg ADDED
qwen2509_object_removal_512/samples/1763751440727__000001000_1.jpg ADDED
qwen2509_object_removal_512/samples/1763751914145__000001000_2.jpg ADDED
qwen2509_object_removal_512/samples/1763755708153__000001250_0.jpg ADDED
qwen2509_object_removal_512/samples/1763756150890__000001250_1.jpg ADDED
qwen2509_object_removal_512/samples/1763756594297__000001250_2.jpg ADDED
qwen2509_object_removal_512/samples/1763760376880__000001500_0.jpg ADDED
qwen2509_object_removal_512/samples/1763760820202__000001500_1.jpg ADDED
qwen2509_object_removal_512/samples/1763761263792__000001500_2.jpg ADDED
qwen2509_object_removal_512/samples/1763765063596__000001750_0.jpg ADDED
qwen2509_object_removal_512/samples/1763765510827__000001750_1.jpg ADDED
qwen2509_object_removal_512/samples/1763765958366__000001750_2.jpg ADDED
qwen2509_object_removal_512/samples/1763769758775__000002000_0.jpg ADDED
qwen2509_object_removal_512/samples/1763770203788__000002000_1.jpg ADDED
qwen2509_object_removal_512/samples/1763770648127__000002000_2.jpg ADDED
qwen2509_object_removal_512/samples/1763774448325__000002250_0.jpg ADDED
qwen2509_object_removal_512/samples/1763774890888__000002250_1.jpg ADDED
qwen2509_object_removal_512/samples/1763775332817__000002250_2.jpg ADDED
qwen2509_object_removal_512/samples/1763779130716__000002500_0.jpg ADDED
qwen2509_object_removal_512/samples/1763779574800__000002500_1.jpg ADDED
qwen2509_object_removal_512/samples/1763780018302__000002500_2.jpg ADDED
qwen2509_object_removal_512/samples/1763783822810__000002750_0.jpg ADDED
qwen2509_object_removal_512/samples/1763784268071__000002750_1.jpg ADDED