Danrisi commited on
Commit
aeb006c
·
verified ·
1 Parent(s): edf43b0

Upload folder using huggingface_hub

Browse files
checkpoint-12000/config.yaml ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ pretrained_model_name_or_path: Qwen/Qwen-Image
2
+ data_config:
3
+ train_batch_size: 1
4
+ num_workers: 4
5
+ img_size: 1024
6
+ caption_dropout_rate: 0.1
7
+ img_dir: /workspace/FinalDataset_Qwen
8
+ random_ratio: true
9
+ caption_type: txt
10
+ train_batch_size: 8
11
+ output_dir: ./output_full_training
12
+ max_train_steps: 20000
13
+ num_train_epochs: 100
14
+ learning_rate: 1.0e-05
15
+ use_8bit_adam: true
16
+ adam_beta1: 0.9
17
+ adam_beta2: 0.999
18
+ adam_weight_decay: 0.01
19
+ adam_epsilon: 1.0e-08
20
+ lr_scheduler: cosine_with_restarts
21
+ lr_warmup_steps: 1000
22
+ max_grad_norm: 1.0
23
+ gradient_accumulation_steps: 4
24
+ mixed_precision: bf16
25
+ freeze_text_encoder: true
26
+ logging_dir: logs
27
+ report_to: null
28
+ checkpointing_steps: 2000
29
+ checkpoints_total_limit: 100
30
+ tracker_project_name: qwen_ultrareal
31
+ resume_from_checkpoint: latest
checkpoint-12000/transformer/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "QwenImageTransformer2DModel",
3
+ "_diffusers_version": "0.36.0.dev0",
4
+ "_name_or_path": "Qwen/Qwen-Image",
5
+ "attention_head_dim": 128,
6
+ "axes_dims_rope": [
7
+ 16,
8
+ 56,
9
+ 56
10
+ ],
11
+ "guidance_embeds": false,
12
+ "in_channels": 64,
13
+ "joint_attention_dim": 3584,
14
+ "num_attention_heads": 24,
15
+ "num_layers": 60,
16
+ "out_channels": 16,
17
+ "patch_size": 2,
18
+ "pooled_projection_dim": 768
19
+ }
checkpoint-12000/transformer/diffusion_pytorch_model-00001-of-00005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e3be158dda409815668d8bc9974a756a33094f2c4cf609285db10f1a141a543
3
+ size 9973578592
checkpoint-12000/transformer/diffusion_pytorch_model-00002-of-00005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13f12b7971f5083de9b224a868aa0f6c18e038bd067ac9a284864d713234c77b
3
+ size 9987326072
checkpoint-12000/transformer/diffusion_pytorch_model-00003-of-00005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c047e0ebd01acefcfee864ab1e763c8e37bb7eee7000ca01489e01e8a595864
3
+ size 9987307440
checkpoint-12000/transformer/diffusion_pytorch_model-00004-of-00005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10325474f64aff06db5f4612306594519291037accde85ff0163394552779331
3
+ size 9930685712
checkpoint-12000/transformer/diffusion_pytorch_model-00005-of-00005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:072638cadcf8149a2ef0305b9872cc6b96d54a21f95961b105ca9adf65954b2e
3
+ size 982130472
checkpoint-12000/transformer/diffusion_pytorch_model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-14000/config.yaml ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ pretrained_model_name_or_path: Qwen/Qwen-Image
2
+ data_config:
3
+ train_batch_size: 1
4
+ num_workers: 4
5
+ img_size: 1024
6
+ caption_dropout_rate: 0.1
7
+ img_dir: /workspace/FinalDataset_Qwen
8
+ random_ratio: true
9
+ caption_type: txt
10
+ train_batch_size: 8
11
+ output_dir: ./output_full_training
12
+ max_train_steps: 20000
13
+ num_train_epochs: 100
14
+ learning_rate: 1.0e-05
15
+ use_8bit_adam: true
16
+ adam_beta1: 0.9
17
+ adam_beta2: 0.999
18
+ adam_weight_decay: 0.01
19
+ adam_epsilon: 1.0e-08
20
+ lr_scheduler: cosine_with_restarts
21
+ lr_warmup_steps: 1000
22
+ max_grad_norm: 1.0
23
+ gradient_accumulation_steps: 4
24
+ mixed_precision: bf16
25
+ freeze_text_encoder: true
26
+ logging_dir: logs
27
+ report_to: null
28
+ checkpointing_steps: 2000
29
+ checkpoints_total_limit: 100
30
+ tracker_project_name: qwen_ultrareal
31
+ resume_from_checkpoint: latest
checkpoint-14000/transformer/config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "QwenImageTransformer2DModel",
3
+ "_diffusers_version": "0.36.0.dev0",
4
+ "_name_or_path": "Qwen/Qwen-Image",
5
+ "attention_head_dim": 128,
6
+ "axes_dims_rope": [
7
+ 16,
8
+ 56,
9
+ 56
10
+ ],
11
+ "guidance_embeds": false,
12
+ "in_channels": 64,
13
+ "joint_attention_dim": 3584,
14
+ "num_attention_heads": 24,
15
+ "num_layers": 60,
16
+ "out_channels": 16,
17
+ "patch_size": 2,
18
+ "pooled_projection_dim": 768
19
+ }
checkpoint-14000/transformer/diffusion_pytorch_model-00001-of-00005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc995f891320bc62d976b5a8e1d0baee1a897b233079c3889e658765ba8184a8
3
+ size 9973578592
checkpoint-14000/transformer/diffusion_pytorch_model-00002-of-00005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e799e650635db92a15d7ffc257debbf0bc0a3cd45de400f33b91f4acfd33c403
3
+ size 9987326072
checkpoint-14000/transformer/diffusion_pytorch_model-00003-of-00005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ebe11db456db13ccc6b38a997d39f6aea9735e56bf3fa83163c3e8ab801d49a0
3
+ size 9987307440
checkpoint-14000/transformer/diffusion_pytorch_model-00004-of-00005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7047256e47c65441bfce4671fcfeaa10d65286b34933dfa3e2354548ce9c1c8a
3
+ size 9930685712
checkpoint-14000/transformer/diffusion_pytorch_model-00005-of-00005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44143509cab14dad07aa497df474b51cded5909e0e8bb95499546e3902c679b3
3
+ size 982130472
checkpoint-14000/transformer/diffusion_pytorch_model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
logs/training.log CHANGED
@@ -22,3 +22,7 @@
22
  2025-09-06 11:57:57.360 | INFO | __main__:save_full_model:116 - Model saved successfully to ./output_full_training/checkpoint-8000
23
  2025-09-06 14:15:11.529 | INFO | __main__:save_full_model:99 - Saving full model to ./output_full_training/checkpoint-10000
24
  2025-09-06 14:15:33.474 | INFO | __main__:save_full_model:116 - Model saved successfully to ./output_full_training/checkpoint-10000
 
 
 
 
 
22
  2025-09-06 11:57:57.360 | INFO | __main__:save_full_model:116 - Model saved successfully to ./output_full_training/checkpoint-8000
23
  2025-09-06 14:15:11.529 | INFO | __main__:save_full_model:99 - Saving full model to ./output_full_training/checkpoint-10000
24
  2025-09-06 14:15:33.474 | INFO | __main__:save_full_model:116 - Model saved successfully to ./output_full_training/checkpoint-10000
25
+ 2025-09-06 16:32:57.914 | INFO | __main__:save_full_model:99 - Saving full model to ./output_full_training/checkpoint-12000
26
+ 2025-09-06 16:33:20.020 | INFO | __main__:save_full_model:116 - Model saved successfully to ./output_full_training/checkpoint-12000
27
+ 2025-09-06 18:51:00.642 | INFO | __main__:save_full_model:99 - Saving full model to ./output_full_training/checkpoint-14000
28
+ 2025-09-06 18:51:23.758 | INFO | __main__:save_full_model:116 - Model saved successfully to ./output_full_training/checkpoint-14000