diff --git a/.gitattributes b/.gitattributes
index 8a608b4272f5b5c5abb5a2d36f19f102be0cc2db..d018acd0dad2382c575331cc665696357968f04e 100644
--- a/.gitattributes
+++ b/.gitattributes
@@ -289,3 +289,39 @@ ACE_plus/examples/exp_example/20250704031317/eval_probe/step-10/BaseSolver_eval_
ACE_plus/examples/exp_example/20250704031317/eval_probe/step-10/BaseSolver_eval_image_step_10/9_0_3_probe_0_\[736_1408_3\].jpg filter=lfs diff=lfs merge=lfs -text
ACE_plus/examples/exp_example/20250704031317/eval_probe/step-10/BaseSolver_eval_image_step_10/0_0_3_probe_0_\[736_1408_3\].jpg filter=lfs diff=lfs merge=lfs -text
ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/2_0_5_probe_0_\[736_1408_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/3_0_5_probe_0_\[880_1168_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/1_0_0_probe_0_\[752_1360_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/0_0_5_probe_0_\[848_1232_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/1_0_5_probe_0_\[752_1360_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/6_0_5_probe_0_\[864_1184_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/6_0_3_probe_0_\[864_1184_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/4_0_3_probe_0_\[784_1328_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/7_0_3_probe_0_\[752_1376_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/2_0_3_probe_0_\[736_1408_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/2_0_0_probe_0_\[736_1408_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/9_0_3_probe_0_\[752_1376_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/5_0_5_probe_0_\[864_1200_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/3_0_3_probe_0_\[880_1168_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/0_0_3_probe_0_\[848_1232_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/1_0_3_probe_0_\[752_1360_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/0_0_3_probe_0_\[720_1440_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/0_0_0_probe_0_\[720_1440_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/1_0_5_probe_0_\[736_1392_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/5_0_5_probe_0_\[720_1440_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/6_0_5_probe_0_\[864_1200_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/8_0_5_probe_0_\[768_1344_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/8_0_3_probe_0_\[768_1344_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/1_0_3_probe_0_\[736_1392_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/9_0_3_probe_0_\[720_1424_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/2_0_0_probe_0_\[720_1424_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/2_0_3_probe_0_\[720_1424_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/5_0_3_probe_0_\[720_1440_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/2_0_5_probe_0_\[720_1424_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/7_0_3_probe_0_\[784_1296_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/6_0_3_probe_0_\[864_1200_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/5_0_0_probe_0_\[720_1440_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/3_0_5_probe_0_\[752_1376_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/3_0_3_probe_0_\[752_1376_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/1_0_0_probe_0_\[736_1392_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/3_0_0_probe_0_\[752_1376_3\].jpg filter=lfs diff=lfs merge=lfs -text
+ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/4_0_3_probe_0_\[864_1200_3\].jpg filter=lfs diff=lfs merge=lfs -text
diff --git a/ACE_plus/examples/exp_example/20250602092700/checkpoints/ldm_step-500/README.md b/ACE_plus/examples/exp_example/20250602092700/checkpoints/ldm_step-500/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..2cae90886aa26c06c289f0d1b74a06bb39921dba
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250602092700/checkpoints/ldm_step-500/README.md
@@ -0,0 +1,10 @@
+## Training procedure
+
+### Framework versions
+
+
+- SWIFT 3.4.0
+### Base model information
+
+
+- BaseModel Class LatentDiffusionACEPlus
diff --git a/ACE_plus/examples/exp_example/20250602092700/checkpoints/ldm_step-500/configuration.json b/ACE_plus/examples/exp_example/20250602092700/checkpoints/ldm_step-500/configuration.json
new file mode 100644
index 0000000000000000000000000000000000000000..9e26dfeeb6e641a33dae4961196235bdb965b21b
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250602092700/checkpoints/ldm_step-500/configuration.json
@@ -0,0 +1 @@
+{}
\ No newline at end of file
diff --git a/ACE_plus/examples/exp_example/20250602092700/noise_schedule.png b/ACE_plus/examples/exp_example/20250602092700/noise_schedule.png
new file mode 100644
index 0000000000000000000000000000000000000000..2d08dfab306b571d00dd9cbd03468d2b58de1f7d
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250602092700/noise_schedule.png differ
diff --git a/ACE_plus/examples/exp_example/20250602092700/sampler_schedule.png b/ACE_plus/examples/exp_example/20250602092700/sampler_schedule.png
new file mode 100644
index 0000000000000000000000000000000000000000..f12f547ac8a8c4c5f946276a86a85e60bdd5f107
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250602092700/sampler_schedule.png differ
diff --git a/ACE_plus/examples/exp_example/20250602092700/std_log.txt b/ACE_plus/examples/exp_example/20250602092700/std_log.txt
new file mode 100644
index 0000000000000000000000000000000000000000..eb6aca8eca2316db83df0b2ed741ca6791c2ccfd
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250602092700/std_log.txt
@@ -0,0 +1,620 @@
+scepter [INFO] 2025-06-02 09:27:02,317 [File: logger.py Function: init_logger at line 85] Running task with log file: /home/Ubuntu/Downloads/Unmodel/ACE_plus/./examples/exp_example/20250602092700/std_log.txt
+scepter [WARNING] 2025-06-02 09:27:02,412 [File: import_utils.py Function: import_module at line 325] ('DATASETS', 'ACEPlusDataset') not found in ast index file
+scepter [INFO] 2025-06-02 09:27:02,413 [File: ace_plus_dataset.py Function: read_data_list at line 151] subject has 4 samples.
+scepter [INFO] 2025-06-02 09:27:02,414 [File: registry.py Function: __init__ at line 185] Built dataloader with len 9223372036854775807
+scepter [WARNING] 2025-06-02 09:27:02,414 [File: import_utils.py Function: import_module at line 325] ('DATASETS', 'ACEPlusDataset') not found in ast index file
+scepter [INFO] 2025-06-02 09:27:02,414 [File: ace_plus_dataset.py Function: read_data_list at line 151] subject has 4 samples.
+scepter [INFO] 2025-06-02 09:27:02,415 [File: registry.py Function: __init__ at line 185] Built dataloader with len 4
+scepter [INFO] 2025-06-02 09:28:16,963 [File: flux.py Function: load_pretrained_model at line 450] Restored from /home/Ubuntu/Downloads/Unmodel/Reference_models/flux1-fill-dev.safetensors with 0 missing and 0 unexpected keys
+scepter [INFO] 2025-06-02 09:28:16,982 [File: ace_plus_ldm.py Function: construct_network at line 62] all parameters:11.90B
+scepter [INFO] 2025-06-02 09:28:17,613 [File: ae_module.py Function: construct_model at line 76] AE Module XFORMERS_IS_AVAILBLE : True
+scepter [INFO] 2025-06-02 09:28:18,411 [File: ae_kl.py Function: init_from_ckpt at line 400] Restored from /home/Ubuntu/Downloads/Unmodel/Reference_models/ae.safetensors with 0 missing and 0 unexpected keys
+scepter [INFO] 2025-06-02 09:28:30,434 [File: diffusion_solver.py Function: add_tuner at line 788] [('base_model.model.double_blocks.0.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.0.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.0.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.0.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.0.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.0.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.0.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.0.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.0.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.0.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.0.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.0.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.0.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.0.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.0.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.0.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.0.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.0.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.0.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.0.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.1.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.1.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.1.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.1.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.1.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.1.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.1.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.1.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.1.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.1.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.1.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.1.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.1.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.1.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.1.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.1.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.1.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.1.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.1.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.1.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.2.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.2.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.2.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.2.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.2.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.2.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.2.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.2.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.2.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.2.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.2.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.2.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.2.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.2.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.2.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.2.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.2.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.2.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.2.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.2.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.3.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.3.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.3.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.3.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.3.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.3.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.3.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.3.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.3.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.3.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.3.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.3.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.3.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.3.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.3.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.3.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.3.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.3.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.3.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.3.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.4.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.4.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.4.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.4.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.4.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.4.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.4.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.4.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.4.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.4.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.4.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.4.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.4.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.4.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.4.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.4.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.4.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.4.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.4.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.4.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.5.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.5.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.5.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.5.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.5.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.5.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.5.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.5.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.5.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.5.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.5.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.5.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.5.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.5.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.5.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.5.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.5.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.5.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.5.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.5.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.6.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.6.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.6.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.6.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.6.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.6.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.6.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.6.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.6.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.6.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.6.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.6.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.6.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.6.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.6.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.6.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.6.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.6.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.6.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.6.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.7.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.7.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.7.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.7.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.7.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.7.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.7.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.7.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.7.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.7.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.7.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.7.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.7.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.7.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.7.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.7.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.7.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.7.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.7.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.7.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.8.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.8.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.8.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.8.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.8.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.8.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.8.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.8.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.8.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.8.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.8.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.8.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.8.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.8.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.8.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.8.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.8.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.8.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.8.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.8.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.9.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.9.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.9.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.9.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.9.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.9.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.9.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.9.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.9.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.9.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.9.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.9.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.9.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.9.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.9.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.9.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.9.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.9.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.9.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.9.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.10.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.10.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.10.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.10.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.10.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.10.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.10.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.10.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.10.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.10.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.10.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.10.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.10.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.10.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.10.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.10.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.10.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.10.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.10.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.10.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.11.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.11.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.11.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.11.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.11.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.11.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.11.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.11.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.11.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.11.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.11.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.11.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.11.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.11.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.11.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.11.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.11.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.11.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.11.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.11.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.12.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.12.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.12.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.12.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.12.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.12.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.12.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.12.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.12.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.12.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.12.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.12.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.12.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.12.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.12.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.12.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.12.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.12.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.12.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.12.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.13.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.13.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.13.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.13.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.13.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.13.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.13.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.13.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.13.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.13.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.13.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.13.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.13.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.13.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.13.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.13.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.13.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.13.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.13.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.13.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.14.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.14.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.14.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.14.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.14.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.14.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.14.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.14.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.14.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.14.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.14.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.14.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.14.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.14.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.14.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.14.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.14.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.14.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.14.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.14.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.15.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.15.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.15.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.15.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.15.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.15.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.15.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.15.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.15.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.15.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.15.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.15.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.15.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.15.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.15.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.15.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.15.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.15.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.15.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.15.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.16.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.16.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.16.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.16.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.16.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.16.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.16.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.16.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.16.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.16.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.16.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.16.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.16.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.16.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.16.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.16.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.16.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.16.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.16.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.16.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.17.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.17.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.17.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.17.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.17.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.17.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.17.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.17.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.17.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.17.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.17.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.17.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.17.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.17.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.17.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.17.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.17.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.17.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.17.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.17.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.18.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.18.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.18.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.18.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.18.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.18.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.18.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.18.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.18.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.18.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.18.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.18.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.18.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.18.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.18.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.18.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.18.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.18.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.18.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.18.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.0.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.0.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.0.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.0.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.0.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.0.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.1.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.1.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.1.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.1.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.1.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.1.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.2.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.2.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.2.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.2.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.2.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.2.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.3.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.3.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.3.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.3.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.3.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.3.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.4.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.4.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.4.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.4.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.4.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.4.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.5.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.5.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.5.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.5.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.5.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.5.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.6.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.6.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.6.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.6.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.6.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.6.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.7.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.7.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.7.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.7.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.7.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.7.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.8.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.8.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.8.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.8.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.8.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.8.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.9.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.9.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.9.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.9.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.9.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.9.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.10.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.10.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.10.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.10.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.10.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.10.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.11.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.11.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.11.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.11.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.11.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.11.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.12.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.12.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.12.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.12.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.12.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.12.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.13.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.13.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.13.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.13.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.13.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.13.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.14.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.14.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.14.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.14.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.14.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.14.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.15.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.15.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.15.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.15.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.15.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.15.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.16.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.16.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.16.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.16.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.16.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.16.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.17.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.17.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.17.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.17.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.17.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.17.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.18.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.18.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.18.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.18.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.18.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.18.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.19.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.19.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.19.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.19.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.19.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.19.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.20.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.20.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.20.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.20.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.20.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.20.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.21.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.21.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.21.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.21.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.21.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.21.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.22.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.22.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.22.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.22.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.22.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.22.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.23.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.23.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.23.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.23.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.23.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.23.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.24.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.24.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.24.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.24.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.24.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.24.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.25.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.25.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.25.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.25.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.25.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.25.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.26.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.26.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.26.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.26.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.26.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.26.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.27.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.27.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.27.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.27.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.27.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.27.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.28.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.28.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.28.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.28.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.28.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.28.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.29.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.29.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.29.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.29.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.29.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.29.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.30.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.30.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.30.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.30.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.30.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.30.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.31.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.31.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.31.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.31.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.31.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.31.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.32.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.32.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.32.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.32.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.32.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.32.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.33.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.33.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.33.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.33.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.33.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.33.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.34.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.34.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.34.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.34.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.34.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.34.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.35.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.35.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.35.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.35.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.35.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.35.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.36.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.36.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.36.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.36.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.36.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.36.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.37.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.37.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.37.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.37.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.37.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.37.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64]))]
+scepter [INFO] 2025-06-02 09:28:30,450 [File: diffusion_solver.py Function: print_model_params_status at line 996] Load trainable params 306315264 / 17178094051 = 1.78%, train part: {'model.double_blocks': 171835392, 'model.single_blocks': 134479872}.
+scepter [INFO] 2025-06-02 09:28:30,450 [File: diffusion_solver.py Function: print_model_params_status at line 1000] Load frozen params 16871778787 / 17178094051 = 98.22%, frozen part: {'model': 11902587968, 'first_stage_model': 83819683, 'cond_stage_model': 4885371136}.
+scepter [INFO] 2025-06-02 09:29:07,140 [File: diffusion_solver.py Function: set_up at line 230] SwiftModel(
+ (base_model): LatentDiffusionACEPlus LatentDiffusionACEPlus(
+ (model): FluxMRModiACEPlus FluxMRModiACEPlus(
+ (pe_embedder): EmbedND()
+ (img_in): Linear(in_features=448, out_features=3072, bias=True)
+ (time_in): MLPEmbedder(
+ (in_layer): Linear(in_features=256, out_features=3072, bias=True)
+ (silu): SiLU()
+ (out_layer): Linear(in_features=3072, out_features=3072, bias=True)
+ )
+ (vector_in): MLPEmbedder(
+ (in_layer): Linear(in_features=768, out_features=3072, bias=True)
+ (silu): SiLU()
+ (out_layer): Linear(in_features=3072, out_features=3072, bias=True)
+ )
+ (guidance_in): MLPEmbedder(
+ (in_layer): Linear(in_features=256, out_features=3072, bias=True)
+ (silu): SiLU()
+ (out_layer): Linear(in_features=3072, out_features=3072, bias=True)
+ )
+ (txt_in): Linear(in_features=4096, out_features=3072, bias=True)
+ (double_blocks): ModuleList(
+ (0-18): 19 x DoubleStreamBlock(
+ (img_mod): Modulation(
+ (lin): lora.Linear(
+ (base_layer): Linear(in_features=3072, out_features=18432, bias=True)
+ (lora_dropout): ModuleDict(
+ (0_SwiftLoRA): Identity()
+ )
+ (lora_A): ModuleDict(
+ (0_SwiftLoRA): Linear(in_features=3072, out_features=64, bias=False)
+ )
+ (lora_B): ModuleDict(
+ (0_SwiftLoRA): Linear(in_features=64, out_features=18432, bias=False)
+ )
+ (lora_embedding_A): ParameterDict()
+ (lora_embedding_B): ParameterDict()
+ (lora_magnitude_vector): ModuleDict()
+ )
+ )
+ (img_norm1): LayerNorm((3072,), eps=1e-06, elementwise_affine=False)
+ (img_attn): SelfAttention(
+ (qkv): lora.Linear(
+ (base_layer): Linear(in_features=3072, out_features=9216, bias=True)
+ (lora_dropout): ModuleDict(
+ (0_SwiftLoRA): Identity()
+ )
+ (lora_A): ModuleDict(
+ (0_SwiftLoRA): Linear(in_features=3072, out_features=64, bias=False)
+ )
+ (lora_B): ModuleDict(
+ (0_SwiftLoRA): Linear(in_features=64, out_features=9216, bias=False)
+ )
+ (lora_embedding_A): ParameterDict()
+ (lora_embedding_B): ParameterDict()
+ (lora_magnitude_vector): ModuleDict()
+ )
+ (norm): QKNorm(
+ (query_norm): RMSNorm()
+ (key_norm): RMSNorm()
+ )
+ (proj): lora.Linear(
+ (base_layer): Linear(in_features=3072, out_features=3072, bias=True)
+ (lora_dropout): ModuleDict(
+ (0_SwiftLoRA): Identity()
+ )
+ (lora_A): ModuleDict(
+ (0_SwiftLoRA): Linear(in_features=3072, out_features=64, bias=False)
+ )
+ (lora_B): ModuleDict(
+ (0_SwiftLoRA): Linear(in_features=64, out_features=3072, bias=False)
+ )
+ (lora_embedding_A): ParameterDict()
+ (lora_embedding_B): ParameterDict()
+ (lora_magnitude_vector): ModuleDict()
+ )
+ )
+ (img_norm2): LayerNorm((3072,), eps=1e-06, elementwise_affine=False)
+ (img_mlp): Sequential(
+ (0): lora.Linear(
+ (base_layer): Linear(in_features=3072, out_features=12288, bias=True)
+ (lora_dropout): ModuleDict(
+ (0_SwiftLoRA): Identity()
+ )
+ (lora_A): ModuleDict(
+ (0_SwiftLoRA): Linear(in_features=3072, out_features=64, bias=False)
+ )
+ (lora_B): ModuleDict(
+ (0_SwiftLoRA): Linear(in_features=64, out_features=12288, bias=False)
+ )
+ (lora_embedding_A): ParameterDict()
+ (lora_embedding_B): ParameterDict()
+ (lora_magnitude_vector): ModuleDict()
+ )
+ (1): GELU(approximate='tanh')
+ (2): lora.Linear(
+ (base_layer): Linear(in_features=12288, out_features=3072, bias=True)
+ (lora_dropout): ModuleDict(
+ (0_SwiftLoRA): Identity()
+ )
+ (lora_A): ModuleDict(
+ (0_SwiftLoRA): Linear(in_features=12288, out_features=64, bias=False)
+ )
+ (lora_B): ModuleDict(
+ (0_SwiftLoRA): Linear(in_features=64, out_features=3072, bias=False)
+ )
+ (lora_embedding_A): ParameterDict()
+ (lora_embedding_B): ParameterDict()
+ (lora_magnitude_vector): ModuleDict()
+ )
+ )
+ (txt_mod): Modulation(
+ (lin): lora.Linear(
+ (base_layer): Linear(in_features=3072, out_features=18432, bias=True)
+ (lora_dropout): ModuleDict(
+ (0_SwiftLoRA): Identity()
+ )
+ (lora_A): ModuleDict(
+ (0_SwiftLoRA): Linear(in_features=3072, out_features=64, bias=False)
+ )
+ (lora_B): ModuleDict(
+ (0_SwiftLoRA): Linear(in_features=64, out_features=18432, bias=False)
+ )
+ (lora_embedding_A): ParameterDict()
+ (lora_embedding_B): ParameterDict()
+ (lora_magnitude_vector): ModuleDict()
+ )
+ )
+ (txt_norm1): LayerNorm((3072,), eps=1e-06, elementwise_affine=False)
+ (txt_attn): SelfAttention(
+ (qkv): lora.Linear(
+ (base_layer): Linear(in_features=3072, out_features=9216, bias=True)
+ (lora_dropout): ModuleDict(
+ (0_SwiftLoRA): Identity()
+ )
+ (lora_A): ModuleDict(
+ (0_SwiftLoRA): Linear(in_features=3072, out_features=64, bias=False)
+ )
+ (lora_B): ModuleDict(
+ (0_SwiftLoRA): Linear(in_features=64, out_features=9216, bias=False)
+ )
+ (lora_embedding_A): ParameterDict()
+ (lora_embedding_B): ParameterDict()
+ (lora_magnitude_vector): ModuleDict()
+ )
+ (norm): QKNorm(
+ (query_norm): RMSNorm()
+ (key_norm): RMSNorm()
+ )
+ (proj): lora.Linear(
+ (base_layer): Linear(in_features=3072, out_features=3072, bias=True)
+ (lora_dropout): ModuleDict(
+ (0_SwiftLoRA): Identity()
+ )
+ (lora_A): ModuleDict(
+ (0_SwiftLoRA): Linear(in_features=3072, out_features=64, bias=False)
+ )
+ (lora_B): ModuleDict(
+ (0_SwiftLoRA): Linear(in_features=64, out_features=3072, bias=False)
+ )
+ (lora_embedding_A): ParameterDict()
+ (lora_embedding_B): ParameterDict()
+ (lora_magnitude_vector): ModuleDict()
+ )
+ )
+ (txt_norm2): LayerNorm((3072,), eps=1e-06, elementwise_affine=False)
+ (txt_mlp): Sequential(
+ (0): lora.Linear(
+ (base_layer): Linear(in_features=3072, out_features=12288, bias=True)
+ (lora_dropout): ModuleDict(
+ (0_SwiftLoRA): Identity()
+ )
+ (lora_A): ModuleDict(
+ (0_SwiftLoRA): Linear(in_features=3072, out_features=64, bias=False)
+ )
+ (lora_B): ModuleDict(
+ (0_SwiftLoRA): Linear(in_features=64, out_features=12288, bias=False)
+ )
+ (lora_embedding_A): ParameterDict()
+ (lora_embedding_B): ParameterDict()
+ (lora_magnitude_vector): ModuleDict()
+ )
+ (1): GELU(approximate='tanh')
+ (2): lora.Linear(
+ (base_layer): Linear(in_features=12288, out_features=3072, bias=True)
+ (lora_dropout): ModuleDict(
+ (0_SwiftLoRA): Identity()
+ )
+ (lora_A): ModuleDict(
+ (0_SwiftLoRA): Linear(in_features=12288, out_features=64, bias=False)
+ )
+ (lora_B): ModuleDict(
+ (0_SwiftLoRA): Linear(in_features=64, out_features=3072, bias=False)
+ )
+ (lora_embedding_A): ParameterDict()
+ (lora_embedding_B): ParameterDict()
+ (lora_magnitude_vector): ModuleDict()
+ )
+ )
+ )
+ )
+ (single_blocks): ModuleList(
+ (0-37): 38 x SingleStreamBlock(
+ (linear1): lora.Linear(
+ (base_layer): Linear(in_features=3072, out_features=21504, bias=True)
+ (lora_dropout): ModuleDict(
+ (0_SwiftLoRA): Identity()
+ )
+ (lora_A): ModuleDict(
+ (0_SwiftLoRA): Linear(in_features=3072, out_features=64, bias=False)
+ )
+ (lora_B): ModuleDict(
+ (0_SwiftLoRA): Linear(in_features=64, out_features=21504, bias=False)
+ )
+ (lora_embedding_A): ParameterDict()
+ (lora_embedding_B): ParameterDict()
+ (lora_magnitude_vector): ModuleDict()
+ )
+ (linear2): lora.Linear(
+ (base_layer): Linear(in_features=15360, out_features=3072, bias=True)
+ (lora_dropout): ModuleDict(
+ (0_SwiftLoRA): Identity()
+ )
+ (lora_A): ModuleDict(
+ (0_SwiftLoRA): Linear(in_features=15360, out_features=64, bias=False)
+ )
+ (lora_B): ModuleDict(
+ (0_SwiftLoRA): Linear(in_features=64, out_features=3072, bias=False)
+ )
+ (lora_embedding_A): ParameterDict()
+ (lora_embedding_B): ParameterDict()
+ (lora_magnitude_vector): ModuleDict()
+ )
+ (norm): QKNorm(
+ (query_norm): RMSNorm()
+ (key_norm): RMSNorm()
+ )
+ (pre_norm): LayerNorm((3072,), eps=1e-06, elementwise_affine=False)
+ (mlp_act): GELU(approximate='tanh')
+ (modulation): Modulation(
+ (lin): lora.Linear(
+ (base_layer): Linear(in_features=3072, out_features=9216, bias=True)
+ (lora_dropout): ModuleDict(
+ (0_SwiftLoRA): Identity()
+ )
+ (lora_A): ModuleDict(
+ (0_SwiftLoRA): Linear(in_features=3072, out_features=64, bias=False)
+ )
+ (lora_B): ModuleDict(
+ (0_SwiftLoRA): Linear(in_features=64, out_features=9216, bias=False)
+ )
+ (lora_embedding_A): ParameterDict()
+ (lora_embedding_B): ParameterDict()
+ (lora_magnitude_vector): ModuleDict()
+ )
+ )
+ )
+ )
+ (final_layer): LastLayer(
+ (norm_final): LayerNorm((3072,), eps=1e-06, elementwise_affine=False)
+ (linear): Linear(in_features=3072, out_features=64, bias=True)
+ (adaLN_modulation): Sequential(
+ (0): SiLU()
+ (1): Linear(in_features=3072, out_features=6144, bias=True)
+ )
+ )
+ )
+ (first_stage_model): AutoencoderKLFlux AutoencoderKLFlux(
+ (encoder): Encoder Encoder(
+ (conv_in): Conv2d(3, 128, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ (down): ModuleList(
+ (0): Module(
+ (block): ModuleList(
+ (0-1): 2 x ResnetBlock(
+ (norm1): GroupNorm(32, 128, eps=1e-06, affine=True)
+ (conv1): Conv2d(128, 128, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ (norm2): GroupNorm(32, 128, eps=1e-06, affine=True)
+ (dropout): Dropout(p=0.0, inplace=False)
+ (conv2): Conv2d(128, 128, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ )
+ )
+ (attn): ModuleList()
+ (downsample): Downsample(
+ (conv): Conv2d(128, 128, kernel_size=(3, 3), stride=(2, 2))
+ )
+ )
+ (1): Module(
+ (block): ModuleList(
+ (0): ResnetBlock(
+ (norm1): GroupNorm(32, 128, eps=1e-06, affine=True)
+ (conv1): Conv2d(128, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ (norm2): GroupNorm(32, 256, eps=1e-06, affine=True)
+ (dropout): Dropout(p=0.0, inplace=False)
+ (conv2): Conv2d(256, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ (nin_shortcut): Conv2d(128, 256, kernel_size=(1, 1), stride=(1, 1))
+ )
+ (1): ResnetBlock(
+ (norm1): GroupNorm(32, 256, eps=1e-06, affine=True)
+ (conv1): Conv2d(256, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ (norm2): GroupNorm(32, 256, eps=1e-06, affine=True)
+ (dropout): Dropout(p=0.0, inplace=False)
+ (conv2): Conv2d(256, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ )
+ )
+ (attn): ModuleList()
+ (downsample): Downsample(
+ (conv): Conv2d(256, 256, kernel_size=(3, 3), stride=(2, 2))
+ )
+ )
+ (2): Module(
+ (block): ModuleList(
+ (0): ResnetBlock(
+ (norm1): GroupNorm(32, 256, eps=1e-06, affine=True)
+ (conv1): Conv2d(256, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ (norm2): GroupNorm(32, 512, eps=1e-06, affine=True)
+ (dropout): Dropout(p=0.0, inplace=False)
+ (conv2): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ (nin_shortcut): Conv2d(256, 512, kernel_size=(1, 1), stride=(1, 1))
+ )
+ (1): ResnetBlock(
+ (norm1): GroupNorm(32, 512, eps=1e-06, affine=True)
+ (conv1): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ (norm2): GroupNorm(32, 512, eps=1e-06, affine=True)
+ (dropout): Dropout(p=0.0, inplace=False)
+ (conv2): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ )
+ )
+ (attn): ModuleList()
+ (downsample): Downsample(
+ (conv): Conv2d(512, 512, kernel_size=(3, 3), stride=(2, 2))
+ )
+ )
+ (3): Module(
+ (block): ModuleList(
+ (0-1): 2 x ResnetBlock(
+ (norm1): GroupNorm(32, 512, eps=1e-06, affine=True)
+ (conv1): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ (norm2): GroupNorm(32, 512, eps=1e-06, affine=True)
+ (dropout): Dropout(p=0.0, inplace=False)
+ (conv2): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ )
+ )
+ (attn): ModuleList()
+ )
+ )
+ (mid): Module(
+ (block_1): ResnetBlock(
+ (norm1): GroupNorm(32, 512, eps=1e-06, affine=True)
+ (conv1): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ (norm2): GroupNorm(32, 512, eps=1e-06, affine=True)
+ (dropout): Dropout(p=0.0, inplace=False)
+ (conv2): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ )
+ (attn_1): MemoryEfficientAttention(
+ (norm): GroupNorm(32, 512, eps=1e-06, affine=True)
+ (q): Conv2d(512, 512, kernel_size=(1, 1), stride=(1, 1))
+ (k): Conv2d(512, 512, kernel_size=(1, 1), stride=(1, 1))
+ (v): Conv2d(512, 512, kernel_size=(1, 1), stride=(1, 1))
+ (proj_out): Conv2d(512, 512, kernel_size=(1, 1), stride=(1, 1))
+ )
+ (block_2): ResnetBlock(
+ (norm1): GroupNorm(32, 512, eps=1e-06, affine=True)
+ (conv1): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ (norm2): GroupNorm(32, 512, eps=1e-06, affine=True)
+ (dropout): Dropout(p=0.0, inplace=False)
+ (conv2): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ )
+ )
+ (norm_out): GroupNorm(32, 512, eps=1e-06, affine=True)
+ (conv_out): Conv2d(512, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ )
+ (decoder): Decoder Decoder(
+ (conv_in): Conv2d(16, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ (mid): Module(
+ (block_1): ResnetBlock(
+ (norm1): GroupNorm(32, 512, eps=1e-06, affine=True)
+ (conv1): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ (norm2): GroupNorm(32, 512, eps=1e-06, affine=True)
+ (dropout): Dropout(p=0.0, inplace=False)
+ (conv2): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ )
+ (attn_1): MemoryEfficientAttention(
+ (norm): GroupNorm(32, 512, eps=1e-06, affine=True)
+ (q): Conv2d(512, 512, kernel_size=(1, 1), stride=(1, 1))
+ (k): Conv2d(512, 512, kernel_size=(1, 1), stride=(1, 1))
+ (v): Conv2d(512, 512, kernel_size=(1, 1), stride=(1, 1))
+ (proj_out): Conv2d(512, 512, kernel_size=(1, 1), stride=(1, 1))
+ )
+ (block_2): ResnetBlock(
+ (norm1): GroupNorm(32, 512, eps=1e-06, affine=True)
+ (conv1): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ (norm2): GroupNorm(32, 512, eps=1e-06, affine=True)
+ (dropout): Dropout(p=0.0, inplace=False)
+ (conv2): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ )
+ )
+ (up): ModuleList(
+ (0): Module(
+ (block): ModuleList(
+ (0): ResnetBlock(
+ (norm1): GroupNorm(32, 256, eps=1e-06, affine=True)
+ (conv1): Conv2d(256, 128, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ (norm2): GroupNorm(32, 128, eps=1e-06, affine=True)
+ (dropout): Dropout(p=0.0, inplace=False)
+ (conv2): Conv2d(128, 128, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ (nin_shortcut): Conv2d(256, 128, kernel_size=(1, 1), stride=(1, 1))
+ )
+ (1-2): 2 x ResnetBlock(
+ (norm1): GroupNorm(32, 128, eps=1e-06, affine=True)
+ (conv1): Conv2d(128, 128, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ (norm2): GroupNorm(32, 128, eps=1e-06, affine=True)
+ (dropout): Dropout(p=0.0, inplace=False)
+ (conv2): Conv2d(128, 128, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ )
+ )
+ (attn): ModuleList()
+ )
+ (1): Module(
+ (block): ModuleList(
+ (0): ResnetBlock(
+ (norm1): GroupNorm(32, 512, eps=1e-06, affine=True)
+ (conv1): Conv2d(512, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ (norm2): GroupNorm(32, 256, eps=1e-06, affine=True)
+ (dropout): Dropout(p=0.0, inplace=False)
+ (conv2): Conv2d(256, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ (nin_shortcut): Conv2d(512, 256, kernel_size=(1, 1), stride=(1, 1))
+ )
+ (1-2): 2 x ResnetBlock(
+ (norm1): GroupNorm(32, 256, eps=1e-06, affine=True)
+ (conv1): Conv2d(256, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ (norm2): GroupNorm(32, 256, eps=1e-06, affine=True)
+ (dropout): Dropout(p=0.0, inplace=False)
+ (conv2): Conv2d(256, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ )
+ )
+ (attn): ModuleList()
+ (upsample): Upsample(
+ (conv): Conv2d(256, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ )
+ )
+ (2-3): 2 x Module(
+ (block): ModuleList(
+ (0-2): 3 x ResnetBlock(
+ (norm1): GroupNorm(32, 512, eps=1e-06, affine=True)
+ (conv1): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ (norm2): GroupNorm(32, 512, eps=1e-06, affine=True)
+ (dropout): Dropout(p=0.0, inplace=False)
+ (conv2): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ )
+ )
+ (attn): ModuleList()
+ (upsample): Upsample(
+ (conv): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ )
+ )
+ )
+ (norm_out): GroupNorm(32, 128, eps=1e-06, affine=True)
+ (conv_out): Conv2d(128, 3, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
+ )
+ (conv1): Identity()
+ (conv2): Identity()
+ )
+ (cond_stage_model): T5ACEPlusClipFluxEmbedder T5ACEPlusClipFluxEmbedder(
+ (t5_model): ACEHFEmbedder ACEHFEmbedder(
+ (hf_module): T5EncoderModel(
+ (shared): Embedding(32128, 4096)
+ (encoder): T5Stack(
+ (embed_tokens): Embedding(32128, 4096)
+ (block): ModuleList(
+ (0): T5Block(
+ (layer): ModuleList(
+ (0): T5LayerSelfAttention(
+ (SelfAttention): T5Attention(
+ (q): Linear(in_features=4096, out_features=4096, bias=False)
+ (k): Linear(in_features=4096, out_features=4096, bias=False)
+ (v): Linear(in_features=4096, out_features=4096, bias=False)
+ (o): Linear(in_features=4096, out_features=4096, bias=False)
+ (relative_attention_bias): Embedding(32, 64)
+ )
+ (layer_norm): T5LayerNorm()
+ (dropout): Dropout(p=0.1, inplace=False)
+ )
+ (1): T5LayerFF(
+ (DenseReluDense): T5DenseGatedActDense(
+ (wi_0): Linear(in_features=4096, out_features=10240, bias=False)
+ (wi_1): Linear(in_features=4096, out_features=10240, bias=False)
+ (wo): Linear(in_features=10240, out_features=4096, bias=False)
+ (dropout): Dropout(p=0.1, inplace=False)
+ (act): NewGELUActivation()
+ )
+ (layer_norm): T5LayerNorm()
+ (dropout): Dropout(p=0.1, inplace=False)
+ )
+ )
+ )
+ (1-23): 23 x T5Block(
+ (layer): ModuleList(
+ (0): T5LayerSelfAttention(
+ (SelfAttention): T5Attention(
+ (q): Linear(in_features=4096, out_features=4096, bias=False)
+ (k): Linear(in_features=4096, out_features=4096, bias=False)
+ (v): Linear(in_features=4096, out_features=4096, bias=False)
+ (o): Linear(in_features=4096, out_features=4096, bias=False)
+ )
+ (layer_norm): T5LayerNorm()
+ (dropout): Dropout(p=0.1, inplace=False)
+ )
+ (1): T5LayerFF(
+ (DenseReluDense): T5DenseGatedActDense(
+ (wi_0): Linear(in_features=4096, out_features=10240, bias=False)
+ (wi_1): Linear(in_features=4096, out_features=10240, bias=False)
+ (wo): Linear(in_features=10240, out_features=4096, bias=False)
+ (dropout): Dropout(p=0.1, inplace=False)
+ (act): NewGELUActivation()
+ )
+ (layer_norm): T5LayerNorm()
+ (dropout): Dropout(p=0.1, inplace=False)
+ )
+ )
+ )
+ )
+ (final_layer_norm): T5LayerNorm()
+ (dropout): Dropout(p=0.1, inplace=False)
+ )
+ )
+ )
+ (clip_model): ACEHFEmbedder ACEHFEmbedder(
+ (hf_module): CLIPTextModel(
+ (text_model): CLIPTextTransformer(
+ (embeddings): CLIPTextEmbeddings(
+ (token_embedding): Embedding(49408, 768)
+ (position_embedding): Embedding(77, 768)
+ )
+ (encoder): CLIPEncoder(
+ (layers): ModuleList(
+ (0-11): 12 x CLIPEncoderLayer(
+ (self_attn): CLIPSdpaAttention(
+ (k_proj): Linear(in_features=768, out_features=768, bias=True)
+ (v_proj): Linear(in_features=768, out_features=768, bias=True)
+ (q_proj): Linear(in_features=768, out_features=768, bias=True)
+ (out_proj): Linear(in_features=768, out_features=768, bias=True)
+ )
+ (layer_norm1): LayerNorm((768,), eps=1e-05, elementwise_affine=True)
+ (mlp): CLIPMLP(
+ (activation_fn): QuickGELUActivation()
+ (fc1): Linear(in_features=768, out_features=3072, bias=True)
+ (fc2): Linear(in_features=3072, out_features=768, bias=True)
+ )
+ (layer_norm2): LayerNorm((768,), eps=1e-05, elementwise_affine=True)
+ )
+ )
+ )
+ (final_layer_norm): LayerNorm((768,), eps=1e-05, elementwise_affine=True)
+ )
+ )
+ )
+ )
+ )
+)
+scepter [INFO] 2025-06-02 09:29:07,179 [File: log.py Function: before_solve at line 260] Tensorboard: save to ./examples/exp_example/20250602092700/tensorboard
+scepter [INFO] 2025-06-02 09:51:38,590 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [20/100000], data_time: 42.3201(42.3201), time: 67.5697(67.5697), loss: 0.2030(0.2030), throughput: 5349/day, all_throughput: 20, pg0_lr: 0.001000, scale: 1.000000, [24mins 36secs 0.02%(85days 9hours 50mins 19secs)]
+scepter [INFO] 2025-06-02 10:13:55,895 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [40/100000], data_time: 41.7579(42.0390), time: 66.8652(67.2175), loss: 0.1496(0.1763), throughput: 5343/day, all_throughput: 40, pg0_lr: 0.001000, scale: 1.000000, [46mins 53secs 0.04%(81days 9hours 1mins 35secs)]
+scepter [INFO] 2025-06-02 10:25:05,166 [File: checkpoint.py Function: after_iter at line 109] Saving checkpoint after 50 steps
+scepter [INFO] 2025-06-02 10:36:13,836 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [60/100000], data_time: 41.7211(41.9331), time: 66.8971(67.1107), loss: 0.1511(0.1679), throughput: 5344/day, all_throughput: 60, pg0_lr: 0.001000, scale: 1.000000, [1hours 9mins 11secs 0.06%(80days 48mins 10secs)]
+scepter [INFO] 2025-06-02 10:58:22,386 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [80/100000], data_time: 41.4367(41.8090), time: 66.4275(66.9399), loss: 0.1644(0.1670), throughput: 5352/day, all_throughput: 80, pg0_lr: 0.001000, scale: 1.000000, [1hours 31mins 19secs 0.08%(79days 5hours 14mins 48secs)]
+scepter [INFO] 2025-06-02 11:20:31,032 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [100/100000], data_time: 41.3202(41.7112), time: 66.4323(66.8384), loss: 0.1641(0.1664), throughput: 5354/day, all_throughput: 100, pg0_lr: 0.001000, scale: 1.000000, [1hours 53mins 28secs 0.10%(78days 17hours 23mins 32secs)]
+scepter [INFO] 2025-06-02 11:20:31,033 [File: checkpoint.py Function: after_iter at line 109] Saving checkpoint after 100 steps
+scepter [INFO] 2025-06-02 11:42:39,626 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [120/100000], data_time: 41.4255(41.6636), time: 66.4297(66.7703), loss: 0.1851(0.1695), throughput: 5355/day, all_throughput: 120, pg0_lr: 0.001000, scale: 1.000000, [2hours 15mins 37secs 0.12%(78days 9hours 21mins 15secs)]
+scepter [INFO] 2025-06-02 12:04:51,591 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [140/100000], data_time: 41.7225(41.6720), time: 66.5982(66.7457), loss: 0.2146(0.1760), throughput: 5357/day, all_throughput: 140, pg0_lr: 0.001000, scale: 1.000000, [2hours 37mins 49secs 0.14%(78days 4hours 10mins 30secs)]
+scepter [INFO] 2025-06-02 12:15:54,666 [File: checkpoint.py Function: after_iter at line 109] Saving checkpoint after 150 steps
+scepter [INFO] 2025-06-02 12:26:59,151 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [160/100000], data_time: 41.3753(41.6349), time: 66.3780(66.6997), loss: 0.2634(0.1869), throughput: 5358/day, all_throughput: 160, pg0_lr: 0.001000, scale: 1.000000, [2hours 59mins 56secs 0.16%(77days 23hours 26mins 5secs)]
+scepter [INFO] 2025-06-02 12:49:04,260 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [180/100000], data_time: 41.2891(41.5965), time: 66.2555(66.6504), loss: 0.2875(0.1981), throughput: 5356/day, all_throughput: 180, pg0_lr: 0.001000, scale: 1.000000, [3hours 22mins 1secs 0.18%(77days 19hours 17mins 18secs)]
+scepter [INFO] 2025-06-02 13:11:15,836 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [200/100000], data_time: 41.4977(41.5866), time: 66.5788(66.6432), loss: 0.2561(0.2039), throughput: 5354/day, all_throughput: 200, pg0_lr: 0.001000, scale: 1.000000, [3hours 44mins 13secs 0.20%(77days 16hours 47mins 39secs)]
+scepter [INFO] 2025-06-02 13:11:15,837 [File: checkpoint.py Function: after_iter at line 109] Saving checkpoint after 200 steps
+scepter [INFO] 2025-06-02 13:33:26,818 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [220/100000], data_time: 41.5564(41.5839), time: 66.5491(66.6346), loss: 0.2442(0.2076), throughput: 5358/day, all_throughput: 220, pg0_lr: 0.001000, scale: 1.000000, [4hours 6mins 24secs 0.22%(77days 14hours 36mins 40secs)]
+scepter [INFO] 2025-06-02 13:55:36,662 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [240/100000], data_time: 41.6124(41.5863), time: 66.4922(66.6228), loss: 0.2165(0.2083), throughput: 5359/day, all_throughput: 240, pg0_lr: 0.001000, scale: 1.000000, [4hours 28mins 34secs 0.24%(77days 12hours 35mins 57secs)]
+scepter [INFO] 2025-06-02 14:06:40,396 [File: checkpoint.py Function: after_iter at line 109] Saving checkpoint after 250 steps
+scepter [INFO] 2025-06-02 14:17:47,873 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [260/100000], data_time: 41.6454(41.5908), time: 66.5605(66.6180), loss: 0.1728(0.2056), throughput: 5361/day, all_throughput: 260, pg0_lr: 0.001000, scale: 1.000000, [4hours 50mins 45secs 0.26%(77days 10hours 59mins 7secs)]
+scepter [INFO] 2025-06-02 14:39:46,950 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [280/100000], data_time: 41.1921(41.5623), time: 65.9539(66.5706), loss: 0.1977(0.2050), throughput: 5365/day, all_throughput: 280, pg0_lr: 0.001000, scale: 1.000000, [5hours 12mins 44secs 0.28%(77days 8hours 20mins 56secs)]
+scepter [INFO] 2025-06-02 15:01:46,703 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [300/100000], data_time: 41.1670(41.5360), time: 65.9876(66.5317), loss: 0.1706(0.2027), throughput: 5369/day, all_throughput: 300, pg0_lr: 0.001000, scale: 1.000000, [5hours 34mins 44secs 0.30%(77days 6hours 4mins 39secs)]
+scepter [INFO] 2025-06-02 15:01:46,703 [File: checkpoint.py Function: after_iter at line 109] Saving checkpoint after 300 steps
+scepter [INFO] 2025-06-02 15:23:47,457 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [320/100000], data_time: 41.1280(41.5105), time: 66.0377(66.5008), loss: 0.1706(0.2007), throughput: 5370/day, all_throughput: 320, pg0_lr: 0.001000, scale: 1.000000, [5hours 56mins 45secs 0.32%(77days 4hours 7mins 51secs)]
+scepter [INFO] 2025-06-02 15:45:53,102 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [340/100000], data_time: 41.4600(41.5075), time: 66.2822(66.4880), loss: 0.1437(0.1974), throughput: 5374/day, all_throughput: 340, pg0_lr: 0.001000, scale: 1.000000, [6hours 18mins 50secs 0.34%(77days 2hours 46mins 6secs)]
+scepter [INFO] 2025-06-02 15:56:57,492 [File: checkpoint.py Function: after_iter at line 109] Saving checkpoint after 350 steps
+scepter [INFO] 2025-06-02 16:08:04,405 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [360/100000], data_time: 41.6830(41.5173), time: 66.5652(66.4922), loss: 0.1532(0.1949), throughput: 5375/day, all_throughput: 360, pg0_lr: 0.001000, scale: 1.000000, [6hours 41mins 1secs 0.36%(77days 1hours 57mins 5secs)]
+scepter [INFO] 2025-06-02 16:30:13,447 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [380/100000], data_time: 41.4708(41.5148), time: 66.4521(66.4901), loss: 0.1476(0.1924), throughput: 5375/day, all_throughput: 380, pg0_lr: 0.001000, scale: 1.000000, [7hours 3mins 11secs 0.38%(77days 1hours 1mins 0secs)]
+scepter [INFO] 2025-06-02 16:52:20,557 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [400/100000], data_time: 41.3496(41.5065), time: 66.3555(66.4834), loss: 0.1394(0.1898), throughput: 5375/day, all_throughput: 400, pg0_lr: 0.001000, scale: 1.000000, [7hours 25mins 18secs 0.40%(77days 18secs)]
+scepter [INFO] 2025-06-02 16:52:20,558 [File: checkpoint.py Function: after_iter at line 109] Saving checkpoint after 400 steps
+scepter [INFO] 2025-06-02 17:14:31,006 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [420/100000], data_time: 41.6155(41.5117), time: 66.5224(66.4853), loss: 0.1420(0.1875), throughput: 5375/day, all_throughput: 420, pg0_lr: 0.001000, scale: 1.000000, [7hours 47mins 28secs 0.42%(76days 23hours 16mins 28secs)]
+scepter [INFO] 2025-06-02 17:36:43,227 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [440/100000], data_time: 41.4885(41.5107), time: 66.6110(66.4910), loss: 0.1349(0.1851), throughput: 5374/day, all_throughput: 440, pg0_lr: 0.001000, scale: 1.000000, [8hours 9mins 40secs 0.44%(76days 22hours 41mins 17secs)]
+scepter [INFO] 2025-06-02 17:47:48,583 [File: checkpoint.py Function: after_iter at line 109] Saving checkpoint after 450 steps
+scepter [INFO] 2025-06-02 17:58:55,039 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [460/100000], data_time: 41.6236(41.5156), time: 66.5906(66.4953), loss: 0.1423(0.1832), throughput: 5374/day, all_throughput: 460, pg0_lr: 0.001000, scale: 1.000000, [8hours 31mins 52secs 0.46%(76days 22hours 5mins 45secs)]
+scepter [INFO] 2025-06-02 18:21:09,747 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [480/100000], data_time: 41.6391(41.5207), time: 66.7354(66.5053), loss: 0.1559(0.1821), throughput: 5373/day, all_throughput: 480, pg0_lr: 0.001000, scale: 1.000000, [8hours 54mins 7secs 0.48%(76days 21hours 41mins 20secs)]
+scepter [INFO] 2025-06-02 18:43:26,230 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [500/100000], data_time: 41.8578(41.5342), time: 66.8242(66.5181), loss: 0.1370(0.1803), throughput: 5374/day, all_throughput: 500, pg0_lr: 0.001000, scale: 1.000000, [9hours 16mins 23secs 0.50%(76days 21hours 23mins 0secs)]
+scepter [INFO] 2025-06-02 18:43:26,231 [File: checkpoint.py Function: after_iter at line 109] Saving checkpoint after 500 steps
+scepter [INFO] 2025-06-02 19:05:33,274 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [520/100000], data_time: 41.4453(41.5308), time: 66.3522(66.5117), loss: 0.2116(0.1815), throughput: 5375/day, all_throughput: 520, pg0_lr: 0.001000, scale: 1.000000, [9hours 38mins 30secs 0.52%(76days 20hours 34mins 15secs)]
+scepter [INFO] 2025-06-02 19:27:53,538 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [540/100000], data_time: 41.6550(41.5354), time: 67.0132(66.5303), loss: 0.1492(0.1803), throughput: 5372/day, all_throughput: 540, pg0_lr: 0.001000, scale: 1.000000, [10hours 51secs 0.54%(76days 20hours 28mins 3secs)]
+scepter [INFO] 2025-06-02 19:39:01,432 [File: checkpoint.py Function: after_iter at line 109] Saving checkpoint after 550 steps
+scepter [INFO] 2025-06-02 19:50:06,907 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [560/100000], data_time: 41.6412(41.5392), time: 66.6685(66.5352), loss: 0.2414(0.1825), throughput: 5373/day, all_throughput: 560, pg0_lr: 0.001000, scale: 1.000000, [10hours 23mins 4secs 0.56%(76days 20hours 18secs)]
+scepter [INFO] 2025-06-02 20:12:08,182 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [580/100000], data_time: 41.2246(41.5283), time: 66.0638(66.5189), loss: 0.2961(0.1864), throughput: 5373/day, all_throughput: 580, pg0_lr: 0.001000, scale: 1.000000, [10hours 45mins 5secs 0.58%(76days 18hours 58mins 22secs)]
+scepter [INFO] 2025-06-02 20:34:10,081 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [600/100000], data_time: 41.2012(41.5174), time: 66.0950(66.5048), loss: 0.2136(0.1873), throughput: 5374/day, all_throughput: 600, pg0_lr: 0.001000, scale: 1.000000, [11hours 7mins 7secs 0.60%(76days 18hours 50secs)]
+scepter [INFO] 2025-06-02 20:34:10,082 [File: checkpoint.py Function: after_iter at line 109] Saving checkpoint after 600 steps
+scepter [INFO] 2025-06-02 20:56:21,375 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [620/100000], data_time: 41.6078(41.5203), time: 66.5647(66.5067), loss: 0.1896(0.1874), throughput: 5375/day, all_throughput: 620, pg0_lr: 0.001000, scale: 1.000000, [11hours 29mins 18secs 0.62%(76days 17hours 30mins 41secs)]
+scepter [INFO] 2025-06-02 21:18:27,057 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [640/100000], data_time: 41.2849(41.5130), time: 66.2841(66.4998), loss: 0.1944(0.1876), throughput: 5376/day, all_throughput: 640, pg0_lr: 0.001000, scale: 1.000000, [11hours 51mins 24secs 0.64%(76days 16hours 46mins 31secs)]
+scepter [INFO] 2025-06-02 21:29:27,002 [File: checkpoint.py Function: after_iter at line 109] Saving checkpoint after 650 steps
+scepter [INFO] 2025-06-02 21:40:27,523 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [660/100000], data_time: 41.1508(41.5020), time: 66.0233(66.4853), loss: 0.1546(0.1866), throughput: 5376/day, all_throughput: 660, pg0_lr: 0.001000, scale: 1.000000, [12hours 13mins 25secs 0.66%(76days 15hours 50mins 36secs)]
diff --git a/ACE_plus/examples/exp_example/20250602092700/train.yaml b/ACE_plus/examples/exp_example/20250602092700/train.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..9c8f141686d91b64513fe45a530c356a9f311e72
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250602092700/train.yaml
@@ -0,0 +1,288 @@
+ENV:
+ BACKEND: nccl
+ SEED: 1999
+SOLVER:
+ # NAME DESCRIPTION: TYPE: default: 'LatentUfitSolver'
+ NAME: FormalACEPlusSolver
+ # MAX_STEPS DESCRIPTION: The total steps for training. TYPE: int default: 100000
+ MAX_STEPS: 100000
+ # USE_AMP DESCRIPTION: Use amp to surpport mix precision or not, default is False. TYPE: bool default: False
+ USE_AMP: True
+ # DTYPE DESCRIPTION: The precision for training. TYPE: str default: 'float32'
+ DTYPE: bfloat16
+ ENABLE_GRADSCALER: False
+ # USE_FAIRSCALE DESCRIPTION: Use fairscale as the backend of ddp, default False. TYPE: bool default: False
+ USE_FAIRSCALE: False
+ USE_ORIG_PARAMS: True
+ USE_FSDP: True # lora use ddp(USE_FSDP=False), else use fsdp(USE_FSDP=True)
+ # LOAD_MODEL_ONLY DESCRIPTION: Only load the model rather than the optimizer and schedule, default is False. TYPE: bool default: False
+ LOAD_MODEL_ONLY: False
+ # RESUME_FROM DESCRIPTION: Resume from some state of training! TYPE: str default: ''
+ RESUME_FROM:
+ # WORK_DIR DESCRIPTION: Save dir of the training log or model. TYPE: str default: ''
+ WORK_DIR: ./examples/exp_example/
+ # LOG_FILE DESCRIPTION: Save log path. TYPE: str default: ''
+ LOG_FILE: std_log.txt
+ # LOG_TRAIN_NUM DESCRIPTION: The number samples used to log in training phase. TYPE: int default: -1
+ LOG_TRAIN_NUM: 16
+ # FSDP_REDUCE_DTYPE DESCRIPTION: The dtype of reduce in FSDP. TYPE: str default: 'float16'
+ FSDP_REDUCE_DTYPE: float32
+ # FSDP_BUFFER_DTYPE DESCRIPTION: The dtype of buffer in FSDP. TYPE: str default: 'float16'
+ FSDP_BUFFER_DTYPE: float32
+ # FSDP_SHARD_MODULES DESCRIPTION: The modules to be sharded in FSDP. TYPE: list default: ['model']
+ FSDP_SHARD_MODULES:
+ - MODULE: 'model.model'
+ FSDP_GROUP: [ 'single_blocks', 'double_blocks']
+ - MODULE: 'cond_stage_model.t5_model.hf_module.encoder'
+ FSDP_GROUP: [ 'block' ] #
+ SAVE_MODULES: [ 'model'] #
+ TRAIN_MODULES: ['model']
+
+ #
+ FILE_SYSTEM:
+ - NAME: HuggingfaceFs
+ TEMP_DIR: ./cache
+ - NAME: ModelscopeFs
+ TEMP_DIR: ./cache
+ #
+ MODEL:
+ NAME: LatentDiffusionACEPlus
+ PARAMETERIZATION: rf
+ TIMESTEPS: 1000
+ GUIDE_SCALE: 1.0
+ PRETRAINED_MODEL:
+ IGNORE_KEYS: [ ]
+ USE_EMA: False
+ EVAL_EMA: False
+ SIZE_FACTOR: 8
+ DIFFUSION:
+ NAME: DiffusionFluxRF
+ PREDICTION_TYPE: raw
+ NOISE_NORM: True
+ # NOISE_SCHEDULER DESCRIPTION: TYPE: default: ''
+ NOISE_SCHEDULER:
+ NAME: FlowMatchFluxShiftScheduler
+ SHIFT: False
+ PRE_T_SAMPLE: True
+ PRE_T_SAMPLE_FOLD: 1
+ SIGMOID_SCALE: 1
+ BASE_SHIFT: 0.5
+ MAX_SHIFT: 1.15
+ SAMPLER_SCHEDULER:
+ NAME: FlowMatchFluxShiftScheduler
+ SHIFT: True
+ PRE_T_SAMPLE: False
+ SIGMOID_SCALE: 1
+ BASE_SHIFT: 0.5
+ MAX_SHIFT: 1.15
+
+ #
+ DIFFUSION_MODEL:
+ # NAME DESCRIPTION: TYPE: default: 'Flux'
+ NAME: FluxMRModiACEPlus
+ PRETRAINED_MODEL: /home/Ubuntu/Downloads/Unmodel/Reference_models/flux1-fill-dev.safetensors
+ # IN_CHANNELS DESCRIPTION: model's input channels. TYPE: int default: 64
+ IN_CHANNELS: 448
+ # OUT_CHANNELS DESCRIPTION: model's input channels. TYPE: int default: 64
+ OUT_CHANNELS: 64
+ # HIDDEN_SIZE DESCRIPTION: model's hidden size. TYPE: int default: 1024
+ HIDDEN_SIZE: 3072
+ REDUX_DIM: 1152
+ # NUM_HEADS DESCRIPTION: number of heads in the transformer. TYPE: int default: 16
+ NUM_HEADS: 24
+ # AXES_DIM DESCRIPTION: dimensions of the axes of the positional encoding. TYPE: list default: [16, 56, 56]
+ AXES_DIM: [ 16, 56, 56 ]
+ # THETA DESCRIPTION: theta for positional encoding. TYPE: int default: 10000
+ THETA: 10000
+ # VEC_IN_DIM DESCRIPTION: dimension of the vector input. TYPE: int default: 768
+ VEC_IN_DIM: 768
+ # GUIDANCE_EMBED DESCRIPTION: whether to use guidance embedding. TYPE: bool default: False
+ GUIDANCE_EMBED: True
+ # CONTEXT_IN_DIM DESCRIPTION: dimension of the context input. TYPE: int default: 4096
+ CONTEXT_IN_DIM: 4096
+ # MLP_RATIO DESCRIPTION: ratio of mlp hidden size to hidden size. TYPE: float default: 4.0
+ MLP_RATIO: 4.0
+ # QKV_BIAS DESCRIPTION: whether to use bias in qkv projection. TYPE: bool default: True
+ QKV_BIAS: True
+ # DEPTH DESCRIPTION: number of transformer blocks. TYPE: int default: 19
+ DEPTH: 19
+ # DEPTH_SINGLE_BLOCKS DESCRIPTION: number of transformer blocks in the single stream block. TYPE: int default: 38
+ DEPTH_SINGLE_BLOCKS: 38
+ # ATTN_BACKEND:setting 'flash_attn' to use flash_attn2, if the version of pytorch > 2.4.0, using 'pytorch' to use pytorch's implementation
+ ATTN_BACKEND: flash_attn
+ # USE_GRAD_CHECKPOINT: setting gc to true can decrease the memory usage.
+ USE_GRAD_CHECKPOINT: True
+
+ #
+ FIRST_STAGE_MODEL:
+ NAME: AutoencoderKLFlux
+ EMBED_DIM: 16
+ PRETRAINED_MODEL: /home/Ubuntu/Downloads/Unmodel/Reference_models/ae.safetensors
+ IGNORE_KEYS: [ ]
+ BATCH_SIZE: 8
+ USE_CONV: False
+ SCALE_FACTOR: 0.3611
+ SHIFT_FACTOR: 0.1159
+ #
+ ENCODER:
+ NAME: Encoder
+ CH: 128
+ OUT_CH: 3
+ NUM_RES_BLOCKS: 2
+ IN_CHANNELS: 3
+ ATTN_RESOLUTIONS: [ ]
+ CH_MULT: [ 1, 2, 4, 4 ]
+ Z_CHANNELS: 16
+ DOUBLE_Z: True
+ DROPOUT: 0.0
+ RESAMP_WITH_CONV: True
+ #
+ DECODER:
+ NAME: Decoder
+ CH: 128
+ OUT_CH: 3
+ NUM_RES_BLOCKS: 2
+ IN_CHANNELS: 3
+ ATTN_RESOLUTIONS: [ ]
+ CH_MULT: [ 1, 2, 4, 4 ]
+ Z_CHANNELS: 16
+ DROPOUT: 0.0
+ RESAMP_WITH_CONV: True
+ GIVE_PRE_END: False
+ TANH_OUT: False
+ #
+ COND_STAGE_MODEL:
+ # NAME DESCRIPTION: TYPE: default: 'T5PlusClipFluxEmbedder'
+ NAME: T5ACEPlusClipFluxEmbedder
+ # T5_MODEL DESCRIPTION: TYPE: default: ''
+ T5_MODEL:
+ # NAME DESCRIPTION: TYPE: default: 'HFEmbedder'
+ NAME: ACEHFEmbedder
+ # HF_MODEL_CLS DESCRIPTION: huggingface cls in transfomer TYPE: NoneType default: None
+ HF_MODEL_CLS: T5EncoderModel
+ # MODEL_PATH DESCRIPTION: model folder path TYPE: NoneType default: None
+ MODEL_PATH: /home/Ubuntu/Downloads/Unmodel/Reference_models/t5_xxl/
+ # HF_TOKENIZER_CLS DESCRIPTION: huggingface cls in transfomer TYPE: NoneType default: None
+ HF_TOKENIZER_CLS: T5Tokenizer
+ # TOKENIZER_PATH DESCRIPTION: tokenizer folder path TYPE: NoneType default: None
+ TOKENIZER_PATH: /home/Ubuntu/Downloads/Unmodel/Reference_models/t5_xxl/
+ ADDED_IDENTIFIER: [ '','{image}', '{caption}', '{mask}', '{ref_image}', '{image1}', '{image2}', '{image3}', '{image4}', '{image5}', '{image6}', '{image7}', '{image8}', '{image9}' ]
+ # MAX_LENGTH DESCRIPTION: max length of input TYPE: int default: 77
+ MAX_LENGTH: 512
+ # OUTPUT_KEY DESCRIPTION: output key TYPE: str default: 'last_hidden_state'
+ OUTPUT_KEY: last_hidden_state
+ # D_TYPE DESCRIPTION: dtype TYPE: str default: 'bfloat16'
+ D_TYPE: bfloat16
+ # BATCH_INFER DESCRIPTION: batch infer TYPE: bool default: False
+ BATCH_INFER: False
+ CLEAN: whitespace
+ # CLIP_MODEL DESCRIPTION: TYPE: default: ''
+ CLIP_MODEL:
+ # NAME DESCRIPTION: TYPE: default: 'HFEmbedder'
+ NAME: ACEHFEmbedder
+ # HF_MODEL_CLS DESCRIPTION: huggingface cls in transfomer TYPE: NoneType default: None
+ HF_MODEL_CLS: CLIPTextModel
+ # MODEL_PATH DESCRIPTION: model folder path TYPE: NoneType default: None
+ MODEL_PATH: /home/Ubuntu/Downloads/Unmodel/Reference_models/clip_l/
+ # HF_TOKENIZER_CLS DESCRIPTION: huggingface cls in transfomer TYPE: NoneType default: None
+ HF_TOKENIZER_CLS: CLIPTokenizer
+ # TOKENIZER_PATH DESCRIPTION: tokenizer folder path TYPE: NoneType default: None
+ TOKENIZER_PATH: /home/Ubuntu/Downloads/Unmodel/Reference_models/clip_l/
+ # MAX_LENGTH DESCRIPTION: max length of input TYPE: int default: 77
+ MAX_LENGTH: 77
+ # OUTPUT_KEY DESCRIPTION: output key TYPE: str default: 'last_hidden_state'
+ OUTPUT_KEY: pooler_output
+ # D_TYPE DESCRIPTION: dtype TYPE: str default: 'bfloat16'
+ D_TYPE: bfloat16
+ # BATCH_INFER DESCRIPTION: batch infer TYPE: bool default: False
+ BATCH_INFER: True
+ CLEAN: whitespace
+ TUNER:
+ # THE LORA PARAMETERS
+ - NAME: SwiftLoRA
+ R: 64
+ LORA_ALPHA: 64
+ LORA_DROPOUT: 0.0
+ BIAS: "none"
+ TARGET_MODULES: "(model.double_blocks.*(.qkv|.img_mlp.0|.img_mlp.2|.txt_mlp.0|.txt_mlp.2|.proj|.img_mod.lin|.txt_mod.lin))|(model.single_blocks.*(.linear1|.linear2|.modulation.lin))$"
+ #
+ SAMPLE_ARGS:
+ SAMPLE_STEPS: 28
+ SAMPLER: flow_euler
+ SEED: 42
+ IMAGE_SIZE: [ 2048, 2048 ]
+ #IMAGE_SIZE: [ 1024, 1024 ]
+ GUIDE_SCALE: 50
+
+ LR_SCHEDULER:
+ NAME: StepAnnealingLR
+ WARMUP_STEPS: 0
+ TOTAL_STEPS: 100000
+ DECAY_MODE: 'cosine'
+ #
+ OPTIMIZER:
+ NAME: AdamW
+ LEARNING_RATE: 1e-3
+ BETAS: [ 0.9, 0.999 ]
+ EPS: 1e-6
+ WEIGHT_DECAY: 1e-2
+ AMSGRAD: False
+ #
+ TRAIN_DATA:
+ NAME: ACEPlusDataset
+ MODE: train
+ DATA_LIST: data/train.csv
+ DELIMITER: "#;#"
+ MODIFY_MODE: True
+ # input_image, input_mask, input_reference_image, target_image, instruction, task_type
+ FIELDS: ["edit_image", "edit_mask", "ref_image", "target_image", "prompt", "data_type"]
+ PATH_PREFIX: ""
+ EDIT_TYPE_LIST: []
+ MAX_SEQ_LEN: 4096
+ # MAX_SEQ_LEN: 2048 -Vijay
+ D: 16
+ PIN_MEMORY: True
+ BATCH_SIZE: 1
+ NUM_WORKERS: 4
+ SAMPLER:
+ NAME: LoopSampler
+
+ EVAL_DATA:
+ NAME: ACEPlusDataset
+ MODE: eval
+ DATA_LIST: data/train.csv
+ DELIMITER: "#;#"
+ MODIFY_MODE: True
+ # input_image, input_mask, input_reference_image, target_image, instruction, task_type
+ FIELDS: [ "edit_image", "edit_mask", "ref_image", "target_image", "prompt", "data_type" ]
+ PATH_PREFIX: ""
+ EDIT_TYPE_LIST: [ ]
+ MAX_SEQ_LEN: 4096
+ # MAX_SEQ_LEN: 2048 -Vijay
+ D: 16
+ PIN_MEMORY: True
+ BATCH_SIZE: 1
+ NUM_WORKERS: 4
+
+ TRAIN_HOOKS:
+ - NAME: ACEBackwardHook
+ GRADIENT_CLIP: 1.0
+ PRIORITY: 10
+ - NAME: LogHook
+ LOG_INTERVAL: 20
+ - NAME: ACECheckpointHook
+ INTERVAL: 50
+ #INTERVAL: 250 --Vijay
+ PRIORITY: 200
+ DISABLE_SNAPSHOT: True
+ - NAME: ProbeDataHook
+ PROB_INTERVAL: 10
+ #PROB_INTERVAL: 50 -Vijay
+ PRIORITY: 0
+ - NAME: TensorboardLogHook
+ LOG_INTERVAL: 50
+ EVAL_HOOKS:
+ - NAME: ProbeDataHook
+ PROB_INTERVAL: 10
+ #PROB_INTERVAL: 50 -Vijay
+ PRIORITY: 0
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/0_0_3_probe_0_[848_1232_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/0_0_3_probe_0_[848_1232_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..e63a44aa5aebd19c2dfa9fdae59278b1ed551c19
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/0_0_3_probe_0_[848_1232_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d477a76c92bd8f9f49ec8cc715e8238da792c3e21a170239ce81977fc9e35e11
+size 166553
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/0_0_5_probe_0_[848_1232_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/0_0_5_probe_0_[848_1232_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..7dc6509f4ca916c5d87d062eaa2c362523bd1d80
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/0_0_5_probe_0_[848_1232_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3afcabe9f07713196caf23789bc714153f0facc3d6c11d76f86d229835e9aafc
+size 116582
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/1_0_0_probe_0_[752_1360_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/1_0_0_probe_0_[752_1360_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..bc14b693ddbad2aa969312cb6c19897a9ec14259
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/1_0_0_probe_0_[752_1360_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3924ff0e8ea07a06b939860538ce9b9b0294917f929c0f63b5cd5211223eba0a
+size 117204
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/1_0_3_probe_0_[752_1360_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/1_0_3_probe_0_[752_1360_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..909b08e29538169b034a7d1efd669d6c4a3fbefd
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/1_0_3_probe_0_[752_1360_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f4014dae6fefe302b69608c403145bd8312afa20514fc726bd8f0ed37da71756
+size 121816
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/1_0_5_probe_0_[752_1360_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/1_0_5_probe_0_[752_1360_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..b4eed8d3f97e9ea3ae5d819bf072ae91ebb49d50
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/1_0_5_probe_0_[752_1360_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:28862e4e42f4768b07ebd12cbdc484fed91bbd74a01e3c4e963ab949bbb5609a
+size 118563
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/2_0_0_probe_0_[736_1408_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/2_0_0_probe_0_[736_1408_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..f289b741260d61061b321c295b00a97606005eeb
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/2_0_0_probe_0_[736_1408_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f55b131dfc946b7fdfde51b37c3905e3900d984cfff1df37fd0ee0ed603959af
+size 100487
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/2_0_3_probe_0_[736_1408_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/2_0_3_probe_0_[736_1408_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..dad866e5a748216df759b48d6767d091a1b9991f
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/2_0_3_probe_0_[736_1408_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:26b7f7971dc27181652b46ef4006c890cf0cd83d28b5742bf7130c0844538878
+size 115540
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/3_0_3_probe_0_[880_1168_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/3_0_3_probe_0_[880_1168_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..4cd80ce0325023d8258ca1e808816d9f52f7ed0a
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/3_0_3_probe_0_[880_1168_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2051a1101cc6aa357e9ee91274e0a2739457d663de72bad594a5db9ec6b5358f
+size 150317
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/3_0_5_probe_0_[880_1168_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/3_0_5_probe_0_[880_1168_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..1227bfb5bb194ae08f946ddc42493d93074a69ce
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/3_0_5_probe_0_[880_1168_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7178a18678ca121514e54416d4c7f671f35dae171f7808c045e0158207d284e3
+size 145210
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/4_0_3_probe_0_[784_1328_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/4_0_3_probe_0_[784_1328_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..f9bd8c85a18a7ce79d87eddc86d10417564929f0
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/4_0_3_probe_0_[784_1328_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:22f195a6db09ad5799482281bf13ab060b39f5701bdebc1f0ef66c3698287838
+size 130026
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/5_0_5_probe_0_[864_1200_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/5_0_5_probe_0_[864_1200_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..d4920ca5c8ba784c8da102f1a822b9cfd1fa531c
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/5_0_5_probe_0_[864_1200_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b5118ec0b48db284e68d7b9074f973e022dddc839955c7da918f858bb985e81a
+size 119297
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/6_0_3_probe_0_[864_1184_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/6_0_3_probe_0_[864_1184_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..6e76db2a5202d80e55a08cb4bbdc1979d37b5520
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/6_0_3_probe_0_[864_1184_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7c8c5c058c916044c3f031c8a1e07277309994c5c3ddd882546756151e2b7a85
+size 135589
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/6_0_5_probe_0_[864_1184_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/6_0_5_probe_0_[864_1184_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..7f2c820eccaeb1a17cdc0602267b13fe0396fbfc
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/6_0_5_probe_0_[864_1184_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:38f4e9f6aa3601f4b6614519201acadf7e2dea03400f99ae6bb358b6a422b5cb
+size 142954
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/7_0_3_probe_0_[752_1376_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/7_0_3_probe_0_[752_1376_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..c11c4b070cec0149a748c6f05d3df0cf2643ed82
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/7_0_3_probe_0_[752_1376_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f942a681c0bd6bf5fdb3b282e263300a2ee086b357f659545098e0a12cc32ddc
+size 150592
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/9_0_3_probe_0_[752_1376_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/9_0_3_probe_0_[752_1376_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..52e962a806bb4ad5f42823e5b5fdda53ff7fe6a5
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/9_0_3_probe_0_[752_1376_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d42ff757f4c8d05cd0ba7a66a3b3a330b3582490306ac73afe63f3344e4b233e
+size 108291
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/0_0_0_probe_0_[720_1440_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/0_0_0_probe_0_[720_1440_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..d2e96a3a9370cb9369dba634e3a0dfcc073aa563
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/0_0_0_probe_0_[720_1440_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2c99a14a58db5685c6ce7363e7171da7758e85bdabf474559995b37a8c62932b
+size 122494
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/0_0_3_probe_0_[720_1440_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/0_0_3_probe_0_[720_1440_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..cb04516541bf15a384f5c573e2d5b5c2653e6494
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/0_0_3_probe_0_[720_1440_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:51976685fc057b31e16fcffc752c30e2758a7f4c4daeb91b4312e1115b7abf14
+size 191066
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/1_0_0_probe_0_[736_1392_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/1_0_0_probe_0_[736_1392_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..4bce8d6546360d1c2acdaf21fd6bab109844ffc5
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/1_0_0_probe_0_[736_1392_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:128c10ac4a8543eb11dd8f449672a2f271526fab6002a8c7e691b8354fde0633
+size 121282
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/1_0_3_probe_0_[736_1392_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/1_0_3_probe_0_[736_1392_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..811eafe491df75586c641fe85294daa4d5f67b82
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/1_0_3_probe_0_[736_1392_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:92a7aa6f5d92f597b56c234df9f5544ea391173fc96fb2218c19a5b2f65634a7
+size 126037
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/1_0_5_probe_0_[736_1392_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/1_0_5_probe_0_[736_1392_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..9281176f75dbe6c3900e97e3d38aa7d3fb58edc4
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/1_0_5_probe_0_[736_1392_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d1d4dd81ada872617e08fdb86d5d7757436cdca87fd37c88f03226d9170f8923
+size 110861
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/2_0_0_probe_0_[720_1424_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/2_0_0_probe_0_[720_1424_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..b50f9dbb95324b06bbc789a6806b577819b6cef1
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/2_0_0_probe_0_[720_1424_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:095a6f8fe58d7ff1ec6703cb2e4e8370d45f5e85be009e6f15dc41b16a71d830
+size 100946
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/2_0_3_probe_0_[720_1424_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/2_0_3_probe_0_[720_1424_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..a04b89f35b75ad107057dfe3bdc1c1f43df01460
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/2_0_3_probe_0_[720_1424_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a7e383f00177452076369fcd1f5321eabe8171be34d4dc1186d0cd6bcbfcc886
+size 115149
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/2_0_5_probe_0_[720_1424_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/2_0_5_probe_0_[720_1424_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..aa9f0b9ed70fc9094f7e254d3fe9c3964d9517c6
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/2_0_5_probe_0_[720_1424_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c70d610e5174dd8f8e389df8b72c5445e86831073ea5b2097aa479f449461dcf
+size 112340
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/3_0_0_probe_0_[752_1376_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/3_0_0_probe_0_[752_1376_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..780a14346a59e0ee4b3ccad589ad194ad24474f1
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/3_0_0_probe_0_[752_1376_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:266e8368c78d9c5735351fbaeeaa48a3102a48bdfea6f0039caa718030fa93b8
+size 116402
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/3_0_3_probe_0_[752_1376_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/3_0_3_probe_0_[752_1376_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..892e0fc17aaf5f5214b86ab6f6fd03cadba731d2
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/3_0_3_probe_0_[752_1376_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59441773db34f6d9c8efb86d1f8018f9f5cad84a4db5e9f50916827ab39f813f
+size 171609
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/3_0_5_probe_0_[752_1376_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/3_0_5_probe_0_[752_1376_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..6116595fa447471df6cf7437774ec59fbcbc5ea4
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/3_0_5_probe_0_[752_1376_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:965a93067a979e26e18b7b6e854a05de300cb196bac801e05e497cc4c7445511
+size 141659
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/4_0_3_probe_0_[864_1200_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/4_0_3_probe_0_[864_1200_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..5908d3ceeffc8fad31a27e413ac885deafe30ba0
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/4_0_3_probe_0_[864_1200_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:489403a28ef18020ee97a1d8c531c415a0efd1dc481ecd543193582bd65a9b31
+size 119880
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/5_0_0_probe_0_[720_1440_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/5_0_0_probe_0_[720_1440_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..790fbeea94124d9d883c0eecea629792640237f6
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/5_0_0_probe_0_[720_1440_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:53c399ae0588e7c13f01b7e60bf0e623405e759713d3de2912028832e48d0507
+size 101226
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/5_0_3_probe_0_[720_1440_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/5_0_3_probe_0_[720_1440_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..c8942f96e79b113bb1afa7f19b28ee4e03a4c147
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/5_0_3_probe_0_[720_1440_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f97aaf3cfdfb049980725dbca0e5806fb4c38b98925b86546cfcffc6a4260d88
+size 104852
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/5_0_5_probe_0_[720_1440_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/5_0_5_probe_0_[720_1440_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..afdc19ab5b6acd117314a3b89fd6eb50cd4de2bb
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/5_0_5_probe_0_[720_1440_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:451827f4baabd281444a49ac87e7d9768a942d8141d8c58987f1299cbe17e8a9
+size 102769
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/6_0_3_probe_0_[864_1200_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/6_0_3_probe_0_[864_1200_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..46e0c07a16b2dfec7bc9f3fc8faca5654a87047a
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/6_0_3_probe_0_[864_1200_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:183457d8de36f34aac31facb57d3869cbbbe94c2cadc24b35c8d72374cee44f0
+size 137250
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/6_0_5_probe_0_[864_1200_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/6_0_5_probe_0_[864_1200_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..b12bdd3816f7d7eeae0e55e1e38eeada25834c99
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/6_0_5_probe_0_[864_1200_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:63d5d6ab6bc91ccff577eacc00151b830327eb0b07d941570a40082f9c20bd68
+size 117642
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/7_0_3_probe_0_[784_1296_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/7_0_3_probe_0_[784_1296_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..af39997e5f5b8b25f3e8c923000ed5e909915d44
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/7_0_3_probe_0_[784_1296_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d5c8cf2a7e1c4072dd2adc0f99cb11f0221f8db2f322afae520f129d4c373f04
+size 138552
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/8_0_3_probe_0_[768_1344_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/8_0_3_probe_0_[768_1344_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..9770e4507f9a2f6418e357f8d4625d5f7785ce5a
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/8_0_3_probe_0_[768_1344_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:89336da690dc396a408cba319ee01aa7e69c2f40e27f7b0e70cbf6a0a113ba75
+size 102141
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/8_0_5_probe_0_[768_1344_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/8_0_5_probe_0_[768_1344_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..9cf5246690fbe13906cd2cbbef4dee7400c490cf
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/8_0_5_probe_0_[768_1344_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2184b0a2fadca12a0e41c80fb816371e69dec05b182a284566a3e992a6cc8201
+size 110141
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/9_0_3_probe_0_[720_1424_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/9_0_3_probe_0_[720_1424_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..2afe34cb343b85c2f12328016cd07d0816ceb09a
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/9_0_3_probe_0_[720_1424_3].jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e987b1064226bad218ee08087411b9bc0bc65e5f7eb1a19e9f464c1b38fd8873
+size 114265
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/0_0_1_probe_0_[864_1184_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/0_0_1_probe_0_[864_1184_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..1c4605520cced30c37ce4572cc352e040e45864a
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/0_0_1_probe_0_[864_1184_3].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/10_0_0_probe_0_[768_1344_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/10_0_0_probe_0_[768_1344_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..9b493f2687478bc9d7d1dadd8a064a66bc582b66
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/10_0_0_probe_0_[768_1344_3].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/10_0_4_probe_0_[768_1344_1].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/10_0_4_probe_0_[768_1344_1].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..a37d102113f06e4b8103f87327ace8087c274745
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/10_0_4_probe_0_[768_1344_1].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/11_0_0_probe_0_[736_1408_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/11_0_0_probe_0_[736_1408_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..0f6e407f5df6012ca8175b680019fa1989f6ae9f
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/11_0_0_probe_0_[736_1408_3].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/11_0_4_probe_0_[736_1408_1].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/11_0_4_probe_0_[736_1408_1].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..85794bbbe1ace5e6cf245b6e74c15da7fd256e11
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/11_0_4_probe_0_[736_1408_1].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/1_0_1_probe_0_[864_1200_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/1_0_1_probe_0_[864_1200_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..f5626a5fc1e5fd9a651ac8a98e8e0afc889b6792
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/1_0_1_probe_0_[864_1200_3].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/1_0_4_probe_0_[864_1200_1].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/1_0_4_probe_0_[864_1200_1].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..9fe81c3b753dd5a784c40455e4a735023a125851
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/1_0_4_probe_0_[864_1200_1].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/2_0_0_probe_0_[800_1296_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/2_0_0_probe_0_[800_1296_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..c7d545296837b46c5dc84042aace05d180362fdc
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/2_0_0_probe_0_[800_1296_3].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/2_0_1_probe_0_[800_1296_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/2_0_1_probe_0_[800_1296_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..80c3f689e2817f1212f93f5cbf45be0003047930
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/2_0_1_probe_0_[800_1296_3].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/2_0_2_probe_0_[800_1296_1].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/2_0_2_probe_0_[800_1296_1].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..1b27aeb0614fb1bda6f01d597247d39ba88b4d00
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/2_0_2_probe_0_[800_1296_1].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/2_0_4_probe_0_[800_1296_1].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/2_0_4_probe_0_[800_1296_1].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..1b27aeb0614fb1bda6f01d597247d39ba88b4d00
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/2_0_4_probe_0_[800_1296_1].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/3_0_1_probe_0_[752_1392_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/3_0_1_probe_0_[752_1392_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..7c163afda0a78b351cf4c2e9e649b4959591e693
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/3_0_1_probe_0_[752_1392_3].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/3_0_4_probe_0_[752_1392_1].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/3_0_4_probe_0_[752_1392_1].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..a34e5d71393dcebaee1bd481cccbfc7a93370baf
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/3_0_4_probe_0_[752_1392_1].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/4_0_1_probe_0_[864_1200_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/4_0_1_probe_0_[864_1200_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..d9ac4ec344a936586ca3b2c81bc13a503b93e974
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/4_0_1_probe_0_[864_1200_3].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/4_0_2_probe_0_[864_1200_1].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/4_0_2_probe_0_[864_1200_1].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..c2bb4b6fdcf48d8750ad87f613cd3ef2c0df5a07
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/4_0_2_probe_0_[864_1200_1].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/4_0_4_probe_0_[864_1200_1].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/4_0_4_probe_0_[864_1200_1].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..c2bb4b6fdcf48d8750ad87f613cd3ef2c0df5a07
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/4_0_4_probe_0_[864_1200_1].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/4_0_5_probe_0_[864_1200_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/4_0_5_probe_0_[864_1200_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..6e19f6f9529f20959008b432b34dbc6163bef2a6
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/4_0_5_probe_0_[864_1200_3].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/5_0_0_probe_0_[864_1184_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/5_0_0_probe_0_[864_1184_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..5766ec6b06cf95a1626da7cc13d1c62d246da8c1
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/5_0_0_probe_0_[864_1184_3].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/5_0_3_probe_0_[864_1184_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/5_0_3_probe_0_[864_1184_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..673170e87c51f464e3e71f3cccf6782b868dd835
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/5_0_3_probe_0_[864_1184_3].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/5_0_4_probe_0_[864_1184_1].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/5_0_4_probe_0_[864_1184_1].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..f910508b46d333b2e633bcfd8fdb68c9955b990f
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/5_0_4_probe_0_[864_1184_1].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/6_0_1_probe_0_[768_1328_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/6_0_1_probe_0_[768_1328_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..ab886d5f496833002589a4a9e12d3ae1f4fac9cf
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/6_0_1_probe_0_[768_1328_3].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/6_0_4_probe_0_[768_1328_1].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/6_0_4_probe_0_[768_1328_1].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..c75ceeea3064222b1427994e6a491e9dc624c301
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/6_0_4_probe_0_[768_1328_1].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/7_0_0_probe_0_[752_1392_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/7_0_0_probe_0_[752_1392_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..4b21d0f314b06fb4f9fcdde26d5eecdecd8e2e57
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/7_0_0_probe_0_[752_1392_3].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/7_0_1_probe_0_[752_1392_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/7_0_1_probe_0_[752_1392_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..a82c446e6502debb03c74a3982d573f68787145c
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/7_0_1_probe_0_[752_1392_3].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/7_0_4_probe_0_[752_1392_1].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/7_0_4_probe_0_[752_1392_1].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..f679fa46b2a6313aefb2061f62d999dd5042ab73
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/7_0_4_probe_0_[752_1392_1].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/8_0_1_probe_0_[736_1392_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/8_0_1_probe_0_[736_1392_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..66b323b31075ab2eb791a2d559b1cdbe2871b319
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/8_0_1_probe_0_[736_1392_3].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/9_0_0_probe_0_[816_1264_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/9_0_0_probe_0_[816_1264_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..da3f31b195e718b124728eea1c65319fac8ffbc8
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/9_0_0_probe_0_[816_1264_3].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/9_0_4_probe_0_[816_1264_1].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/9_0_4_probe_0_[816_1264_1].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..2a78c7640fe7c630af91fafd4ed5bcec91fdd262
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/9_0_4_probe_0_[816_1264_1].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_0_probe_0_[848_1232_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_0_probe_0_[848_1232_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..82a4cdb460b9edc6d0812a2e3ab532d9efd32fdc
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_0_probe_0_[848_1232_3].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_1_probe_0_[848_1232_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_1_probe_0_[848_1232_3].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..49d3e9e1ccef428320601e418b674e5ec7692965
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_1_probe_0_[848_1232_3].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_2_probe_0_[848_1232_1].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_2_probe_0_[848_1232_1].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..540af239af51be0fa6431ad9b661d3e1a230a094
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_2_probe_0_[848_1232_1].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_4_probe_0_[848_1232_1].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_4_probe_0_[848_1232_1].jpg
new file mode 100644
index 0000000000000000000000000000000000000000..540af239af51be0fa6431ad9b661d3e1a230a094
Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_4_probe_0_[848_1232_1].jpg differ
diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0_probe.html b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0_probe.html
new file mode 100644
index 0000000000000000000000000000000000000000..daaa1a3c49f1a3c828c5dc1f2795419f22ce5838
--- /dev/null
+++ b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0_probe.html
@@ -0,0 +1,4 @@
+
+
+
0-0|edit_image0; | 0-1|modify_image0; | 0-2|edit_mask0; | 0-3|target_image; | 0-4|target_mask; | 0-5 | 0-5 |
0-0|edit_image0; | 0-1|modify_image0; | 0-2|edit_mask0; | 0-3|target_image; | 0-4|target_mask; | 0-5 | 0-5 |
0-0|edit_image0; | 0-1|modify_image0; | 0-2|edit_mask0; | 0-3|target_image; | 0-4|target_mask; | 0-5 | 0-5 |
0-0|edit_image0; | 0-1|modify_image0; | 0-2|edit_mask0; | 0-3|target_image; | 0-4|target_mask; | 0-5 | 0-5 |
0-0|edit_image0; | 0-1|modify_image0; | 0-2|edit_mask0; | 0-3|target_image; | 0-4|target_mask; | 0-5 | 0-5 |
0-0|edit_image0; | 0-1|modify_image0; | 0-2|edit_mask0; | 0-3|target_image; | 0-4|target_mask; | 0-5 | 0-5 |
0-0|edit_image0; | 0-1|modify_image0; | 0-2|edit_mask0; | 0-3|target_image; | 0-4|target_mask; | 0-5 | 0-5 |
0-0|edit_image0; | 0-1|modify_image0; | 0-2|edit_mask0; | 0-3|target_image; | 0-4|target_mask; | 0-5 | 0-5 |
0-0|edit_image0; | 0-1|modify_image0; | 0-2|edit_mask0; | 0-3|target_image; | 0-4|target_mask; | 0-5 | 0-5 |
0-0|edit_image0; | 0-1|modify_image0; | 0-2|edit_mask0; | 0-3|target_image; | 0-4|target_mask; | 0-5 | 0-5 |
0-0|edit_image0; | 0-1|modify_image0; | 0-2|edit_mask0; | 0-3|target_image; | 0-4|target_mask; | 0-5 | 0-5 |
0-0|edit_image0; | 0-1|modify_image0; | 0-2|edit_mask0; | 0-3|target_image; | 0-4|target_mask; | 0-5 | 0-5 |