diff --git a/.gitattributes b/.gitattributes index 8a608b4272f5b5c5abb5a2d36f19f102be0cc2db..d018acd0dad2382c575331cc665696357968f04e 100644 --- a/.gitattributes +++ b/.gitattributes @@ -289,3 +289,39 @@ ACE_plus/examples/exp_example/20250704031317/eval_probe/step-10/BaseSolver_eval_ ACE_plus/examples/exp_example/20250704031317/eval_probe/step-10/BaseSolver_eval_image_step_10/9_0_3_probe_0_\[736_1408_3\].jpg filter=lfs diff=lfs merge=lfs -text ACE_plus/examples/exp_example/20250704031317/eval_probe/step-10/BaseSolver_eval_image_step_10/0_0_3_probe_0_\[736_1408_3\].jpg filter=lfs diff=lfs merge=lfs -text ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/2_0_5_probe_0_\[736_1408_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/3_0_5_probe_0_\[880_1168_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/1_0_0_probe_0_\[752_1360_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/0_0_5_probe_0_\[848_1232_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/1_0_5_probe_0_\[752_1360_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/6_0_5_probe_0_\[864_1184_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/6_0_3_probe_0_\[864_1184_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/4_0_3_probe_0_\[784_1328_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/7_0_3_probe_0_\[752_1376_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/2_0_3_probe_0_\[736_1408_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/2_0_0_probe_0_\[736_1408_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/9_0_3_probe_0_\[752_1376_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/5_0_5_probe_0_\[864_1200_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/3_0_3_probe_0_\[880_1168_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/0_0_3_probe_0_\[848_1232_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/1_0_3_probe_0_\[752_1360_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/0_0_3_probe_0_\[720_1440_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/0_0_0_probe_0_\[720_1440_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/1_0_5_probe_0_\[736_1392_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/5_0_5_probe_0_\[720_1440_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/6_0_5_probe_0_\[864_1200_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/8_0_5_probe_0_\[768_1344_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/8_0_3_probe_0_\[768_1344_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/1_0_3_probe_0_\[736_1392_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/9_0_3_probe_0_\[720_1424_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/2_0_0_probe_0_\[720_1424_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/2_0_3_probe_0_\[720_1424_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/5_0_3_probe_0_\[720_1440_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/2_0_5_probe_0_\[720_1424_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/7_0_3_probe_0_\[784_1296_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/6_0_3_probe_0_\[864_1200_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/5_0_0_probe_0_\[720_1440_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/3_0_5_probe_0_\[752_1376_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/3_0_3_probe_0_\[752_1376_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/1_0_0_probe_0_\[736_1392_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/3_0_0_probe_0_\[752_1376_3\].jpg filter=lfs diff=lfs merge=lfs -text +ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/4_0_3_probe_0_\[864_1200_3\].jpg filter=lfs diff=lfs merge=lfs -text diff --git a/ACE_plus/examples/exp_example/20250602092700/checkpoints/ldm_step-500/README.md b/ACE_plus/examples/exp_example/20250602092700/checkpoints/ldm_step-500/README.md new file mode 100644 index 0000000000000000000000000000000000000000..2cae90886aa26c06c289f0d1b74a06bb39921dba --- /dev/null +++ b/ACE_plus/examples/exp_example/20250602092700/checkpoints/ldm_step-500/README.md @@ -0,0 +1,10 @@ +## Training procedure + +### Framework versions + + +- SWIFT 3.4.0 +### Base model information + + +- BaseModel Class LatentDiffusionACEPlus diff --git a/ACE_plus/examples/exp_example/20250602092700/checkpoints/ldm_step-500/configuration.json b/ACE_plus/examples/exp_example/20250602092700/checkpoints/ldm_step-500/configuration.json new file mode 100644 index 0000000000000000000000000000000000000000..9e26dfeeb6e641a33dae4961196235bdb965b21b --- /dev/null +++ b/ACE_plus/examples/exp_example/20250602092700/checkpoints/ldm_step-500/configuration.json @@ -0,0 +1 @@ +{} \ No newline at end of file diff --git a/ACE_plus/examples/exp_example/20250602092700/noise_schedule.png b/ACE_plus/examples/exp_example/20250602092700/noise_schedule.png new file mode 100644 index 0000000000000000000000000000000000000000..2d08dfab306b571d00dd9cbd03468d2b58de1f7d Binary files /dev/null and b/ACE_plus/examples/exp_example/20250602092700/noise_schedule.png differ diff --git a/ACE_plus/examples/exp_example/20250602092700/sampler_schedule.png b/ACE_plus/examples/exp_example/20250602092700/sampler_schedule.png new file mode 100644 index 0000000000000000000000000000000000000000..f12f547ac8a8c4c5f946276a86a85e60bdd5f107 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250602092700/sampler_schedule.png differ diff --git a/ACE_plus/examples/exp_example/20250602092700/std_log.txt b/ACE_plus/examples/exp_example/20250602092700/std_log.txt new file mode 100644 index 0000000000000000000000000000000000000000..eb6aca8eca2316db83df0b2ed741ca6791c2ccfd --- /dev/null +++ b/ACE_plus/examples/exp_example/20250602092700/std_log.txt @@ -0,0 +1,620 @@ +scepter [INFO] 2025-06-02 09:27:02,317 [File: logger.py Function: init_logger at line 85] Running task with log file: /home/Ubuntu/Downloads/Unmodel/ACE_plus/./examples/exp_example/20250602092700/std_log.txt +scepter [WARNING] 2025-06-02 09:27:02,412 [File: import_utils.py Function: import_module at line 325] ('DATASETS', 'ACEPlusDataset') not found in ast index file +scepter [INFO] 2025-06-02 09:27:02,413 [File: ace_plus_dataset.py Function: read_data_list at line 151] subject has 4 samples. +scepter [INFO] 2025-06-02 09:27:02,414 [File: registry.py Function: __init__ at line 185] Built dataloader with len 9223372036854775807 +scepter [WARNING] 2025-06-02 09:27:02,414 [File: import_utils.py Function: import_module at line 325] ('DATASETS', 'ACEPlusDataset') not found in ast index file +scepter [INFO] 2025-06-02 09:27:02,414 [File: ace_plus_dataset.py Function: read_data_list at line 151] subject has 4 samples. +scepter [INFO] 2025-06-02 09:27:02,415 [File: registry.py Function: __init__ at line 185] Built dataloader with len 4 +scepter [INFO] 2025-06-02 09:28:16,963 [File: flux.py Function: load_pretrained_model at line 450] Restored from /home/Ubuntu/Downloads/Unmodel/Reference_models/flux1-fill-dev.safetensors with 0 missing and 0 unexpected keys +scepter [INFO] 2025-06-02 09:28:16,982 [File: ace_plus_ldm.py Function: construct_network at line 62] all parameters:11.90B +scepter [INFO] 2025-06-02 09:28:17,613 [File: ae_module.py Function: construct_model at line 76] AE Module XFORMERS_IS_AVAILBLE : True +scepter [INFO] 2025-06-02 09:28:18,411 [File: ae_kl.py Function: init_from_ckpt at line 400] Restored from /home/Ubuntu/Downloads/Unmodel/Reference_models/ae.safetensors with 0 missing and 0 unexpected keys +scepter [INFO] 2025-06-02 09:28:30,434 [File: diffusion_solver.py Function: add_tuner at line 788] [('base_model.model.double_blocks.0.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.0.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.0.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.0.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.0.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.0.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.0.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.0.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.0.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.0.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.0.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.0.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.0.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.0.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.0.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.0.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.0.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.0.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.0.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.0.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.1.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.1.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.1.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.1.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.1.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.1.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.1.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.1.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.1.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.1.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.1.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.1.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.1.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.1.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.1.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.1.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.1.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.1.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.1.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.1.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.2.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.2.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.2.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.2.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.2.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.2.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.2.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.2.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.2.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.2.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.2.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.2.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.2.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.2.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.2.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.2.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.2.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.2.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.2.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.2.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.3.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.3.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.3.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.3.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.3.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.3.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.3.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.3.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.3.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.3.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.3.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.3.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.3.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.3.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.3.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.3.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.3.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.3.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.3.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.3.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.4.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.4.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.4.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.4.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.4.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.4.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.4.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.4.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.4.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.4.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.4.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.4.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.4.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.4.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.4.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.4.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.4.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.4.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.4.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.4.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.5.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.5.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.5.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.5.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.5.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.5.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.5.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.5.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.5.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.5.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.5.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.5.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.5.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.5.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.5.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.5.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.5.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.5.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.5.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.5.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.6.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.6.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.6.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.6.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.6.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.6.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.6.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.6.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.6.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.6.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.6.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.6.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.6.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.6.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.6.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.6.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.6.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.6.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.6.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.6.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.7.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.7.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.7.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.7.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.7.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.7.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.7.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.7.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.7.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.7.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.7.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.7.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.7.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.7.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.7.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.7.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.7.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.7.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.7.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.7.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.8.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.8.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.8.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.8.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.8.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.8.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.8.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.8.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.8.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.8.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.8.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.8.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.8.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.8.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.8.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.8.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.8.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.8.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.8.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.8.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.9.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.9.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.9.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.9.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.9.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.9.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.9.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.9.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.9.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.9.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.9.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.9.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.9.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.9.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.9.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.9.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.9.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.9.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.9.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.9.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.10.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.10.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.10.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.10.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.10.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.10.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.10.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.10.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.10.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.10.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.10.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.10.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.10.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.10.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.10.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.10.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.10.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.10.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.10.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.10.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.11.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.11.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.11.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.11.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.11.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.11.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.11.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.11.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.11.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.11.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.11.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.11.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.11.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.11.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.11.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.11.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.11.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.11.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.11.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.11.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.12.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.12.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.12.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.12.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.12.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.12.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.12.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.12.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.12.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.12.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.12.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.12.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.12.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.12.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.12.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.12.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.12.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.12.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.12.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.12.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.13.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.13.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.13.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.13.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.13.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.13.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.13.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.13.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.13.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.13.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.13.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.13.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.13.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.13.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.13.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.13.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.13.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.13.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.13.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.13.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.14.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.14.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.14.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.14.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.14.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.14.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.14.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.14.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.14.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.14.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.14.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.14.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.14.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.14.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.14.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.14.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.14.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.14.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.14.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.14.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.15.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.15.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.15.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.15.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.15.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.15.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.15.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.15.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.15.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.15.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.15.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.15.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.15.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.15.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.15.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.15.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.15.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.15.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.15.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.15.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.16.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.16.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.16.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.16.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.16.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.16.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.16.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.16.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.16.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.16.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.16.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.16.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.16.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.16.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.16.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.16.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.16.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.16.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.16.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.16.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.17.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.17.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.17.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.17.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.17.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.17.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.17.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.17.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.17.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.17.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.17.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.17.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.17.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.17.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.17.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.17.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.17.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.17.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.17.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.17.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.18.img_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.18.img_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.18.img_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.18.img_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.18.img_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.18.img_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.18.img_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.18.img_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.18.img_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.18.img_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.18.txt_mod.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.18.txt_mod.lin.lora_B.0_SwiftLoRA.weight', torch.Size([18432, 64])), ('base_model.model.double_blocks.18.txt_attn.qkv.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.18.txt_attn.qkv.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.double_blocks.18.txt_attn.proj.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.18.txt_attn.proj.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.double_blocks.18.txt_mlp.0.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.double_blocks.18.txt_mlp.0.lora_B.0_SwiftLoRA.weight', torch.Size([12288, 64])), ('base_model.model.double_blocks.18.txt_mlp.2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 12288])), ('base_model.model.double_blocks.18.txt_mlp.2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.0.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.0.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.0.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.0.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.0.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.0.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.1.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.1.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.1.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.1.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.1.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.1.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.2.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.2.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.2.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.2.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.2.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.2.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.3.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.3.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.3.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.3.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.3.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.3.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.4.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.4.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.4.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.4.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.4.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.4.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.5.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.5.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.5.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.5.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.5.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.5.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.6.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.6.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.6.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.6.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.6.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.6.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.7.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.7.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.7.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.7.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.7.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.7.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.8.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.8.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.8.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.8.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.8.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.8.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.9.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.9.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.9.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.9.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.9.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.9.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.10.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.10.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.10.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.10.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.10.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.10.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.11.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.11.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.11.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.11.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.11.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.11.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.12.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.12.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.12.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.12.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.12.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.12.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.13.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.13.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.13.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.13.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.13.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.13.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.14.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.14.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.14.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.14.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.14.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.14.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.15.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.15.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.15.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.15.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.15.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.15.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.16.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.16.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.16.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.16.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.16.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.16.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.17.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.17.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.17.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.17.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.17.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.17.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.18.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.18.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.18.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.18.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.18.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.18.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.19.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.19.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.19.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.19.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.19.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.19.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.20.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.20.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.20.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.20.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.20.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.20.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.21.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.21.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.21.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.21.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.21.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.21.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.22.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.22.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.22.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.22.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.22.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.22.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.23.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.23.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.23.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.23.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.23.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.23.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.24.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.24.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.24.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.24.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.24.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.24.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.25.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.25.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.25.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.25.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.25.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.25.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.26.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.26.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.26.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.26.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.26.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.26.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.27.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.27.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.27.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.27.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.27.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.27.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.28.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.28.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.28.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.28.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.28.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.28.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.29.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.29.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.29.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.29.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.29.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.29.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.30.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.30.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.30.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.30.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.30.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.30.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.31.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.31.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.31.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.31.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.31.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.31.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.32.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.32.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.32.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.32.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.32.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.32.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.33.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.33.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.33.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.33.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.33.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.33.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.34.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.34.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.34.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.34.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.34.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.34.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.35.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.35.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.35.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.35.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.35.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.35.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.36.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.36.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.36.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.36.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.36.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.36.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64])), ('base_model.model.single_blocks.37.linear1.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.37.linear1.lora_B.0_SwiftLoRA.weight', torch.Size([21504, 64])), ('base_model.model.single_blocks.37.linear2.lora_A.0_SwiftLoRA.weight', torch.Size([64, 15360])), ('base_model.model.single_blocks.37.linear2.lora_B.0_SwiftLoRA.weight', torch.Size([3072, 64])), ('base_model.model.single_blocks.37.modulation.lin.lora_A.0_SwiftLoRA.weight', torch.Size([64, 3072])), ('base_model.model.single_blocks.37.modulation.lin.lora_B.0_SwiftLoRA.weight', torch.Size([9216, 64]))] +scepter [INFO] 2025-06-02 09:28:30,450 [File: diffusion_solver.py Function: print_model_params_status at line 996] Load trainable params 306315264 / 17178094051 = 1.78%, train part: {'model.double_blocks': 171835392, 'model.single_blocks': 134479872}. +scepter [INFO] 2025-06-02 09:28:30,450 [File: diffusion_solver.py Function: print_model_params_status at line 1000] Load frozen params 16871778787 / 17178094051 = 98.22%, frozen part: {'model': 11902587968, 'first_stage_model': 83819683, 'cond_stage_model': 4885371136}. +scepter [INFO] 2025-06-02 09:29:07,140 [File: diffusion_solver.py Function: set_up at line 230] SwiftModel( + (base_model): LatentDiffusionACEPlus LatentDiffusionACEPlus( + (model): FluxMRModiACEPlus FluxMRModiACEPlus( + (pe_embedder): EmbedND() + (img_in): Linear(in_features=448, out_features=3072, bias=True) + (time_in): MLPEmbedder( + (in_layer): Linear(in_features=256, out_features=3072, bias=True) + (silu): SiLU() + (out_layer): Linear(in_features=3072, out_features=3072, bias=True) + ) + (vector_in): MLPEmbedder( + (in_layer): Linear(in_features=768, out_features=3072, bias=True) + (silu): SiLU() + (out_layer): Linear(in_features=3072, out_features=3072, bias=True) + ) + (guidance_in): MLPEmbedder( + (in_layer): Linear(in_features=256, out_features=3072, bias=True) + (silu): SiLU() + (out_layer): Linear(in_features=3072, out_features=3072, bias=True) + ) + (txt_in): Linear(in_features=4096, out_features=3072, bias=True) + (double_blocks): ModuleList( + (0-18): 19 x DoubleStreamBlock( + (img_mod): Modulation( + (lin): lora.Linear( + (base_layer): Linear(in_features=3072, out_features=18432, bias=True) + (lora_dropout): ModuleDict( + (0_SwiftLoRA): Identity() + ) + (lora_A): ModuleDict( + (0_SwiftLoRA): Linear(in_features=3072, out_features=64, bias=False) + ) + (lora_B): ModuleDict( + (0_SwiftLoRA): Linear(in_features=64, out_features=18432, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + ) + (img_norm1): LayerNorm((3072,), eps=1e-06, elementwise_affine=False) + (img_attn): SelfAttention( + (qkv): lora.Linear( + (base_layer): Linear(in_features=3072, out_features=9216, bias=True) + (lora_dropout): ModuleDict( + (0_SwiftLoRA): Identity() + ) + (lora_A): ModuleDict( + (0_SwiftLoRA): Linear(in_features=3072, out_features=64, bias=False) + ) + (lora_B): ModuleDict( + (0_SwiftLoRA): Linear(in_features=64, out_features=9216, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (norm): QKNorm( + (query_norm): RMSNorm() + (key_norm): RMSNorm() + ) + (proj): lora.Linear( + (base_layer): Linear(in_features=3072, out_features=3072, bias=True) + (lora_dropout): ModuleDict( + (0_SwiftLoRA): Identity() + ) + (lora_A): ModuleDict( + (0_SwiftLoRA): Linear(in_features=3072, out_features=64, bias=False) + ) + (lora_B): ModuleDict( + (0_SwiftLoRA): Linear(in_features=64, out_features=3072, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + ) + (img_norm2): LayerNorm((3072,), eps=1e-06, elementwise_affine=False) + (img_mlp): Sequential( + (0): lora.Linear( + (base_layer): Linear(in_features=3072, out_features=12288, bias=True) + (lora_dropout): ModuleDict( + (0_SwiftLoRA): Identity() + ) + (lora_A): ModuleDict( + (0_SwiftLoRA): Linear(in_features=3072, out_features=64, bias=False) + ) + (lora_B): ModuleDict( + (0_SwiftLoRA): Linear(in_features=64, out_features=12288, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (1): GELU(approximate='tanh') + (2): lora.Linear( + (base_layer): Linear(in_features=12288, out_features=3072, bias=True) + (lora_dropout): ModuleDict( + (0_SwiftLoRA): Identity() + ) + (lora_A): ModuleDict( + (0_SwiftLoRA): Linear(in_features=12288, out_features=64, bias=False) + ) + (lora_B): ModuleDict( + (0_SwiftLoRA): Linear(in_features=64, out_features=3072, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + ) + (txt_mod): Modulation( + (lin): lora.Linear( + (base_layer): Linear(in_features=3072, out_features=18432, bias=True) + (lora_dropout): ModuleDict( + (0_SwiftLoRA): Identity() + ) + (lora_A): ModuleDict( + (0_SwiftLoRA): Linear(in_features=3072, out_features=64, bias=False) + ) + (lora_B): ModuleDict( + (0_SwiftLoRA): Linear(in_features=64, out_features=18432, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + ) + (txt_norm1): LayerNorm((3072,), eps=1e-06, elementwise_affine=False) + (txt_attn): SelfAttention( + (qkv): lora.Linear( + (base_layer): Linear(in_features=3072, out_features=9216, bias=True) + (lora_dropout): ModuleDict( + (0_SwiftLoRA): Identity() + ) + (lora_A): ModuleDict( + (0_SwiftLoRA): Linear(in_features=3072, out_features=64, bias=False) + ) + (lora_B): ModuleDict( + (0_SwiftLoRA): Linear(in_features=64, out_features=9216, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (norm): QKNorm( + (query_norm): RMSNorm() + (key_norm): RMSNorm() + ) + (proj): lora.Linear( + (base_layer): Linear(in_features=3072, out_features=3072, bias=True) + (lora_dropout): ModuleDict( + (0_SwiftLoRA): Identity() + ) + (lora_A): ModuleDict( + (0_SwiftLoRA): Linear(in_features=3072, out_features=64, bias=False) + ) + (lora_B): ModuleDict( + (0_SwiftLoRA): Linear(in_features=64, out_features=3072, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + ) + (txt_norm2): LayerNorm((3072,), eps=1e-06, elementwise_affine=False) + (txt_mlp): Sequential( + (0): lora.Linear( + (base_layer): Linear(in_features=3072, out_features=12288, bias=True) + (lora_dropout): ModuleDict( + (0_SwiftLoRA): Identity() + ) + (lora_A): ModuleDict( + (0_SwiftLoRA): Linear(in_features=3072, out_features=64, bias=False) + ) + (lora_B): ModuleDict( + (0_SwiftLoRA): Linear(in_features=64, out_features=12288, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (1): GELU(approximate='tanh') + (2): lora.Linear( + (base_layer): Linear(in_features=12288, out_features=3072, bias=True) + (lora_dropout): ModuleDict( + (0_SwiftLoRA): Identity() + ) + (lora_A): ModuleDict( + (0_SwiftLoRA): Linear(in_features=12288, out_features=64, bias=False) + ) + (lora_B): ModuleDict( + (0_SwiftLoRA): Linear(in_features=64, out_features=3072, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + ) + ) + ) + (single_blocks): ModuleList( + (0-37): 38 x SingleStreamBlock( + (linear1): lora.Linear( + (base_layer): Linear(in_features=3072, out_features=21504, bias=True) + (lora_dropout): ModuleDict( + (0_SwiftLoRA): Identity() + ) + (lora_A): ModuleDict( + (0_SwiftLoRA): Linear(in_features=3072, out_features=64, bias=False) + ) + (lora_B): ModuleDict( + (0_SwiftLoRA): Linear(in_features=64, out_features=21504, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (linear2): lora.Linear( + (base_layer): Linear(in_features=15360, out_features=3072, bias=True) + (lora_dropout): ModuleDict( + (0_SwiftLoRA): Identity() + ) + (lora_A): ModuleDict( + (0_SwiftLoRA): Linear(in_features=15360, out_features=64, bias=False) + ) + (lora_B): ModuleDict( + (0_SwiftLoRA): Linear(in_features=64, out_features=3072, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + (norm): QKNorm( + (query_norm): RMSNorm() + (key_norm): RMSNorm() + ) + (pre_norm): LayerNorm((3072,), eps=1e-06, elementwise_affine=False) + (mlp_act): GELU(approximate='tanh') + (modulation): Modulation( + (lin): lora.Linear( + (base_layer): Linear(in_features=3072, out_features=9216, bias=True) + (lora_dropout): ModuleDict( + (0_SwiftLoRA): Identity() + ) + (lora_A): ModuleDict( + (0_SwiftLoRA): Linear(in_features=3072, out_features=64, bias=False) + ) + (lora_B): ModuleDict( + (0_SwiftLoRA): Linear(in_features=64, out_features=9216, bias=False) + ) + (lora_embedding_A): ParameterDict() + (lora_embedding_B): ParameterDict() + (lora_magnitude_vector): ModuleDict() + ) + ) + ) + ) + (final_layer): LastLayer( + (norm_final): LayerNorm((3072,), eps=1e-06, elementwise_affine=False) + (linear): Linear(in_features=3072, out_features=64, bias=True) + (adaLN_modulation): Sequential( + (0): SiLU() + (1): Linear(in_features=3072, out_features=6144, bias=True) + ) + ) + ) + (first_stage_model): AutoencoderKLFlux AutoencoderKLFlux( + (encoder): Encoder Encoder( + (conv_in): Conv2d(3, 128, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + (down): ModuleList( + (0): Module( + (block): ModuleList( + (0-1): 2 x ResnetBlock( + (norm1): GroupNorm(32, 128, eps=1e-06, affine=True) + (conv1): Conv2d(128, 128, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + (norm2): GroupNorm(32, 128, eps=1e-06, affine=True) + (dropout): Dropout(p=0.0, inplace=False) + (conv2): Conv2d(128, 128, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + ) + ) + (attn): ModuleList() + (downsample): Downsample( + (conv): Conv2d(128, 128, kernel_size=(3, 3), stride=(2, 2)) + ) + ) + (1): Module( + (block): ModuleList( + (0): ResnetBlock( + (norm1): GroupNorm(32, 128, eps=1e-06, affine=True) + (conv1): Conv2d(128, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + (norm2): GroupNorm(32, 256, eps=1e-06, affine=True) + (dropout): Dropout(p=0.0, inplace=False) + (conv2): Conv2d(256, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + (nin_shortcut): Conv2d(128, 256, kernel_size=(1, 1), stride=(1, 1)) + ) + (1): ResnetBlock( + (norm1): GroupNorm(32, 256, eps=1e-06, affine=True) + (conv1): Conv2d(256, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + (norm2): GroupNorm(32, 256, eps=1e-06, affine=True) + (dropout): Dropout(p=0.0, inplace=False) + (conv2): Conv2d(256, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + ) + ) + (attn): ModuleList() + (downsample): Downsample( + (conv): Conv2d(256, 256, kernel_size=(3, 3), stride=(2, 2)) + ) + ) + (2): Module( + (block): ModuleList( + (0): ResnetBlock( + (norm1): GroupNorm(32, 256, eps=1e-06, affine=True) + (conv1): Conv2d(256, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + (norm2): GroupNorm(32, 512, eps=1e-06, affine=True) + (dropout): Dropout(p=0.0, inplace=False) + (conv2): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + (nin_shortcut): Conv2d(256, 512, kernel_size=(1, 1), stride=(1, 1)) + ) + (1): ResnetBlock( + (norm1): GroupNorm(32, 512, eps=1e-06, affine=True) + (conv1): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + (norm2): GroupNorm(32, 512, eps=1e-06, affine=True) + (dropout): Dropout(p=0.0, inplace=False) + (conv2): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + ) + ) + (attn): ModuleList() + (downsample): Downsample( + (conv): Conv2d(512, 512, kernel_size=(3, 3), stride=(2, 2)) + ) + ) + (3): Module( + (block): ModuleList( + (0-1): 2 x ResnetBlock( + (norm1): GroupNorm(32, 512, eps=1e-06, affine=True) + (conv1): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + (norm2): GroupNorm(32, 512, eps=1e-06, affine=True) + (dropout): Dropout(p=0.0, inplace=False) + (conv2): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + ) + ) + (attn): ModuleList() + ) + ) + (mid): Module( + (block_1): ResnetBlock( + (norm1): GroupNorm(32, 512, eps=1e-06, affine=True) + (conv1): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + (norm2): GroupNorm(32, 512, eps=1e-06, affine=True) + (dropout): Dropout(p=0.0, inplace=False) + (conv2): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + ) + (attn_1): MemoryEfficientAttention( + (norm): GroupNorm(32, 512, eps=1e-06, affine=True) + (q): Conv2d(512, 512, kernel_size=(1, 1), stride=(1, 1)) + (k): Conv2d(512, 512, kernel_size=(1, 1), stride=(1, 1)) + (v): Conv2d(512, 512, kernel_size=(1, 1), stride=(1, 1)) + (proj_out): Conv2d(512, 512, kernel_size=(1, 1), stride=(1, 1)) + ) + (block_2): ResnetBlock( + (norm1): GroupNorm(32, 512, eps=1e-06, affine=True) + (conv1): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + (norm2): GroupNorm(32, 512, eps=1e-06, affine=True) + (dropout): Dropout(p=0.0, inplace=False) + (conv2): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + ) + ) + (norm_out): GroupNorm(32, 512, eps=1e-06, affine=True) + (conv_out): Conv2d(512, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + ) + (decoder): Decoder Decoder( + (conv_in): Conv2d(16, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + (mid): Module( + (block_1): ResnetBlock( + (norm1): GroupNorm(32, 512, eps=1e-06, affine=True) + (conv1): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + (norm2): GroupNorm(32, 512, eps=1e-06, affine=True) + (dropout): Dropout(p=0.0, inplace=False) + (conv2): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + ) + (attn_1): MemoryEfficientAttention( + (norm): GroupNorm(32, 512, eps=1e-06, affine=True) + (q): Conv2d(512, 512, kernel_size=(1, 1), stride=(1, 1)) + (k): Conv2d(512, 512, kernel_size=(1, 1), stride=(1, 1)) + (v): Conv2d(512, 512, kernel_size=(1, 1), stride=(1, 1)) + (proj_out): Conv2d(512, 512, kernel_size=(1, 1), stride=(1, 1)) + ) + (block_2): ResnetBlock( + (norm1): GroupNorm(32, 512, eps=1e-06, affine=True) + (conv1): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + (norm2): GroupNorm(32, 512, eps=1e-06, affine=True) + (dropout): Dropout(p=0.0, inplace=False) + (conv2): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + ) + ) + (up): ModuleList( + (0): Module( + (block): ModuleList( + (0): ResnetBlock( + (norm1): GroupNorm(32, 256, eps=1e-06, affine=True) + (conv1): Conv2d(256, 128, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + (norm2): GroupNorm(32, 128, eps=1e-06, affine=True) + (dropout): Dropout(p=0.0, inplace=False) + (conv2): Conv2d(128, 128, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + (nin_shortcut): Conv2d(256, 128, kernel_size=(1, 1), stride=(1, 1)) + ) + (1-2): 2 x ResnetBlock( + (norm1): GroupNorm(32, 128, eps=1e-06, affine=True) + (conv1): Conv2d(128, 128, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + (norm2): GroupNorm(32, 128, eps=1e-06, affine=True) + (dropout): Dropout(p=0.0, inplace=False) + (conv2): Conv2d(128, 128, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + ) + ) + (attn): ModuleList() + ) + (1): Module( + (block): ModuleList( + (0): ResnetBlock( + (norm1): GroupNorm(32, 512, eps=1e-06, affine=True) + (conv1): Conv2d(512, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + (norm2): GroupNorm(32, 256, eps=1e-06, affine=True) + (dropout): Dropout(p=0.0, inplace=False) + (conv2): Conv2d(256, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + (nin_shortcut): Conv2d(512, 256, kernel_size=(1, 1), stride=(1, 1)) + ) + (1-2): 2 x ResnetBlock( + (norm1): GroupNorm(32, 256, eps=1e-06, affine=True) + (conv1): Conv2d(256, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + (norm2): GroupNorm(32, 256, eps=1e-06, affine=True) + (dropout): Dropout(p=0.0, inplace=False) + (conv2): Conv2d(256, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + ) + ) + (attn): ModuleList() + (upsample): Upsample( + (conv): Conv2d(256, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + ) + ) + (2-3): 2 x Module( + (block): ModuleList( + (0-2): 3 x ResnetBlock( + (norm1): GroupNorm(32, 512, eps=1e-06, affine=True) + (conv1): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + (norm2): GroupNorm(32, 512, eps=1e-06, affine=True) + (dropout): Dropout(p=0.0, inplace=False) + (conv2): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + ) + ) + (attn): ModuleList() + (upsample): Upsample( + (conv): Conv2d(512, 512, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + ) + ) + ) + (norm_out): GroupNorm(32, 128, eps=1e-06, affine=True) + (conv_out): Conv2d(128, 3, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1)) + ) + (conv1): Identity() + (conv2): Identity() + ) + (cond_stage_model): T5ACEPlusClipFluxEmbedder T5ACEPlusClipFluxEmbedder( + (t5_model): ACEHFEmbedder ACEHFEmbedder( + (hf_module): T5EncoderModel( + (shared): Embedding(32128, 4096) + (encoder): T5Stack( + (embed_tokens): Embedding(32128, 4096) + (block): ModuleList( + (0): T5Block( + (layer): ModuleList( + (0): T5LayerSelfAttention( + (SelfAttention): T5Attention( + (q): Linear(in_features=4096, out_features=4096, bias=False) + (k): Linear(in_features=4096, out_features=4096, bias=False) + (v): Linear(in_features=4096, out_features=4096, bias=False) + (o): Linear(in_features=4096, out_features=4096, bias=False) + (relative_attention_bias): Embedding(32, 64) + ) + (layer_norm): T5LayerNorm() + (dropout): Dropout(p=0.1, inplace=False) + ) + (1): T5LayerFF( + (DenseReluDense): T5DenseGatedActDense( + (wi_0): Linear(in_features=4096, out_features=10240, bias=False) + (wi_1): Linear(in_features=4096, out_features=10240, bias=False) + (wo): Linear(in_features=10240, out_features=4096, bias=False) + (dropout): Dropout(p=0.1, inplace=False) + (act): NewGELUActivation() + ) + (layer_norm): T5LayerNorm() + (dropout): Dropout(p=0.1, inplace=False) + ) + ) + ) + (1-23): 23 x T5Block( + (layer): ModuleList( + (0): T5LayerSelfAttention( + (SelfAttention): T5Attention( + (q): Linear(in_features=4096, out_features=4096, bias=False) + (k): Linear(in_features=4096, out_features=4096, bias=False) + (v): Linear(in_features=4096, out_features=4096, bias=False) + (o): Linear(in_features=4096, out_features=4096, bias=False) + ) + (layer_norm): T5LayerNorm() + (dropout): Dropout(p=0.1, inplace=False) + ) + (1): T5LayerFF( + (DenseReluDense): T5DenseGatedActDense( + (wi_0): Linear(in_features=4096, out_features=10240, bias=False) + (wi_1): Linear(in_features=4096, out_features=10240, bias=False) + (wo): Linear(in_features=10240, out_features=4096, bias=False) + (dropout): Dropout(p=0.1, inplace=False) + (act): NewGELUActivation() + ) + (layer_norm): T5LayerNorm() + (dropout): Dropout(p=0.1, inplace=False) + ) + ) + ) + ) + (final_layer_norm): T5LayerNorm() + (dropout): Dropout(p=0.1, inplace=False) + ) + ) + ) + (clip_model): ACEHFEmbedder ACEHFEmbedder( + (hf_module): CLIPTextModel( + (text_model): CLIPTextTransformer( + (embeddings): CLIPTextEmbeddings( + (token_embedding): Embedding(49408, 768) + (position_embedding): Embedding(77, 768) + ) + (encoder): CLIPEncoder( + (layers): ModuleList( + (0-11): 12 x CLIPEncoderLayer( + (self_attn): CLIPSdpaAttention( + (k_proj): Linear(in_features=768, out_features=768, bias=True) + (v_proj): Linear(in_features=768, out_features=768, bias=True) + (q_proj): Linear(in_features=768, out_features=768, bias=True) + (out_proj): Linear(in_features=768, out_features=768, bias=True) + ) + (layer_norm1): LayerNorm((768,), eps=1e-05, elementwise_affine=True) + (mlp): CLIPMLP( + (activation_fn): QuickGELUActivation() + (fc1): Linear(in_features=768, out_features=3072, bias=True) + (fc2): Linear(in_features=3072, out_features=768, bias=True) + ) + (layer_norm2): LayerNorm((768,), eps=1e-05, elementwise_affine=True) + ) + ) + ) + (final_layer_norm): LayerNorm((768,), eps=1e-05, elementwise_affine=True) + ) + ) + ) + ) + ) +) +scepter [INFO] 2025-06-02 09:29:07,179 [File: log.py Function: before_solve at line 260] Tensorboard: save to ./examples/exp_example/20250602092700/tensorboard +scepter [INFO] 2025-06-02 09:51:38,590 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [20/100000], data_time: 42.3201(42.3201), time: 67.5697(67.5697), loss: 0.2030(0.2030), throughput: 5349/day, all_throughput: 20, pg0_lr: 0.001000, scale: 1.000000, [24mins 36secs 0.02%(85days 9hours 50mins 19secs)] +scepter [INFO] 2025-06-02 10:13:55,895 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [40/100000], data_time: 41.7579(42.0390), time: 66.8652(67.2175), loss: 0.1496(0.1763), throughput: 5343/day, all_throughput: 40, pg0_lr: 0.001000, scale: 1.000000, [46mins 53secs 0.04%(81days 9hours 1mins 35secs)] +scepter [INFO] 2025-06-02 10:25:05,166 [File: checkpoint.py Function: after_iter at line 109] Saving checkpoint after 50 steps +scepter [INFO] 2025-06-02 10:36:13,836 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [60/100000], data_time: 41.7211(41.9331), time: 66.8971(67.1107), loss: 0.1511(0.1679), throughput: 5344/day, all_throughput: 60, pg0_lr: 0.001000, scale: 1.000000, [1hours 9mins 11secs 0.06%(80days 48mins 10secs)] +scepter [INFO] 2025-06-02 10:58:22,386 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [80/100000], data_time: 41.4367(41.8090), time: 66.4275(66.9399), loss: 0.1644(0.1670), throughput: 5352/day, all_throughput: 80, pg0_lr: 0.001000, scale: 1.000000, [1hours 31mins 19secs 0.08%(79days 5hours 14mins 48secs)] +scepter [INFO] 2025-06-02 11:20:31,032 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [100/100000], data_time: 41.3202(41.7112), time: 66.4323(66.8384), loss: 0.1641(0.1664), throughput: 5354/day, all_throughput: 100, pg0_lr: 0.001000, scale: 1.000000, [1hours 53mins 28secs 0.10%(78days 17hours 23mins 32secs)] +scepter [INFO] 2025-06-02 11:20:31,033 [File: checkpoint.py Function: after_iter at line 109] Saving checkpoint after 100 steps +scepter [INFO] 2025-06-02 11:42:39,626 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [120/100000], data_time: 41.4255(41.6636), time: 66.4297(66.7703), loss: 0.1851(0.1695), throughput: 5355/day, all_throughput: 120, pg0_lr: 0.001000, scale: 1.000000, [2hours 15mins 37secs 0.12%(78days 9hours 21mins 15secs)] +scepter [INFO] 2025-06-02 12:04:51,591 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [140/100000], data_time: 41.7225(41.6720), time: 66.5982(66.7457), loss: 0.2146(0.1760), throughput: 5357/day, all_throughput: 140, pg0_lr: 0.001000, scale: 1.000000, [2hours 37mins 49secs 0.14%(78days 4hours 10mins 30secs)] +scepter [INFO] 2025-06-02 12:15:54,666 [File: checkpoint.py Function: after_iter at line 109] Saving checkpoint after 150 steps +scepter [INFO] 2025-06-02 12:26:59,151 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [160/100000], data_time: 41.3753(41.6349), time: 66.3780(66.6997), loss: 0.2634(0.1869), throughput: 5358/day, all_throughput: 160, pg0_lr: 0.001000, scale: 1.000000, [2hours 59mins 56secs 0.16%(77days 23hours 26mins 5secs)] +scepter [INFO] 2025-06-02 12:49:04,260 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [180/100000], data_time: 41.2891(41.5965), time: 66.2555(66.6504), loss: 0.2875(0.1981), throughput: 5356/day, all_throughput: 180, pg0_lr: 0.001000, scale: 1.000000, [3hours 22mins 1secs 0.18%(77days 19hours 17mins 18secs)] +scepter [INFO] 2025-06-02 13:11:15,836 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [200/100000], data_time: 41.4977(41.5866), time: 66.5788(66.6432), loss: 0.2561(0.2039), throughput: 5354/day, all_throughput: 200, pg0_lr: 0.001000, scale: 1.000000, [3hours 44mins 13secs 0.20%(77days 16hours 47mins 39secs)] +scepter [INFO] 2025-06-02 13:11:15,837 [File: checkpoint.py Function: after_iter at line 109] Saving checkpoint after 200 steps +scepter [INFO] 2025-06-02 13:33:26,818 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [220/100000], data_time: 41.5564(41.5839), time: 66.5491(66.6346), loss: 0.2442(0.2076), throughput: 5358/day, all_throughput: 220, pg0_lr: 0.001000, scale: 1.000000, [4hours 6mins 24secs 0.22%(77days 14hours 36mins 40secs)] +scepter [INFO] 2025-06-02 13:55:36,662 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [240/100000], data_time: 41.6124(41.5863), time: 66.4922(66.6228), loss: 0.2165(0.2083), throughput: 5359/day, all_throughput: 240, pg0_lr: 0.001000, scale: 1.000000, [4hours 28mins 34secs 0.24%(77days 12hours 35mins 57secs)] +scepter [INFO] 2025-06-02 14:06:40,396 [File: checkpoint.py Function: after_iter at line 109] Saving checkpoint after 250 steps +scepter [INFO] 2025-06-02 14:17:47,873 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [260/100000], data_time: 41.6454(41.5908), time: 66.5605(66.6180), loss: 0.1728(0.2056), throughput: 5361/day, all_throughput: 260, pg0_lr: 0.001000, scale: 1.000000, [4hours 50mins 45secs 0.26%(77days 10hours 59mins 7secs)] +scepter [INFO] 2025-06-02 14:39:46,950 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [280/100000], data_time: 41.1921(41.5623), time: 65.9539(66.5706), loss: 0.1977(0.2050), throughput: 5365/day, all_throughput: 280, pg0_lr: 0.001000, scale: 1.000000, [5hours 12mins 44secs 0.28%(77days 8hours 20mins 56secs)] +scepter [INFO] 2025-06-02 15:01:46,703 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [300/100000], data_time: 41.1670(41.5360), time: 65.9876(66.5317), loss: 0.1706(0.2027), throughput: 5369/day, all_throughput: 300, pg0_lr: 0.001000, scale: 1.000000, [5hours 34mins 44secs 0.30%(77days 6hours 4mins 39secs)] +scepter [INFO] 2025-06-02 15:01:46,703 [File: checkpoint.py Function: after_iter at line 109] Saving checkpoint after 300 steps +scepter [INFO] 2025-06-02 15:23:47,457 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [320/100000], data_time: 41.1280(41.5105), time: 66.0377(66.5008), loss: 0.1706(0.2007), throughput: 5370/day, all_throughput: 320, pg0_lr: 0.001000, scale: 1.000000, [5hours 56mins 45secs 0.32%(77days 4hours 7mins 51secs)] +scepter [INFO] 2025-06-02 15:45:53,102 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [340/100000], data_time: 41.4600(41.5075), time: 66.2822(66.4880), loss: 0.1437(0.1974), throughput: 5374/day, all_throughput: 340, pg0_lr: 0.001000, scale: 1.000000, [6hours 18mins 50secs 0.34%(77days 2hours 46mins 6secs)] +scepter [INFO] 2025-06-02 15:56:57,492 [File: checkpoint.py Function: after_iter at line 109] Saving checkpoint after 350 steps +scepter [INFO] 2025-06-02 16:08:04,405 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [360/100000], data_time: 41.6830(41.5173), time: 66.5652(66.4922), loss: 0.1532(0.1949), throughput: 5375/day, all_throughput: 360, pg0_lr: 0.001000, scale: 1.000000, [6hours 41mins 1secs 0.36%(77days 1hours 57mins 5secs)] +scepter [INFO] 2025-06-02 16:30:13,447 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [380/100000], data_time: 41.4708(41.5148), time: 66.4521(66.4901), loss: 0.1476(0.1924), throughput: 5375/day, all_throughput: 380, pg0_lr: 0.001000, scale: 1.000000, [7hours 3mins 11secs 0.38%(77days 1hours 1mins 0secs)] +scepter [INFO] 2025-06-02 16:52:20,557 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [400/100000], data_time: 41.3496(41.5065), time: 66.3555(66.4834), loss: 0.1394(0.1898), throughput: 5375/day, all_throughput: 400, pg0_lr: 0.001000, scale: 1.000000, [7hours 25mins 18secs 0.40%(77days 18secs)] +scepter [INFO] 2025-06-02 16:52:20,558 [File: checkpoint.py Function: after_iter at line 109] Saving checkpoint after 400 steps +scepter [INFO] 2025-06-02 17:14:31,006 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [420/100000], data_time: 41.6155(41.5117), time: 66.5224(66.4853), loss: 0.1420(0.1875), throughput: 5375/day, all_throughput: 420, pg0_lr: 0.001000, scale: 1.000000, [7hours 47mins 28secs 0.42%(76days 23hours 16mins 28secs)] +scepter [INFO] 2025-06-02 17:36:43,227 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [440/100000], data_time: 41.4885(41.5107), time: 66.6110(66.4910), loss: 0.1349(0.1851), throughput: 5374/day, all_throughput: 440, pg0_lr: 0.001000, scale: 1.000000, [8hours 9mins 40secs 0.44%(76days 22hours 41mins 17secs)] +scepter [INFO] 2025-06-02 17:47:48,583 [File: checkpoint.py Function: after_iter at line 109] Saving checkpoint after 450 steps +scepter [INFO] 2025-06-02 17:58:55,039 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [460/100000], data_time: 41.6236(41.5156), time: 66.5906(66.4953), loss: 0.1423(0.1832), throughput: 5374/day, all_throughput: 460, pg0_lr: 0.001000, scale: 1.000000, [8hours 31mins 52secs 0.46%(76days 22hours 5mins 45secs)] +scepter [INFO] 2025-06-02 18:21:09,747 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [480/100000], data_time: 41.6391(41.5207), time: 66.7354(66.5053), loss: 0.1559(0.1821), throughput: 5373/day, all_throughput: 480, pg0_lr: 0.001000, scale: 1.000000, [8hours 54mins 7secs 0.48%(76days 21hours 41mins 20secs)] +scepter [INFO] 2025-06-02 18:43:26,230 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [500/100000], data_time: 41.8578(41.5342), time: 66.8242(66.5181), loss: 0.1370(0.1803), throughput: 5374/day, all_throughput: 500, pg0_lr: 0.001000, scale: 1.000000, [9hours 16mins 23secs 0.50%(76days 21hours 23mins 0secs)] +scepter [INFO] 2025-06-02 18:43:26,231 [File: checkpoint.py Function: after_iter at line 109] Saving checkpoint after 500 steps +scepter [INFO] 2025-06-02 19:05:33,274 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [520/100000], data_time: 41.4453(41.5308), time: 66.3522(66.5117), loss: 0.2116(0.1815), throughput: 5375/day, all_throughput: 520, pg0_lr: 0.001000, scale: 1.000000, [9hours 38mins 30secs 0.52%(76days 20hours 34mins 15secs)] +scepter [INFO] 2025-06-02 19:27:53,538 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [540/100000], data_time: 41.6550(41.5354), time: 67.0132(66.5303), loss: 0.1492(0.1803), throughput: 5372/day, all_throughput: 540, pg0_lr: 0.001000, scale: 1.000000, [10hours 51secs 0.54%(76days 20hours 28mins 3secs)] +scepter [INFO] 2025-06-02 19:39:01,432 [File: checkpoint.py Function: after_iter at line 109] Saving checkpoint after 550 steps +scepter [INFO] 2025-06-02 19:50:06,907 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [560/100000], data_time: 41.6412(41.5392), time: 66.6685(66.5352), loss: 0.2414(0.1825), throughput: 5373/day, all_throughput: 560, pg0_lr: 0.001000, scale: 1.000000, [10hours 23mins 4secs 0.56%(76days 20hours 18secs)] +scepter [INFO] 2025-06-02 20:12:08,182 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [580/100000], data_time: 41.2246(41.5283), time: 66.0638(66.5189), loss: 0.2961(0.1864), throughput: 5373/day, all_throughput: 580, pg0_lr: 0.001000, scale: 1.000000, [10hours 45mins 5secs 0.58%(76days 18hours 58mins 22secs)] +scepter [INFO] 2025-06-02 20:34:10,081 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [600/100000], data_time: 41.2012(41.5174), time: 66.0950(66.5048), loss: 0.2136(0.1873), throughput: 5374/day, all_throughput: 600, pg0_lr: 0.001000, scale: 1.000000, [11hours 7mins 7secs 0.60%(76days 18hours 50secs)] +scepter [INFO] 2025-06-02 20:34:10,082 [File: checkpoint.py Function: after_iter at line 109] Saving checkpoint after 600 steps +scepter [INFO] 2025-06-02 20:56:21,375 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [620/100000], data_time: 41.6078(41.5203), time: 66.5647(66.5067), loss: 0.1896(0.1874), throughput: 5375/day, all_throughput: 620, pg0_lr: 0.001000, scale: 1.000000, [11hours 29mins 18secs 0.62%(76days 17hours 30mins 41secs)] +scepter [INFO] 2025-06-02 21:18:27,057 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [640/100000], data_time: 41.2849(41.5130), time: 66.2841(66.4998), loss: 0.1944(0.1876), throughput: 5376/day, all_throughput: 640, pg0_lr: 0.001000, scale: 1.000000, [11hours 51mins 24secs 0.64%(76days 16hours 46mins 31secs)] +scepter [INFO] 2025-06-02 21:29:27,002 [File: checkpoint.py Function: after_iter at line 109] Saving checkpoint after 650 steps +scepter [INFO] 2025-06-02 21:40:27,523 [File: log.py Function: _print_iter_log at line 71] Stage [train] iter: [660/100000], data_time: 41.1508(41.5020), time: 66.0233(66.4853), loss: 0.1546(0.1866), throughput: 5376/day, all_throughput: 660, pg0_lr: 0.001000, scale: 1.000000, [12hours 13mins 25secs 0.66%(76days 15hours 50mins 36secs)] diff --git a/ACE_plus/examples/exp_example/20250602092700/train.yaml b/ACE_plus/examples/exp_example/20250602092700/train.yaml new file mode 100644 index 0000000000000000000000000000000000000000..9c8f141686d91b64513fe45a530c356a9f311e72 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250602092700/train.yaml @@ -0,0 +1,288 @@ +ENV: + BACKEND: nccl + SEED: 1999 +SOLVER: + # NAME DESCRIPTION: TYPE: default: 'LatentUfitSolver' + NAME: FormalACEPlusSolver + # MAX_STEPS DESCRIPTION: The total steps for training. TYPE: int default: 100000 + MAX_STEPS: 100000 + # USE_AMP DESCRIPTION: Use amp to surpport mix precision or not, default is False. TYPE: bool default: False + USE_AMP: True + # DTYPE DESCRIPTION: The precision for training. TYPE: str default: 'float32' + DTYPE: bfloat16 + ENABLE_GRADSCALER: False + # USE_FAIRSCALE DESCRIPTION: Use fairscale as the backend of ddp, default False. TYPE: bool default: False + USE_FAIRSCALE: False + USE_ORIG_PARAMS: True + USE_FSDP: True # lora use ddp(USE_FSDP=False), else use fsdp(USE_FSDP=True) + # LOAD_MODEL_ONLY DESCRIPTION: Only load the model rather than the optimizer and schedule, default is False. TYPE: bool default: False + LOAD_MODEL_ONLY: False + # RESUME_FROM DESCRIPTION: Resume from some state of training! TYPE: str default: '' + RESUME_FROM: + # WORK_DIR DESCRIPTION: Save dir of the training log or model. TYPE: str default: '' + WORK_DIR: ./examples/exp_example/ + # LOG_FILE DESCRIPTION: Save log path. TYPE: str default: '' + LOG_FILE: std_log.txt + # LOG_TRAIN_NUM DESCRIPTION: The number samples used to log in training phase. TYPE: int default: -1 + LOG_TRAIN_NUM: 16 + # FSDP_REDUCE_DTYPE DESCRIPTION: The dtype of reduce in FSDP. TYPE: str default: 'float16' + FSDP_REDUCE_DTYPE: float32 + # FSDP_BUFFER_DTYPE DESCRIPTION: The dtype of buffer in FSDP. TYPE: str default: 'float16' + FSDP_BUFFER_DTYPE: float32 + # FSDP_SHARD_MODULES DESCRIPTION: The modules to be sharded in FSDP. TYPE: list default: ['model'] + FSDP_SHARD_MODULES: + - MODULE: 'model.model' + FSDP_GROUP: [ 'single_blocks', 'double_blocks'] + - MODULE: 'cond_stage_model.t5_model.hf_module.encoder' + FSDP_GROUP: [ 'block' ] # + SAVE_MODULES: [ 'model'] # + TRAIN_MODULES: ['model'] + + # + FILE_SYSTEM: + - NAME: HuggingfaceFs + TEMP_DIR: ./cache + - NAME: ModelscopeFs + TEMP_DIR: ./cache + # + MODEL: + NAME: LatentDiffusionACEPlus + PARAMETERIZATION: rf + TIMESTEPS: 1000 + GUIDE_SCALE: 1.0 + PRETRAINED_MODEL: + IGNORE_KEYS: [ ] + USE_EMA: False + EVAL_EMA: False + SIZE_FACTOR: 8 + DIFFUSION: + NAME: DiffusionFluxRF + PREDICTION_TYPE: raw + NOISE_NORM: True + # NOISE_SCHEDULER DESCRIPTION: TYPE: default: '' + NOISE_SCHEDULER: + NAME: FlowMatchFluxShiftScheduler + SHIFT: False + PRE_T_SAMPLE: True + PRE_T_SAMPLE_FOLD: 1 + SIGMOID_SCALE: 1 + BASE_SHIFT: 0.5 + MAX_SHIFT: 1.15 + SAMPLER_SCHEDULER: + NAME: FlowMatchFluxShiftScheduler + SHIFT: True + PRE_T_SAMPLE: False + SIGMOID_SCALE: 1 + BASE_SHIFT: 0.5 + MAX_SHIFT: 1.15 + + # + DIFFUSION_MODEL: + # NAME DESCRIPTION: TYPE: default: 'Flux' + NAME: FluxMRModiACEPlus + PRETRAINED_MODEL: /home/Ubuntu/Downloads/Unmodel/Reference_models/flux1-fill-dev.safetensors + # IN_CHANNELS DESCRIPTION: model's input channels. TYPE: int default: 64 + IN_CHANNELS: 448 + # OUT_CHANNELS DESCRIPTION: model's input channels. TYPE: int default: 64 + OUT_CHANNELS: 64 + # HIDDEN_SIZE DESCRIPTION: model's hidden size. TYPE: int default: 1024 + HIDDEN_SIZE: 3072 + REDUX_DIM: 1152 + # NUM_HEADS DESCRIPTION: number of heads in the transformer. TYPE: int default: 16 + NUM_HEADS: 24 + # AXES_DIM DESCRIPTION: dimensions of the axes of the positional encoding. TYPE: list default: [16, 56, 56] + AXES_DIM: [ 16, 56, 56 ] + # THETA DESCRIPTION: theta for positional encoding. TYPE: int default: 10000 + THETA: 10000 + # VEC_IN_DIM DESCRIPTION: dimension of the vector input. TYPE: int default: 768 + VEC_IN_DIM: 768 + # GUIDANCE_EMBED DESCRIPTION: whether to use guidance embedding. TYPE: bool default: False + GUIDANCE_EMBED: True + # CONTEXT_IN_DIM DESCRIPTION: dimension of the context input. TYPE: int default: 4096 + CONTEXT_IN_DIM: 4096 + # MLP_RATIO DESCRIPTION: ratio of mlp hidden size to hidden size. TYPE: float default: 4.0 + MLP_RATIO: 4.0 + # QKV_BIAS DESCRIPTION: whether to use bias in qkv projection. TYPE: bool default: True + QKV_BIAS: True + # DEPTH DESCRIPTION: number of transformer blocks. TYPE: int default: 19 + DEPTH: 19 + # DEPTH_SINGLE_BLOCKS DESCRIPTION: number of transformer blocks in the single stream block. TYPE: int default: 38 + DEPTH_SINGLE_BLOCKS: 38 + # ATTN_BACKEND:setting 'flash_attn' to use flash_attn2, if the version of pytorch > 2.4.0, using 'pytorch' to use pytorch's implementation + ATTN_BACKEND: flash_attn + # USE_GRAD_CHECKPOINT: setting gc to true can decrease the memory usage. + USE_GRAD_CHECKPOINT: True + + # + FIRST_STAGE_MODEL: + NAME: AutoencoderKLFlux + EMBED_DIM: 16 + PRETRAINED_MODEL: /home/Ubuntu/Downloads/Unmodel/Reference_models/ae.safetensors + IGNORE_KEYS: [ ] + BATCH_SIZE: 8 + USE_CONV: False + SCALE_FACTOR: 0.3611 + SHIFT_FACTOR: 0.1159 + # + ENCODER: + NAME: Encoder + CH: 128 + OUT_CH: 3 + NUM_RES_BLOCKS: 2 + IN_CHANNELS: 3 + ATTN_RESOLUTIONS: [ ] + CH_MULT: [ 1, 2, 4, 4 ] + Z_CHANNELS: 16 + DOUBLE_Z: True + DROPOUT: 0.0 + RESAMP_WITH_CONV: True + # + DECODER: + NAME: Decoder + CH: 128 + OUT_CH: 3 + NUM_RES_BLOCKS: 2 + IN_CHANNELS: 3 + ATTN_RESOLUTIONS: [ ] + CH_MULT: [ 1, 2, 4, 4 ] + Z_CHANNELS: 16 + DROPOUT: 0.0 + RESAMP_WITH_CONV: True + GIVE_PRE_END: False + TANH_OUT: False + # + COND_STAGE_MODEL: + # NAME DESCRIPTION: TYPE: default: 'T5PlusClipFluxEmbedder' + NAME: T5ACEPlusClipFluxEmbedder + # T5_MODEL DESCRIPTION: TYPE: default: '' + T5_MODEL: + # NAME DESCRIPTION: TYPE: default: 'HFEmbedder' + NAME: ACEHFEmbedder + # HF_MODEL_CLS DESCRIPTION: huggingface cls in transfomer TYPE: NoneType default: None + HF_MODEL_CLS: T5EncoderModel + # MODEL_PATH DESCRIPTION: model folder path TYPE: NoneType default: None + MODEL_PATH: /home/Ubuntu/Downloads/Unmodel/Reference_models/t5_xxl/ + # HF_TOKENIZER_CLS DESCRIPTION: huggingface cls in transfomer TYPE: NoneType default: None + HF_TOKENIZER_CLS: T5Tokenizer + # TOKENIZER_PATH DESCRIPTION: tokenizer folder path TYPE: NoneType default: None + TOKENIZER_PATH: /home/Ubuntu/Downloads/Unmodel/Reference_models/t5_xxl/ + ADDED_IDENTIFIER: [ '','{image}', '{caption}', '{mask}', '{ref_image}', '{image1}', '{image2}', '{image3}', '{image4}', '{image5}', '{image6}', '{image7}', '{image8}', '{image9}' ] + # MAX_LENGTH DESCRIPTION: max length of input TYPE: int default: 77 + MAX_LENGTH: 512 + # OUTPUT_KEY DESCRIPTION: output key TYPE: str default: 'last_hidden_state' + OUTPUT_KEY: last_hidden_state + # D_TYPE DESCRIPTION: dtype TYPE: str default: 'bfloat16' + D_TYPE: bfloat16 + # BATCH_INFER DESCRIPTION: batch infer TYPE: bool default: False + BATCH_INFER: False + CLEAN: whitespace + # CLIP_MODEL DESCRIPTION: TYPE: default: '' + CLIP_MODEL: + # NAME DESCRIPTION: TYPE: default: 'HFEmbedder' + NAME: ACEHFEmbedder + # HF_MODEL_CLS DESCRIPTION: huggingface cls in transfomer TYPE: NoneType default: None + HF_MODEL_CLS: CLIPTextModel + # MODEL_PATH DESCRIPTION: model folder path TYPE: NoneType default: None + MODEL_PATH: /home/Ubuntu/Downloads/Unmodel/Reference_models/clip_l/ + # HF_TOKENIZER_CLS DESCRIPTION: huggingface cls in transfomer TYPE: NoneType default: None + HF_TOKENIZER_CLS: CLIPTokenizer + # TOKENIZER_PATH DESCRIPTION: tokenizer folder path TYPE: NoneType default: None + TOKENIZER_PATH: /home/Ubuntu/Downloads/Unmodel/Reference_models/clip_l/ + # MAX_LENGTH DESCRIPTION: max length of input TYPE: int default: 77 + MAX_LENGTH: 77 + # OUTPUT_KEY DESCRIPTION: output key TYPE: str default: 'last_hidden_state' + OUTPUT_KEY: pooler_output + # D_TYPE DESCRIPTION: dtype TYPE: str default: 'bfloat16' + D_TYPE: bfloat16 + # BATCH_INFER DESCRIPTION: batch infer TYPE: bool default: False + BATCH_INFER: True + CLEAN: whitespace + TUNER: + # THE LORA PARAMETERS + - NAME: SwiftLoRA + R: 64 + LORA_ALPHA: 64 + LORA_DROPOUT: 0.0 + BIAS: "none" + TARGET_MODULES: "(model.double_blocks.*(.qkv|.img_mlp.0|.img_mlp.2|.txt_mlp.0|.txt_mlp.2|.proj|.img_mod.lin|.txt_mod.lin))|(model.single_blocks.*(.linear1|.linear2|.modulation.lin))$" + # + SAMPLE_ARGS: + SAMPLE_STEPS: 28 + SAMPLER: flow_euler + SEED: 42 + IMAGE_SIZE: [ 2048, 2048 ] + #IMAGE_SIZE: [ 1024, 1024 ] + GUIDE_SCALE: 50 + + LR_SCHEDULER: + NAME: StepAnnealingLR + WARMUP_STEPS: 0 + TOTAL_STEPS: 100000 + DECAY_MODE: 'cosine' + # + OPTIMIZER: + NAME: AdamW + LEARNING_RATE: 1e-3 + BETAS: [ 0.9, 0.999 ] + EPS: 1e-6 + WEIGHT_DECAY: 1e-2 + AMSGRAD: False + # + TRAIN_DATA: + NAME: ACEPlusDataset + MODE: train + DATA_LIST: data/train.csv + DELIMITER: "#;#" + MODIFY_MODE: True + # input_image, input_mask, input_reference_image, target_image, instruction, task_type + FIELDS: ["edit_image", "edit_mask", "ref_image", "target_image", "prompt", "data_type"] + PATH_PREFIX: "" + EDIT_TYPE_LIST: [] + MAX_SEQ_LEN: 4096 + # MAX_SEQ_LEN: 2048 -Vijay + D: 16 + PIN_MEMORY: True + BATCH_SIZE: 1 + NUM_WORKERS: 4 + SAMPLER: + NAME: LoopSampler + + EVAL_DATA: + NAME: ACEPlusDataset + MODE: eval + DATA_LIST: data/train.csv + DELIMITER: "#;#" + MODIFY_MODE: True + # input_image, input_mask, input_reference_image, target_image, instruction, task_type + FIELDS: [ "edit_image", "edit_mask", "ref_image", "target_image", "prompt", "data_type" ] + PATH_PREFIX: "" + EDIT_TYPE_LIST: [ ] + MAX_SEQ_LEN: 4096 + # MAX_SEQ_LEN: 2048 -Vijay + D: 16 + PIN_MEMORY: True + BATCH_SIZE: 1 + NUM_WORKERS: 4 + + TRAIN_HOOKS: + - NAME: ACEBackwardHook + GRADIENT_CLIP: 1.0 + PRIORITY: 10 + - NAME: LogHook + LOG_INTERVAL: 20 + - NAME: ACECheckpointHook + INTERVAL: 50 + #INTERVAL: 250 --Vijay + PRIORITY: 200 + DISABLE_SNAPSHOT: True + - NAME: ProbeDataHook + PROB_INTERVAL: 10 + #PROB_INTERVAL: 50 -Vijay + PRIORITY: 0 + - NAME: TensorboardLogHook + LOG_INTERVAL: 50 + EVAL_HOOKS: + - NAME: ProbeDataHook + PROB_INTERVAL: 10 + #PROB_INTERVAL: 50 -Vijay + PRIORITY: 0 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/0_0_3_probe_0_[848_1232_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/0_0_3_probe_0_[848_1232_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..e63a44aa5aebd19c2dfa9fdae59278b1ed551c19 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/0_0_3_probe_0_[848_1232_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d477a76c92bd8f9f49ec8cc715e8238da792c3e21a170239ce81977fc9e35e11 +size 166553 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/0_0_5_probe_0_[848_1232_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/0_0_5_probe_0_[848_1232_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..7dc6509f4ca916c5d87d062eaa2c362523bd1d80 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/0_0_5_probe_0_[848_1232_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3afcabe9f07713196caf23789bc714153f0facc3d6c11d76f86d229835e9aafc +size 116582 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/1_0_0_probe_0_[752_1360_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/1_0_0_probe_0_[752_1360_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..bc14b693ddbad2aa969312cb6c19897a9ec14259 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/1_0_0_probe_0_[752_1360_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3924ff0e8ea07a06b939860538ce9b9b0294917f929c0f63b5cd5211223eba0a +size 117204 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/1_0_3_probe_0_[752_1360_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/1_0_3_probe_0_[752_1360_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..909b08e29538169b034a7d1efd669d6c4a3fbefd --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/1_0_3_probe_0_[752_1360_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f4014dae6fefe302b69608c403145bd8312afa20514fc726bd8f0ed37da71756 +size 121816 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/1_0_5_probe_0_[752_1360_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/1_0_5_probe_0_[752_1360_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..b4eed8d3f97e9ea3ae5d819bf072ae91ebb49d50 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/1_0_5_probe_0_[752_1360_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:28862e4e42f4768b07ebd12cbdc484fed91bbd74a01e3c4e963ab949bbb5609a +size 118563 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/2_0_0_probe_0_[736_1408_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/2_0_0_probe_0_[736_1408_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..f289b741260d61061b321c295b00a97606005eeb --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/2_0_0_probe_0_[736_1408_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f55b131dfc946b7fdfde51b37c3905e3900d984cfff1df37fd0ee0ed603959af +size 100487 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/2_0_3_probe_0_[736_1408_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/2_0_3_probe_0_[736_1408_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..dad866e5a748216df759b48d6767d091a1b9991f --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/2_0_3_probe_0_[736_1408_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:26b7f7971dc27181652b46ef4006c890cf0cd83d28b5742bf7130c0844538878 +size 115540 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/3_0_3_probe_0_[880_1168_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/3_0_3_probe_0_[880_1168_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..4cd80ce0325023d8258ca1e808816d9f52f7ed0a --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/3_0_3_probe_0_[880_1168_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2051a1101cc6aa357e9ee91274e0a2739457d663de72bad594a5db9ec6b5358f +size 150317 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/3_0_5_probe_0_[880_1168_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/3_0_5_probe_0_[880_1168_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..1227bfb5bb194ae08f946ddc42493d93074a69ce --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/3_0_5_probe_0_[880_1168_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7178a18678ca121514e54416d4c7f671f35dae171f7808c045e0158207d284e3 +size 145210 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/4_0_3_probe_0_[784_1328_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/4_0_3_probe_0_[784_1328_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..f9bd8c85a18a7ce79d87eddc86d10417564929f0 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/4_0_3_probe_0_[784_1328_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:22f195a6db09ad5799482281bf13ab060b39f5701bdebc1f0ef66c3698287838 +size 130026 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/5_0_5_probe_0_[864_1200_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/5_0_5_probe_0_[864_1200_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..d4920ca5c8ba784c8da102f1a822b9cfd1fa531c --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/5_0_5_probe_0_[864_1200_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b5118ec0b48db284e68d7b9074f973e022dddc839955c7da918f858bb985e81a +size 119297 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/6_0_3_probe_0_[864_1184_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/6_0_3_probe_0_[864_1184_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..6e76db2a5202d80e55a08cb4bbdc1979d37b5520 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/6_0_3_probe_0_[864_1184_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7c8c5c058c916044c3f031c8a1e07277309994c5c3ddd882546756151e2b7a85 +size 135589 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/6_0_5_probe_0_[864_1184_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/6_0_5_probe_0_[864_1184_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..7f2c820eccaeb1a17cdc0602267b13fe0396fbfc --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/6_0_5_probe_0_[864_1184_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:38f4e9f6aa3601f4b6614519201acadf7e2dea03400f99ae6bb358b6a422b5cb +size 142954 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/7_0_3_probe_0_[752_1376_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/7_0_3_probe_0_[752_1376_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..c11c4b070cec0149a748c6f05d3df0cf2643ed82 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/7_0_3_probe_0_[752_1376_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f942a681c0bd6bf5fdb3b282e263300a2ee086b357f659545098e0a12cc32ddc +size 150592 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/9_0_3_probe_0_[752_1376_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/9_0_3_probe_0_[752_1376_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..52e962a806bb4ad5f42823e5b5fdda53ff7fe6a5 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-0/BaseSolver_eval_image_step_0/9_0_3_probe_0_[752_1376_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d42ff757f4c8d05cd0ba7a66a3b3a330b3582490306ac73afe63f3344e4b233e +size 108291 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/0_0_0_probe_0_[720_1440_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/0_0_0_probe_0_[720_1440_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..d2e96a3a9370cb9369dba634e3a0dfcc073aa563 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/0_0_0_probe_0_[720_1440_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2c99a14a58db5685c6ce7363e7171da7758e85bdabf474559995b37a8c62932b +size 122494 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/0_0_3_probe_0_[720_1440_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/0_0_3_probe_0_[720_1440_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..cb04516541bf15a384f5c573e2d5b5c2653e6494 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/0_0_3_probe_0_[720_1440_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:51976685fc057b31e16fcffc752c30e2758a7f4c4daeb91b4312e1115b7abf14 +size 191066 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/1_0_0_probe_0_[736_1392_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/1_0_0_probe_0_[736_1392_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..4bce8d6546360d1c2acdaf21fd6bab109844ffc5 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/1_0_0_probe_0_[736_1392_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:128c10ac4a8543eb11dd8f449672a2f271526fab6002a8c7e691b8354fde0633 +size 121282 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/1_0_3_probe_0_[736_1392_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/1_0_3_probe_0_[736_1392_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..811eafe491df75586c641fe85294daa4d5f67b82 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/1_0_3_probe_0_[736_1392_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:92a7aa6f5d92f597b56c234df9f5544ea391173fc96fb2218c19a5b2f65634a7 +size 126037 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/1_0_5_probe_0_[736_1392_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/1_0_5_probe_0_[736_1392_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..9281176f75dbe6c3900e97e3d38aa7d3fb58edc4 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/1_0_5_probe_0_[736_1392_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d1d4dd81ada872617e08fdb86d5d7757436cdca87fd37c88f03226d9170f8923 +size 110861 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/2_0_0_probe_0_[720_1424_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/2_0_0_probe_0_[720_1424_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..b50f9dbb95324b06bbc789a6806b577819b6cef1 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/2_0_0_probe_0_[720_1424_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:095a6f8fe58d7ff1ec6703cb2e4e8370d45f5e85be009e6f15dc41b16a71d830 +size 100946 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/2_0_3_probe_0_[720_1424_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/2_0_3_probe_0_[720_1424_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..a04b89f35b75ad107057dfe3bdc1c1f43df01460 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/2_0_3_probe_0_[720_1424_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a7e383f00177452076369fcd1f5321eabe8171be34d4dc1186d0cd6bcbfcc886 +size 115149 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/2_0_5_probe_0_[720_1424_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/2_0_5_probe_0_[720_1424_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..aa9f0b9ed70fc9094f7e254d3fe9c3964d9517c6 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/2_0_5_probe_0_[720_1424_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c70d610e5174dd8f8e389df8b72c5445e86831073ea5b2097aa479f449461dcf +size 112340 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/3_0_0_probe_0_[752_1376_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/3_0_0_probe_0_[752_1376_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..780a14346a59e0ee4b3ccad589ad194ad24474f1 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/3_0_0_probe_0_[752_1376_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:266e8368c78d9c5735351fbaeeaa48a3102a48bdfea6f0039caa718030fa93b8 +size 116402 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/3_0_3_probe_0_[752_1376_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/3_0_3_probe_0_[752_1376_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..892e0fc17aaf5f5214b86ab6f6fd03cadba731d2 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/3_0_3_probe_0_[752_1376_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59441773db34f6d9c8efb86d1f8018f9f5cad84a4db5e9f50916827ab39f813f +size 171609 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/3_0_5_probe_0_[752_1376_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/3_0_5_probe_0_[752_1376_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..6116595fa447471df6cf7437774ec59fbcbc5ea4 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/3_0_5_probe_0_[752_1376_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:965a93067a979e26e18b7b6e854a05de300cb196bac801e05e497cc4c7445511 +size 141659 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/4_0_3_probe_0_[864_1200_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/4_0_3_probe_0_[864_1200_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..5908d3ceeffc8fad31a27e413ac885deafe30ba0 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/4_0_3_probe_0_[864_1200_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:489403a28ef18020ee97a1d8c531c415a0efd1dc481ecd543193582bd65a9b31 +size 119880 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/5_0_0_probe_0_[720_1440_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/5_0_0_probe_0_[720_1440_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..790fbeea94124d9d883c0eecea629792640237f6 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/5_0_0_probe_0_[720_1440_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:53c399ae0588e7c13f01b7e60bf0e623405e759713d3de2912028832e48d0507 +size 101226 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/5_0_3_probe_0_[720_1440_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/5_0_3_probe_0_[720_1440_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..c8942f96e79b113bb1afa7f19b28ee4e03a4c147 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/5_0_3_probe_0_[720_1440_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f97aaf3cfdfb049980725dbca0e5806fb4c38b98925b86546cfcffc6a4260d88 +size 104852 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/5_0_5_probe_0_[720_1440_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/5_0_5_probe_0_[720_1440_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..afdc19ab5b6acd117314a3b89fd6eb50cd4de2bb --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/5_0_5_probe_0_[720_1440_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:451827f4baabd281444a49ac87e7d9768a942d8141d8c58987f1299cbe17e8a9 +size 102769 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/6_0_3_probe_0_[864_1200_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/6_0_3_probe_0_[864_1200_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..46e0c07a16b2dfec7bc9f3fc8faca5654a87047a --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/6_0_3_probe_0_[864_1200_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:183457d8de36f34aac31facb57d3869cbbbe94c2cadc24b35c8d72374cee44f0 +size 137250 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/6_0_5_probe_0_[864_1200_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/6_0_5_probe_0_[864_1200_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..b12bdd3816f7d7eeae0e55e1e38eeada25834c99 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/6_0_5_probe_0_[864_1200_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:63d5d6ab6bc91ccff577eacc00151b830327eb0b07d941570a40082f9c20bd68 +size 117642 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/7_0_3_probe_0_[784_1296_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/7_0_3_probe_0_[784_1296_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..af39997e5f5b8b25f3e8c923000ed5e909915d44 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/7_0_3_probe_0_[784_1296_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d5c8cf2a7e1c4072dd2adc0f99cb11f0221f8db2f322afae520f129d4c373f04 +size 138552 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/8_0_3_probe_0_[768_1344_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/8_0_3_probe_0_[768_1344_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..9770e4507f9a2f6418e357f8d4625d5f7785ce5a --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/8_0_3_probe_0_[768_1344_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:89336da690dc396a408cba319ee01aa7e69c2f40e27f7b0e70cbf6a0a113ba75 +size 102141 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/8_0_5_probe_0_[768_1344_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/8_0_5_probe_0_[768_1344_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..9cf5246690fbe13906cd2cbbef4dee7400c490cf --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/8_0_5_probe_0_[768_1344_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2184b0a2fadca12a0e41c80fb816371e69dec05b182a284566a3e992a6cc8201 +size 110141 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/9_0_3_probe_0_[720_1424_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/9_0_3_probe_0_[720_1424_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..2afe34cb343b85c2f12328016cd07d0816ceb09a --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-40/BaseSolver_eval_image_step_40/9_0_3_probe_0_[720_1424_3].jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e987b1064226bad218ee08087411b9bc0bc65e5f7eb1a19e9f464c1b38fd8873 +size 114265 diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/0_0_1_probe_0_[864_1184_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/0_0_1_probe_0_[864_1184_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..1c4605520cced30c37ce4572cc352e040e45864a Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/0_0_1_probe_0_[864_1184_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/10_0_0_probe_0_[768_1344_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/10_0_0_probe_0_[768_1344_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..9b493f2687478bc9d7d1dadd8a064a66bc582b66 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/10_0_0_probe_0_[768_1344_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/10_0_4_probe_0_[768_1344_1].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/10_0_4_probe_0_[768_1344_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..a37d102113f06e4b8103f87327ace8087c274745 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/10_0_4_probe_0_[768_1344_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/11_0_0_probe_0_[736_1408_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/11_0_0_probe_0_[736_1408_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..0f6e407f5df6012ca8175b680019fa1989f6ae9f Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/11_0_0_probe_0_[736_1408_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/11_0_4_probe_0_[736_1408_1].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/11_0_4_probe_0_[736_1408_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..85794bbbe1ace5e6cf245b6e74c15da7fd256e11 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/11_0_4_probe_0_[736_1408_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/1_0_1_probe_0_[864_1200_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/1_0_1_probe_0_[864_1200_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..f5626a5fc1e5fd9a651ac8a98e8e0afc889b6792 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/1_0_1_probe_0_[864_1200_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/1_0_4_probe_0_[864_1200_1].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/1_0_4_probe_0_[864_1200_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..9fe81c3b753dd5a784c40455e4a735023a125851 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/1_0_4_probe_0_[864_1200_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/2_0_0_probe_0_[800_1296_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/2_0_0_probe_0_[800_1296_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..c7d545296837b46c5dc84042aace05d180362fdc Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/2_0_0_probe_0_[800_1296_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/2_0_1_probe_0_[800_1296_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/2_0_1_probe_0_[800_1296_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..80c3f689e2817f1212f93f5cbf45be0003047930 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/2_0_1_probe_0_[800_1296_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/2_0_2_probe_0_[800_1296_1].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/2_0_2_probe_0_[800_1296_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..1b27aeb0614fb1bda6f01d597247d39ba88b4d00 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/2_0_2_probe_0_[800_1296_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/2_0_4_probe_0_[800_1296_1].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/2_0_4_probe_0_[800_1296_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..1b27aeb0614fb1bda6f01d597247d39ba88b4d00 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/2_0_4_probe_0_[800_1296_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/3_0_1_probe_0_[752_1392_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/3_0_1_probe_0_[752_1392_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..7c163afda0a78b351cf4c2e9e649b4959591e693 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/3_0_1_probe_0_[752_1392_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/3_0_4_probe_0_[752_1392_1].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/3_0_4_probe_0_[752_1392_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..a34e5d71393dcebaee1bd481cccbfc7a93370baf Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/3_0_4_probe_0_[752_1392_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/4_0_1_probe_0_[864_1200_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/4_0_1_probe_0_[864_1200_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..d9ac4ec344a936586ca3b2c81bc13a503b93e974 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/4_0_1_probe_0_[864_1200_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/4_0_2_probe_0_[864_1200_1].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/4_0_2_probe_0_[864_1200_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..c2bb4b6fdcf48d8750ad87f613cd3ef2c0df5a07 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/4_0_2_probe_0_[864_1200_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/4_0_4_probe_0_[864_1200_1].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/4_0_4_probe_0_[864_1200_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..c2bb4b6fdcf48d8750ad87f613cd3ef2c0df5a07 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/4_0_4_probe_0_[864_1200_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/4_0_5_probe_0_[864_1200_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/4_0_5_probe_0_[864_1200_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..6e19f6f9529f20959008b432b34dbc6163bef2a6 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/4_0_5_probe_0_[864_1200_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/5_0_0_probe_0_[864_1184_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/5_0_0_probe_0_[864_1184_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..5766ec6b06cf95a1626da7cc13d1c62d246da8c1 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/5_0_0_probe_0_[864_1184_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/5_0_3_probe_0_[864_1184_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/5_0_3_probe_0_[864_1184_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..673170e87c51f464e3e71f3cccf6782b868dd835 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/5_0_3_probe_0_[864_1184_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/5_0_4_probe_0_[864_1184_1].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/5_0_4_probe_0_[864_1184_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..f910508b46d333b2e633bcfd8fdb68c9955b990f Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/5_0_4_probe_0_[864_1184_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/6_0_1_probe_0_[768_1328_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/6_0_1_probe_0_[768_1328_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..ab886d5f496833002589a4a9e12d3ae1f4fac9cf Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/6_0_1_probe_0_[768_1328_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/6_0_4_probe_0_[768_1328_1].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/6_0_4_probe_0_[768_1328_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..c75ceeea3064222b1427994e6a491e9dc624c301 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/6_0_4_probe_0_[768_1328_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/7_0_0_probe_0_[752_1392_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/7_0_0_probe_0_[752_1392_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..4b21d0f314b06fb4f9fcdde26d5eecdecd8e2e57 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/7_0_0_probe_0_[752_1392_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/7_0_1_probe_0_[752_1392_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/7_0_1_probe_0_[752_1392_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..a82c446e6502debb03c74a3982d573f68787145c Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/7_0_1_probe_0_[752_1392_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/7_0_4_probe_0_[752_1392_1].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/7_0_4_probe_0_[752_1392_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..f679fa46b2a6313aefb2061f62d999dd5042ab73 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/7_0_4_probe_0_[752_1392_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/8_0_1_probe_0_[736_1392_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/8_0_1_probe_0_[736_1392_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..66b323b31075ab2eb791a2d559b1cdbe2871b319 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/8_0_1_probe_0_[736_1392_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/9_0_0_probe_0_[816_1264_3].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/9_0_0_probe_0_[816_1264_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..da3f31b195e718b124728eea1c65319fac8ffbc8 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/9_0_0_probe_0_[816_1264_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/9_0_4_probe_0_[816_1264_1].jpg b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/9_0_4_probe_0_[816_1264_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..2a78c7640fe7c630af91fafd4ed5bcec91fdd262 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/eval_probe/step-90/BaseSolver_eval_image_step_90/9_0_4_probe_0_[816_1264_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_0_probe_0_[848_1232_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_0_probe_0_[848_1232_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..82a4cdb460b9edc6d0812a2e3ab532d9efd32fdc Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_0_probe_0_[848_1232_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_1_probe_0_[848_1232_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_1_probe_0_[848_1232_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..49d3e9e1ccef428320601e418b674e5ec7692965 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_1_probe_0_[848_1232_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_2_probe_0_[848_1232_1].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_2_probe_0_[848_1232_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..540af239af51be0fa6431ad9b661d3e1a230a094 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_2_probe_0_[848_1232_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_4_probe_0_[848_1232_1].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_4_probe_0_[848_1232_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..540af239af51be0fa6431ad9b661d3e1a230a094 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_4_probe_0_[848_1232_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0_probe.html b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0_probe.html new file mode 100644 index 0000000000000000000000000000000000000000..daaa1a3c49f1a3c828c5dc1f2795419f22ce5838 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0_probe.html @@ -0,0 +1,4 @@ + + +

+

0-0|edit_image0;

0-1|modify_image0;

0-2|edit_mask0;

0-3|target_image;

0-4|target_mask;

0-5

0-5

\ No newline at end of file diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/meta.json b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/meta.json new file mode 100644 index 0000000000000000000000000000000000000000..2563546a0970efc7fd19d73910ad19b93a07acdc --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-0/meta.json @@ -0,0 +1 @@ +{"BaseSolver_prompt": ["The Woman wears this outfit consisting of a pair of denim shorts. The shorts are a light blue color and have a high-waisted fit. They have a button and zipper closure at the waist and two pockets on the front. The front of the shorts has a frayed hemline with tassels hanging down from the sides. The denim appears to be slightly faded and has a slightly distressed look."], "BaseSolver_batch_size": 1, "BaseSolver_train_image": {"ori_file": [["./examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_0_probe_0_[848_1232_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_1_probe_0_[848_1232_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_2_probe_0_[848_1232_1].jpg", "./examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_3_probe_0_[848_1232_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_4_probe_0_[848_1232_1].jpg", "./examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0/0_0_5_probe_0_[848_1232_3].jpg"]], "html": "./examples/exp_example/20250704031317/train_probe/step-0/BaseSolver_train_image_step_0_probe.html"}, "BaseSolver_train_label": [["edit_image0; ", "modify_image0; ", "edit_mask0; ", "target_image; ", "target_mask; ", "The Woman wears this outfit consisting of a pair of denim shorts. The shorts are a light blue color and have a high-waisted fit. They have a button and zipper closure at the waist and two pockets on the front. The front of the shorts has a frayed hemline with tassels hanging down from the sides. The denim appears to be slightly faded and has a slightly distressed look."]]} \ No newline at end of file diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10/0_0_0_probe_0_[784_1328_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10/0_0_0_probe_0_[784_1328_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..4daa77a696afbed9cfc73dd41a1e02e0704a89dc Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10/0_0_0_probe_0_[784_1328_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10/0_0_1_probe_0_[784_1328_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10/0_0_1_probe_0_[784_1328_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..1a0fa0abaac8828ceeafffc5fc9d77b1de491f92 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10/0_0_1_probe_0_[784_1328_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10/0_0_2_probe_0_[784_1328_1].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10/0_0_2_probe_0_[784_1328_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..edbb47f22e054ad135810f55c3f6655e646f542a Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10/0_0_2_probe_0_[784_1328_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10/0_0_3_probe_0_[784_1328_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10/0_0_3_probe_0_[784_1328_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..2efc0a2f2e2d3fc2228d31d50277e65655b89595 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10/0_0_3_probe_0_[784_1328_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10/0_0_4_probe_0_[784_1328_1].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10/0_0_4_probe_0_[784_1328_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..edbb47f22e054ad135810f55c3f6655e646f542a Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10/0_0_4_probe_0_[784_1328_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10/0_0_5_probe_0_[784_1328_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10/0_0_5_probe_0_[784_1328_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..29c03e1df5cf9b28e55b602a52c3dc2519eb12f4 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10/0_0_5_probe_0_[784_1328_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10_probe.html b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10_probe.html new file mode 100644 index 0000000000000000000000000000000000000000..f51cb242ed63667e3d80313a0aca509efb727dfa --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10_probe.html @@ -0,0 +1,4 @@ + + +

+

0-0|edit_image0;

0-1|modify_image0;

0-2|edit_mask0;

0-3|target_image;

0-4|target_mask;

0-5

0-5

\ No newline at end of file diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-10/meta.json b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-10/meta.json new file mode 100644 index 0000000000000000000000000000000000000000..2be1456811de391cb34f79aaef42c17de729003b --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-10/meta.json @@ -0,0 +1 @@ +{"BaseSolver_prompt": ["The Woman wears this outfit consisting of a bikini bottom with a floral print. The bikini bottom is made of a soft, stretchy fabric and has a low-rise waistband. The print is a mix of blue and white flowers with green leaves and stems. The flowers are arranged in a symmetrical pattern, with some overlapping each other. The fabric appears to be lightweight and comfortable."], "BaseSolver_batch_size": 1, "BaseSolver_train_image": {"ori_file": [["./examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10/0_0_0_probe_0_[784_1328_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10/0_0_1_probe_0_[784_1328_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10/0_0_2_probe_0_[784_1328_1].jpg", "./examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10/0_0_3_probe_0_[784_1328_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10/0_0_4_probe_0_[784_1328_1].jpg", "./examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10/0_0_5_probe_0_[784_1328_3].jpg"]], "html": "./examples/exp_example/20250704031317/train_probe/step-10/BaseSolver_train_image_step_10_probe.html"}, "BaseSolver_train_label": [["edit_image0; ", "modify_image0; ", "edit_mask0; ", "target_image; ", "target_mask; ", "The Woman wears this outfit consisting of a bikini bottom with a floral print. The bikini bottom is made of a soft, stretchy fabric and has a low-rise waistband. The print is a mix of blue and white flowers with green leaves and stems. The flowers are arranged in a symmetrical pattern, with some overlapping each other. The fabric appears to be lightweight and comfortable."]]} \ No newline at end of file diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100/0_0_0_probe_0_[736_1392_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100/0_0_0_probe_0_[736_1392_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..f5c3c9d2536c7a07df674f6268bb0b11a0cddc14 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100/0_0_0_probe_0_[736_1392_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100/0_0_1_probe_0_[736_1392_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100/0_0_1_probe_0_[736_1392_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..66b323b31075ab2eb791a2d559b1cdbe2871b319 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100/0_0_1_probe_0_[736_1392_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100/0_0_2_probe_0_[736_1392_1].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100/0_0_2_probe_0_[736_1392_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..9b3c702f6edd91308a788e8850b8664d5ead95c9 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100/0_0_2_probe_0_[736_1392_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100/0_0_3_probe_0_[736_1392_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100/0_0_3_probe_0_[736_1392_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..60db069b2b4382c33899c784c0ee55b84dab7f7a Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100/0_0_3_probe_0_[736_1392_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100/0_0_4_probe_0_[736_1392_1].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100/0_0_4_probe_0_[736_1392_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..9b3c702f6edd91308a788e8850b8664d5ead95c9 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100/0_0_4_probe_0_[736_1392_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100/0_0_5_probe_0_[736_1392_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100/0_0_5_probe_0_[736_1392_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..c70689def7fc1c3e348ae2848229ce3f22d4524b Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100/0_0_5_probe_0_[736_1392_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100_probe.html b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100_probe.html new file mode 100644 index 0000000000000000000000000000000000000000..6d77b33746b32f921b0922039720892db0e937c6 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100_probe.html @@ -0,0 +1,4 @@ + + +

+

0-0|edit_image0;

0-1|modify_image0;

0-2|edit_mask0;

0-3|target_image;

0-4|target_mask;

0-5

0-5

\ No newline at end of file diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-100/meta.json b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-100/meta.json new file mode 100644 index 0000000000000000000000000000000000000000..96ed8b639c2b3cc561ce3d7b5d1b4715d3324a48 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-100/meta.json @@ -0,0 +1 @@ +{"BaseSolver_prompt": ["The Woman wears this outfit consisting of a Sleeveless top.The sleeveless top in a light purple color. The top has a square neckline with a ruched design on the front. The ruffles on the shoulders are made of a soft, ruffled fabric and are tied in a bow at the back. The neckline is square and the top appears to be fitted at the waist."], "BaseSolver_batch_size": 1, "BaseSolver_train_image": {"ori_file": [["./examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100/0_0_0_probe_0_[736_1392_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100/0_0_1_probe_0_[736_1392_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100/0_0_2_probe_0_[736_1392_1].jpg", "./examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100/0_0_3_probe_0_[736_1392_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100/0_0_4_probe_0_[736_1392_1].jpg", "./examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100/0_0_5_probe_0_[736_1392_3].jpg"]], "html": "./examples/exp_example/20250704031317/train_probe/step-100/BaseSolver_train_image_step_100_probe.html"}, "BaseSolver_train_label": [["edit_image0; ", "modify_image0; ", "edit_mask0; ", "target_image; ", "target_mask; ", "The Woman wears this outfit consisting of a Sleeveless top.The sleeveless top in a light purple color. The top has a square neckline with a ruched design on the front. The ruffles on the shoulders are made of a soft, ruffled fabric and are tied in a bow at the back. The neckline is square and the top appears to be fitted at the waist."]]} \ No newline at end of file diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-110/BaseSolver_train_image_step_110/0_0_0_probe_0_[800_1296_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-110/BaseSolver_train_image_step_110/0_0_0_probe_0_[800_1296_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..297a809607fabeac733e164958e67223efdfad12 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-110/BaseSolver_train_image_step_110/0_0_0_probe_0_[800_1296_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-110/BaseSolver_train_image_step_110/0_0_1_probe_0_[800_1296_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-110/BaseSolver_train_image_step_110/0_0_1_probe_0_[800_1296_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..971fb100df52f86c55a4abf5f66b83019fb96054 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-110/BaseSolver_train_image_step_110/0_0_1_probe_0_[800_1296_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-110/BaseSolver_train_image_step_110/0_0_2_probe_0_[800_1296_1].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-110/BaseSolver_train_image_step_110/0_0_2_probe_0_[800_1296_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..7ad14874f8ba192a545bcafd9dc28cca052157f4 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-110/BaseSolver_train_image_step_110/0_0_2_probe_0_[800_1296_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-110/BaseSolver_train_image_step_110/0_0_3_probe_0_[800_1296_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-110/BaseSolver_train_image_step_110/0_0_3_probe_0_[800_1296_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..0a03a33ba3d4115afe44c371d1634375e989b6bd Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-110/BaseSolver_train_image_step_110/0_0_3_probe_0_[800_1296_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-110/BaseSolver_train_image_step_110/0_0_4_probe_0_[800_1296_1].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-110/BaseSolver_train_image_step_110/0_0_4_probe_0_[800_1296_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..7ad14874f8ba192a545bcafd9dc28cca052157f4 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-110/BaseSolver_train_image_step_110/0_0_4_probe_0_[800_1296_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-110/BaseSolver_train_image_step_110_probe.html b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-110/BaseSolver_train_image_step_110_probe.html new file mode 100644 index 0000000000000000000000000000000000000000..914bd0c5bc43ee31bd42fdeaef67a45a6f105781 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-110/BaseSolver_train_image_step_110_probe.html @@ -0,0 +1,4 @@ + + +

+

0-0|edit_image0;

0-1|modify_image0;

0-2|edit_mask0;

0-3|target_image;

0-4|target_mask;

0-5

0-5

\ No newline at end of file diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-110/meta.json b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-110/meta.json new file mode 100644 index 0000000000000000000000000000000000000000..de23e609a9ab15ff95684f630cad74ab3e67de05 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-110/meta.json @@ -0,0 +1 @@ +{"BaseSolver_prompt": ["The Woman wears this outfit consisting of a bikini top. The top is a triangle shape with a deep V-neckline. It is made of a lightweight fabric and has a floral print in shades of blue, green, and white. The print features a variety of flowers in different sizes and shapes, including roses, daisies, and lilies. The straps of the top are thin and adjustable, allowing for a comfortable fit. The overall design is modern and stylish."], "BaseSolver_batch_size": 1, "BaseSolver_train_image": {"ori_file": [["./examples/exp_example/20250704031317/train_probe/step-110/BaseSolver_train_image_step_110/0_0_0_probe_0_[800_1296_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-110/BaseSolver_train_image_step_110/0_0_1_probe_0_[800_1296_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-110/BaseSolver_train_image_step_110/0_0_2_probe_0_[800_1296_1].jpg", "./examples/exp_example/20250704031317/train_probe/step-110/BaseSolver_train_image_step_110/0_0_3_probe_0_[800_1296_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-110/BaseSolver_train_image_step_110/0_0_4_probe_0_[800_1296_1].jpg", "./examples/exp_example/20250704031317/train_probe/step-110/BaseSolver_train_image_step_110/0_0_5_probe_0_[800_1296_3].jpg"]], "html": "./examples/exp_example/20250704031317/train_probe/step-110/BaseSolver_train_image_step_110_probe.html"}, "BaseSolver_train_label": [["edit_image0; ", "modify_image0; ", "edit_mask0; ", "target_image; ", "target_mask; ", "The Woman wears this outfit consisting of a bikini top. The top is a triangle shape with a deep V-neckline. It is made of a lightweight fabric and has a floral print in shades of blue, green, and white. The print features a variety of flowers in different sizes and shapes, including roses, daisies, and lilies. The straps of the top are thin and adjustable, allowing for a comfortable fit. The overall design is modern and stylish."]]} \ No newline at end of file diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-20/BaseSolver_train_image_step_20/0_0_0_probe_0_[800_1280_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-20/BaseSolver_train_image_step_20/0_0_0_probe_0_[800_1280_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..fccaf60136802f5665612754d562c02088c6bb5e Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-20/BaseSolver_train_image_step_20/0_0_0_probe_0_[800_1280_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-20/BaseSolver_train_image_step_20/0_0_1_probe_0_[800_1280_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-20/BaseSolver_train_image_step_20/0_0_1_probe_0_[800_1280_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..27952ba12fd42081588bcf3f0151ca720e10c12a Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-20/BaseSolver_train_image_step_20/0_0_1_probe_0_[800_1280_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-20/BaseSolver_train_image_step_20/0_0_2_probe_0_[800_1280_1].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-20/BaseSolver_train_image_step_20/0_0_2_probe_0_[800_1280_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..42d573d9436f30133a8504815d74e829a0e10897 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-20/BaseSolver_train_image_step_20/0_0_2_probe_0_[800_1280_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-20/BaseSolver_train_image_step_20/0_0_4_probe_0_[800_1280_1].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-20/BaseSolver_train_image_step_20/0_0_4_probe_0_[800_1280_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..42d573d9436f30133a8504815d74e829a0e10897 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-20/BaseSolver_train_image_step_20/0_0_4_probe_0_[800_1280_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-20/BaseSolver_train_image_step_20/0_0_5_probe_0_[800_1280_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-20/BaseSolver_train_image_step_20/0_0_5_probe_0_[800_1280_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..6795d61375f4e5a510f2d95e00fbe6d0a0014a89 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-20/BaseSolver_train_image_step_20/0_0_5_probe_0_[800_1280_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-20/BaseSolver_train_image_step_20_probe.html b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-20/BaseSolver_train_image_step_20_probe.html new file mode 100644 index 0000000000000000000000000000000000000000..d6e9d303bada0280f8c8d2f17d6b45b2f0deebc9 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-20/BaseSolver_train_image_step_20_probe.html @@ -0,0 +1,4 @@ + + +

+

0-0|edit_image0;

0-1|modify_image0;

0-2|edit_mask0;

0-3|target_image;

0-4|target_mask;

0-5

0-5

\ No newline at end of file diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-20/meta.json b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-20/meta.json new file mode 100644 index 0000000000000000000000000000000000000000..612ae04b7225b1f28e52671c69b64665a1b79fd6 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-20/meta.json @@ -0,0 +1 @@ +{"BaseSolver_prompt": ["The Woman wears this outfit consisting of long denim shorts.The shorts are a pair of blue denim shorts. The shorts are more than the knee-length and have a high-waisted fit. They have a button and zipper closure at the waist and two pockets on the front. The fabric appears to be a medium wash denim with a slightly faded look. The waistband is adjustable with a belt loop and there are two buttons on the left side of the waistband. The legs of the shorts are slightly flared at the bottom, creating a flattering silhouette."], "BaseSolver_batch_size": 1, "BaseSolver_train_image": {"ori_file": [["./examples/exp_example/20250704031317/train_probe/step-20/BaseSolver_train_image_step_20/0_0_0_probe_0_[800_1280_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-20/BaseSolver_train_image_step_20/0_0_1_probe_0_[800_1280_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-20/BaseSolver_train_image_step_20/0_0_2_probe_0_[800_1280_1].jpg", "./examples/exp_example/20250704031317/train_probe/step-20/BaseSolver_train_image_step_20/0_0_3_probe_0_[800_1280_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-20/BaseSolver_train_image_step_20/0_0_4_probe_0_[800_1280_1].jpg", "./examples/exp_example/20250704031317/train_probe/step-20/BaseSolver_train_image_step_20/0_0_5_probe_0_[800_1280_3].jpg"]], "html": "./examples/exp_example/20250704031317/train_probe/step-20/BaseSolver_train_image_step_20_probe.html"}, "BaseSolver_train_label": [["edit_image0; ", "modify_image0; ", "edit_mask0; ", "target_image; ", "target_mask; ", "The Woman wears this outfit consisting of long denim shorts.The shorts are a pair of blue denim shorts. The shorts are more than the knee-length and have a high-waisted fit. They have a button and zipper closure at the waist and two pockets on the front. The fabric appears to be a medium wash denim with a slightly faded look. The waistband is adjustable with a belt loop and there are two buttons on the left side of the waistband. The legs of the shorts are slightly flared at the bottom, creating a flattering silhouette."]]} \ No newline at end of file diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-30/BaseSolver_train_image_step_30/0_0_0_probe_0_[736_1408_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-30/BaseSolver_train_image_step_30/0_0_0_probe_0_[736_1408_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..5a6fbda97e57ed9f8be16398fff5283ccb9cda77 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-30/BaseSolver_train_image_step_30/0_0_0_probe_0_[736_1408_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-30/BaseSolver_train_image_step_30/0_0_1_probe_0_[736_1408_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-30/BaseSolver_train_image_step_30/0_0_1_probe_0_[736_1408_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..dbf9c67f71e6c4801685f5d7d17ffc0561399727 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-30/BaseSolver_train_image_step_30/0_0_1_probe_0_[736_1408_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-30/BaseSolver_train_image_step_30/0_0_2_probe_0_[736_1408_1].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-30/BaseSolver_train_image_step_30/0_0_2_probe_0_[736_1408_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..a456077f0a3a99da6a1b18fcee6fff842d00fdf9 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-30/BaseSolver_train_image_step_30/0_0_2_probe_0_[736_1408_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-30/BaseSolver_train_image_step_30/0_0_4_probe_0_[736_1408_1].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-30/BaseSolver_train_image_step_30/0_0_4_probe_0_[736_1408_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..a456077f0a3a99da6a1b18fcee6fff842d00fdf9 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-30/BaseSolver_train_image_step_30/0_0_4_probe_0_[736_1408_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-30/BaseSolver_train_image_step_30_probe.html b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-30/BaseSolver_train_image_step_30_probe.html new file mode 100644 index 0000000000000000000000000000000000000000..12031f2bfc3d98deaeae5a6dd3922064d44378c4 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-30/BaseSolver_train_image_step_30_probe.html @@ -0,0 +1,4 @@ + + +

+

0-0|edit_image0;

0-1|modify_image0;

0-2|edit_mask0;

0-3|target_image;

0-4|target_mask;

0-5

0-5

\ No newline at end of file diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-30/meta.json b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-30/meta.json new file mode 100644 index 0000000000000000000000000000000000000000..6d1293385705963413cc4ee8f5fa9f4ad195fe54 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-30/meta.json @@ -0,0 +1 @@ +{"BaseSolver_prompt": ["The Woman wears this outfit consisting of a bikini bottom with a floral print. The bikini bottom is made of a soft, stretchy fabric and has a low-rise waistband. The print is a mix of blue and white flowers with green leaves and stems. The flowers are arranged in a symmetrical pattern, with some overlapping each other. The fabric appears to be lightweight and comfortable."], "BaseSolver_batch_size": 1, "BaseSolver_train_image": {"ori_file": [["./examples/exp_example/20250704031317/train_probe/step-30/BaseSolver_train_image_step_30/0_0_0_probe_0_[736_1408_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-30/BaseSolver_train_image_step_30/0_0_1_probe_0_[736_1408_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-30/BaseSolver_train_image_step_30/0_0_2_probe_0_[736_1408_1].jpg", "./examples/exp_example/20250704031317/train_probe/step-30/BaseSolver_train_image_step_30/0_0_3_probe_0_[736_1408_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-30/BaseSolver_train_image_step_30/0_0_4_probe_0_[736_1408_1].jpg", "./examples/exp_example/20250704031317/train_probe/step-30/BaseSolver_train_image_step_30/0_0_5_probe_0_[736_1408_3].jpg"]], "html": "./examples/exp_example/20250704031317/train_probe/step-30/BaseSolver_train_image_step_30_probe.html"}, "BaseSolver_train_label": [["edit_image0; ", "modify_image0; ", "edit_mask0; ", "target_image; ", "target_mask; ", "The Woman wears this outfit consisting of a bikini bottom with a floral print. The bikini bottom is made of a soft, stretchy fabric and has a low-rise waistband. The print is a mix of blue and white flowers with green leaves and stems. The flowers are arranged in a symmetrical pattern, with some overlapping each other. The fabric appears to be lightweight and comfortable."]]} \ No newline at end of file diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40/0_0_0_probe_0_[736_1408_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40/0_0_0_probe_0_[736_1408_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..cc02a716930005b2cbe2464b88c21209d74bfef5 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40/0_0_0_probe_0_[736_1408_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40/0_0_1_probe_0_[736_1408_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40/0_0_1_probe_0_[736_1408_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..869ba04fd696845ee761b6fb3d4d994e341f85e9 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40/0_0_1_probe_0_[736_1408_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40/0_0_2_probe_0_[736_1408_1].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40/0_0_2_probe_0_[736_1408_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..98db454f4ebbff93823c7d7324a4be29ddf3cdc3 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40/0_0_2_probe_0_[736_1408_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40/0_0_3_probe_0_[736_1408_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40/0_0_3_probe_0_[736_1408_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..f38d6b6bc8428649e1d414503090108b0d5e8106 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40/0_0_3_probe_0_[736_1408_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40/0_0_4_probe_0_[736_1408_1].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40/0_0_4_probe_0_[736_1408_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..98db454f4ebbff93823c7d7324a4be29ddf3cdc3 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40/0_0_4_probe_0_[736_1408_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40/0_0_5_probe_0_[736_1408_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40/0_0_5_probe_0_[736_1408_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..d98fce2badcb701a9586c15f93d832c95739f854 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40/0_0_5_probe_0_[736_1408_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40_probe.html b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40_probe.html new file mode 100644 index 0000000000000000000000000000000000000000..f5bb41051e4f2a36b52c8660aef8b10a2dc64c5d --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40_probe.html @@ -0,0 +1,4 @@ + + +

+

0-0|edit_image0;

0-1|modify_image0;

0-2|edit_mask0;

0-3|target_image;

0-4|target_mask;

0-5

0-5

\ No newline at end of file diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-40/meta.json b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-40/meta.json new file mode 100644 index 0000000000000000000000000000000000000000..9313517b9bedc22eab995a6c7b3259378baea7b1 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-40/meta.json @@ -0,0 +1 @@ +{"BaseSolver_prompt": ["The Woman wears this outfit consisting of a Sleeveless top.The sleeveless top in a light purple color. The top has a square neckline with a ruched design on the front. The ruffles on the shoulders are made of a soft, ruffled fabric and are tied in a bow at the back. The neckline is square and the top appears to be fitted at the waist."], "BaseSolver_batch_size": 1, "BaseSolver_train_image": {"ori_file": [["./examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40/0_0_0_probe_0_[736_1408_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40/0_0_1_probe_0_[736_1408_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40/0_0_2_probe_0_[736_1408_1].jpg", "./examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40/0_0_3_probe_0_[736_1408_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40/0_0_4_probe_0_[736_1408_1].jpg", "./examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40/0_0_5_probe_0_[736_1408_3].jpg"]], "html": "./examples/exp_example/20250704031317/train_probe/step-40/BaseSolver_train_image_step_40_probe.html"}, "BaseSolver_train_label": [["edit_image0; ", "modify_image0; ", "edit_mask0; ", "target_image; ", "target_mask; ", "The Woman wears this outfit consisting of a Sleeveless top.The sleeveless top in a light purple color. The top has a square neckline with a ruched design on the front. The ruffles on the shoulders are made of a soft, ruffled fabric and are tied in a bow at the back. The neckline is square and the top appears to be fitted at the waist."]]} \ No newline at end of file diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-50/BaseSolver_train_image_step_50/0_0_1_probe_0_[720_1424_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-50/BaseSolver_train_image_step_50/0_0_1_probe_0_[720_1424_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..9cd0f7a8e8d263142ddf213a835d844ccd100502 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-50/BaseSolver_train_image_step_50/0_0_1_probe_0_[720_1424_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-50/BaseSolver_train_image_step_50/0_0_2_probe_0_[720_1424_1].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-50/BaseSolver_train_image_step_50/0_0_2_probe_0_[720_1424_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..f2bc479be8bcd2dfef929bd1982fab01521888b4 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-50/BaseSolver_train_image_step_50/0_0_2_probe_0_[720_1424_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-50/BaseSolver_train_image_step_50/0_0_4_probe_0_[720_1424_1].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-50/BaseSolver_train_image_step_50/0_0_4_probe_0_[720_1424_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..f2bc479be8bcd2dfef929bd1982fab01521888b4 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-50/BaseSolver_train_image_step_50/0_0_4_probe_0_[720_1424_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-50/BaseSolver_train_image_step_50_probe.html b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-50/BaseSolver_train_image_step_50_probe.html new file mode 100644 index 0000000000000000000000000000000000000000..4e6775c4b110040570f9e7b64612166110250ddb --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-50/BaseSolver_train_image_step_50_probe.html @@ -0,0 +1,4 @@ + + +

+

0-0|edit_image0;

0-1|modify_image0;

0-2|edit_mask0;

0-3|target_image;

0-4|target_mask;

0-5

0-5

\ No newline at end of file diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-50/meta.json b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-50/meta.json new file mode 100644 index 0000000000000000000000000000000000000000..596d56ad70ad57357b4bff5a13af7579129181c1 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-50/meta.json @@ -0,0 +1 @@ +{"BaseSolver_prompt": ["The Woman wears this outfit consisting of long denim shorts.The shorts are a pair of blue denim shorts. The shorts are more than the knee-length and have a high-waisted fit. They have a button and zipper closure at the waist and two pockets on the front. The fabric appears to be a medium wash denim with a slightly faded look. The waistband is adjustable with a belt loop and there are two buttons on the left side of the waistband. The legs of the shorts are slightly flared at the bottom, creating a flattering silhouette."], "BaseSolver_batch_size": 1, "BaseSolver_train_image": {"ori_file": [["./examples/exp_example/20250704031317/train_probe/step-50/BaseSolver_train_image_step_50/0_0_0_probe_0_[720_1424_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-50/BaseSolver_train_image_step_50/0_0_1_probe_0_[720_1424_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-50/BaseSolver_train_image_step_50/0_0_2_probe_0_[720_1424_1].jpg", "./examples/exp_example/20250704031317/train_probe/step-50/BaseSolver_train_image_step_50/0_0_3_probe_0_[720_1424_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-50/BaseSolver_train_image_step_50/0_0_4_probe_0_[720_1424_1].jpg", "./examples/exp_example/20250704031317/train_probe/step-50/BaseSolver_train_image_step_50/0_0_5_probe_0_[720_1424_3].jpg"]], "html": "./examples/exp_example/20250704031317/train_probe/step-50/BaseSolver_train_image_step_50_probe.html"}, "BaseSolver_train_label": [["edit_image0; ", "modify_image0; ", "edit_mask0; ", "target_image; ", "target_mask; ", "The Woman wears this outfit consisting of long denim shorts.The shorts are a pair of blue denim shorts. The shorts are more than the knee-length and have a high-waisted fit. They have a button and zipper closure at the waist and two pockets on the front. The fabric appears to be a medium wash denim with a slightly faded look. The waistband is adjustable with a belt loop and there are two buttons on the left side of the waistband. The legs of the shorts are slightly flared at the bottom, creating a flattering silhouette."]]} \ No newline at end of file diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-60/BaseSolver_train_image_step_60/0_0_0_probe_0_[768_1344_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-60/BaseSolver_train_image_step_60/0_0_0_probe_0_[768_1344_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..1c05d8e2394251d03b421553a61a2d215eb40377 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-60/BaseSolver_train_image_step_60/0_0_0_probe_0_[768_1344_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-60/BaseSolver_train_image_step_60/0_0_1_probe_0_[768_1344_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-60/BaseSolver_train_image_step_60/0_0_1_probe_0_[768_1344_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..f6569d866acaa9847e33c660ec6146892efbbfe1 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-60/BaseSolver_train_image_step_60/0_0_1_probe_0_[768_1344_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-60/BaseSolver_train_image_step_60/0_0_2_probe_0_[768_1344_1].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-60/BaseSolver_train_image_step_60/0_0_2_probe_0_[768_1344_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..ab4b1e95c9383844675190ae125d35d72d131346 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-60/BaseSolver_train_image_step_60/0_0_2_probe_0_[768_1344_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-60/BaseSolver_train_image_step_60/0_0_4_probe_0_[768_1344_1].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-60/BaseSolver_train_image_step_60/0_0_4_probe_0_[768_1344_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..ab4b1e95c9383844675190ae125d35d72d131346 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-60/BaseSolver_train_image_step_60/0_0_4_probe_0_[768_1344_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-60/BaseSolver_train_image_step_60/0_0_5_probe_0_[768_1344_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-60/BaseSolver_train_image_step_60/0_0_5_probe_0_[768_1344_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..4c0bef635ff0d99037cc6b66d2cf05b9075f68d0 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-60/BaseSolver_train_image_step_60/0_0_5_probe_0_[768_1344_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-60/BaseSolver_train_image_step_60_probe.html b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-60/BaseSolver_train_image_step_60_probe.html new file mode 100644 index 0000000000000000000000000000000000000000..985570f28bcddadcfc740ed8e32f16312b8d7c03 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-60/BaseSolver_train_image_step_60_probe.html @@ -0,0 +1,4 @@ + + +

+

0-0|edit_image0;

0-1|modify_image0;

0-2|edit_mask0;

0-3|target_image;

0-4|target_mask;

0-5

0-5

\ No newline at end of file diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-60/meta.json b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-60/meta.json new file mode 100644 index 0000000000000000000000000000000000000000..b6cc01419c8c6738c3e941e80c64a225937202cc --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-60/meta.json @@ -0,0 +1 @@ +{"BaseSolver_prompt": ["The Woman wears this outfit consisting of a sleeveless top with a square neckline. The top is made of a yellow fabric with a floral print in shades of orange, pink, and blue. The print features large flowers and leaves in various shades of pink, blue, and green. The flowers are arranged in a symmetrical pattern, with the largest flower in the center and the smaller ones on either side. The neckline is square and has a ruffled detail at the top."], "BaseSolver_batch_size": 1, "BaseSolver_train_image": {"ori_file": [["./examples/exp_example/20250704031317/train_probe/step-60/BaseSolver_train_image_step_60/0_0_0_probe_0_[768_1344_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-60/BaseSolver_train_image_step_60/0_0_1_probe_0_[768_1344_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-60/BaseSolver_train_image_step_60/0_0_2_probe_0_[768_1344_1].jpg", "./examples/exp_example/20250704031317/train_probe/step-60/BaseSolver_train_image_step_60/0_0_3_probe_0_[768_1344_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-60/BaseSolver_train_image_step_60/0_0_4_probe_0_[768_1344_1].jpg", "./examples/exp_example/20250704031317/train_probe/step-60/BaseSolver_train_image_step_60/0_0_5_probe_0_[768_1344_3].jpg"]], "html": "./examples/exp_example/20250704031317/train_probe/step-60/BaseSolver_train_image_step_60_probe.html"}, "BaseSolver_train_label": [["edit_image0; ", "modify_image0; ", "edit_mask0; ", "target_image; ", "target_mask; ", "The Woman wears this outfit consisting of a sleeveless top with a square neckline. The top is made of a yellow fabric with a floral print in shades of orange, pink, and blue. The print features large flowers and leaves in various shades of pink, blue, and green. The flowers are arranged in a symmetrical pattern, with the largest flower in the center and the smaller ones on either side. The neckline is square and has a ruffled detail at the top."]]} \ No newline at end of file diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-70/BaseSolver_train_image_step_70/0_0_0_probe_0_[864_1200_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-70/BaseSolver_train_image_step_70/0_0_0_probe_0_[864_1200_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..ce9e62e6f69dbe7653d86bcf0d65963af185e4fe Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-70/BaseSolver_train_image_step_70/0_0_0_probe_0_[864_1200_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-70/BaseSolver_train_image_step_70/0_0_1_probe_0_[864_1200_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-70/BaseSolver_train_image_step_70/0_0_1_probe_0_[864_1200_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..67bf97da91819294098b51cda2115ebba44d0b3b Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-70/BaseSolver_train_image_step_70/0_0_1_probe_0_[864_1200_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-70/BaseSolver_train_image_step_70/0_0_2_probe_0_[864_1200_1].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-70/BaseSolver_train_image_step_70/0_0_2_probe_0_[864_1200_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..b251548a14915f550c49a23b4271f8f5f637b5a5 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-70/BaseSolver_train_image_step_70/0_0_2_probe_0_[864_1200_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-70/BaseSolver_train_image_step_70/0_0_4_probe_0_[864_1200_1].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-70/BaseSolver_train_image_step_70/0_0_4_probe_0_[864_1200_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..b251548a14915f550c49a23b4271f8f5f637b5a5 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-70/BaseSolver_train_image_step_70/0_0_4_probe_0_[864_1200_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-70/BaseSolver_train_image_step_70/0_0_5_probe_0_[864_1200_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-70/BaseSolver_train_image_step_70/0_0_5_probe_0_[864_1200_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..2b153800f89791e27333f2ebdabe5ebe9dc7f269 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-70/BaseSolver_train_image_step_70/0_0_5_probe_0_[864_1200_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-70/BaseSolver_train_image_step_70_probe.html b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-70/BaseSolver_train_image_step_70_probe.html new file mode 100644 index 0000000000000000000000000000000000000000..558e847d10ad8d08e71cdedafcda6ebe9977ff04 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-70/BaseSolver_train_image_step_70_probe.html @@ -0,0 +1,4 @@ + + +

+

0-0|edit_image0;

0-1|modify_image0;

0-2|edit_mask0;

0-3|target_image;

0-4|target_mask;

0-5

0-5

\ No newline at end of file diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-70/meta.json b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-70/meta.json new file mode 100644 index 0000000000000000000000000000000000000000..6bc1e6f9a58d8f1433c4257badb4fa79a9f95b2b --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-70/meta.json @@ -0,0 +1 @@ +{"BaseSolver_prompt": ["The Woman wears this outfit consisting of a pair of denim shorts. The shorts are a light blue color and have a high-waisted fit. They have a button and zipper closure at the waist and two pockets on the front. The front of the shorts has a frayed hemline with tassels hanging down from the sides. The denim appears to be slightly faded and has a slightly distressed look."], "BaseSolver_batch_size": 1, "BaseSolver_train_image": {"ori_file": [["./examples/exp_example/20250704031317/train_probe/step-70/BaseSolver_train_image_step_70/0_0_0_probe_0_[864_1200_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-70/BaseSolver_train_image_step_70/0_0_1_probe_0_[864_1200_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-70/BaseSolver_train_image_step_70/0_0_2_probe_0_[864_1200_1].jpg", "./examples/exp_example/20250704031317/train_probe/step-70/BaseSolver_train_image_step_70/0_0_3_probe_0_[864_1200_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-70/BaseSolver_train_image_step_70/0_0_4_probe_0_[864_1200_1].jpg", "./examples/exp_example/20250704031317/train_probe/step-70/BaseSolver_train_image_step_70/0_0_5_probe_0_[864_1200_3].jpg"]], "html": "./examples/exp_example/20250704031317/train_probe/step-70/BaseSolver_train_image_step_70_probe.html"}, "BaseSolver_train_label": [["edit_image0; ", "modify_image0; ", "edit_mask0; ", "target_image; ", "target_mask; ", "The Woman wears this outfit consisting of a pair of denim shorts. The shorts are a light blue color and have a high-waisted fit. They have a button and zipper closure at the waist and two pockets on the front. The front of the shorts has a frayed hemline with tassels hanging down from the sides. The denim appears to be slightly faded and has a slightly distressed look."]]} \ No newline at end of file diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-80/BaseSolver_train_image_step_80/0_0_0_probe_0_[800_1296_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-80/BaseSolver_train_image_step_80/0_0_0_probe_0_[800_1296_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..b5e60957c46b33734f83a34ac04d523dc2e03680 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-80/BaseSolver_train_image_step_80/0_0_0_probe_0_[800_1296_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-80/BaseSolver_train_image_step_80/0_0_1_probe_0_[800_1296_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-80/BaseSolver_train_image_step_80/0_0_1_probe_0_[800_1296_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..af77302ad0767e07719b47602e42e614cd5684e9 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-80/BaseSolver_train_image_step_80/0_0_1_probe_0_[800_1296_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-80/BaseSolver_train_image_step_80/0_0_2_probe_0_[800_1296_1].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-80/BaseSolver_train_image_step_80/0_0_2_probe_0_[800_1296_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..243039fdda6a955c0a7f332944bd95017bf750fd Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-80/BaseSolver_train_image_step_80/0_0_2_probe_0_[800_1296_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-80/BaseSolver_train_image_step_80/0_0_4_probe_0_[800_1296_1].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-80/BaseSolver_train_image_step_80/0_0_4_probe_0_[800_1296_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..243039fdda6a955c0a7f332944bd95017bf750fd Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-80/BaseSolver_train_image_step_80/0_0_4_probe_0_[800_1296_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-80/BaseSolver_train_image_step_80_probe.html b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-80/BaseSolver_train_image_step_80_probe.html new file mode 100644 index 0000000000000000000000000000000000000000..968e6aae99405dde16f839bd4525930a97e39dee --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-80/BaseSolver_train_image_step_80_probe.html @@ -0,0 +1,4 @@ + + +

+

0-0|edit_image0;

0-1|modify_image0;

0-2|edit_mask0;

0-3|target_image;

0-4|target_mask;

0-5

0-5

\ No newline at end of file diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-80/meta.json b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-80/meta.json new file mode 100644 index 0000000000000000000000000000000000000000..e005ee84fd1379de08637512493669c03c70b6b1 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-80/meta.json @@ -0,0 +1 @@ +{"BaseSolver_prompt": ["The Woman wears this outfit consisting of a one-piece swimsuit.This one-piece swimsuit is with a colorful and intricate design. The swimsuit has a scoop neckline and thin straps that cross over the shoulders. The body of the swimsuit is predominantly blue with a zigzag pattern in the center. The pattern is made up of various shapes and colors, including pink, yellow, green, blue, and orange. The design is intricate and detailed, with a mix of geometric shapes and patterns. The overall effect is a vibrant and eye-catching piece of swimwear."], "BaseSolver_batch_size": 1, "BaseSolver_train_image": {"ori_file": [["./examples/exp_example/20250704031317/train_probe/step-80/BaseSolver_train_image_step_80/0_0_0_probe_0_[800_1296_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-80/BaseSolver_train_image_step_80/0_0_1_probe_0_[800_1296_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-80/BaseSolver_train_image_step_80/0_0_2_probe_0_[800_1296_1].jpg", "./examples/exp_example/20250704031317/train_probe/step-80/BaseSolver_train_image_step_80/0_0_3_probe_0_[800_1296_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-80/BaseSolver_train_image_step_80/0_0_4_probe_0_[800_1296_1].jpg", "./examples/exp_example/20250704031317/train_probe/step-80/BaseSolver_train_image_step_80/0_0_5_probe_0_[800_1296_3].jpg"]], "html": "./examples/exp_example/20250704031317/train_probe/step-80/BaseSolver_train_image_step_80_probe.html"}, "BaseSolver_train_label": [["edit_image0; ", "modify_image0; ", "edit_mask0; ", "target_image; ", "target_mask; ", "The Woman wears this outfit consisting of a one-piece swimsuit.This one-piece swimsuit is with a colorful and intricate design. The swimsuit has a scoop neckline and thin straps that cross over the shoulders. The body of the swimsuit is predominantly blue with a zigzag pattern in the center. The pattern is made up of various shapes and colors, including pink, yellow, green, blue, and orange. The design is intricate and detailed, with a mix of geometric shapes and patterns. The overall effect is a vibrant and eye-catching piece of swimwear."]]} \ No newline at end of file diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-90/BaseSolver_train_image_step_90/0_0_1_probe_0_[832_1248_3].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-90/BaseSolver_train_image_step_90/0_0_1_probe_0_[832_1248_3].jpg new file mode 100644 index 0000000000000000000000000000000000000000..74912a30d78d2702924be66834c99f31376884f9 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-90/BaseSolver_train_image_step_90/0_0_1_probe_0_[832_1248_3].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-90/BaseSolver_train_image_step_90/0_0_2_probe_0_[832_1248_1].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-90/BaseSolver_train_image_step_90/0_0_2_probe_0_[832_1248_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..0e5cafac680c03aab92ac71a5762c68ec8cf82f1 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-90/BaseSolver_train_image_step_90/0_0_2_probe_0_[832_1248_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-90/BaseSolver_train_image_step_90/0_0_4_probe_0_[832_1248_1].jpg b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-90/BaseSolver_train_image_step_90/0_0_4_probe_0_[832_1248_1].jpg new file mode 100644 index 0000000000000000000000000000000000000000..0e5cafac680c03aab92ac71a5762c68ec8cf82f1 Binary files /dev/null and b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-90/BaseSolver_train_image_step_90/0_0_4_probe_0_[832_1248_1].jpg differ diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-90/BaseSolver_train_image_step_90_probe.html b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-90/BaseSolver_train_image_step_90_probe.html new file mode 100644 index 0000000000000000000000000000000000000000..f08c1f4f4f703278a4548a89b1eb023323405623 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-90/BaseSolver_train_image_step_90_probe.html @@ -0,0 +1,4 @@ + + +

+

0-0|edit_image0;

0-1|modify_image0;

0-2|edit_mask0;

0-3|target_image;

0-4|target_mask;

0-5

0-5

\ No newline at end of file diff --git a/ACE_plus/examples/exp_example/20250704031317/train_probe/step-90/meta.json b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-90/meta.json new file mode 100644 index 0000000000000000000000000000000000000000..2cb0cfe27d7c2e470d62cbb4f6ef41d4a8d30255 --- /dev/null +++ b/ACE_plus/examples/exp_example/20250704031317/train_probe/step-90/meta.json @@ -0,0 +1 @@ +{"BaseSolver_prompt": ["The Woman wears this outfit consisting of a bikini bottom.The bikini bottom is with a leopard print design. The bikini bottom is made of a soft, stretchy fabric and has a low-rise waistband. The print is a combination of black and blue leopard spots, creating a bold and eye-catching look. The fabric appears to be lightweight and comfortable."], "BaseSolver_batch_size": 1, "BaseSolver_train_image": {"ori_file": [["./examples/exp_example/20250704031317/train_probe/step-90/BaseSolver_train_image_step_90/0_0_0_probe_0_[832_1248_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-90/BaseSolver_train_image_step_90/0_0_1_probe_0_[832_1248_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-90/BaseSolver_train_image_step_90/0_0_2_probe_0_[832_1248_1].jpg", "./examples/exp_example/20250704031317/train_probe/step-90/BaseSolver_train_image_step_90/0_0_3_probe_0_[832_1248_3].jpg", "./examples/exp_example/20250704031317/train_probe/step-90/BaseSolver_train_image_step_90/0_0_4_probe_0_[832_1248_1].jpg", "./examples/exp_example/20250704031317/train_probe/step-90/BaseSolver_train_image_step_90/0_0_5_probe_0_[832_1248_3].jpg"]], "html": "./examples/exp_example/20250704031317/train_probe/step-90/BaseSolver_train_image_step_90_probe.html"}, "BaseSolver_train_label": [["edit_image0; ", "modify_image0; ", "edit_mask0; ", "target_image; ", "target_mask; ", "The Woman wears this outfit consisting of a bikini bottom.The bikini bottom is with a leopard print design. The bikini bottom is made of a soft, stretchy fabric and has a low-rise waistband. The print is a combination of black and blue leopard spots, creating a bold and eye-catching look. The fabric appears to be lightweight and comfortable."]]} \ No newline at end of file