vaskers5 commited on
Commit
7789128
·
verified ·
1 Parent(s): a43f408

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. 01_11_2025/31/models/net_g_40000.pth +3 -0
  2. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_100000_lq_up_gt.png +3 -0
  3. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_10000_lq_up_gt.png +3 -0
  4. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_105000_lq_up_gt.png +3 -0
  5. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_110000_lq_up_gt.png +3 -0
  6. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_20000_lq_up_gt.png +3 -0
  7. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_25000_lq_up_gt.png +3 -0
  8. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_40000_lq_up_gt.png +3 -0
  9. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_55000_lq_up_gt.png +3 -0
  10. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_60000_lq_up_gt.png +3 -0
  11. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_65000_lq_up_gt.png +3 -0
  12. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_75000_lq_up_gt.png +3 -0
  13. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_80000_lq_up_gt.png +3 -0
  14. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_85000_lq_up_gt.png +3 -0
  15. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_90000_lq_up_gt.png +3 -0
  16. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_002cfe2087f432e0/juggernaut-xl-v9_002cfe2087f432e0_x2_105000_lq_up_gt.png +3 -0
  17. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_002cfe2087f432e0/juggernaut-xl-v9_002cfe2087f432e0_x2_15000_lq_up_gt.png +3 -0
  18. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_002cfe2087f432e0/juggernaut-xl-v9_002cfe2087f432e0_x2_35000_lq_up_gt.png +3 -0
  19. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_002cfe2087f432e0/juggernaut-xl-v9_002cfe2087f432e0_x2_45000_lq_up_gt.png +3 -0
  20. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_002cfe2087f432e0/juggernaut-xl-v9_002cfe2087f432e0_x2_50000_lq_up_gt.png +3 -0
  21. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_002cfe2087f432e0/juggernaut-xl-v9_002cfe2087f432e0_x2_60000_lq_up_gt.png +3 -0
  22. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_002cfe2087f432e0/juggernaut-xl-v9_002cfe2087f432e0_x2_70000_lq_up_gt.png +3 -0
  23. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_002cfe2087f432e0/juggernaut-xl-v9_002cfe2087f432e0_x2_75000_lq_up_gt.png +3 -0
  24. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_002cfe2087f432e0/juggernaut-xl-v9_002cfe2087f432e0_x2_95000_lq_up_gt.png +3 -0
  25. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_00b562abdf5766d3/juggernaut-xl-v9_00b562abdf5766d3_x2_30000_lq_up_gt.png +3 -0
  26. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_00b562abdf5766d3/juggernaut-xl-v9_00b562abdf5766d3_x2_45000_lq_up_gt.png +3 -0
  27. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_00b562abdf5766d3/juggernaut-xl-v9_00b562abdf5766d3_x2_70000_lq_up_gt.png +3 -0
  28. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_00b562abdf5766d3/juggernaut-xl-v9_00b562abdf5766d3_x2_80000_lq_up_gt.png +3 -0
  29. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_00d2a568dd4afd0d/juggernaut-xl-v9_00d2a568dd4afd0d_x2_15000_lq_up_gt.png +3 -0
  30. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_00d2a568dd4afd0d/juggernaut-xl-v9_00d2a568dd4afd0d_x2_75000_lq_up_gt.png +3 -0
  31. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_00d33e05424ff8cc/juggernaut-xl-v9_00d33e05424ff8cc_x2_20000_lq_up_gt.png +3 -0
  32. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_00e65d13159b498b/juggernaut-xl-v9_00e65d13159b498b_x2_100000_lq_up_gt.png +3 -0
  33. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_00e65d13159b498b/juggernaut-xl-v9_00e65d13159b498b_x2_10000_lq_up_gt.png +3 -0
  34. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_0121a497c2b82f2b/juggernaut-xl-v9_0121a497c2b82f2b_x2_40000_lq_up_gt.png +3 -0
  35. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_0121a497c2b82f2b/juggernaut-xl-v9_0121a497c2b82f2b_x2_80000_lq_up_gt.png +3 -0
  36. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_0172351a70034840/juggernaut-xl-v9_0172351a70034840_x2_100000_lq_up_gt.png +3 -0
  37. 01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_0172351a70034840/juggernaut-xl-v9_0172351a70034840_x2_45000_lq_up_gt.png +3 -0
  38. 01_11_2025/31_archived_20251101_183720/train_31_20251101_182408.log +246 -0
  39. 01_11_2025/32/basicsr_options.yaml +220 -0
  40. 01_11_2025/32/train_32_20251101_185531.log +0 -0
  41. 01_11_2025/32_4_archived_20251101_173315/basicsr_options.yaml +220 -0
  42. 01_11_2025/32_5/basicsr_options.yaml +220 -0
  43. 01_11_2025/32_6/train_32_6_20251101_175325.log +569 -0
  44. 01_11_2025/32_7/basicsr_options.yaml +220 -0
  45. 01_11_2025/32_7/train_32_7_20251101_183702.log +571 -0
  46. 01_11_2025/32_archived_20251101_095943/basicsr_options.yaml +205 -0
  47. 01_11_2025/32_archived_20251101_095943/train_32_20251101_095839.log +548 -0
  48. 01_11_2025/32_archived_20251101_104933/basicsr_options.yaml +205 -0
  49. 01_11_2025/32_archived_20251101_104933/train_32_20251101_095943.log +548 -0
  50. 01_11_2025/32_archived_20251101_161010/basicsr_options.yaml +213 -0
01_11_2025/31/models/net_g_40000.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f998f444707892e357e3ab32099e937a7e2afadc60dc2eb0f32212ae3049e8d
3
+ size 61082991
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_100000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 72a9c43f8affe57b3cec73940da683ae1a64ad2b65ec4c806b5cc86891c47752
  • Pointer size: 132 Bytes
  • Size of remote file: 3.78 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_10000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 3a84833f4ebc1c927b9e359bd814f64d3988bb53168a52a8f9f6d4720916fab2
  • Pointer size: 132 Bytes
  • Size of remote file: 3.92 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_105000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 04619f481e17af97f238d5e19b3d2967e0ef07049bebda51dca6f593b081d5fd
  • Pointer size: 132 Bytes
  • Size of remote file: 3.78 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_110000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: ec8a57670ac1564a728118786a06cfc3375232cdbcef2c554dee2ea13dc5e5aa
  • Pointer size: 132 Bytes
  • Size of remote file: 3.78 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_20000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 512e0420661d9c627dd890e74740d3ad30023de28016ee46307f6a77a5287ac0
  • Pointer size: 132 Bytes
  • Size of remote file: 3.88 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_25000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 233d1c0f524a959ae74a3060424c7dc6820f7adde842f8127e0031800eba9da8
  • Pointer size: 132 Bytes
  • Size of remote file: 3.85 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_40000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 834e11d1538e4dd57e51f099bc2658964cd884060f9b5ca666d7210ac0d71744
  • Pointer size: 132 Bytes
  • Size of remote file: 3.83 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_55000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: fdf64074fb0571c77f701b43e125ed2e150567602ea41f858af13bf1e9a735e7
  • Pointer size: 132 Bytes
  • Size of remote file: 3.77 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_60000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 345cc04af46602886d29b42bb1a6b721b576979719a4d884f8f416c4476b7898
  • Pointer size: 132 Bytes
  • Size of remote file: 3.78 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_65000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: e66d756ea8233c94f624653cef5116aa5ffa514e3c3a033911291eca8b862ff2
  • Pointer size: 132 Bytes
  • Size of remote file: 3.77 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_75000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 778a336bed3e791db4fe01f88befa913073ba5353fd7a416b0fc1748d888f5c2
  • Pointer size: 132 Bytes
  • Size of remote file: 3.77 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_80000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 1f866707a6f7fd57c5b45b0bf517fa2e06fb02447c619d08d28ce276b4ba334f
  • Pointer size: 132 Bytes
  • Size of remote file: 3.77 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_85000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 6d23d4dd9bea7cd99ad8477ac03dae3b2af36d8dbe6939cb1323f6b6a99f04d9
  • Pointer size: 132 Bytes
  • Size of remote file: 3.78 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_001ffaceaff5f33f/juggernaut-xl-v9_001ffaceaff5f33f_x2_90000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 48a17c1e49f55a79e86d79aff8575c49f7ae660fa99c28b92bc5d67080e5a354
  • Pointer size: 132 Bytes
  • Size of remote file: 3.77 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_002cfe2087f432e0/juggernaut-xl-v9_002cfe2087f432e0_x2_105000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 9095d1f274a782ac0f2409e20f4fe68191a4b5c11f319cff2d0807eb645df1e2
  • Pointer size: 132 Bytes
  • Size of remote file: 4.16 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_002cfe2087f432e0/juggernaut-xl-v9_002cfe2087f432e0_x2_15000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 639d66b675ac569e351cfe4260ac1782bf6e6c442b93b0e12160a3749d5c5f17
  • Pointer size: 132 Bytes
  • Size of remote file: 4.16 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_002cfe2087f432e0/juggernaut-xl-v9_002cfe2087f432e0_x2_35000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 0886db25351cce9f11a43a15dbf384ad2bb676f2d94ba42914c49d9beeb15614
  • Pointer size: 132 Bytes
  • Size of remote file: 4.17 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_002cfe2087f432e0/juggernaut-xl-v9_002cfe2087f432e0_x2_45000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 7ad57b911cd28233131503a7c8d09d1bc2fcc29e071427a802093edc6c251882
  • Pointer size: 132 Bytes
  • Size of remote file: 4.17 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_002cfe2087f432e0/juggernaut-xl-v9_002cfe2087f432e0_x2_50000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 5ab28b86f3cee4c6e5e38788a7416aee62d4d0e1e154d8b1598f0d9001f49fcf
  • Pointer size: 132 Bytes
  • Size of remote file: 4.17 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_002cfe2087f432e0/juggernaut-xl-v9_002cfe2087f432e0_x2_60000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 3e38455fc9d427a7b52d94a442f4fa73a8db40949d7a19c5cfeef04c6c1fc4cf
  • Pointer size: 132 Bytes
  • Size of remote file: 4.17 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_002cfe2087f432e0/juggernaut-xl-v9_002cfe2087f432e0_x2_70000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: eb9df435f0d226a5945f5735c13bffd6da4c82c2f222bd83e0d78f80d4290455
  • Pointer size: 132 Bytes
  • Size of remote file: 4.16 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_002cfe2087f432e0/juggernaut-xl-v9_002cfe2087f432e0_x2_75000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 0b765104c74a630389404f843a6faebd8798b271a0f9b09e473df64572a35b26
  • Pointer size: 132 Bytes
  • Size of remote file: 4.16 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_002cfe2087f432e0/juggernaut-xl-v9_002cfe2087f432e0_x2_95000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 0f764302c1d8aa0b6d59089ecb7e526aa88f649668860a682fb404c94280c5d7
  • Pointer size: 132 Bytes
  • Size of remote file: 4.16 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_00b562abdf5766d3/juggernaut-xl-v9_00b562abdf5766d3_x2_30000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 67f775871a962771df004c14bc1afe610e491c69009d2bdea60b68b3d06e864a
  • Pointer size: 132 Bytes
  • Size of remote file: 4.9 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_00b562abdf5766d3/juggernaut-xl-v9_00b562abdf5766d3_x2_45000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 0446b38991ece106b5bc94358106950c5ddcd657bb265986df01985c15f550c9
  • Pointer size: 132 Bytes
  • Size of remote file: 4.9 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_00b562abdf5766d3/juggernaut-xl-v9_00b562abdf5766d3_x2_70000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: e1c39a54d0d6369b6b4215e76adc4e4f749d5c02c58f86235884b1596f1512e5
  • Pointer size: 132 Bytes
  • Size of remote file: 4.91 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_00b562abdf5766d3/juggernaut-xl-v9_00b562abdf5766d3_x2_80000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 4746bf2d824e889a3e0bebc354a1fa2f49d9fdfd86f202cf09f253a6cf02ca22
  • Pointer size: 132 Bytes
  • Size of remote file: 4.91 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_00d2a568dd4afd0d/juggernaut-xl-v9_00d2a568dd4afd0d_x2_15000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: e3a3e29e9440ac5df911ec84f0f6dc149d08fd1fc1ded047b2d2753c0adf4439
  • Pointer size: 132 Bytes
  • Size of remote file: 4.17 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_00d2a568dd4afd0d/juggernaut-xl-v9_00d2a568dd4afd0d_x2_75000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: e2711d9e932674b511d89b8e91ac29f9135dff8050f0383d4cb6c54c58a1c3ff
  • Pointer size: 132 Bytes
  • Size of remote file: 4.14 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_00d33e05424ff8cc/juggernaut-xl-v9_00d33e05424ff8cc_x2_20000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 7eeb8707e158d31f5bf67a6bcb55df87c4655f462493e7389da914f7e0ee062e
  • Pointer size: 132 Bytes
  • Size of remote file: 5.65 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_00e65d13159b498b/juggernaut-xl-v9_00e65d13159b498b_x2_100000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: ae10c884c72e608161d79a9ce1e24259114fa44ed5ffb4ceb502dde45e552f15
  • Pointer size: 132 Bytes
  • Size of remote file: 3.22 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_00e65d13159b498b/juggernaut-xl-v9_00e65d13159b498b_x2_10000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: cbc7a29aa153513e4a7bf17794681545ca6dd2d43276b5b1431682c93e9fb441
  • Pointer size: 132 Bytes
  • Size of remote file: 3.22 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_0121a497c2b82f2b/juggernaut-xl-v9_0121a497c2b82f2b_x2_40000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 79cf6694023f6502b4802b30141367795ab0f1d285e97bdfda66bc8dbb3bfea5
  • Pointer size: 132 Bytes
  • Size of remote file: 4.95 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_0121a497c2b82f2b/juggernaut-xl-v9_0121a497c2b82f2b_x2_80000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 83e37706eea2ba4c39ce7365aab63aa554176a046114dc97270517f5befe304a
  • Pointer size: 132 Bytes
  • Size of remote file: 4.94 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_0172351a70034840/juggernaut-xl-v9_0172351a70034840_x2_100000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 26543ddda99cb4275f7efd6e76d4b6297d754c44cdececb0fb52134982d86314
  • Pointer size: 132 Bytes
  • Size of remote file: 4.93 MB
01_11_2025/31/visualization/val_x2/juggernaut-xl-v9_0172351a70034840/juggernaut-xl-v9_0172351a70034840_x2_45000_lq_up_gt.png ADDED

Git LFS Details

  • SHA256: 796641a918b223dfd8da26181f685f4730ba46f02ed3934ee26d4524edb79d95
  • Pointer size: 132 Bytes
  • Size of remote file: 4.94 MB
01_11_2025/31_archived_20251101_183720/train_31_20251101_182408.log ADDED
@@ -0,0 +1,246 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-11-01 18:24:08,512 INFO:
2
+ ____ _ _____ ____
3
+ / __ ) ____ _ _____ (_)_____/ ___/ / __ \
4
+ / __ |/ __ `// ___// // ___/\__ \ / /_/ /
5
+ / /_/ // /_/ /(__ )/ // /__ ___/ // _, _/
6
+ /_____/ \__,_//____//_/ \___//____//_/ |_|
7
+ ______ __ __ __ __
8
+ / ____/____ ____ ____/ / / / __ __ _____ / /__ / /
9
+ / / __ / __ \ / __ \ / __ / / / / / / // ___// //_/ / /
10
+ / /_/ // /_/ // /_/ // /_/ / / /___/ /_/ // /__ / /< /_/
11
+ \____/ \____/ \____/ \____/ /_____/\____/ \___//_/|_| (_)
12
+
13
+ Version Information:
14
+ BasicSR: 1.4.2
15
+ PyTorch: 2.9.0+cu129
16
+ TorchVision: 0.24.0+cpu
17
+ 2025-11-01 18:24:08,512 INFO:
18
+ model_type: SwinIRLatentModelMultiHead
19
+ primary_head: x4
20
+ scale: 4
21
+ num_gpu: 3
22
+ manual_seed: 0
23
+ find_unused_parameters: False
24
+ vae_sources:[
25
+ flux_vae:[
26
+ hf_repo: wolfgangblack/flux_vae
27
+ vae_kind: kl
28
+ ]
29
+ ]
30
+ datasets:[
31
+ train:[
32
+ name: gpt4_nanobana_midjourney_recraft_CropsFluxVAE
33
+ type: MultiScaleLatentCacheDataset
34
+ scales: [128, 256, 512]
35
+ cache_dirs: ['/data/kazanplova/datasets/full_latent_upscale_dataset_train/nano_banana_crops/embeddings/flux_vae', '/data/kazanplova/datasets/full_latent_upscale_dataset_train/midjourney_full_dataset_crops_new/embeddings/flux_vae', '/data/kazanplova/datasets/full_latent_upscale_dataset_train/gpt4_crops/embeddings/flux_vae', '/data/kazanplova/datasets/full_latent_upscale_dataset_train/recraft_data_crops/embeddings/flux_vae', '/data/kazanplova/datasets/full_latent_upscale_dataset_train/mjv5_crops_fix/embeddings/flux_vae']
36
+ vae_names: ['flux_vae']
37
+ phase: train
38
+ filename_tmpl: {}
39
+ io_backend:[
40
+ type: disk
41
+ ]
42
+ scale: 4
43
+ mean: None
44
+ std: None
45
+ num_worker_per_gpu: 16
46
+ batch_size_per_gpu: 256
47
+ pin_memory: True
48
+ persistent_workers: True
49
+ ]
50
+ val:[
51
+ name: sdxk_120_1024x1024
52
+ type: MultiScaleLatentCacheDataset
53
+ scales: [256, 512, 1024]
54
+ cache_dirs: ['/data/kazanplova/datasets/latent_upscale_validation_120_samples/embeddings/flux_vae']
55
+ vae_names: ['flux_vae']
56
+ phase: val
57
+ io_backend:[
58
+ type: disk
59
+ ]
60
+ scale: 4
61
+ mean: None
62
+ std: None
63
+ batch_size_per_gpu: 16
64
+ num_worker_per_gpu: 4
65
+ pin_memory: True
66
+ ]
67
+ ]
68
+ network_g:[
69
+ type: SwinIRMultiHead
70
+ in_chans: 16
71
+ img_size: 32
72
+ window_size: 8
73
+ img_range: 1.0
74
+ depths: [6, 6, 6, 6, 6, 6]
75
+ embed_dim: 180
76
+ num_heads: [6, 6, 6, 6, 6, 6]
77
+ mlp_ratio: 2
78
+ resi_connection: 1conv
79
+ head_num_feat: 128
80
+ primary_head: x4
81
+ heads: [OrderedDict({'name': 'x2', 'scale': 2, 'out_chans': 16}), OrderedDict({'name': 'x4', 'scale': 4, 'out_chans': 16, 'primary': True})]
82
+ ]
83
+ path:[
84
+ pretrain_network_g: None
85
+ strict_load_g: False
86
+ experiments_root: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/31
87
+ models: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/31/models
88
+ training_states: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/31/training_states
89
+ log: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/31
90
+ visualization: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/31/visualization
91
+ ]
92
+ compile:[
93
+ enabled: False
94
+ mode: max-autotune
95
+ dynamic: True
96
+ fullgraph: False
97
+ backend: None
98
+ ]
99
+ train:[
100
+ ema_decay: 0.999
101
+ head_inputs:[
102
+ x2:[
103
+ lq: 256
104
+ gt: 512
105
+ ]
106
+ x4:[
107
+ lq: 128
108
+ gt: 512
109
+ ]
110
+ ]
111
+ optim_g:[
112
+ type: Adam
113
+ lr: 0.0002
114
+ weight_decay: 0
115
+ betas: [0.9, 0.995]
116
+ ]
117
+ grad_clip:[
118
+ enabled: True
119
+ generator:[
120
+ type: norm
121
+ max_norm: 0.4
122
+ norm_type: 2.0
123
+ ]
124
+ ]
125
+ scheduler:[
126
+ type: MultiStepLR
127
+ milestones: [62500, 93750, 112500]
128
+ gamma: 0.5
129
+ ]
130
+ total_steps: 125000
131
+ warmup_iter: -1
132
+ l1_latent_x2_opt:[
133
+ type: L1Loss
134
+ loss_weight: 1.0
135
+ reduction: mean
136
+ space: latent
137
+ target: x2
138
+ ]
139
+ l1_latent_x4_opt:[
140
+ type: L1Loss
141
+ loss_weight: 1.0
142
+ reduction: mean
143
+ space: latent
144
+ target: x4
145
+ ]
146
+ fft_latent_x2_opt:[
147
+ type: FFTFrequencyLoss
148
+ loss_weight: 0.1
149
+ reduction: mean
150
+ space: latent
151
+ target: x2
152
+ norm: ortho
153
+ use_log_amplitude: False
154
+ alpha: 0.0
155
+ normalize_weight: True
156
+ eps: 1e-8
157
+ ]
158
+ fft_latent_x4_opt:[
159
+ type: FFTFrequencyLoss
160
+ loss_weight: 0.1
161
+ reduction: mean
162
+ space: latent
163
+ target: x4
164
+ norm: ortho
165
+ use_log_amplitude: False
166
+ alpha: 0.0
167
+ normalize_weight: True
168
+ eps: 1e-8
169
+ ]
170
+ ]
171
+ val:[
172
+ val_freq: 5000
173
+ save_img: True
174
+ head_evals:[
175
+ x2:[
176
+ save_img: True
177
+ label: val_x2
178
+ val_sizes:[
179
+ lq: 512
180
+ gt: 1024
181
+ ]
182
+ metrics:[
183
+ l1_latent:[
184
+ type: L1Loss
185
+ space: latent
186
+ ]
187
+ pixel_psnr_pt:[
188
+ type: calculate_psnr_pt
189
+ space: pixel
190
+ crop_border: 2
191
+ test_y_channel: False
192
+ ]
193
+ ]
194
+ ]
195
+ x4:[
196
+ save_img: True
197
+ label: val_x4
198
+ val_sizes:[
199
+ lq: 256
200
+ gt: 1024
201
+ ]
202
+ metrics:[
203
+ l1_latent:[
204
+ type: L1Loss
205
+ space: latent
206
+ ]
207
+ l2_latent:[
208
+ type: MSELoss
209
+ space: latent
210
+ ]
211
+ pixel_psnr_pt:[
212
+ type: calculate_psnr_pt
213
+ space: pixel
214
+ crop_border: 2
215
+ test_y_channel: False
216
+ ]
217
+ ]
218
+ ]
219
+ ]
220
+ ]
221
+ logger:[
222
+ print_freq: 100
223
+ save_checkpoint_freq: 5000
224
+ use_tb_logger: True
225
+ wandb:[
226
+ project: Swin2SR-Latent-SR
227
+ entity: kazanplova-it-more
228
+ resume_id: None
229
+ max_val_images: 10
230
+ ]
231
+ ]
232
+ dist_params:[
233
+ backend: nccl
234
+ port: 29500
235
+ dist: True
236
+ ]
237
+ load_networks_only: False
238
+ exp_name: 31
239
+ name: 31
240
+ dist: True
241
+ rank: 0
242
+ world_size: 3
243
+ auto_resume: False
244
+ is_train: True
245
+ root_path: /data/kazanplova/latent_vae_upscale_train
246
+
01_11_2025/32/basicsr_options.yaml ADDED
@@ -0,0 +1,220 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # GENERATE TIME: Sat Nov 1 18:55:31 2025
2
+ # CMD:
3
+ # train_vae.py -opt /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/basicsr_options.yaml --launcher pytorch --local_rank 0
4
+
5
+ model_type: SwinIRLatentModelMultiHead
6
+ primary_head: x4
7
+ scale: 4
8
+ num_gpu: auto
9
+ manual_seed: 0
10
+ find_unused_parameters: false
11
+ vae_sources:
12
+ flux_vae:
13
+ hf_repo: wolfgangblack/flux_vae
14
+ vae_kind: kl
15
+ datasets:
16
+ train:
17
+ name: gpt4_nanobana_midjourney_recraft_CropsFluxVAE
18
+ type: MultiScaleLatentCacheDataset
19
+ scales:
20
+ - 128
21
+ - 256
22
+ - 512
23
+ cache_dirs:
24
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/nano_banana_crops/embeddings/flux_vae
25
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/midjourney_full_dataset_crops_new/embeddings/flux_vae
26
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/gpt4_crops/embeddings/flux_vae
27
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/recraft_data_crops/embeddings/flux_vae
28
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/mjv5_crops_fix/embeddings/flux_vae
29
+ vae_names:
30
+ - flux_vae
31
+ phase: train
32
+ filename_tmpl: '{}'
33
+ io_backend:
34
+ type: disk
35
+ scale: 4
36
+ mean: null
37
+ std: null
38
+ num_worker_per_gpu: 32
39
+ batch_size_per_gpu: 64
40
+ pin_memory: true
41
+ persistent_workers: true
42
+ val:
43
+ name: sdxk_120_1024x1024
44
+ type: MultiScaleLatentCacheDataset
45
+ scales:
46
+ - 256
47
+ - 512
48
+ - 1024
49
+ cache_dirs:
50
+ - /data/kazanplova/datasets/latent_upscale_validation_120_samples/embeddings/flux_vae
51
+ vae_names:
52
+ - flux_vae
53
+ phase: val
54
+ io_backend:
55
+ type: disk
56
+ scale: 4
57
+ mean: null
58
+ std: null
59
+ batch_size_per_gpu: 16
60
+ num_worker_per_gpu: 4
61
+ pin_memory: true
62
+ network_g:
63
+ type: SwinIRMultiHead
64
+ in_chans: 16
65
+ img_size: 32
66
+ window_size: 16
67
+ img_range: 1.0
68
+ depths:
69
+ - 6
70
+ - 6
71
+ - 6
72
+ - 6
73
+ - 6
74
+ - 6
75
+ embed_dim: 360
76
+ num_heads:
77
+ - 12
78
+ - 12
79
+ - 12
80
+ - 12
81
+ - 12
82
+ - 12
83
+ mlp_ratio: 2
84
+ resi_connection: 1conv
85
+ primary_head: x4
86
+ head_num_feat: 256
87
+ heads:
88
+ - name: x2
89
+ scale: 2
90
+ out_chans: 16
91
+ - name: x4
92
+ scale: 4
93
+ out_chans: 16
94
+ primary: true
95
+ compile:
96
+ enabled: false
97
+ mode: max-autotune
98
+ dynamic: true
99
+ fullgraph: false
100
+ backend: null
101
+ train:
102
+ ema_decay: 0.999
103
+ head_inputs:
104
+ x2:
105
+ lq: 256
106
+ gt: 512
107
+ x4:
108
+ lq: 128
109
+ gt: 512
110
+ optim_g:
111
+ type: Adam
112
+ lr: 0.0002
113
+ weight_decay: 0
114
+ betas:
115
+ - 0.9
116
+ - 0.995
117
+ grad_clip:
118
+ enabled: true
119
+ generator:
120
+ type: norm
121
+ max_norm: 0.4
122
+ norm_type: 2.0
123
+ scheduler:
124
+ type: MultiStepLR
125
+ milestones:
126
+ - 62500
127
+ - 93750
128
+ - 112500
129
+ gamma: 0.5
130
+ total_steps: 125000
131
+ warmup_iter: -1
132
+ l1_latent_x2_opt:
133
+ type: L1Loss
134
+ loss_weight: 1.0
135
+ reduction: mean
136
+ space: latent
137
+ target: x2
138
+ l1_latent_x4_opt:
139
+ type: L1Loss
140
+ loss_weight: 1.0
141
+ reduction: mean
142
+ space: latent
143
+ target: x4
144
+ fft_latent_x2_opt:
145
+ type: FFTFrequencyLoss
146
+ loss_weight: 0.1
147
+ reduction: mean
148
+ space: latent
149
+ target: x2
150
+ norm: ortho
151
+ use_log_amplitude: false
152
+ alpha: 0.0
153
+ normalize_weight: true
154
+ eps: 1e-8
155
+ fft_latent_x4_opt:
156
+ type: FFTFrequencyLoss
157
+ loss_weight: 0.1
158
+ reduction: mean
159
+ space: latent
160
+ target: x4
161
+ norm: ortho
162
+ use_log_amplitude: false
163
+ alpha: 0.0
164
+ normalize_weight: true
165
+ eps: 1e-8
166
+ val:
167
+ val_freq: 2500
168
+ save_img: true
169
+ head_evals:
170
+ x2:
171
+ save_img: true
172
+ label: val_x2
173
+ val_sizes:
174
+ lq: 512
175
+ gt: 1024
176
+ metrics:
177
+ l1_latent:
178
+ type: L1Loss
179
+ space: latent
180
+ pixel_psnr_pt:
181
+ type: calculate_psnr_pt
182
+ space: pixel
183
+ crop_border: 2
184
+ test_y_channel: false
185
+ x4:
186
+ save_img: true
187
+ label: val_x4
188
+ val_sizes:
189
+ lq: 256
190
+ gt: 1024
191
+ metrics:
192
+ l1_latent:
193
+ type: L1Loss
194
+ space: latent
195
+ l2_latent:
196
+ type: MSELoss
197
+ space: latent
198
+ pixel_psnr_pt:
199
+ type: calculate_psnr_pt
200
+ space: pixel
201
+ crop_border: 2
202
+ test_y_channel: false
203
+ logger:
204
+ print_freq: 100
205
+ save_checkpoint_freq: 2500
206
+ use_tb_logger: true
207
+ wandb:
208
+ project: Swin2SR-Latent-SR
209
+ entity: kazanplova-it-more
210
+ resume_id: null
211
+ max_val_images: 10
212
+ dist_params:
213
+ backend: nccl
214
+ port: 29500
215
+ dist: true
216
+ load_networks_only: false
217
+ exp_name: '32'
218
+ name: '32'
219
+ path:
220
+ experiments_root: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025
01_11_2025/32/train_32_20251101_185531.log ADDED
The diff for this file is too large to render. See raw diff
 
01_11_2025/32_4_archived_20251101_173315/basicsr_options.yaml ADDED
@@ -0,0 +1,220 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # GENERATE TIME: Sat Nov 1 17:33:12 2025
2
+ # CMD:
3
+ # train_vae.py -opt /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/basicsr_options.yaml
4
+
5
+ model_type: SwinIRLatentModelMultiHead
6
+ primary_head: x4
7
+ scale: 4
8
+ num_gpu: auto
9
+ manual_seed: 0
10
+ find_unused_parameters: false
11
+ vae_sources:
12
+ flux_vae:
13
+ hf_repo: wolfgangblack/flux_vae
14
+ vae_kind: kl
15
+ datasets:
16
+ train:
17
+ name: gpt4_nanobana_midjourney_recraft_CropsFluxVAE
18
+ type: MultiScaleLatentCacheDataset
19
+ scales:
20
+ - 128
21
+ - 256
22
+ - 512
23
+ cache_dirs:
24
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/nano_banana_crops/embeddings/flux_vae
25
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/midjourney_full_dataset_crops_new/embeddings/flux_vae
26
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/gpt4_crops/embeddings/flux_vae
27
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/recraft_data_crops/embeddings/flux_vae
28
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/mjv5_crops_fix/embeddings/flux_vae
29
+ vae_names:
30
+ - flux_vae
31
+ phase: train
32
+ filename_tmpl: '{}'
33
+ io_backend:
34
+ type: disk
35
+ scale: 4
36
+ mean: null
37
+ std: null
38
+ num_worker_per_gpu: 12
39
+ batch_size_per_gpu: 48
40
+ pin_memory: true
41
+ persistent_workers: true
42
+ val:
43
+ name: sdxk_120_1024x1024
44
+ type: MultiScaleLatentCacheDataset
45
+ scales:
46
+ - 256
47
+ - 512
48
+ - 1024
49
+ cache_dirs:
50
+ - /data/kazanplova/datasets/latent_upscale_validation_120_samples/embeddings/flux_vae
51
+ vae_names:
52
+ - flux_vae
53
+ phase: val
54
+ io_backend:
55
+ type: disk
56
+ scale: 4
57
+ mean: null
58
+ std: null
59
+ batch_size_per_gpu: 16
60
+ num_worker_per_gpu: 4
61
+ pin_memory: true
62
+ network_g:
63
+ type: SwinIRMultiHead
64
+ in_chans: 16
65
+ img_size: 32
66
+ window_size: 16
67
+ img_range: 1.0
68
+ depths:
69
+ - 6
70
+ - 6
71
+ - 6
72
+ - 6
73
+ - 6
74
+ - 6
75
+ embed_dim: 360
76
+ num_heads:
77
+ - 12
78
+ - 12
79
+ - 12
80
+ - 12
81
+ - 12
82
+ - 12
83
+ mlp_ratio: 2
84
+ resi_connection: 1conv
85
+ primary_head: x4
86
+ head_num_feat: 256
87
+ heads:
88
+ - name: x2
89
+ scale: 2
90
+ out_chans: 16
91
+ - name: x4
92
+ scale: 4
93
+ out_chans: 16
94
+ primary: true
95
+ compile:
96
+ enabled: false
97
+ mode: max-autotune
98
+ dynamic: true
99
+ fullgraph: false
100
+ backend: null
101
+ train:
102
+ ema_decay: 0.999
103
+ head_inputs:
104
+ x2:
105
+ lq: 256
106
+ gt: 512
107
+ x4:
108
+ lq: 128
109
+ gt: 512
110
+ optim_g:
111
+ type: Adam
112
+ lr: 0.0002
113
+ weight_decay: 0
114
+ betas:
115
+ - 0.9
116
+ - 0.995
117
+ grad_clip:
118
+ enabled: true
119
+ generator:
120
+ type: norm
121
+ max_norm: 0.4
122
+ norm_type: 2.0
123
+ scheduler:
124
+ type: MultiStepLR
125
+ milestones:
126
+ - 62500
127
+ - 93750
128
+ - 112500
129
+ gamma: 0.5
130
+ total_steps: 125000
131
+ warmup_iter: -1
132
+ l1_latent_x2_opt:
133
+ type: L1Loss
134
+ loss_weight: 1.0
135
+ reduction: mean
136
+ space: latent
137
+ target: x2
138
+ l1_latent_x4_opt:
139
+ type: L1Loss
140
+ loss_weight: 1.0
141
+ reduction: mean
142
+ space: latent
143
+ target: x4
144
+ fft_latent_x2_opt:
145
+ type: FFTFrequencyLoss
146
+ loss_weight: 0.1
147
+ reduction: mean
148
+ space: latent
149
+ target: x2
150
+ norm: ortho
151
+ use_log_amplitude: false
152
+ alpha: 0.0
153
+ normalize_weight: true
154
+ eps: 1e-8
155
+ fft_latent_x4_opt:
156
+ type: FFTFrequencyLoss
157
+ loss_weight: 0.1
158
+ reduction: mean
159
+ space: latent
160
+ target: x4
161
+ norm: ortho
162
+ use_log_amplitude: false
163
+ alpha: 0.0
164
+ normalize_weight: true
165
+ eps: 1e-8
166
+ val:
167
+ val_freq: 5000
168
+ save_img: true
169
+ head_evals:
170
+ x2:
171
+ save_img: true
172
+ label: val_x2
173
+ val_sizes:
174
+ lq: 512
175
+ gt: 1024
176
+ metrics:
177
+ l1_latent:
178
+ type: L1Loss
179
+ space: latent
180
+ pixel_psnr_pt:
181
+ type: calculate_psnr_pt
182
+ space: pixel
183
+ crop_border: 2
184
+ test_y_channel: false
185
+ x4:
186
+ save_img: true
187
+ label: val_x4
188
+ val_sizes:
189
+ lq: 256
190
+ gt: 1024
191
+ metrics:
192
+ l1_latent:
193
+ type: L1Loss
194
+ space: latent
195
+ l2_latent:
196
+ type: MSELoss
197
+ space: latent
198
+ pixel_psnr_pt:
199
+ type: calculate_psnr_pt
200
+ space: pixel
201
+ crop_border: 2
202
+ test_y_channel: false
203
+ logger:
204
+ print_freq: 100
205
+ save_checkpoint_freq: 5000
206
+ use_tb_logger: true
207
+ wandb:
208
+ project: Swin2SR-Latent-SR
209
+ entity: kazanplova-it-more
210
+ resume_id: null
211
+ max_val_images: 10
212
+ dist_params:
213
+ backend: nccl
214
+ port: 29500
215
+ dist: true
216
+ load_networks_only: false
217
+ exp_name: '32'
218
+ name: '32_4'
219
+ path:
220
+ experiments_root: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025
01_11_2025/32_5/basicsr_options.yaml ADDED
@@ -0,0 +1,220 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # GENERATE TIME: Sat Nov 1 17:52:16 2025
2
+ # CMD:
3
+ # train_vae.py -opt /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/basicsr_options.yaml
4
+
5
+ model_type: SwinIRLatentModelMultiHead
6
+ primary_head: x4
7
+ scale: 4
8
+ num_gpu: auto
9
+ manual_seed: 0
10
+ find_unused_parameters: false
11
+ vae_sources:
12
+ flux_vae:
13
+ hf_repo: wolfgangblack/flux_vae
14
+ vae_kind: kl
15
+ datasets:
16
+ train:
17
+ name: gpt4_nanobana_midjourney_recraft_CropsFluxVAE
18
+ type: MultiScaleLatentCacheDataset
19
+ scales:
20
+ - 128
21
+ - 256
22
+ - 512
23
+ cache_dirs:
24
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/nano_banana_crops/embeddings/flux_vae
25
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/midjourney_full_dataset_crops_new/embeddings/flux_vae
26
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/gpt4_crops/embeddings/flux_vae
27
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/recraft_data_crops/embeddings/flux_vae
28
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/mjv5_crops_fix/embeddings/flux_vae
29
+ vae_names:
30
+ - flux_vae
31
+ phase: train
32
+ filename_tmpl: '{}'
33
+ io_backend:
34
+ type: disk
35
+ scale: 4
36
+ mean: null
37
+ std: null
38
+ num_worker_per_gpu: 32
39
+ batch_size_per_gpu: 128
40
+ pin_memory: true
41
+ persistent_workers: true
42
+ val:
43
+ name: sdxk_120_1024x1024
44
+ type: MultiScaleLatentCacheDataset
45
+ scales:
46
+ - 256
47
+ - 512
48
+ - 1024
49
+ cache_dirs:
50
+ - /data/kazanplova/datasets/latent_upscale_validation_120_samples/embeddings/flux_vae
51
+ vae_names:
52
+ - flux_vae
53
+ phase: val
54
+ io_backend:
55
+ type: disk
56
+ scale: 4
57
+ mean: null
58
+ std: null
59
+ batch_size_per_gpu: 16
60
+ num_worker_per_gpu: 4
61
+ pin_memory: true
62
+ network_g:
63
+ type: SwinIRMultiHead
64
+ in_chans: 16
65
+ img_size: 32
66
+ window_size: 16
67
+ img_range: 1.0
68
+ depths:
69
+ - 6
70
+ - 6
71
+ - 6
72
+ - 6
73
+ - 6
74
+ - 6
75
+ embed_dim: 360
76
+ num_heads:
77
+ - 12
78
+ - 12
79
+ - 12
80
+ - 12
81
+ - 12
82
+ - 12
83
+ mlp_ratio: 2
84
+ resi_connection: 1conv
85
+ primary_head: x4
86
+ head_num_feat: 256
87
+ heads:
88
+ - name: x2
89
+ scale: 2
90
+ out_chans: 16
91
+ - name: x4
92
+ scale: 4
93
+ out_chans: 16
94
+ primary: true
95
+ compile:
96
+ enabled: false
97
+ mode: max-autotune
98
+ dynamic: true
99
+ fullgraph: false
100
+ backend: null
101
+ train:
102
+ ema_decay: 0.999
103
+ head_inputs:
104
+ x2:
105
+ lq: 256
106
+ gt: 512
107
+ x4:
108
+ lq: 128
109
+ gt: 512
110
+ optim_g:
111
+ type: Adam
112
+ lr: 0.0002
113
+ weight_decay: 0
114
+ betas:
115
+ - 0.9
116
+ - 0.995
117
+ grad_clip:
118
+ enabled: true
119
+ generator:
120
+ type: norm
121
+ max_norm: 0.4
122
+ norm_type: 2.0
123
+ scheduler:
124
+ type: MultiStepLR
125
+ milestones:
126
+ - 62500
127
+ - 93750
128
+ - 112500
129
+ gamma: 0.5
130
+ total_steps: 125000
131
+ warmup_iter: -1
132
+ l1_latent_x2_opt:
133
+ type: L1Loss
134
+ loss_weight: 1.0
135
+ reduction: mean
136
+ space: latent
137
+ target: x2
138
+ l1_latent_x4_opt:
139
+ type: L1Loss
140
+ loss_weight: 1.0
141
+ reduction: mean
142
+ space: latent
143
+ target: x4
144
+ fft_latent_x2_opt:
145
+ type: FFTFrequencyLoss
146
+ loss_weight: 0.1
147
+ reduction: mean
148
+ space: latent
149
+ target: x2
150
+ norm: ortho
151
+ use_log_amplitude: false
152
+ alpha: 0.0
153
+ normalize_weight: true
154
+ eps: 1e-8
155
+ fft_latent_x4_opt:
156
+ type: FFTFrequencyLoss
157
+ loss_weight: 0.1
158
+ reduction: mean
159
+ space: latent
160
+ target: x4
161
+ norm: ortho
162
+ use_log_amplitude: false
163
+ alpha: 0.0
164
+ normalize_weight: true
165
+ eps: 1e-8
166
+ val:
167
+ val_freq: 5000
168
+ save_img: true
169
+ head_evals:
170
+ x2:
171
+ save_img: true
172
+ label: val_x2
173
+ val_sizes:
174
+ lq: 512
175
+ gt: 1024
176
+ metrics:
177
+ l1_latent:
178
+ type: L1Loss
179
+ space: latent
180
+ pixel_psnr_pt:
181
+ type: calculate_psnr_pt
182
+ space: pixel
183
+ crop_border: 2
184
+ test_y_channel: false
185
+ x4:
186
+ save_img: true
187
+ label: val_x4
188
+ val_sizes:
189
+ lq: 256
190
+ gt: 1024
191
+ metrics:
192
+ l1_latent:
193
+ type: L1Loss
194
+ space: latent
195
+ l2_latent:
196
+ type: MSELoss
197
+ space: latent
198
+ pixel_psnr_pt:
199
+ type: calculate_psnr_pt
200
+ space: pixel
201
+ crop_border: 2
202
+ test_y_channel: false
203
+ logger:
204
+ print_freq: 100
205
+ save_checkpoint_freq: 5000
206
+ use_tb_logger: true
207
+ wandb:
208
+ project: Swin2SR-Latent-SR
209
+ entity: kazanplova-it-more
210
+ resume_id: null
211
+ max_val_images: 10
212
+ dist_params:
213
+ backend: nccl
214
+ port: 29500
215
+ dist: true
216
+ load_networks_only: false
217
+ exp_name: '32'
218
+ name: '32_5'
219
+ path:
220
+ experiments_root: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025
01_11_2025/32_6/train_32_6_20251101_175325.log ADDED
@@ -0,0 +1,569 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-11-01 17:53:25,678 INFO:
2
+ ____ _ _____ ____
3
+ / __ ) ____ _ _____ (_)_____/ ___/ / __ \
4
+ / __ |/ __ `// ___// // ___/\__ \ / /_/ /
5
+ / /_/ // /_/ /(__ )/ // /__ ___/ // _, _/
6
+ /_____/ \__,_//____//_/ \___//____//_/ |_|
7
+ ______ __ __ __ __
8
+ / ____/____ ____ ____/ / / / __ __ _____ / /__ / /
9
+ / / __ / __ \ / __ \ / __ / / / / / / // ___// //_/ / /
10
+ / /_/ // /_/ // /_/ // /_/ / / /___/ /_/ // /__ / /< /_/
11
+ \____/ \____/ \____/ \____/ /_____/\____/ \___//_/|_| (_)
12
+
13
+ Version Information:
14
+ BasicSR: 1.4.2
15
+ PyTorch: 2.9.0+cu129
16
+ TorchVision: 0.24.0+cpu
17
+ 2025-11-01 17:53:25,679 INFO:
18
+ model_type: SwinIRLatentModelMultiHead
19
+ primary_head: x4
20
+ scale: 4
21
+ num_gpu: 1
22
+ manual_seed: 0
23
+ find_unused_parameters: False
24
+ vae_sources:[
25
+ flux_vae:[
26
+ hf_repo: wolfgangblack/flux_vae
27
+ vae_kind: kl
28
+ ]
29
+ ]
30
+ datasets:[
31
+ train:[
32
+ name: gpt4_nanobana_midjourney_recraft_CropsFluxVAE
33
+ type: MultiScaleLatentCacheDataset
34
+ scales: [128, 256, 512]
35
+ cache_dirs: ['/data/kazanplova/datasets/full_latent_upscale_dataset_train/nano_banana_crops/embeddings/flux_vae', '/data/kazanplova/datasets/full_latent_upscale_dataset_train/midjourney_full_dataset_crops_new/embeddings/flux_vae', '/data/kazanplova/datasets/full_latent_upscale_dataset_train/gpt4_crops/embeddings/flux_vae', '/data/kazanplova/datasets/full_latent_upscale_dataset_train/recraft_data_crops/embeddings/flux_vae', '/data/kazanplova/datasets/full_latent_upscale_dataset_train/mjv5_crops_fix/embeddings/flux_vae']
36
+ vae_names: ['flux_vae']
37
+ phase: train
38
+ filename_tmpl: {}
39
+ io_backend:[
40
+ type: disk
41
+ ]
42
+ scale: 4
43
+ mean: None
44
+ std: None
45
+ num_worker_per_gpu: 16
46
+ batch_size_per_gpu: 64
47
+ pin_memory: True
48
+ persistent_workers: True
49
+ ]
50
+ val:[
51
+ name: sdxk_120_1024x1024
52
+ type: MultiScaleLatentCacheDataset
53
+ scales: [256, 512, 1024]
54
+ cache_dirs: ['/data/kazanplova/datasets/latent_upscale_validation_120_samples/embeddings/flux_vae']
55
+ vae_names: ['flux_vae']
56
+ phase: val
57
+ io_backend:[
58
+ type: disk
59
+ ]
60
+ scale: 4
61
+ mean: None
62
+ std: None
63
+ batch_size_per_gpu: 16
64
+ num_worker_per_gpu: 4
65
+ pin_memory: True
66
+ ]
67
+ ]
68
+ network_g:[
69
+ type: SwinIRMultiHead
70
+ in_chans: 16
71
+ img_size: 32
72
+ window_size: 16
73
+ img_range: 1.0
74
+ depths: [6, 6, 6, 6, 6, 6]
75
+ embed_dim: 360
76
+ num_heads: [12, 12, 12, 12, 12, 12]
77
+ mlp_ratio: 2
78
+ resi_connection: 1conv
79
+ primary_head: x4
80
+ head_num_feat: 256
81
+ heads: [OrderedDict({'name': 'x2', 'scale': 2, 'out_chans': 16}), OrderedDict({'name': 'x4', 'scale': 4, 'out_chans': 16, 'primary': True})]
82
+ ]
83
+ compile:[
84
+ enabled: False
85
+ mode: max-autotune
86
+ dynamic: True
87
+ fullgraph: False
88
+ backend: None
89
+ ]
90
+ train:[
91
+ ema_decay: 0.999
92
+ head_inputs:[
93
+ x2:[
94
+ lq: 256
95
+ gt: 512
96
+ ]
97
+ x4:[
98
+ lq: 128
99
+ gt: 512
100
+ ]
101
+ ]
102
+ optim_g:[
103
+ type: Adam
104
+ lr: 0.0002
105
+ weight_decay: 0
106
+ betas: [0.9, 0.995]
107
+ ]
108
+ grad_clip:[
109
+ enabled: True
110
+ generator:[
111
+ type: norm
112
+ max_norm: 0.4
113
+ norm_type: 2.0
114
+ ]
115
+ ]
116
+ scheduler:[
117
+ type: MultiStepLR
118
+ milestones: [62500, 93750, 112500]
119
+ gamma: 0.5
120
+ ]
121
+ total_steps: 125000
122
+ warmup_iter: -1
123
+ l1_latent_x2_opt:[
124
+ type: L1Loss
125
+ loss_weight: 1.0
126
+ reduction: mean
127
+ space: latent
128
+ target: x2
129
+ ]
130
+ l1_latent_x4_opt:[
131
+ type: L1Loss
132
+ loss_weight: 1.0
133
+ reduction: mean
134
+ space: latent
135
+ target: x4
136
+ ]
137
+ fft_latent_x2_opt:[
138
+ type: FFTFrequencyLoss
139
+ loss_weight: 0.1
140
+ reduction: mean
141
+ space: latent
142
+ target: x2
143
+ norm: ortho
144
+ use_log_amplitude: False
145
+ alpha: 0.0
146
+ normalize_weight: True
147
+ eps: 1e-8
148
+ ]
149
+ fft_latent_x4_opt:[
150
+ type: FFTFrequencyLoss
151
+ loss_weight: 0.1
152
+ reduction: mean
153
+ space: latent
154
+ target: x4
155
+ norm: ortho
156
+ use_log_amplitude: False
157
+ alpha: 0.0
158
+ normalize_weight: True
159
+ eps: 1e-8
160
+ ]
161
+ ]
162
+ val:[
163
+ val_freq: 5000
164
+ save_img: True
165
+ head_evals:[
166
+ x2:[
167
+ save_img: True
168
+ label: val_x2
169
+ val_sizes:[
170
+ lq: 512
171
+ gt: 1024
172
+ ]
173
+ metrics:[
174
+ l1_latent:[
175
+ type: L1Loss
176
+ space: latent
177
+ ]
178
+ pixel_psnr_pt:[
179
+ type: calculate_psnr_pt
180
+ space: pixel
181
+ crop_border: 2
182
+ test_y_channel: False
183
+ ]
184
+ ]
185
+ ]
186
+ x4:[
187
+ save_img: True
188
+ label: val_x4
189
+ val_sizes:[
190
+ lq: 256
191
+ gt: 1024
192
+ ]
193
+ metrics:[
194
+ l1_latent:[
195
+ type: L1Loss
196
+ space: latent
197
+ ]
198
+ l2_latent:[
199
+ type: MSELoss
200
+ space: latent
201
+ ]
202
+ pixel_psnr_pt:[
203
+ type: calculate_psnr_pt
204
+ space: pixel
205
+ crop_border: 2
206
+ test_y_channel: False
207
+ ]
208
+ ]
209
+ ]
210
+ ]
211
+ ]
212
+ logger:[
213
+ print_freq: 100
214
+ save_checkpoint_freq: 5000
215
+ use_tb_logger: True
216
+ wandb:[
217
+ project: Swin2SR-Latent-SR
218
+ entity: kazanplova-it-more
219
+ resume_id: None
220
+ max_val_images: 10
221
+ ]
222
+ ]
223
+ dist_params:[
224
+ backend: nccl
225
+ port: 29500
226
+ dist: True
227
+ ]
228
+ load_networks_only: False
229
+ exp_name: 32
230
+ name: 32_6
231
+ path:[
232
+ experiments_root: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/32_6
233
+ models: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/32_6/models
234
+ training_states: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/32_6/training_states
235
+ log: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/32_6
236
+ visualization: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/32_6/visualization
237
+ ]
238
+ dist: False
239
+ rank: 0
240
+ world_size: 1
241
+ auto_resume: False
242
+ is_train: True
243
+ root_path: /data/kazanplova/latent_vae_upscale_train
244
+
245
+ 2025-11-01 17:53:27,295 INFO: Use wandb logger with id=lip43i3d; project=Swin2SR-Latent-SR.
246
+ 2025-11-01 17:53:40,017 INFO: Dataset [MultiScaleLatentCacheDataset] - gpt4_nanobana_midjourney_recraft_CropsFluxVAE is built.
247
+ 2025-11-01 17:53:40,018 INFO: Training statistics:
248
+ Number of train images: 4858507
249
+ Dataset enlarge ratio: 1
250
+ Batch size per gpu: 64
251
+ World size (gpu number): 1
252
+ Steps per epoch: 75915
253
+ Configured training steps: 125000
254
+ Approximate epochs to cover: 2.
255
+ 2025-11-01 17:53:40,023 INFO: Dataset [MultiScaleLatentCacheDataset] - sdxk_120_1024x1024 is built.
256
+ 2025-11-01 17:53:40,023 INFO: Number of val images/folders in sdxk_120_1024x1024: 153
257
+ 2025-11-01 17:53:40,629 INFO: Network [SwinIRMultiHead] is created.
258
+ 2025-11-01 17:53:40,891 INFO: Network: SwinIRMultiHead, with parameters: 54,917,584
259
+ 2025-11-01 17:53:40,891 INFO: SwinIRMultiHead(
260
+ (conv_first): Conv2d(16, 360, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
261
+ (patch_embed): PatchEmbed(
262
+ (norm): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
263
+ )
264
+ (patch_unembed): PatchUnEmbed()
265
+ (pos_drop): Dropout(p=0.0, inplace=False)
266
+ (layers): ModuleList(
267
+ (0): RSTB(
268
+ (residual_group): BasicLayer(
269
+ dim=360, input_resolution=(32, 32), depth=6
270
+ (blocks): ModuleList(
271
+ (0): SwinTransformerBlock(
272
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=0, mlp_ratio=2.0
273
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
274
+ (attn): WindowAttention(
275
+ dim=360, window_size=(16, 16), num_heads=12
276
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
277
+ (attn_drop): Dropout(p=0.0, inplace=False)
278
+ (proj): Linear(in_features=360, out_features=360, bias=True)
279
+ (proj_drop): Dropout(p=0.0, inplace=False)
280
+ (softmax): Softmax(dim=-1)
281
+ )
282
+ (drop_path): Identity()
283
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
284
+ (mlp): Mlp(
285
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
286
+ (act): GELU(approximate='none')
287
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
288
+ (drop): Dropout(p=0.0, inplace=False)
289
+ )
290
+ )
291
+ (1): SwinTransformerBlock(
292
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=8, mlp_ratio=2.0
293
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
294
+ (attn): WindowAttention(
295
+ dim=360, window_size=(16, 16), num_heads=12
296
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
297
+ (attn_drop): Dropout(p=0.0, inplace=False)
298
+ (proj): Linear(in_features=360, out_features=360, bias=True)
299
+ (proj_drop): Dropout(p=0.0, inplace=False)
300
+ (softmax): Softmax(dim=-1)
301
+ )
302
+ (drop_path): DropPath()
303
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
304
+ (mlp): Mlp(
305
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
306
+ (act): GELU(approximate='none')
307
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
308
+ (drop): Dropout(p=0.0, inplace=False)
309
+ )
310
+ )
311
+ (2): SwinTransformerBlock(
312
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=0, mlp_ratio=2.0
313
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
314
+ (attn): WindowAttention(
315
+ dim=360, window_size=(16, 16), num_heads=12
316
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
317
+ (attn_drop): Dropout(p=0.0, inplace=False)
318
+ (proj): Linear(in_features=360, out_features=360, bias=True)
319
+ (proj_drop): Dropout(p=0.0, inplace=False)
320
+ (softmax): Softmax(dim=-1)
321
+ )
322
+ (drop_path): DropPath()
323
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
324
+ (mlp): Mlp(
325
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
326
+ (act): GELU(approximate='none')
327
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
328
+ (drop): Dropout(p=0.0, inplace=False)
329
+ )
330
+ )
331
+ (3): SwinTransformerBlock(
332
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=8, mlp_ratio=2.0
333
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
334
+ (attn): WindowAttention(
335
+ dim=360, window_size=(16, 16), num_heads=12
336
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
337
+ (attn_drop): Dropout(p=0.0, inplace=False)
338
+ (proj): Linear(in_features=360, out_features=360, bias=True)
339
+ (proj_drop): Dropout(p=0.0, inplace=False)
340
+ (softmax): Softmax(dim=-1)
341
+ )
342
+ (drop_path): DropPath()
343
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
344
+ (mlp): Mlp(
345
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
346
+ (act): GELU(approximate='none')
347
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
348
+ (drop): Dropout(p=0.0, inplace=False)
349
+ )
350
+ )
351
+ (4): SwinTransformerBlock(
352
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=0, mlp_ratio=2.0
353
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
354
+ (attn): WindowAttention(
355
+ dim=360, window_size=(16, 16), num_heads=12
356
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
357
+ (attn_drop): Dropout(p=0.0, inplace=False)
358
+ (proj): Linear(in_features=360, out_features=360, bias=True)
359
+ (proj_drop): Dropout(p=0.0, inplace=False)
360
+ (softmax): Softmax(dim=-1)
361
+ )
362
+ (drop_path): DropPath()
363
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
364
+ (mlp): Mlp(
365
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
366
+ (act): GELU(approximate='none')
367
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
368
+ (drop): Dropout(p=0.0, inplace=False)
369
+ )
370
+ )
371
+ (5): SwinTransformerBlock(
372
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=8, mlp_ratio=2.0
373
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
374
+ (attn): WindowAttention(
375
+ dim=360, window_size=(16, 16), num_heads=12
376
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
377
+ (attn_drop): Dropout(p=0.0, inplace=False)
378
+ (proj): Linear(in_features=360, out_features=360, bias=True)
379
+ (proj_drop): Dropout(p=0.0, inplace=False)
380
+ (softmax): Softmax(dim=-1)
381
+ )
382
+ (drop_path): DropPath()
383
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
384
+ (mlp): Mlp(
385
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
386
+ (act): GELU(approximate='none')
387
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
388
+ (drop): Dropout(p=0.0, inplace=False)
389
+ )
390
+ )
391
+ )
392
+ )
393
+ (conv): Conv2d(360, 360, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
394
+ (patch_embed): PatchEmbed()
395
+ (patch_unembed): PatchUnEmbed()
396
+ )
397
+ (1-5): 5 x RSTB(
398
+ (residual_group): BasicLayer(
399
+ dim=360, input_resolution=(32, 32), depth=6
400
+ (blocks): ModuleList(
401
+ (0): SwinTransformerBlock(
402
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=0, mlp_ratio=2.0
403
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
404
+ (attn): WindowAttention(
405
+ dim=360, window_size=(16, 16), num_heads=12
406
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
407
+ (attn_drop): Dropout(p=0.0, inplace=False)
408
+ (proj): Linear(in_features=360, out_features=360, bias=True)
409
+ (proj_drop): Dropout(p=0.0, inplace=False)
410
+ (softmax): Softmax(dim=-1)
411
+ )
412
+ (drop_path): DropPath()
413
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
414
+ (mlp): Mlp(
415
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
416
+ (act): GELU(approximate='none')
417
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
418
+ (drop): Dropout(p=0.0, inplace=False)
419
+ )
420
+ )
421
+ (1): SwinTransformerBlock(
422
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=8, mlp_ratio=2.0
423
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
424
+ (attn): WindowAttention(
425
+ dim=360, window_size=(16, 16), num_heads=12
426
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
427
+ (attn_drop): Dropout(p=0.0, inplace=False)
428
+ (proj): Linear(in_features=360, out_features=360, bias=True)
429
+ (proj_drop): Dropout(p=0.0, inplace=False)
430
+ (softmax): Softmax(dim=-1)
431
+ )
432
+ (drop_path): DropPath()
433
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
434
+ (mlp): Mlp(
435
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
436
+ (act): GELU(approximate='none')
437
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
438
+ (drop): Dropout(p=0.0, inplace=False)
439
+ )
440
+ )
441
+ (2): SwinTransformerBlock(
442
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=0, mlp_ratio=2.0
443
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
444
+ (attn): WindowAttention(
445
+ dim=360, window_size=(16, 16), num_heads=12
446
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
447
+ (attn_drop): Dropout(p=0.0, inplace=False)
448
+ (proj): Linear(in_features=360, out_features=360, bias=True)
449
+ (proj_drop): Dropout(p=0.0, inplace=False)
450
+ (softmax): Softmax(dim=-1)
451
+ )
452
+ (drop_path): DropPath()
453
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
454
+ (mlp): Mlp(
455
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
456
+ (act): GELU(approximate='none')
457
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
458
+ (drop): Dropout(p=0.0, inplace=False)
459
+ )
460
+ )
461
+ (3): SwinTransformerBlock(
462
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=8, mlp_ratio=2.0
463
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
464
+ (attn): WindowAttention(
465
+ dim=360, window_size=(16, 16), num_heads=12
466
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
467
+ (attn_drop): Dropout(p=0.0, inplace=False)
468
+ (proj): Linear(in_features=360, out_features=360, bias=True)
469
+ (proj_drop): Dropout(p=0.0, inplace=False)
470
+ (softmax): Softmax(dim=-1)
471
+ )
472
+ (drop_path): DropPath()
473
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
474
+ (mlp): Mlp(
475
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
476
+ (act): GELU(approximate='none')
477
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
478
+ (drop): Dropout(p=0.0, inplace=False)
479
+ )
480
+ )
481
+ (4): SwinTransformerBlock(
482
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=0, mlp_ratio=2.0
483
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
484
+ (attn): WindowAttention(
485
+ dim=360, window_size=(16, 16), num_heads=12
486
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
487
+ (attn_drop): Dropout(p=0.0, inplace=False)
488
+ (proj): Linear(in_features=360, out_features=360, bias=True)
489
+ (proj_drop): Dropout(p=0.0, inplace=False)
490
+ (softmax): Softmax(dim=-1)
491
+ )
492
+ (drop_path): DropPath()
493
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
494
+ (mlp): Mlp(
495
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
496
+ (act): GELU(approximate='none')
497
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
498
+ (drop): Dropout(p=0.0, inplace=False)
499
+ )
500
+ )
501
+ (5): SwinTransformerBlock(
502
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=8, mlp_ratio=2.0
503
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
504
+ (attn): WindowAttention(
505
+ dim=360, window_size=(16, 16), num_heads=12
506
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
507
+ (attn_drop): Dropout(p=0.0, inplace=False)
508
+ (proj): Linear(in_features=360, out_features=360, bias=True)
509
+ (proj_drop): Dropout(p=0.0, inplace=False)
510
+ (softmax): Softmax(dim=-1)
511
+ )
512
+ (drop_path): DropPath()
513
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
514
+ (mlp): Mlp(
515
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
516
+ (act): GELU(approximate='none')
517
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
518
+ (drop): Dropout(p=0.0, inplace=False)
519
+ )
520
+ )
521
+ )
522
+ )
523
+ (conv): Conv2d(360, 360, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
524
+ (patch_embed): PatchEmbed()
525
+ (patch_unembed): PatchUnEmbed()
526
+ )
527
+ )
528
+ (norm): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
529
+ (conv_after_body): Conv2d(360, 360, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
530
+ (heads): ModuleDict(
531
+ (x2): _SwinIRPixelShuffleHead(
532
+ (conv_before): Sequential(
533
+ (0): Conv2d(360, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
534
+ (1): LeakyReLU(negative_slope=0.01, inplace=True)
535
+ )
536
+ (upsample): Upsample(
537
+ (0): Conv2d(256, 1024, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
538
+ (1): PixelShuffle(upscale_factor=2)
539
+ )
540
+ (conv_last): Conv2d(256, 16, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
541
+ )
542
+ (x4): _SwinIRPixelShuffleHead(
543
+ (conv_before): Sequential(
544
+ (0): Conv2d(360, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
545
+ (1): LeakyReLU(negative_slope=0.01, inplace=True)
546
+ )
547
+ (upsample): Upsample(
548
+ (0): Conv2d(256, 1024, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
549
+ (1): PixelShuffle(upscale_factor=2)
550
+ (2): Conv2d(256, 1024, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
551
+ (3): PixelShuffle(upscale_factor=2)
552
+ )
553
+ (conv_last): Conv2d(256, 16, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
554
+ )
555
+ )
556
+ )
557
+ 2025-11-01 17:53:40,894 INFO: Use EMA with decay: 0.999
558
+ 2025-11-01 17:53:41,442 INFO: Network [SwinIRMultiHead] is created.
559
+ 2025-11-01 17:53:41,509 INFO: Loss [L1Loss] is created.
560
+ 2025-11-01 17:53:41,510 INFO: Initialized l1_latent_x2_opt in latent space (w=1.0).
561
+ 2025-11-01 17:53:41,511 INFO: Loss [L1Loss] is created.
562
+ 2025-11-01 17:53:41,511 INFO: Initialized l1_latent_x4_opt in latent space (w=1.0).
563
+ 2025-11-01 17:53:41,511 INFO: Loss [FFTFrequencyLoss] is created.
564
+ 2025-11-01 17:53:41,512 INFO: Initialized fft_latent_x2_opt in latent space (w=0.1).
565
+ 2025-11-01 17:53:41,513 INFO: Loss [FFTFrequencyLoss] is created.
566
+ 2025-11-01 17:53:41,514 INFO: Initialized fft_latent_x4_opt in latent space (w=0.1).
567
+ 2025-11-01 17:53:41,516 INFO: Gradient clipping enabled (G:norm(max_norm=0.4, norm_type=2.0)).
568
+ 2025-11-01 17:53:41,517 INFO: Model [SwinIRLatentModelMultiHead] is created.
569
+ 2025-11-01 17:53:42,378 INFO: Start training from epoch: 0, step: 0
01_11_2025/32_7/basicsr_options.yaml ADDED
@@ -0,0 +1,220 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # GENERATE TIME: Sat Nov 1 18:37:02 2025
2
+ # CMD:
3
+ # train_vae.py -opt /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/basicsr_options.yaml
4
+
5
+ model_type: SwinIRLatentModelMultiHead
6
+ primary_head: x4
7
+ scale: 4
8
+ num_gpu: auto
9
+ manual_seed: 0
10
+ find_unused_parameters: false
11
+ vae_sources:
12
+ flux_vae:
13
+ hf_repo: wolfgangblack/flux_vae
14
+ vae_kind: kl
15
+ datasets:
16
+ train:
17
+ name: gpt4_nanobana_midjourney_recraft_CropsFluxVAE
18
+ type: MultiScaleLatentCacheDataset
19
+ scales:
20
+ - 128
21
+ - 256
22
+ - 512
23
+ cache_dirs:
24
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/nano_banana_crops/embeddings/flux_vae
25
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/midjourney_full_dataset_crops_new/embeddings/flux_vae
26
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/gpt4_crops/embeddings/flux_vae
27
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/recraft_data_crops/embeddings/flux_vae
28
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/mjv5_crops_fix/embeddings/flux_vae
29
+ vae_names:
30
+ - flux_vae
31
+ phase: train
32
+ filename_tmpl: '{}'
33
+ io_backend:
34
+ type: disk
35
+ scale: 4
36
+ mean: null
37
+ std: null
38
+ num_worker_per_gpu: 32
39
+ batch_size_per_gpu: 64
40
+ pin_memory: true
41
+ persistent_workers: true
42
+ val:
43
+ name: sdxk_120_1024x1024
44
+ type: MultiScaleLatentCacheDataset
45
+ scales:
46
+ - 256
47
+ - 512
48
+ - 1024
49
+ cache_dirs:
50
+ - /data/kazanplova/datasets/latent_upscale_validation_120_samples/embeddings/flux_vae
51
+ vae_names:
52
+ - flux_vae
53
+ phase: val
54
+ io_backend:
55
+ type: disk
56
+ scale: 4
57
+ mean: null
58
+ std: null
59
+ batch_size_per_gpu: 16
60
+ num_worker_per_gpu: 4
61
+ pin_memory: true
62
+ network_g:
63
+ type: SwinIRMultiHead
64
+ in_chans: 16
65
+ img_size: 32
66
+ window_size: 16
67
+ img_range: 1.0
68
+ depths:
69
+ - 6
70
+ - 6
71
+ - 6
72
+ - 6
73
+ - 6
74
+ - 6
75
+ embed_dim: 360
76
+ num_heads:
77
+ - 12
78
+ - 12
79
+ - 12
80
+ - 12
81
+ - 12
82
+ - 12
83
+ mlp_ratio: 2
84
+ resi_connection: 1conv
85
+ primary_head: x4
86
+ head_num_feat: 256
87
+ heads:
88
+ - name: x2
89
+ scale: 2
90
+ out_chans: 16
91
+ - name: x4
92
+ scale: 4
93
+ out_chans: 16
94
+ primary: true
95
+ compile:
96
+ enabled: false
97
+ mode: max-autotune
98
+ dynamic: true
99
+ fullgraph: false
100
+ backend: null
101
+ train:
102
+ ema_decay: 0.999
103
+ head_inputs:
104
+ x2:
105
+ lq: 256
106
+ gt: 512
107
+ x4:
108
+ lq: 128
109
+ gt: 512
110
+ optim_g:
111
+ type: Adam
112
+ lr: 0.0002
113
+ weight_decay: 0
114
+ betas:
115
+ - 0.9
116
+ - 0.995
117
+ grad_clip:
118
+ enabled: true
119
+ generator:
120
+ type: norm
121
+ max_norm: 0.4
122
+ norm_type: 2.0
123
+ scheduler:
124
+ type: MultiStepLR
125
+ milestones:
126
+ - 62500
127
+ - 93750
128
+ - 112500
129
+ gamma: 0.5
130
+ total_steps: 125000
131
+ warmup_iter: -1
132
+ l1_latent_x2_opt:
133
+ type: L1Loss
134
+ loss_weight: 1.0
135
+ reduction: mean
136
+ space: latent
137
+ target: x2
138
+ l1_latent_x4_opt:
139
+ type: L1Loss
140
+ loss_weight: 1.0
141
+ reduction: mean
142
+ space: latent
143
+ target: x4
144
+ fft_latent_x2_opt:
145
+ type: FFTFrequencyLoss
146
+ loss_weight: 0.1
147
+ reduction: mean
148
+ space: latent
149
+ target: x2
150
+ norm: ortho
151
+ use_log_amplitude: false
152
+ alpha: 0.0
153
+ normalize_weight: true
154
+ eps: 1e-8
155
+ fft_latent_x4_opt:
156
+ type: FFTFrequencyLoss
157
+ loss_weight: 0.1
158
+ reduction: mean
159
+ space: latent
160
+ target: x4
161
+ norm: ortho
162
+ use_log_amplitude: false
163
+ alpha: 0.0
164
+ normalize_weight: true
165
+ eps: 1e-8
166
+ val:
167
+ val_freq: 5000
168
+ save_img: true
169
+ head_evals:
170
+ x2:
171
+ save_img: true
172
+ label: val_x2
173
+ val_sizes:
174
+ lq: 512
175
+ gt: 1024
176
+ metrics:
177
+ l1_latent:
178
+ type: L1Loss
179
+ space: latent
180
+ pixel_psnr_pt:
181
+ type: calculate_psnr_pt
182
+ space: pixel
183
+ crop_border: 2
184
+ test_y_channel: false
185
+ x4:
186
+ save_img: true
187
+ label: val_x4
188
+ val_sizes:
189
+ lq: 256
190
+ gt: 1024
191
+ metrics:
192
+ l1_latent:
193
+ type: L1Loss
194
+ space: latent
195
+ l2_latent:
196
+ type: MSELoss
197
+ space: latent
198
+ pixel_psnr_pt:
199
+ type: calculate_psnr_pt
200
+ space: pixel
201
+ crop_border: 2
202
+ test_y_channel: false
203
+ logger:
204
+ print_freq: 100
205
+ save_checkpoint_freq: 5000
206
+ use_tb_logger: true
207
+ wandb:
208
+ project: Swin2SR-Latent-SR
209
+ entity: kazanplova-it-more
210
+ resume_id: null
211
+ max_val_images: 10
212
+ dist_params:
213
+ backend: nccl
214
+ port: 29500
215
+ dist: true
216
+ load_networks_only: false
217
+ exp_name: '32'
218
+ name: '32_7'
219
+ path:
220
+ experiments_root: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025
01_11_2025/32_7/train_32_7_20251101_183702.log ADDED
@@ -0,0 +1,571 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-11-01 18:37:02,704 INFO:
2
+ ____ _ _____ ____
3
+ / __ ) ____ _ _____ (_)_____/ ___/ / __ \
4
+ / __ |/ __ `// ___// // ___/\__ \ / /_/ /
5
+ / /_/ // /_/ /(__ )/ // /__ ___/ // _, _/
6
+ /_____/ \__,_//____//_/ \___//____//_/ |_|
7
+ ______ __ __ __ __
8
+ / ____/____ ____ ____/ / / / __ __ _____ / /__ / /
9
+ / / __ / __ \ / __ \ / __ / / / / / / // ___// //_/ / /
10
+ / /_/ // /_/ // /_/ // /_/ / / /___/ /_/ // /__ / /< /_/
11
+ \____/ \____/ \____/ \____/ /_____/\____/ \___//_/|_| (_)
12
+
13
+ Version Information:
14
+ BasicSR: 1.4.2
15
+ PyTorch: 2.9.0+cu129
16
+ TorchVision: 0.24.0+cpu
17
+ 2025-11-01 18:37:02,705 INFO:
18
+ model_type: SwinIRLatentModelMultiHead
19
+ primary_head: x4
20
+ scale: 4
21
+ num_gpu: 1
22
+ manual_seed: 0
23
+ find_unused_parameters: False
24
+ vae_sources:[
25
+ flux_vae:[
26
+ hf_repo: wolfgangblack/flux_vae
27
+ vae_kind: kl
28
+ ]
29
+ ]
30
+ datasets:[
31
+ train:[
32
+ name: gpt4_nanobana_midjourney_recraft_CropsFluxVAE
33
+ type: MultiScaleLatentCacheDataset
34
+ scales: [128, 256, 512]
35
+ cache_dirs: ['/data/kazanplova/datasets/full_latent_upscale_dataset_train/nano_banana_crops/embeddings/flux_vae', '/data/kazanplova/datasets/full_latent_upscale_dataset_train/midjourney_full_dataset_crops_new/embeddings/flux_vae', '/data/kazanplova/datasets/full_latent_upscale_dataset_train/gpt4_crops/embeddings/flux_vae', '/data/kazanplova/datasets/full_latent_upscale_dataset_train/recraft_data_crops/embeddings/flux_vae', '/data/kazanplova/datasets/full_latent_upscale_dataset_train/mjv5_crops_fix/embeddings/flux_vae']
36
+ vae_names: ['flux_vae']
37
+ phase: train
38
+ filename_tmpl: {}
39
+ io_backend:[
40
+ type: disk
41
+ ]
42
+ scale: 4
43
+ mean: None
44
+ std: None
45
+ num_worker_per_gpu: 32
46
+ batch_size_per_gpu: 64
47
+ pin_memory: True
48
+ persistent_workers: True
49
+ ]
50
+ val:[
51
+ name: sdxk_120_1024x1024
52
+ type: MultiScaleLatentCacheDataset
53
+ scales: [256, 512, 1024]
54
+ cache_dirs: ['/data/kazanplova/datasets/latent_upscale_validation_120_samples/embeddings/flux_vae']
55
+ vae_names: ['flux_vae']
56
+ phase: val
57
+ io_backend:[
58
+ type: disk
59
+ ]
60
+ scale: 4
61
+ mean: None
62
+ std: None
63
+ batch_size_per_gpu: 16
64
+ num_worker_per_gpu: 4
65
+ pin_memory: True
66
+ ]
67
+ ]
68
+ network_g:[
69
+ type: SwinIRMultiHead
70
+ in_chans: 16
71
+ img_size: 32
72
+ window_size: 16
73
+ img_range: 1.0
74
+ depths: [6, 6, 6, 6, 6, 6]
75
+ embed_dim: 360
76
+ num_heads: [12, 12, 12, 12, 12, 12]
77
+ mlp_ratio: 2
78
+ resi_connection: 1conv
79
+ primary_head: x4
80
+ head_num_feat: 256
81
+ heads: [OrderedDict({'name': 'x2', 'scale': 2, 'out_chans': 16}), OrderedDict({'name': 'x4', 'scale': 4, 'out_chans': 16, 'primary': True})]
82
+ ]
83
+ compile:[
84
+ enabled: False
85
+ mode: max-autotune
86
+ dynamic: True
87
+ fullgraph: False
88
+ backend: None
89
+ ]
90
+ train:[
91
+ ema_decay: 0.999
92
+ head_inputs:[
93
+ x2:[
94
+ lq: 256
95
+ gt: 512
96
+ ]
97
+ x4:[
98
+ lq: 128
99
+ gt: 512
100
+ ]
101
+ ]
102
+ optim_g:[
103
+ type: Adam
104
+ lr: 0.0002
105
+ weight_decay: 0
106
+ betas: [0.9, 0.995]
107
+ ]
108
+ grad_clip:[
109
+ enabled: True
110
+ generator:[
111
+ type: norm
112
+ max_norm: 0.4
113
+ norm_type: 2.0
114
+ ]
115
+ ]
116
+ scheduler:[
117
+ type: MultiStepLR
118
+ milestones: [62500, 93750, 112500]
119
+ gamma: 0.5
120
+ ]
121
+ total_steps: 125000
122
+ warmup_iter: -1
123
+ l1_latent_x2_opt:[
124
+ type: L1Loss
125
+ loss_weight: 1.0
126
+ reduction: mean
127
+ space: latent
128
+ target: x2
129
+ ]
130
+ l1_latent_x4_opt:[
131
+ type: L1Loss
132
+ loss_weight: 1.0
133
+ reduction: mean
134
+ space: latent
135
+ target: x4
136
+ ]
137
+ fft_latent_x2_opt:[
138
+ type: FFTFrequencyLoss
139
+ loss_weight: 0.1
140
+ reduction: mean
141
+ space: latent
142
+ target: x2
143
+ norm: ortho
144
+ use_log_amplitude: False
145
+ alpha: 0.0
146
+ normalize_weight: True
147
+ eps: 1e-8
148
+ ]
149
+ fft_latent_x4_opt:[
150
+ type: FFTFrequencyLoss
151
+ loss_weight: 0.1
152
+ reduction: mean
153
+ space: latent
154
+ target: x4
155
+ norm: ortho
156
+ use_log_amplitude: False
157
+ alpha: 0.0
158
+ normalize_weight: True
159
+ eps: 1e-8
160
+ ]
161
+ ]
162
+ val:[
163
+ val_freq: 5000
164
+ save_img: True
165
+ head_evals:[
166
+ x2:[
167
+ save_img: True
168
+ label: val_x2
169
+ val_sizes:[
170
+ lq: 512
171
+ gt: 1024
172
+ ]
173
+ metrics:[
174
+ l1_latent:[
175
+ type: L1Loss
176
+ space: latent
177
+ ]
178
+ pixel_psnr_pt:[
179
+ type: calculate_psnr_pt
180
+ space: pixel
181
+ crop_border: 2
182
+ test_y_channel: False
183
+ ]
184
+ ]
185
+ ]
186
+ x4:[
187
+ save_img: True
188
+ label: val_x4
189
+ val_sizes:[
190
+ lq: 256
191
+ gt: 1024
192
+ ]
193
+ metrics:[
194
+ l1_latent:[
195
+ type: L1Loss
196
+ space: latent
197
+ ]
198
+ l2_latent:[
199
+ type: MSELoss
200
+ space: latent
201
+ ]
202
+ pixel_psnr_pt:[
203
+ type: calculate_psnr_pt
204
+ space: pixel
205
+ crop_border: 2
206
+ test_y_channel: False
207
+ ]
208
+ ]
209
+ ]
210
+ ]
211
+ ]
212
+ logger:[
213
+ print_freq: 100
214
+ save_checkpoint_freq: 5000
215
+ use_tb_logger: True
216
+ wandb:[
217
+ project: Swin2SR-Latent-SR
218
+ entity: kazanplova-it-more
219
+ resume_id: None
220
+ max_val_images: 10
221
+ ]
222
+ ]
223
+ dist_params:[
224
+ backend: nccl
225
+ port: 29500
226
+ dist: True
227
+ ]
228
+ load_networks_only: False
229
+ exp_name: 32
230
+ name: 32_7
231
+ path:[
232
+ experiments_root: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/32_7
233
+ models: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/32_7/models
234
+ training_states: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/32_7/training_states
235
+ log: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/32_7
236
+ visualization: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/32_7/visualization
237
+ ]
238
+ dist: False
239
+ rank: 0
240
+ world_size: 1
241
+ auto_resume: False
242
+ is_train: True
243
+ root_path: /data/kazanplova/latent_vae_upscale_train
244
+
245
+ 2025-11-01 18:37:04,406 INFO: Use wandb logger with id=1x0wne5b; project=Swin2SR-Latent-SR.
246
+ 2025-11-01 18:37:16,610 INFO: Dataset [MultiScaleLatentCacheDataset] - gpt4_nanobana_midjourney_recraft_CropsFluxVAE is built.
247
+ 2025-11-01 18:37:16,611 INFO: Training statistics:
248
+ Number of train images: 4858507
249
+ Dataset enlarge ratio: 1
250
+ Batch size per gpu: 64
251
+ World size (gpu number): 1
252
+ Steps per epoch: 75915
253
+ Configured training steps: 125000
254
+ Approximate epochs to cover: 2.
255
+ 2025-11-01 18:37:16,614 INFO: Dataset [MultiScaleLatentCacheDataset] - sdxk_120_1024x1024 is built.
256
+ 2025-11-01 18:37:16,614 INFO: Number of val images/folders in sdxk_120_1024x1024: 153
257
+ 2025-11-01 18:37:16,615 INFO: Enabled find_unused_parameters=True for multi-head training overrides.
258
+ 2025-11-01 18:37:17,515 INFO: Network [SwinIRMultiHead] is created.
259
+ 2025-11-01 18:37:17,724 INFO: Network: SwinIRMultiHead, with parameters: 54,917,584
260
+ 2025-11-01 18:37:17,725 INFO: SwinIRMultiHead(
261
+ (conv_first): Conv2d(16, 360, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
262
+ (patch_embed): PatchEmbed(
263
+ (norm): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
264
+ )
265
+ (patch_unembed): PatchUnEmbed()
266
+ (pos_drop): Dropout(p=0.0, inplace=False)
267
+ (layers): ModuleList(
268
+ (0): RSTB(
269
+ (residual_group): BasicLayer(
270
+ dim=360, input_resolution=(32, 32), depth=6
271
+ (blocks): ModuleList(
272
+ (0): SwinTransformerBlock(
273
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=0, mlp_ratio=2.0
274
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
275
+ (attn): WindowAttention(
276
+ dim=360, window_size=(16, 16), num_heads=12
277
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
278
+ (attn_drop): Dropout(p=0.0, inplace=False)
279
+ (proj): Linear(in_features=360, out_features=360, bias=True)
280
+ (proj_drop): Dropout(p=0.0, inplace=False)
281
+ (softmax): Softmax(dim=-1)
282
+ )
283
+ (drop_path): Identity()
284
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
285
+ (mlp): Mlp(
286
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
287
+ (act): GELU(approximate='none')
288
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
289
+ (drop): Dropout(p=0.0, inplace=False)
290
+ )
291
+ )
292
+ (1): SwinTransformerBlock(
293
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=8, mlp_ratio=2.0
294
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
295
+ (attn): WindowAttention(
296
+ dim=360, window_size=(16, 16), num_heads=12
297
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
298
+ (attn_drop): Dropout(p=0.0, inplace=False)
299
+ (proj): Linear(in_features=360, out_features=360, bias=True)
300
+ (proj_drop): Dropout(p=0.0, inplace=False)
301
+ (softmax): Softmax(dim=-1)
302
+ )
303
+ (drop_path): DropPath()
304
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
305
+ (mlp): Mlp(
306
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
307
+ (act): GELU(approximate='none')
308
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
309
+ (drop): Dropout(p=0.0, inplace=False)
310
+ )
311
+ )
312
+ (2): SwinTransformerBlock(
313
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=0, mlp_ratio=2.0
314
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
315
+ (attn): WindowAttention(
316
+ dim=360, window_size=(16, 16), num_heads=12
317
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
318
+ (attn_drop): Dropout(p=0.0, inplace=False)
319
+ (proj): Linear(in_features=360, out_features=360, bias=True)
320
+ (proj_drop): Dropout(p=0.0, inplace=False)
321
+ (softmax): Softmax(dim=-1)
322
+ )
323
+ (drop_path): DropPath()
324
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
325
+ (mlp): Mlp(
326
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
327
+ (act): GELU(approximate='none')
328
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
329
+ (drop): Dropout(p=0.0, inplace=False)
330
+ )
331
+ )
332
+ (3): SwinTransformerBlock(
333
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=8, mlp_ratio=2.0
334
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
335
+ (attn): WindowAttention(
336
+ dim=360, window_size=(16, 16), num_heads=12
337
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
338
+ (attn_drop): Dropout(p=0.0, inplace=False)
339
+ (proj): Linear(in_features=360, out_features=360, bias=True)
340
+ (proj_drop): Dropout(p=0.0, inplace=False)
341
+ (softmax): Softmax(dim=-1)
342
+ )
343
+ (drop_path): DropPath()
344
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
345
+ (mlp): Mlp(
346
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
347
+ (act): GELU(approximate='none')
348
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
349
+ (drop): Dropout(p=0.0, inplace=False)
350
+ )
351
+ )
352
+ (4): SwinTransformerBlock(
353
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=0, mlp_ratio=2.0
354
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
355
+ (attn): WindowAttention(
356
+ dim=360, window_size=(16, 16), num_heads=12
357
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
358
+ (attn_drop): Dropout(p=0.0, inplace=False)
359
+ (proj): Linear(in_features=360, out_features=360, bias=True)
360
+ (proj_drop): Dropout(p=0.0, inplace=False)
361
+ (softmax): Softmax(dim=-1)
362
+ )
363
+ (drop_path): DropPath()
364
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
365
+ (mlp): Mlp(
366
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
367
+ (act): GELU(approximate='none')
368
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
369
+ (drop): Dropout(p=0.0, inplace=False)
370
+ )
371
+ )
372
+ (5): SwinTransformerBlock(
373
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=8, mlp_ratio=2.0
374
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
375
+ (attn): WindowAttention(
376
+ dim=360, window_size=(16, 16), num_heads=12
377
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
378
+ (attn_drop): Dropout(p=0.0, inplace=False)
379
+ (proj): Linear(in_features=360, out_features=360, bias=True)
380
+ (proj_drop): Dropout(p=0.0, inplace=False)
381
+ (softmax): Softmax(dim=-1)
382
+ )
383
+ (drop_path): DropPath()
384
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
385
+ (mlp): Mlp(
386
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
387
+ (act): GELU(approximate='none')
388
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
389
+ (drop): Dropout(p=0.0, inplace=False)
390
+ )
391
+ )
392
+ )
393
+ )
394
+ (conv): Conv2d(360, 360, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
395
+ (patch_embed): PatchEmbed()
396
+ (patch_unembed): PatchUnEmbed()
397
+ )
398
+ (1-5): 5 x RSTB(
399
+ (residual_group): BasicLayer(
400
+ dim=360, input_resolution=(32, 32), depth=6
401
+ (blocks): ModuleList(
402
+ (0): SwinTransformerBlock(
403
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=0, mlp_ratio=2.0
404
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
405
+ (attn): WindowAttention(
406
+ dim=360, window_size=(16, 16), num_heads=12
407
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
408
+ (attn_drop): Dropout(p=0.0, inplace=False)
409
+ (proj): Linear(in_features=360, out_features=360, bias=True)
410
+ (proj_drop): Dropout(p=0.0, inplace=False)
411
+ (softmax): Softmax(dim=-1)
412
+ )
413
+ (drop_path): DropPath()
414
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
415
+ (mlp): Mlp(
416
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
417
+ (act): GELU(approximate='none')
418
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
419
+ (drop): Dropout(p=0.0, inplace=False)
420
+ )
421
+ )
422
+ (1): SwinTransformerBlock(
423
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=8, mlp_ratio=2.0
424
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
425
+ (attn): WindowAttention(
426
+ dim=360, window_size=(16, 16), num_heads=12
427
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
428
+ (attn_drop): Dropout(p=0.0, inplace=False)
429
+ (proj): Linear(in_features=360, out_features=360, bias=True)
430
+ (proj_drop): Dropout(p=0.0, inplace=False)
431
+ (softmax): Softmax(dim=-1)
432
+ )
433
+ (drop_path): DropPath()
434
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
435
+ (mlp): Mlp(
436
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
437
+ (act): GELU(approximate='none')
438
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
439
+ (drop): Dropout(p=0.0, inplace=False)
440
+ )
441
+ )
442
+ (2): SwinTransformerBlock(
443
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=0, mlp_ratio=2.0
444
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
445
+ (attn): WindowAttention(
446
+ dim=360, window_size=(16, 16), num_heads=12
447
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
448
+ (attn_drop): Dropout(p=0.0, inplace=False)
449
+ (proj): Linear(in_features=360, out_features=360, bias=True)
450
+ (proj_drop): Dropout(p=0.0, inplace=False)
451
+ (softmax): Softmax(dim=-1)
452
+ )
453
+ (drop_path): DropPath()
454
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
455
+ (mlp): Mlp(
456
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
457
+ (act): GELU(approximate='none')
458
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
459
+ (drop): Dropout(p=0.0, inplace=False)
460
+ )
461
+ )
462
+ (3): SwinTransformerBlock(
463
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=8, mlp_ratio=2.0
464
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
465
+ (attn): WindowAttention(
466
+ dim=360, window_size=(16, 16), num_heads=12
467
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
468
+ (attn_drop): Dropout(p=0.0, inplace=False)
469
+ (proj): Linear(in_features=360, out_features=360, bias=True)
470
+ (proj_drop): Dropout(p=0.0, inplace=False)
471
+ (softmax): Softmax(dim=-1)
472
+ )
473
+ (drop_path): DropPath()
474
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
475
+ (mlp): Mlp(
476
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
477
+ (act): GELU(approximate='none')
478
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
479
+ (drop): Dropout(p=0.0, inplace=False)
480
+ )
481
+ )
482
+ (4): SwinTransformerBlock(
483
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=0, mlp_ratio=2.0
484
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
485
+ (attn): WindowAttention(
486
+ dim=360, window_size=(16, 16), num_heads=12
487
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
488
+ (attn_drop): Dropout(p=0.0, inplace=False)
489
+ (proj): Linear(in_features=360, out_features=360, bias=True)
490
+ (proj_drop): Dropout(p=0.0, inplace=False)
491
+ (softmax): Softmax(dim=-1)
492
+ )
493
+ (drop_path): DropPath()
494
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
495
+ (mlp): Mlp(
496
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
497
+ (act): GELU(approximate='none')
498
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
499
+ (drop): Dropout(p=0.0, inplace=False)
500
+ )
501
+ )
502
+ (5): SwinTransformerBlock(
503
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=8, mlp_ratio=2.0
504
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
505
+ (attn): WindowAttention(
506
+ dim=360, window_size=(16, 16), num_heads=12
507
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
508
+ (attn_drop): Dropout(p=0.0, inplace=False)
509
+ (proj): Linear(in_features=360, out_features=360, bias=True)
510
+ (proj_drop): Dropout(p=0.0, inplace=False)
511
+ (softmax): Softmax(dim=-1)
512
+ )
513
+ (drop_path): DropPath()
514
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
515
+ (mlp): Mlp(
516
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
517
+ (act): GELU(approximate='none')
518
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
519
+ (drop): Dropout(p=0.0, inplace=False)
520
+ )
521
+ )
522
+ )
523
+ )
524
+ (conv): Conv2d(360, 360, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
525
+ (patch_embed): PatchEmbed()
526
+ (patch_unembed): PatchUnEmbed()
527
+ )
528
+ )
529
+ (norm): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
530
+ (conv_after_body): Conv2d(360, 360, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
531
+ (heads): ModuleDict(
532
+ (x2): _SwinIRPixelShuffleHead(
533
+ (conv_before): Sequential(
534
+ (0): Conv2d(360, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
535
+ (1): LeakyReLU(negative_slope=0.01, inplace=True)
536
+ )
537
+ (upsample): Upsample(
538
+ (0): Conv2d(256, 1024, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
539
+ (1): PixelShuffle(upscale_factor=2)
540
+ )
541
+ (conv_last): Conv2d(256, 16, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
542
+ )
543
+ (x4): _SwinIRPixelShuffleHead(
544
+ (conv_before): Sequential(
545
+ (0): Conv2d(360, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
546
+ (1): LeakyReLU(negative_slope=0.01, inplace=True)
547
+ )
548
+ (upsample): Upsample(
549
+ (0): Conv2d(256, 1024, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
550
+ (1): PixelShuffle(upscale_factor=2)
551
+ (2): Conv2d(256, 1024, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
552
+ (3): PixelShuffle(upscale_factor=2)
553
+ )
554
+ (conv_last): Conv2d(256, 16, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
555
+ )
556
+ )
557
+ )
558
+ 2025-11-01 18:37:17,728 INFO: Use EMA with decay: 0.999
559
+ 2025-11-01 18:37:18,275 INFO: Network [SwinIRMultiHead] is created.
560
+ 2025-11-01 18:37:18,341 INFO: Loss [L1Loss] is created.
561
+ 2025-11-01 18:37:18,342 INFO: Initialized l1_latent_x2_opt in latent space (w=1.0).
562
+ 2025-11-01 18:37:18,343 INFO: Loss [L1Loss] is created.
563
+ 2025-11-01 18:37:18,343 INFO: Initialized l1_latent_x4_opt in latent space (w=1.0).
564
+ 2025-11-01 18:37:18,343 INFO: Loss [FFTFrequencyLoss] is created.
565
+ 2025-11-01 18:37:18,344 INFO: Initialized fft_latent_x2_opt in latent space (w=0.1).
566
+ 2025-11-01 18:37:18,345 INFO: Loss [FFTFrequencyLoss] is created.
567
+ 2025-11-01 18:37:18,346 INFO: Initialized fft_latent_x4_opt in latent space (w=0.1).
568
+ 2025-11-01 18:37:18,347 INFO: Gradient clipping enabled (G:norm(max_norm=0.4, norm_type=2.0)).
569
+ 2025-11-01 18:37:18,348 INFO: Model [SwinIRLatentModelMultiHead] is created.
570
+ 2025-11-01 18:37:19,528 INFO: Start training from epoch: 0, step: 0
571
+ 2025-11-01 18:39:15,803 INFO: [32_7..][epoch: 0, step: 100, lr:(2.000e-04,)] [eta: 1 day, 13:56:24, time (data): 1.163 (0.022)] l1_latent_x2_opt: 9.2823e-01 fft_latent_x2_opt: 8.1074e-01 l1_latent_x4_opt: 1.0574e+00 fft_latent_x4_opt: 9.1827e-01
01_11_2025/32_archived_20251101_095943/basicsr_options.yaml ADDED
@@ -0,0 +1,205 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # GENERATE TIME: Sat Nov 1 09:58:39 2025
2
+ # CMD:
3
+ # train_vae.py -opt /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/basicsr_options.yaml --launcher pytorch --local_rank 0
4
+
5
+ model_type: SwinIRLatentModelMultiHead
6
+ primary_head: x4
7
+ scale: 4
8
+ num_gpu: auto
9
+ manual_seed: 0
10
+ find_unused_parameters: false
11
+ vae_sources:
12
+ flux_vae:
13
+ hf_repo: wolfgangblack/flux_vae
14
+ vae_kind: kl
15
+ datasets:
16
+ train:
17
+ name: gpt4_nanobana_midjourney_recraft_CropsFluxVAE
18
+ type: MultiScaleLatentCacheDataset
19
+ scales:
20
+ - 128
21
+ - 256
22
+ - 512
23
+ cache_dirs:
24
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/nano_banana_crops/embeddings/flux_vae
25
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/midjourney_full_dataset_crops_new/embeddings/flux_vae
26
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/gpt4_crops/embeddings/flux_vae
27
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/recraft_data_crops/embeddings/flux_vae
28
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/mjv5_crops_fix/embeddings/flux_vae
29
+ vae_names:
30
+ - flux_vae
31
+ phase: train
32
+ filename_tmpl: '{}'
33
+ io_backend:
34
+ type: disk
35
+ scale: 4
36
+ mean: null
37
+ std: null
38
+ num_worker_per_gpu: 32
39
+ batch_size_per_gpu: 256
40
+ pin_memory: true
41
+ persistent_workers: true
42
+ val:
43
+ name: sdxk_120_1024x1024
44
+ type: MultiScaleLatentCacheDataset
45
+ scales:
46
+ - 256
47
+ - 512
48
+ - 1024
49
+ cache_dirs:
50
+ - /data/kazanplova/datasets/latent_upscale_validation_120_samples/embeddings/flux_vae
51
+ vae_names:
52
+ - flux_vae
53
+ phase: val
54
+ io_backend:
55
+ type: disk
56
+ scale: 4
57
+ mean: null
58
+ std: null
59
+ batch_size_per_gpu: 16
60
+ num_worker_per_gpu: 4
61
+ pin_memory: true
62
+ network_g:
63
+ type: SwinIRMultiHead
64
+ in_chans: 16
65
+ img_size: 32
66
+ window_size: 16
67
+ img_range: 1.0
68
+ depths:
69
+ - 6
70
+ - 6
71
+ - 6
72
+ - 6
73
+ - 6
74
+ - 6
75
+ embed_dim: 360
76
+ num_heads:
77
+ - 12
78
+ - 12
79
+ - 12
80
+ - 12
81
+ - 12
82
+ - 12
83
+ mlp_ratio: 2
84
+ resi_connection: 1conv
85
+ primary_head: x4
86
+ head_num_feat: 256
87
+ heads:
88
+ - name: x2
89
+ scale: 2
90
+ out_chans: 16
91
+ - name: x4
92
+ scale: 4
93
+ out_chans: 16
94
+ primary: true
95
+ compile:
96
+ enabled: false
97
+ mode: max-autotune
98
+ dynamic: true
99
+ fullgraph: false
100
+ backend: null
101
+ train:
102
+ ema_decay: 0.999
103
+ optim_g:
104
+ type: Adam
105
+ lr: 0.00015
106
+ weight_decay: 0
107
+ betas:
108
+ - 0.9
109
+ - 0.995
110
+ grad_clip:
111
+ enabled: true
112
+ generator:
113
+ type: norm
114
+ max_norm: 0.4
115
+ norm_type: 2.0
116
+ scheduler:
117
+ type: MultiStepLR
118
+ milestones:
119
+ - 75000
120
+ - 90000
121
+ - 110000
122
+ gamma: 0.5
123
+ total_steps: 125000
124
+ warmup_iter: -1
125
+ l1_latent_x2_opt:
126
+ type: L1Loss
127
+ loss_weight: 0.5
128
+ reduction: mean
129
+ space: latent
130
+ target: x2
131
+ fft_latent_x2_opt:
132
+ type: FFTFrequencyLoss
133
+ loss_weight: 0.1
134
+ reduction: mean
135
+ space: latent
136
+ target: x2
137
+ norm: ortho
138
+ use_log_amplitude: false
139
+ alpha: 0.0
140
+ normalize_weight: true
141
+ l1_latent_x4_opt:
142
+ type: L1Loss
143
+ loss_weight: 0.5
144
+ reduction: mean
145
+ space: latent
146
+ target: x4
147
+ fft_latent_x4_opt:
148
+ type: FFTFrequencyLoss
149
+ loss_weight: 0.1
150
+ reduction: mean
151
+ space: latent
152
+ target: x4
153
+ norm: ortho
154
+ use_log_amplitude: false
155
+ alpha: 0.0
156
+ normalize_weight: true
157
+ val:
158
+ val_freq: 5000
159
+ save_img: true
160
+ head_evals:
161
+ x2:
162
+ save_img: true
163
+ label: val_x2
164
+ metrics:
165
+ l1_latent:
166
+ type: L1Loss
167
+ space: latent
168
+ pixel_psnr_pt:
169
+ type: calculate_psnr_pt
170
+ space: pixel
171
+ crop_border: 2
172
+ test_y_channel: false
173
+ x4:
174
+ save_img: true
175
+ label: val_x4
176
+ metrics:
177
+ l1_latent:
178
+ type: L1Loss
179
+ space: latent
180
+ l2_latent:
181
+ type: MSELoss
182
+ space: latent
183
+ pixel_psnr_pt:
184
+ type: calculate_psnr_pt
185
+ space: pixel
186
+ crop_border: 2
187
+ test_y_channel: false
188
+ logger:
189
+ print_freq: 100
190
+ save_checkpoint_freq: 5000
191
+ use_tb_logger: true
192
+ wandb:
193
+ project: Swin2SR-Latent-SR
194
+ entity: kazanplova-it-more
195
+ resume_id: null
196
+ max_val_images: 10
197
+ dist_params:
198
+ backend: nccl
199
+ port: 29500
200
+ dist: true
201
+ load_networks_only: false
202
+ exp_name: '32'
203
+ name: '32'
204
+ path:
205
+ experiments_root: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025
01_11_2025/32_archived_20251101_095943/train_32_20251101_095839.log ADDED
@@ -0,0 +1,548 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-11-01 09:58:39,616 INFO:
2
+ ____ _ _____ ____
3
+ / __ ) ____ _ _____ (_)_____/ ___/ / __ \
4
+ / __ |/ __ `// ___// // ___/\__ \ / /_/ /
5
+ / /_/ // /_/ /(__ )/ // /__ ___/ // _, _/
6
+ /_____/ \__,_//____//_/ \___//____//_/ |_|
7
+ ______ __ __ __ __
8
+ / ____/____ ____ ____/ / / / __ __ _____ / /__ / /
9
+ / / __ / __ \ / __ \ / __ / / / / / / // ___// //_/ / /
10
+ / /_/ // /_/ // /_/ // /_/ / / /___/ /_/ // /__ / /< /_/
11
+ \____/ \____/ \____/ \____/ /_____/\____/ \___//_/|_| (_)
12
+
13
+ Version Information:
14
+ BasicSR: 1.4.2
15
+ PyTorch: 2.9.0+cu129
16
+ TorchVision: 0.24.0+cpu
17
+ 2025-11-01 09:58:39,616 INFO:
18
+ model_type: SwinIRLatentModelMultiHead
19
+ primary_head: x4
20
+ scale: 4
21
+ num_gpu: 3
22
+ manual_seed: 0
23
+ find_unused_parameters: False
24
+ vae_sources:[
25
+ flux_vae:[
26
+ hf_repo: wolfgangblack/flux_vae
27
+ vae_kind: kl
28
+ ]
29
+ ]
30
+ datasets:[
31
+ train:[
32
+ name: gpt4_nanobana_midjourney_recraft_CropsFluxVAE
33
+ type: MultiScaleLatentCacheDataset
34
+ scales: [128, 256, 512]
35
+ cache_dirs: ['/data/kazanplova/datasets/full_latent_upscale_dataset_train/nano_banana_crops/embeddings/flux_vae', '/data/kazanplova/datasets/full_latent_upscale_dataset_train/midjourney_full_dataset_crops_new/embeddings/flux_vae', '/data/kazanplova/datasets/full_latent_upscale_dataset_train/gpt4_crops/embeddings/flux_vae', '/data/kazanplova/datasets/full_latent_upscale_dataset_train/recraft_data_crops/embeddings/flux_vae', '/data/kazanplova/datasets/full_latent_upscale_dataset_train/mjv5_crops_fix/embeddings/flux_vae']
36
+ vae_names: ['flux_vae']
37
+ phase: train
38
+ filename_tmpl: {}
39
+ io_backend:[
40
+ type: disk
41
+ ]
42
+ scale: 4
43
+ mean: None
44
+ std: None
45
+ num_worker_per_gpu: 32
46
+ batch_size_per_gpu: 256
47
+ pin_memory: True
48
+ persistent_workers: True
49
+ ]
50
+ val:[
51
+ name: sdxk_120_1024x1024
52
+ type: MultiScaleLatentCacheDataset
53
+ scales: [256, 512, 1024]
54
+ cache_dirs: ['/data/kazanplova/datasets/latent_upscale_validation_120_samples/embeddings/flux_vae']
55
+ vae_names: ['flux_vae']
56
+ phase: val
57
+ io_backend:[
58
+ type: disk
59
+ ]
60
+ scale: 4
61
+ mean: None
62
+ std: None
63
+ batch_size_per_gpu: 16
64
+ num_worker_per_gpu: 4
65
+ pin_memory: True
66
+ ]
67
+ ]
68
+ network_g:[
69
+ type: SwinIRMultiHead
70
+ in_chans: 16
71
+ img_size: 32
72
+ window_size: 16
73
+ img_range: 1.0
74
+ depths: [6, 6, 6, 6, 6, 6]
75
+ embed_dim: 360
76
+ num_heads: [12, 12, 12, 12, 12, 12]
77
+ mlp_ratio: 2
78
+ resi_connection: 1conv
79
+ primary_head: x4
80
+ head_num_feat: 256
81
+ heads: [OrderedDict({'name': 'x2', 'scale': 2, 'out_chans': 16}), OrderedDict({'name': 'x4', 'scale': 4, 'out_chans': 16, 'primary': True})]
82
+ ]
83
+ compile:[
84
+ enabled: False
85
+ mode: max-autotune
86
+ dynamic: True
87
+ fullgraph: False
88
+ backend: None
89
+ ]
90
+ train:[
91
+ ema_decay: 0.999
92
+ optim_g:[
93
+ type: Adam
94
+ lr: 0.00015
95
+ weight_decay: 0
96
+ betas: [0.9, 0.995]
97
+ ]
98
+ grad_clip:[
99
+ enabled: True
100
+ generator:[
101
+ type: norm
102
+ max_norm: 0.4
103
+ norm_type: 2.0
104
+ ]
105
+ ]
106
+ scheduler:[
107
+ type: MultiStepLR
108
+ milestones: [75000, 90000, 110000]
109
+ gamma: 0.5
110
+ ]
111
+ total_steps: 125000
112
+ warmup_iter: -1
113
+ l1_latent_x2_opt:[
114
+ type: L1Loss
115
+ loss_weight: 0.5
116
+ reduction: mean
117
+ space: latent
118
+ target: x2
119
+ ]
120
+ fft_latent_x2_opt:[
121
+ type: FFTFrequencyLoss
122
+ loss_weight: 0.1
123
+ reduction: mean
124
+ space: latent
125
+ target: x2
126
+ norm: ortho
127
+ use_log_amplitude: False
128
+ alpha: 0.0
129
+ normalize_weight: True
130
+ ]
131
+ l1_latent_x4_opt:[
132
+ type: L1Loss
133
+ loss_weight: 0.5
134
+ reduction: mean
135
+ space: latent
136
+ target: x4
137
+ ]
138
+ fft_latent_x4_opt:[
139
+ type: FFTFrequencyLoss
140
+ loss_weight: 0.1
141
+ reduction: mean
142
+ space: latent
143
+ target: x4
144
+ norm: ortho
145
+ use_log_amplitude: False
146
+ alpha: 0.0
147
+ normalize_weight: True
148
+ ]
149
+ ]
150
+ val:[
151
+ val_freq: 5000
152
+ save_img: True
153
+ head_evals:[
154
+ x2:[
155
+ save_img: True
156
+ label: val_x2
157
+ metrics:[
158
+ l1_latent:[
159
+ type: L1Loss
160
+ space: latent
161
+ ]
162
+ pixel_psnr_pt:[
163
+ type: calculate_psnr_pt
164
+ space: pixel
165
+ crop_border: 2
166
+ test_y_channel: False
167
+ ]
168
+ ]
169
+ ]
170
+ x4:[
171
+ save_img: True
172
+ label: val_x4
173
+ metrics:[
174
+ l1_latent:[
175
+ type: L1Loss
176
+ space: latent
177
+ ]
178
+ l2_latent:[
179
+ type: MSELoss
180
+ space: latent
181
+ ]
182
+ pixel_psnr_pt:[
183
+ type: calculate_psnr_pt
184
+ space: pixel
185
+ crop_border: 2
186
+ test_y_channel: False
187
+ ]
188
+ ]
189
+ ]
190
+ ]
191
+ ]
192
+ logger:[
193
+ print_freq: 100
194
+ save_checkpoint_freq: 5000
195
+ use_tb_logger: True
196
+ wandb:[
197
+ project: Swin2SR-Latent-SR
198
+ entity: kazanplova-it-more
199
+ resume_id: None
200
+ max_val_images: 10
201
+ ]
202
+ ]
203
+ dist_params:[
204
+ backend: nccl
205
+ port: 29500
206
+ dist: True
207
+ ]
208
+ load_networks_only: False
209
+ exp_name: 32
210
+ name: 32
211
+ path:[
212
+ experiments_root: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/32
213
+ models: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/32/models
214
+ training_states: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/32/training_states
215
+ log: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/32
216
+ visualization: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/32/visualization
217
+ ]
218
+ dist: True
219
+ rank: 0
220
+ world_size: 3
221
+ auto_resume: False
222
+ is_train: True
223
+ root_path: /data/kazanplova/latent_vae_upscale_train
224
+
225
+ 2025-11-01 09:58:41,357 INFO: Use wandb logger with id=st3iv9sy; project=Swin2SR-Latent-SR.
226
+ 2025-11-01 09:58:54,358 INFO: Dataset [MultiScaleLatentCacheDataset] - gpt4_nanobana_midjourney_recraft_CropsFluxVAE is built.
227
+ 2025-11-01 09:58:54,359 INFO: Training statistics:
228
+ Number of train images: 4858507
229
+ Dataset enlarge ratio: 1
230
+ Batch size per gpu: 256
231
+ World size (gpu number): 3
232
+ Steps per epoch: 6327
233
+ Configured training steps: 125000
234
+ Approximate epochs to cover: 20.
235
+ 2025-11-01 09:58:54,362 INFO: Dataset [MultiScaleLatentCacheDataset] - sdxk_120_1024x1024 is built.
236
+ 2025-11-01 09:58:54,363 INFO: Number of val images/folders in sdxk_120_1024x1024: 153
237
+ 2025-11-01 09:58:54,831 INFO: Network [SwinIRMultiHead] is created.
238
+ 2025-11-01 09:58:56,417 INFO: Network: DistributedDataParallel - SwinIRMultiHead, with parameters: 54,917,584
239
+ 2025-11-01 09:58:56,418 INFO: SwinIRMultiHead(
240
+ (conv_first): Conv2d(16, 360, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
241
+ (patch_embed): PatchEmbed(
242
+ (norm): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
243
+ )
244
+ (patch_unembed): PatchUnEmbed()
245
+ (pos_drop): Dropout(p=0.0, inplace=False)
246
+ (layers): ModuleList(
247
+ (0): RSTB(
248
+ (residual_group): BasicLayer(
249
+ dim=360, input_resolution=(32, 32), depth=6
250
+ (blocks): ModuleList(
251
+ (0): SwinTransformerBlock(
252
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=0, mlp_ratio=2.0
253
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
254
+ (attn): WindowAttention(
255
+ dim=360, window_size=(16, 16), num_heads=12
256
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
257
+ (attn_drop): Dropout(p=0.0, inplace=False)
258
+ (proj): Linear(in_features=360, out_features=360, bias=True)
259
+ (proj_drop): Dropout(p=0.0, inplace=False)
260
+ (softmax): Softmax(dim=-1)
261
+ )
262
+ (drop_path): Identity()
263
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
264
+ (mlp): Mlp(
265
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
266
+ (act): GELU(approximate='none')
267
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
268
+ (drop): Dropout(p=0.0, inplace=False)
269
+ )
270
+ )
271
+ (1): SwinTransformerBlock(
272
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=8, mlp_ratio=2.0
273
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
274
+ (attn): WindowAttention(
275
+ dim=360, window_size=(16, 16), num_heads=12
276
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
277
+ (attn_drop): Dropout(p=0.0, inplace=False)
278
+ (proj): Linear(in_features=360, out_features=360, bias=True)
279
+ (proj_drop): Dropout(p=0.0, inplace=False)
280
+ (softmax): Softmax(dim=-1)
281
+ )
282
+ (drop_path): DropPath()
283
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
284
+ (mlp): Mlp(
285
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
286
+ (act): GELU(approximate='none')
287
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
288
+ (drop): Dropout(p=0.0, inplace=False)
289
+ )
290
+ )
291
+ (2): SwinTransformerBlock(
292
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=0, mlp_ratio=2.0
293
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
294
+ (attn): WindowAttention(
295
+ dim=360, window_size=(16, 16), num_heads=12
296
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
297
+ (attn_drop): Dropout(p=0.0, inplace=False)
298
+ (proj): Linear(in_features=360, out_features=360, bias=True)
299
+ (proj_drop): Dropout(p=0.0, inplace=False)
300
+ (softmax): Softmax(dim=-1)
301
+ )
302
+ (drop_path): DropPath()
303
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
304
+ (mlp): Mlp(
305
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
306
+ (act): GELU(approximate='none')
307
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
308
+ (drop): Dropout(p=0.0, inplace=False)
309
+ )
310
+ )
311
+ (3): SwinTransformerBlock(
312
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=8, mlp_ratio=2.0
313
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
314
+ (attn): WindowAttention(
315
+ dim=360, window_size=(16, 16), num_heads=12
316
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
317
+ (attn_drop): Dropout(p=0.0, inplace=False)
318
+ (proj): Linear(in_features=360, out_features=360, bias=True)
319
+ (proj_drop): Dropout(p=0.0, inplace=False)
320
+ (softmax): Softmax(dim=-1)
321
+ )
322
+ (drop_path): DropPath()
323
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
324
+ (mlp): Mlp(
325
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
326
+ (act): GELU(approximate='none')
327
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
328
+ (drop): Dropout(p=0.0, inplace=False)
329
+ )
330
+ )
331
+ (4): SwinTransformerBlock(
332
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=0, mlp_ratio=2.0
333
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
334
+ (attn): WindowAttention(
335
+ dim=360, window_size=(16, 16), num_heads=12
336
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
337
+ (attn_drop): Dropout(p=0.0, inplace=False)
338
+ (proj): Linear(in_features=360, out_features=360, bias=True)
339
+ (proj_drop): Dropout(p=0.0, inplace=False)
340
+ (softmax): Softmax(dim=-1)
341
+ )
342
+ (drop_path): DropPath()
343
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
344
+ (mlp): Mlp(
345
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
346
+ (act): GELU(approximate='none')
347
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
348
+ (drop): Dropout(p=0.0, inplace=False)
349
+ )
350
+ )
351
+ (5): SwinTransformerBlock(
352
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=8, mlp_ratio=2.0
353
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
354
+ (attn): WindowAttention(
355
+ dim=360, window_size=(16, 16), num_heads=12
356
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
357
+ (attn_drop): Dropout(p=0.0, inplace=False)
358
+ (proj): Linear(in_features=360, out_features=360, bias=True)
359
+ (proj_drop): Dropout(p=0.0, inplace=False)
360
+ (softmax): Softmax(dim=-1)
361
+ )
362
+ (drop_path): DropPath()
363
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
364
+ (mlp): Mlp(
365
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
366
+ (act): GELU(approximate='none')
367
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
368
+ (drop): Dropout(p=0.0, inplace=False)
369
+ )
370
+ )
371
+ )
372
+ )
373
+ (conv): Conv2d(360, 360, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
374
+ (patch_embed): PatchEmbed()
375
+ (patch_unembed): PatchUnEmbed()
376
+ )
377
+ (1-5): 5 x RSTB(
378
+ (residual_group): BasicLayer(
379
+ dim=360, input_resolution=(32, 32), depth=6
380
+ (blocks): ModuleList(
381
+ (0): SwinTransformerBlock(
382
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=0, mlp_ratio=2.0
383
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
384
+ (attn): WindowAttention(
385
+ dim=360, window_size=(16, 16), num_heads=12
386
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
387
+ (attn_drop): Dropout(p=0.0, inplace=False)
388
+ (proj): Linear(in_features=360, out_features=360, bias=True)
389
+ (proj_drop): Dropout(p=0.0, inplace=False)
390
+ (softmax): Softmax(dim=-1)
391
+ )
392
+ (drop_path): DropPath()
393
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
394
+ (mlp): Mlp(
395
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
396
+ (act): GELU(approximate='none')
397
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
398
+ (drop): Dropout(p=0.0, inplace=False)
399
+ )
400
+ )
401
+ (1): SwinTransformerBlock(
402
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=8, mlp_ratio=2.0
403
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
404
+ (attn): WindowAttention(
405
+ dim=360, window_size=(16, 16), num_heads=12
406
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
407
+ (attn_drop): Dropout(p=0.0, inplace=False)
408
+ (proj): Linear(in_features=360, out_features=360, bias=True)
409
+ (proj_drop): Dropout(p=0.0, inplace=False)
410
+ (softmax): Softmax(dim=-1)
411
+ )
412
+ (drop_path): DropPath()
413
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
414
+ (mlp): Mlp(
415
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
416
+ (act): GELU(approximate='none')
417
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
418
+ (drop): Dropout(p=0.0, inplace=False)
419
+ )
420
+ )
421
+ (2): SwinTransformerBlock(
422
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=0, mlp_ratio=2.0
423
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
424
+ (attn): WindowAttention(
425
+ dim=360, window_size=(16, 16), num_heads=12
426
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
427
+ (attn_drop): Dropout(p=0.0, inplace=False)
428
+ (proj): Linear(in_features=360, out_features=360, bias=True)
429
+ (proj_drop): Dropout(p=0.0, inplace=False)
430
+ (softmax): Softmax(dim=-1)
431
+ )
432
+ (drop_path): DropPath()
433
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
434
+ (mlp): Mlp(
435
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
436
+ (act): GELU(approximate='none')
437
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
438
+ (drop): Dropout(p=0.0, inplace=False)
439
+ )
440
+ )
441
+ (3): SwinTransformerBlock(
442
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=8, mlp_ratio=2.0
443
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
444
+ (attn): WindowAttention(
445
+ dim=360, window_size=(16, 16), num_heads=12
446
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
447
+ (attn_drop): Dropout(p=0.0, inplace=False)
448
+ (proj): Linear(in_features=360, out_features=360, bias=True)
449
+ (proj_drop): Dropout(p=0.0, inplace=False)
450
+ (softmax): Softmax(dim=-1)
451
+ )
452
+ (drop_path): DropPath()
453
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
454
+ (mlp): Mlp(
455
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
456
+ (act): GELU(approximate='none')
457
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
458
+ (drop): Dropout(p=0.0, inplace=False)
459
+ )
460
+ )
461
+ (4): SwinTransformerBlock(
462
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=0, mlp_ratio=2.0
463
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
464
+ (attn): WindowAttention(
465
+ dim=360, window_size=(16, 16), num_heads=12
466
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
467
+ (attn_drop): Dropout(p=0.0, inplace=False)
468
+ (proj): Linear(in_features=360, out_features=360, bias=True)
469
+ (proj_drop): Dropout(p=0.0, inplace=False)
470
+ (softmax): Softmax(dim=-1)
471
+ )
472
+ (drop_path): DropPath()
473
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
474
+ (mlp): Mlp(
475
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
476
+ (act): GELU(approximate='none')
477
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
478
+ (drop): Dropout(p=0.0, inplace=False)
479
+ )
480
+ )
481
+ (5): SwinTransformerBlock(
482
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=8, mlp_ratio=2.0
483
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
484
+ (attn): WindowAttention(
485
+ dim=360, window_size=(16, 16), num_heads=12
486
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
487
+ (attn_drop): Dropout(p=0.0, inplace=False)
488
+ (proj): Linear(in_features=360, out_features=360, bias=True)
489
+ (proj_drop): Dropout(p=0.0, inplace=False)
490
+ (softmax): Softmax(dim=-1)
491
+ )
492
+ (drop_path): DropPath()
493
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
494
+ (mlp): Mlp(
495
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
496
+ (act): GELU(approximate='none')
497
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
498
+ (drop): Dropout(p=0.0, inplace=False)
499
+ )
500
+ )
501
+ )
502
+ )
503
+ (conv): Conv2d(360, 360, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
504
+ (patch_embed): PatchEmbed()
505
+ (patch_unembed): PatchUnEmbed()
506
+ )
507
+ )
508
+ (norm): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
509
+ (conv_after_body): Conv2d(360, 360, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
510
+ (heads): ModuleDict(
511
+ (x2): _SwinIRPixelShuffleHead(
512
+ (conv_before): Sequential(
513
+ (0): Conv2d(360, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
514
+ (1): LeakyReLU(negative_slope=0.01, inplace=True)
515
+ )
516
+ (upsample): Upsample(
517
+ (0): Conv2d(256, 1024, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
518
+ (1): PixelShuffle(upscale_factor=2)
519
+ )
520
+ (conv_last): Conv2d(256, 16, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
521
+ )
522
+ (x4): _SwinIRPixelShuffleHead(
523
+ (conv_before): Sequential(
524
+ (0): Conv2d(360, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
525
+ (1): LeakyReLU(negative_slope=0.01, inplace=True)
526
+ )
527
+ (upsample): Upsample(
528
+ (0): Conv2d(256, 1024, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
529
+ (1): PixelShuffle(upscale_factor=2)
530
+ (2): Conv2d(256, 1024, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
531
+ (3): PixelShuffle(upscale_factor=2)
532
+ )
533
+ (conv_last): Conv2d(256, 16, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
534
+ )
535
+ )
536
+ )
537
+ 2025-11-01 09:58:56,421 INFO: Use EMA with decay: 0.999
538
+ 2025-11-01 09:58:56,865 INFO: Network [SwinIRMultiHead] is created.
539
+ 2025-11-01 09:58:56,933 INFO: Loss [L1Loss] is created.
540
+ 2025-11-01 09:58:56,934 INFO: Initialized l1_latent_x2_opt in latent space (w=0.5).
541
+ 2025-11-01 09:58:56,935 INFO: Loss [FFTFrequencyLoss] is created.
542
+ 2025-11-01 09:58:56,936 INFO: Initialized fft_latent_x2_opt in latent space (w=0.1).
543
+ 2025-11-01 09:58:56,937 INFO: Loss [L1Loss] is created.
544
+ 2025-11-01 09:58:56,937 INFO: Initialized l1_latent_x4_opt in latent space (w=0.5).
545
+ 2025-11-01 09:58:56,938 INFO: Loss [FFTFrequencyLoss] is created.
546
+ 2025-11-01 09:58:56,939 INFO: Initialized fft_latent_x4_opt in latent space (w=0.1).
547
+ 2025-11-01 09:58:56,942 INFO: Gradient clipping enabled (G:norm(max_norm=0.4, norm_type=2.0)).
548
+ 2025-11-01 09:58:56,942 INFO: Model [SwinIRLatentModelMultiHead] is created.
01_11_2025/32_archived_20251101_104933/basicsr_options.yaml ADDED
@@ -0,0 +1,205 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # GENERATE TIME: Sat Nov 1 09:59:43 2025
2
+ # CMD:
3
+ # train_vae.py -opt /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/basicsr_options.yaml --launcher pytorch --local_rank 0
4
+
5
+ model_type: SwinIRLatentModelMultiHead
6
+ primary_head: x4
7
+ scale: 4
8
+ num_gpu: auto
9
+ manual_seed: 0
10
+ find_unused_parameters: false
11
+ vae_sources:
12
+ flux_vae:
13
+ hf_repo: wolfgangblack/flux_vae
14
+ vae_kind: kl
15
+ datasets:
16
+ train:
17
+ name: gpt4_nanobana_midjourney_recraft_CropsFluxVAE
18
+ type: MultiScaleLatentCacheDataset
19
+ scales:
20
+ - 128
21
+ - 256
22
+ - 512
23
+ cache_dirs:
24
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/nano_banana_crops/embeddings/flux_vae
25
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/midjourney_full_dataset_crops_new/embeddings/flux_vae
26
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/gpt4_crops/embeddings/flux_vae
27
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/recraft_data_crops/embeddings/flux_vae
28
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/mjv5_crops_fix/embeddings/flux_vae
29
+ vae_names:
30
+ - flux_vae
31
+ phase: train
32
+ filename_tmpl: '{}'
33
+ io_backend:
34
+ type: disk
35
+ scale: 4
36
+ mean: null
37
+ std: null
38
+ num_worker_per_gpu: 32
39
+ batch_size_per_gpu: 256
40
+ pin_memory: true
41
+ persistent_workers: true
42
+ val:
43
+ name: sdxk_120_1024x1024
44
+ type: MultiScaleLatentCacheDataset
45
+ scales:
46
+ - 256
47
+ - 512
48
+ - 1024
49
+ cache_dirs:
50
+ - /data/kazanplova/datasets/latent_upscale_validation_120_samples/embeddings/flux_vae
51
+ vae_names:
52
+ - flux_vae
53
+ phase: val
54
+ io_backend:
55
+ type: disk
56
+ scale: 4
57
+ mean: null
58
+ std: null
59
+ batch_size_per_gpu: 16
60
+ num_worker_per_gpu: 4
61
+ pin_memory: true
62
+ network_g:
63
+ type: SwinIRMultiHead
64
+ in_chans: 16
65
+ img_size: 32
66
+ window_size: 16
67
+ img_range: 1.0
68
+ depths:
69
+ - 6
70
+ - 6
71
+ - 6
72
+ - 6
73
+ - 6
74
+ - 6
75
+ embed_dim: 360
76
+ num_heads:
77
+ - 12
78
+ - 12
79
+ - 12
80
+ - 12
81
+ - 12
82
+ - 12
83
+ mlp_ratio: 2
84
+ resi_connection: 1conv
85
+ primary_head: x4
86
+ head_num_feat: 256
87
+ heads:
88
+ - name: x2
89
+ scale: 2
90
+ out_chans: 16
91
+ - name: x4
92
+ scale: 4
93
+ out_chans: 16
94
+ primary: true
95
+ compile:
96
+ enabled: false
97
+ mode: max-autotune
98
+ dynamic: true
99
+ fullgraph: false
100
+ backend: null
101
+ train:
102
+ ema_decay: 0.999
103
+ optim_g:
104
+ type: Adam
105
+ lr: 0.00015
106
+ weight_decay: 0
107
+ betas:
108
+ - 0.9
109
+ - 0.995
110
+ grad_clip:
111
+ enabled: true
112
+ generator:
113
+ type: norm
114
+ max_norm: 0.4
115
+ norm_type: 2.0
116
+ scheduler:
117
+ type: MultiStepLR
118
+ milestones:
119
+ - 75000
120
+ - 90000
121
+ - 110000
122
+ gamma: 0.5
123
+ total_steps: 125000
124
+ warmup_iter: -1
125
+ l1_latent_x2_opt:
126
+ type: L1Loss
127
+ loss_weight: 0.5
128
+ reduction: mean
129
+ space: latent
130
+ target: x2
131
+ fft_latent_x2_opt:
132
+ type: FFTFrequencyLoss
133
+ loss_weight: 0.1
134
+ reduction: mean
135
+ space: latent
136
+ target: x2
137
+ norm: ortho
138
+ use_log_amplitude: false
139
+ alpha: 0.0
140
+ normalize_weight: true
141
+ l1_latent_x4_opt:
142
+ type: L1Loss
143
+ loss_weight: 0.5
144
+ reduction: mean
145
+ space: latent
146
+ target: x4
147
+ fft_latent_x4_opt:
148
+ type: FFTFrequencyLoss
149
+ loss_weight: 0.1
150
+ reduction: mean
151
+ space: latent
152
+ target: x4
153
+ norm: ortho
154
+ use_log_amplitude: false
155
+ alpha: 0.0
156
+ normalize_weight: true
157
+ val:
158
+ val_freq: 5000
159
+ save_img: true
160
+ head_evals:
161
+ x2:
162
+ save_img: true
163
+ label: val_x2
164
+ metrics:
165
+ l1_latent:
166
+ type: L1Loss
167
+ space: latent
168
+ pixel_psnr_pt:
169
+ type: calculate_psnr_pt
170
+ space: pixel
171
+ crop_border: 2
172
+ test_y_channel: false
173
+ x4:
174
+ save_img: true
175
+ label: val_x4
176
+ metrics:
177
+ l1_latent:
178
+ type: L1Loss
179
+ space: latent
180
+ l2_latent:
181
+ type: MSELoss
182
+ space: latent
183
+ pixel_psnr_pt:
184
+ type: calculate_psnr_pt
185
+ space: pixel
186
+ crop_border: 2
187
+ test_y_channel: false
188
+ logger:
189
+ print_freq: 100
190
+ save_checkpoint_freq: 5000
191
+ use_tb_logger: true
192
+ wandb:
193
+ project: Swin2SR-Latent-SR
194
+ entity: kazanplova-it-more
195
+ resume_id: null
196
+ max_val_images: 10
197
+ dist_params:
198
+ backend: nccl
199
+ port: 29500
200
+ dist: true
201
+ load_networks_only: false
202
+ exp_name: '32'
203
+ name: '32'
204
+ path:
205
+ experiments_root: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025
01_11_2025/32_archived_20251101_104933/train_32_20251101_095943.log ADDED
@@ -0,0 +1,548 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-11-01 09:59:43,922 INFO:
2
+ ____ _ _____ ____
3
+ / __ ) ____ _ _____ (_)_____/ ___/ / __ \
4
+ / __ |/ __ `// ___// // ___/\__ \ / /_/ /
5
+ / /_/ // /_/ /(__ )/ // /__ ___/ // _, _/
6
+ /_____/ \__,_//____//_/ \___//____//_/ |_|
7
+ ______ __ __ __ __
8
+ / ____/____ ____ ____/ / / / __ __ _____ / /__ / /
9
+ / / __ / __ \ / __ \ / __ / / / / / / // ___// //_/ / /
10
+ / /_/ // /_/ // /_/ // /_/ / / /___/ /_/ // /__ / /< /_/
11
+ \____/ \____/ \____/ \____/ /_____/\____/ \___//_/|_| (_)
12
+
13
+ Version Information:
14
+ BasicSR: 1.4.2
15
+ PyTorch: 2.9.0+cu129
16
+ TorchVision: 0.24.0+cpu
17
+ 2025-11-01 09:59:43,923 INFO:
18
+ model_type: SwinIRLatentModelMultiHead
19
+ primary_head: x4
20
+ scale: 4
21
+ num_gpu: 3
22
+ manual_seed: 0
23
+ find_unused_parameters: False
24
+ vae_sources:[
25
+ flux_vae:[
26
+ hf_repo: wolfgangblack/flux_vae
27
+ vae_kind: kl
28
+ ]
29
+ ]
30
+ datasets:[
31
+ train:[
32
+ name: gpt4_nanobana_midjourney_recraft_CropsFluxVAE
33
+ type: MultiScaleLatentCacheDataset
34
+ scales: [128, 256, 512]
35
+ cache_dirs: ['/data/kazanplova/datasets/full_latent_upscale_dataset_train/nano_banana_crops/embeddings/flux_vae', '/data/kazanplova/datasets/full_latent_upscale_dataset_train/midjourney_full_dataset_crops_new/embeddings/flux_vae', '/data/kazanplova/datasets/full_latent_upscale_dataset_train/gpt4_crops/embeddings/flux_vae', '/data/kazanplova/datasets/full_latent_upscale_dataset_train/recraft_data_crops/embeddings/flux_vae', '/data/kazanplova/datasets/full_latent_upscale_dataset_train/mjv5_crops_fix/embeddings/flux_vae']
36
+ vae_names: ['flux_vae']
37
+ phase: train
38
+ filename_tmpl: {}
39
+ io_backend:[
40
+ type: disk
41
+ ]
42
+ scale: 4
43
+ mean: None
44
+ std: None
45
+ num_worker_per_gpu: 32
46
+ batch_size_per_gpu: 256
47
+ pin_memory: True
48
+ persistent_workers: True
49
+ ]
50
+ val:[
51
+ name: sdxk_120_1024x1024
52
+ type: MultiScaleLatentCacheDataset
53
+ scales: [256, 512, 1024]
54
+ cache_dirs: ['/data/kazanplova/datasets/latent_upscale_validation_120_samples/embeddings/flux_vae']
55
+ vae_names: ['flux_vae']
56
+ phase: val
57
+ io_backend:[
58
+ type: disk
59
+ ]
60
+ scale: 4
61
+ mean: None
62
+ std: None
63
+ batch_size_per_gpu: 16
64
+ num_worker_per_gpu: 4
65
+ pin_memory: True
66
+ ]
67
+ ]
68
+ network_g:[
69
+ type: SwinIRMultiHead
70
+ in_chans: 16
71
+ img_size: 32
72
+ window_size: 16
73
+ img_range: 1.0
74
+ depths: [6, 6, 6, 6, 6, 6]
75
+ embed_dim: 360
76
+ num_heads: [12, 12, 12, 12, 12, 12]
77
+ mlp_ratio: 2
78
+ resi_connection: 1conv
79
+ primary_head: x4
80
+ head_num_feat: 256
81
+ heads: [OrderedDict({'name': 'x2', 'scale': 2, 'out_chans': 16}), OrderedDict({'name': 'x4', 'scale': 4, 'out_chans': 16, 'primary': True})]
82
+ ]
83
+ compile:[
84
+ enabled: False
85
+ mode: max-autotune
86
+ dynamic: True
87
+ fullgraph: False
88
+ backend: None
89
+ ]
90
+ train:[
91
+ ema_decay: 0.999
92
+ optim_g:[
93
+ type: Adam
94
+ lr: 0.00015
95
+ weight_decay: 0
96
+ betas: [0.9, 0.995]
97
+ ]
98
+ grad_clip:[
99
+ enabled: True
100
+ generator:[
101
+ type: norm
102
+ max_norm: 0.4
103
+ norm_type: 2.0
104
+ ]
105
+ ]
106
+ scheduler:[
107
+ type: MultiStepLR
108
+ milestones: [75000, 90000, 110000]
109
+ gamma: 0.5
110
+ ]
111
+ total_steps: 125000
112
+ warmup_iter: -1
113
+ l1_latent_x2_opt:[
114
+ type: L1Loss
115
+ loss_weight: 0.5
116
+ reduction: mean
117
+ space: latent
118
+ target: x2
119
+ ]
120
+ fft_latent_x2_opt:[
121
+ type: FFTFrequencyLoss
122
+ loss_weight: 0.1
123
+ reduction: mean
124
+ space: latent
125
+ target: x2
126
+ norm: ortho
127
+ use_log_amplitude: False
128
+ alpha: 0.0
129
+ normalize_weight: True
130
+ ]
131
+ l1_latent_x4_opt:[
132
+ type: L1Loss
133
+ loss_weight: 0.5
134
+ reduction: mean
135
+ space: latent
136
+ target: x4
137
+ ]
138
+ fft_latent_x4_opt:[
139
+ type: FFTFrequencyLoss
140
+ loss_weight: 0.1
141
+ reduction: mean
142
+ space: latent
143
+ target: x4
144
+ norm: ortho
145
+ use_log_amplitude: False
146
+ alpha: 0.0
147
+ normalize_weight: True
148
+ ]
149
+ ]
150
+ val:[
151
+ val_freq: 5000
152
+ save_img: True
153
+ head_evals:[
154
+ x2:[
155
+ save_img: True
156
+ label: val_x2
157
+ metrics:[
158
+ l1_latent:[
159
+ type: L1Loss
160
+ space: latent
161
+ ]
162
+ pixel_psnr_pt:[
163
+ type: calculate_psnr_pt
164
+ space: pixel
165
+ crop_border: 2
166
+ test_y_channel: False
167
+ ]
168
+ ]
169
+ ]
170
+ x4:[
171
+ save_img: True
172
+ label: val_x4
173
+ metrics:[
174
+ l1_latent:[
175
+ type: L1Loss
176
+ space: latent
177
+ ]
178
+ l2_latent:[
179
+ type: MSELoss
180
+ space: latent
181
+ ]
182
+ pixel_psnr_pt:[
183
+ type: calculate_psnr_pt
184
+ space: pixel
185
+ crop_border: 2
186
+ test_y_channel: False
187
+ ]
188
+ ]
189
+ ]
190
+ ]
191
+ ]
192
+ logger:[
193
+ print_freq: 100
194
+ save_checkpoint_freq: 5000
195
+ use_tb_logger: True
196
+ wandb:[
197
+ project: Swin2SR-Latent-SR
198
+ entity: kazanplova-it-more
199
+ resume_id: None
200
+ max_val_images: 10
201
+ ]
202
+ ]
203
+ dist_params:[
204
+ backend: nccl
205
+ port: 29500
206
+ dist: True
207
+ ]
208
+ load_networks_only: False
209
+ exp_name: 32
210
+ name: 32
211
+ path:[
212
+ experiments_root: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/32
213
+ models: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/32/models
214
+ training_states: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/32/training_states
215
+ log: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/32
216
+ visualization: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/32/visualization
217
+ ]
218
+ dist: True
219
+ rank: 0
220
+ world_size: 3
221
+ auto_resume: False
222
+ is_train: True
223
+ root_path: /data/kazanplova/latent_vae_upscale_train
224
+
225
+ 2025-11-01 09:59:45,536 INFO: Use wandb logger with id=jtopgky0; project=Swin2SR-Latent-SR.
226
+ 2025-11-01 09:59:58,701 INFO: Dataset [MultiScaleLatentCacheDataset] - gpt4_nanobana_midjourney_recraft_CropsFluxVAE is built.
227
+ 2025-11-01 09:59:58,702 INFO: Training statistics:
228
+ Number of train images: 4858507
229
+ Dataset enlarge ratio: 1
230
+ Batch size per gpu: 256
231
+ World size (gpu number): 3
232
+ Steps per epoch: 6327
233
+ Configured training steps: 125000
234
+ Approximate epochs to cover: 20.
235
+ 2025-11-01 09:59:58,705 INFO: Dataset [MultiScaleLatentCacheDataset] - sdxk_120_1024x1024 is built.
236
+ 2025-11-01 09:59:58,706 INFO: Number of val images/folders in sdxk_120_1024x1024: 153
237
+ 2025-11-01 09:59:59,178 INFO: Network [SwinIRMultiHead] is created.
238
+ 2025-11-01 10:00:00,760 INFO: Network: DistributedDataParallel - SwinIRMultiHead, with parameters: 54,917,584
239
+ 2025-11-01 10:00:00,761 INFO: SwinIRMultiHead(
240
+ (conv_first): Conv2d(16, 360, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
241
+ (patch_embed): PatchEmbed(
242
+ (norm): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
243
+ )
244
+ (patch_unembed): PatchUnEmbed()
245
+ (pos_drop): Dropout(p=0.0, inplace=False)
246
+ (layers): ModuleList(
247
+ (0): RSTB(
248
+ (residual_group): BasicLayer(
249
+ dim=360, input_resolution=(32, 32), depth=6
250
+ (blocks): ModuleList(
251
+ (0): SwinTransformerBlock(
252
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=0, mlp_ratio=2.0
253
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
254
+ (attn): WindowAttention(
255
+ dim=360, window_size=(16, 16), num_heads=12
256
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
257
+ (attn_drop): Dropout(p=0.0, inplace=False)
258
+ (proj): Linear(in_features=360, out_features=360, bias=True)
259
+ (proj_drop): Dropout(p=0.0, inplace=False)
260
+ (softmax): Softmax(dim=-1)
261
+ )
262
+ (drop_path): Identity()
263
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
264
+ (mlp): Mlp(
265
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
266
+ (act): GELU(approximate='none')
267
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
268
+ (drop): Dropout(p=0.0, inplace=False)
269
+ )
270
+ )
271
+ (1): SwinTransformerBlock(
272
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=8, mlp_ratio=2.0
273
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
274
+ (attn): WindowAttention(
275
+ dim=360, window_size=(16, 16), num_heads=12
276
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
277
+ (attn_drop): Dropout(p=0.0, inplace=False)
278
+ (proj): Linear(in_features=360, out_features=360, bias=True)
279
+ (proj_drop): Dropout(p=0.0, inplace=False)
280
+ (softmax): Softmax(dim=-1)
281
+ )
282
+ (drop_path): DropPath()
283
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
284
+ (mlp): Mlp(
285
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
286
+ (act): GELU(approximate='none')
287
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
288
+ (drop): Dropout(p=0.0, inplace=False)
289
+ )
290
+ )
291
+ (2): SwinTransformerBlock(
292
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=0, mlp_ratio=2.0
293
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
294
+ (attn): WindowAttention(
295
+ dim=360, window_size=(16, 16), num_heads=12
296
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
297
+ (attn_drop): Dropout(p=0.0, inplace=False)
298
+ (proj): Linear(in_features=360, out_features=360, bias=True)
299
+ (proj_drop): Dropout(p=0.0, inplace=False)
300
+ (softmax): Softmax(dim=-1)
301
+ )
302
+ (drop_path): DropPath()
303
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
304
+ (mlp): Mlp(
305
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
306
+ (act): GELU(approximate='none')
307
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
308
+ (drop): Dropout(p=0.0, inplace=False)
309
+ )
310
+ )
311
+ (3): SwinTransformerBlock(
312
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=8, mlp_ratio=2.0
313
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
314
+ (attn): WindowAttention(
315
+ dim=360, window_size=(16, 16), num_heads=12
316
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
317
+ (attn_drop): Dropout(p=0.0, inplace=False)
318
+ (proj): Linear(in_features=360, out_features=360, bias=True)
319
+ (proj_drop): Dropout(p=0.0, inplace=False)
320
+ (softmax): Softmax(dim=-1)
321
+ )
322
+ (drop_path): DropPath()
323
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
324
+ (mlp): Mlp(
325
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
326
+ (act): GELU(approximate='none')
327
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
328
+ (drop): Dropout(p=0.0, inplace=False)
329
+ )
330
+ )
331
+ (4): SwinTransformerBlock(
332
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=0, mlp_ratio=2.0
333
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
334
+ (attn): WindowAttention(
335
+ dim=360, window_size=(16, 16), num_heads=12
336
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
337
+ (attn_drop): Dropout(p=0.0, inplace=False)
338
+ (proj): Linear(in_features=360, out_features=360, bias=True)
339
+ (proj_drop): Dropout(p=0.0, inplace=False)
340
+ (softmax): Softmax(dim=-1)
341
+ )
342
+ (drop_path): DropPath()
343
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
344
+ (mlp): Mlp(
345
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
346
+ (act): GELU(approximate='none')
347
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
348
+ (drop): Dropout(p=0.0, inplace=False)
349
+ )
350
+ )
351
+ (5): SwinTransformerBlock(
352
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=8, mlp_ratio=2.0
353
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
354
+ (attn): WindowAttention(
355
+ dim=360, window_size=(16, 16), num_heads=12
356
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
357
+ (attn_drop): Dropout(p=0.0, inplace=False)
358
+ (proj): Linear(in_features=360, out_features=360, bias=True)
359
+ (proj_drop): Dropout(p=0.0, inplace=False)
360
+ (softmax): Softmax(dim=-1)
361
+ )
362
+ (drop_path): DropPath()
363
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
364
+ (mlp): Mlp(
365
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
366
+ (act): GELU(approximate='none')
367
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
368
+ (drop): Dropout(p=0.0, inplace=False)
369
+ )
370
+ )
371
+ )
372
+ )
373
+ (conv): Conv2d(360, 360, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
374
+ (patch_embed): PatchEmbed()
375
+ (patch_unembed): PatchUnEmbed()
376
+ )
377
+ (1-5): 5 x RSTB(
378
+ (residual_group): BasicLayer(
379
+ dim=360, input_resolution=(32, 32), depth=6
380
+ (blocks): ModuleList(
381
+ (0): SwinTransformerBlock(
382
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=0, mlp_ratio=2.0
383
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
384
+ (attn): WindowAttention(
385
+ dim=360, window_size=(16, 16), num_heads=12
386
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
387
+ (attn_drop): Dropout(p=0.0, inplace=False)
388
+ (proj): Linear(in_features=360, out_features=360, bias=True)
389
+ (proj_drop): Dropout(p=0.0, inplace=False)
390
+ (softmax): Softmax(dim=-1)
391
+ )
392
+ (drop_path): DropPath()
393
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
394
+ (mlp): Mlp(
395
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
396
+ (act): GELU(approximate='none')
397
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
398
+ (drop): Dropout(p=0.0, inplace=False)
399
+ )
400
+ )
401
+ (1): SwinTransformerBlock(
402
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=8, mlp_ratio=2.0
403
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
404
+ (attn): WindowAttention(
405
+ dim=360, window_size=(16, 16), num_heads=12
406
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
407
+ (attn_drop): Dropout(p=0.0, inplace=False)
408
+ (proj): Linear(in_features=360, out_features=360, bias=True)
409
+ (proj_drop): Dropout(p=0.0, inplace=False)
410
+ (softmax): Softmax(dim=-1)
411
+ )
412
+ (drop_path): DropPath()
413
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
414
+ (mlp): Mlp(
415
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
416
+ (act): GELU(approximate='none')
417
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
418
+ (drop): Dropout(p=0.0, inplace=False)
419
+ )
420
+ )
421
+ (2): SwinTransformerBlock(
422
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=0, mlp_ratio=2.0
423
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
424
+ (attn): WindowAttention(
425
+ dim=360, window_size=(16, 16), num_heads=12
426
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
427
+ (attn_drop): Dropout(p=0.0, inplace=False)
428
+ (proj): Linear(in_features=360, out_features=360, bias=True)
429
+ (proj_drop): Dropout(p=0.0, inplace=False)
430
+ (softmax): Softmax(dim=-1)
431
+ )
432
+ (drop_path): DropPath()
433
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
434
+ (mlp): Mlp(
435
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
436
+ (act): GELU(approximate='none')
437
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
438
+ (drop): Dropout(p=0.0, inplace=False)
439
+ )
440
+ )
441
+ (3): SwinTransformerBlock(
442
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=8, mlp_ratio=2.0
443
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
444
+ (attn): WindowAttention(
445
+ dim=360, window_size=(16, 16), num_heads=12
446
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
447
+ (attn_drop): Dropout(p=0.0, inplace=False)
448
+ (proj): Linear(in_features=360, out_features=360, bias=True)
449
+ (proj_drop): Dropout(p=0.0, inplace=False)
450
+ (softmax): Softmax(dim=-1)
451
+ )
452
+ (drop_path): DropPath()
453
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
454
+ (mlp): Mlp(
455
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
456
+ (act): GELU(approximate='none')
457
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
458
+ (drop): Dropout(p=0.0, inplace=False)
459
+ )
460
+ )
461
+ (4): SwinTransformerBlock(
462
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=0, mlp_ratio=2.0
463
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
464
+ (attn): WindowAttention(
465
+ dim=360, window_size=(16, 16), num_heads=12
466
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
467
+ (attn_drop): Dropout(p=0.0, inplace=False)
468
+ (proj): Linear(in_features=360, out_features=360, bias=True)
469
+ (proj_drop): Dropout(p=0.0, inplace=False)
470
+ (softmax): Softmax(dim=-1)
471
+ )
472
+ (drop_path): DropPath()
473
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
474
+ (mlp): Mlp(
475
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
476
+ (act): GELU(approximate='none')
477
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
478
+ (drop): Dropout(p=0.0, inplace=False)
479
+ )
480
+ )
481
+ (5): SwinTransformerBlock(
482
+ dim=360, input_resolution=(32, 32), num_heads=12, window_size=16, shift_size=8, mlp_ratio=2.0
483
+ (norm1): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
484
+ (attn): WindowAttention(
485
+ dim=360, window_size=(16, 16), num_heads=12
486
+ (qkv): Linear(in_features=360, out_features=1080, bias=True)
487
+ (attn_drop): Dropout(p=0.0, inplace=False)
488
+ (proj): Linear(in_features=360, out_features=360, bias=True)
489
+ (proj_drop): Dropout(p=0.0, inplace=False)
490
+ (softmax): Softmax(dim=-1)
491
+ )
492
+ (drop_path): DropPath()
493
+ (norm2): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
494
+ (mlp): Mlp(
495
+ (fc1): Linear(in_features=360, out_features=720, bias=True)
496
+ (act): GELU(approximate='none')
497
+ (fc2): Linear(in_features=720, out_features=360, bias=True)
498
+ (drop): Dropout(p=0.0, inplace=False)
499
+ )
500
+ )
501
+ )
502
+ )
503
+ (conv): Conv2d(360, 360, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
504
+ (patch_embed): PatchEmbed()
505
+ (patch_unembed): PatchUnEmbed()
506
+ )
507
+ )
508
+ (norm): LayerNorm((360,), eps=1e-05, elementwise_affine=True)
509
+ (conv_after_body): Conv2d(360, 360, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
510
+ (heads): ModuleDict(
511
+ (x2): _SwinIRPixelShuffleHead(
512
+ (conv_before): Sequential(
513
+ (0): Conv2d(360, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
514
+ (1): LeakyReLU(negative_slope=0.01, inplace=True)
515
+ )
516
+ (upsample): Upsample(
517
+ (0): Conv2d(256, 1024, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
518
+ (1): PixelShuffle(upscale_factor=2)
519
+ )
520
+ (conv_last): Conv2d(256, 16, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
521
+ )
522
+ (x4): _SwinIRPixelShuffleHead(
523
+ (conv_before): Sequential(
524
+ (0): Conv2d(360, 256, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
525
+ (1): LeakyReLU(negative_slope=0.01, inplace=True)
526
+ )
527
+ (upsample): Upsample(
528
+ (0): Conv2d(256, 1024, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
529
+ (1): PixelShuffle(upscale_factor=2)
530
+ (2): Conv2d(256, 1024, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
531
+ (3): PixelShuffle(upscale_factor=2)
532
+ )
533
+ (conv_last): Conv2d(256, 16, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
534
+ )
535
+ )
536
+ )
537
+ 2025-11-01 10:00:00,763 INFO: Use EMA with decay: 0.999
538
+ 2025-11-01 10:00:01,169 INFO: Network [SwinIRMultiHead] is created.
539
+ 2025-11-01 10:00:01,236 INFO: Loss [L1Loss] is created.
540
+ 2025-11-01 10:00:01,237 INFO: Initialized l1_latent_x2_opt in latent space (w=0.5).
541
+ 2025-11-01 10:00:01,239 INFO: Loss [FFTFrequencyLoss] is created.
542
+ 2025-11-01 10:00:01,239 INFO: Initialized fft_latent_x2_opt in latent space (w=0.1).
543
+ 2025-11-01 10:00:01,239 INFO: Loss [L1Loss] is created.
544
+ 2025-11-01 10:00:01,240 INFO: Initialized l1_latent_x4_opt in latent space (w=0.5).
545
+ 2025-11-01 10:00:01,241 INFO: Loss [FFTFrequencyLoss] is created.
546
+ 2025-11-01 10:00:01,242 INFO: Initialized fft_latent_x4_opt in latent space (w=0.1).
547
+ 2025-11-01 10:00:01,244 INFO: Gradient clipping enabled (G:norm(max_norm=0.4, norm_type=2.0)).
548
+ 2025-11-01 10:00:01,244 INFO: Model [SwinIRLatentModelMultiHead] is created.
01_11_2025/32_archived_20251101_161010/basicsr_options.yaml ADDED
@@ -0,0 +1,213 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # GENERATE TIME: Sat Nov 1 10:49:33 2025
2
+ # CMD:
3
+ # train_vae.py -opt /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025/basicsr_options.yaml --launcher pytorch --local_rank 0
4
+
5
+ model_type: SwinIRLatentModelMultiHead
6
+ primary_head: x4
7
+ scale: 4
8
+ num_gpu: auto
9
+ manual_seed: 0
10
+ find_unused_parameters: false
11
+ vae_sources:
12
+ flux_vae:
13
+ hf_repo: wolfgangblack/flux_vae
14
+ vae_kind: kl
15
+ datasets:
16
+ train:
17
+ name: gpt4_nanobana_midjourney_recraft_CropsFluxVAE
18
+ type: MultiScaleLatentCacheDataset
19
+ scales:
20
+ - 128
21
+ - 256
22
+ - 512
23
+ cache_dirs:
24
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/nano_banana_crops/embeddings/flux_vae
25
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/midjourney_full_dataset_crops_new/embeddings/flux_vae
26
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/gpt4_crops/embeddings/flux_vae
27
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/recraft_data_crops/embeddings/flux_vae
28
+ - /data/kazanplova/datasets/full_latent_upscale_dataset_train/mjv5_crops_fix/embeddings/flux_vae
29
+ vae_names:
30
+ - flux_vae
31
+ phase: train
32
+ filename_tmpl: '{}'
33
+ io_backend:
34
+ type: disk
35
+ scale: 4
36
+ mean: null
37
+ std: null
38
+ num_worker_per_gpu: 32
39
+ batch_size_per_gpu: 256
40
+ pin_memory: true
41
+ persistent_workers: true
42
+ val:
43
+ name: sdxk_120_1024x1024
44
+ type: MultiScaleLatentCacheDataset
45
+ scales:
46
+ - 256
47
+ - 512
48
+ - 1024
49
+ cache_dirs:
50
+ - /data/kazanplova/datasets/latent_upscale_validation_120_samples/embeddings/flux_vae
51
+ vae_names:
52
+ - flux_vae
53
+ phase: val
54
+ io_backend:
55
+ type: disk
56
+ scale: 4
57
+ mean: null
58
+ std: null
59
+ batch_size_per_gpu: 16
60
+ num_worker_per_gpu: 4
61
+ pin_memory: true
62
+ network_g:
63
+ type: SwinIRMultiHead
64
+ in_chans: 16
65
+ img_size: 32
66
+ window_size: 16
67
+ img_range: 1.0
68
+ depths:
69
+ - 6
70
+ - 6
71
+ - 6
72
+ - 6
73
+ - 6
74
+ - 6
75
+ embed_dim: 360
76
+ num_heads:
77
+ - 12
78
+ - 12
79
+ - 12
80
+ - 12
81
+ - 12
82
+ - 12
83
+ mlp_ratio: 2
84
+ resi_connection: 1conv
85
+ primary_head: x4
86
+ head_num_feat: 256
87
+ heads:
88
+ - name: x2
89
+ scale: 2
90
+ out_chans: 16
91
+ - name: x4
92
+ scale: 4
93
+ out_chans: 16
94
+ primary: true
95
+ compile:
96
+ enabled: false
97
+ mode: max-autotune
98
+ dynamic: true
99
+ fullgraph: false
100
+ backend: null
101
+ train:
102
+ ema_decay: 0.999
103
+ optim_g:
104
+ type: Adam
105
+ lr: 0.0002
106
+ weight_decay: 0
107
+ betas:
108
+ - 0.9
109
+ - 0.995
110
+ grad_clip:
111
+ enabled: true
112
+ generator:
113
+ type: norm
114
+ max_norm: 0.4
115
+ norm_type: 2.0
116
+ scheduler:
117
+ type: MultiStepLR
118
+ milestones:
119
+ - 62500
120
+ - 93750
121
+ - 112500
122
+ gamma: 0.5
123
+ total_steps: 125000
124
+ warmup_iter: -1
125
+ l1_latent_x2_opt:
126
+ type: L1Loss
127
+ loss_weight: 1.0
128
+ reduction: mean
129
+ space: latent
130
+ target: x2
131
+ l1_latent_x4_opt:
132
+ type: L1Loss
133
+ loss_weight: 1.0
134
+ reduction: mean
135
+ space: latent
136
+ target: x4
137
+ fft_latent_x2_opt:
138
+ type: FFTFrequencyLoss
139
+ loss_weight: 0.1
140
+ reduction: mean
141
+ space: latent
142
+ target: x2
143
+ norm: ortho
144
+ use_log_amplitude: false
145
+ alpha: 0.0
146
+ normalize_weight: true
147
+ eps: 1e-8
148
+ fft_latent_x4_opt:
149
+ type: FFTFrequencyLoss
150
+ loss_weight: 0.1
151
+ reduction: mean
152
+ space: latent
153
+ target: x4
154
+ norm: ortho
155
+ use_log_amplitude: false
156
+ alpha: 0.0
157
+ normalize_weight: true
158
+ eps: 1e-8
159
+ val:
160
+ val_freq: 5000
161
+ save_img: true
162
+ head_evals:
163
+ x2:
164
+ save_img: true
165
+ label: val_x2
166
+ val_sizes:
167
+ lq: 512
168
+ gt: 1024
169
+ metrics:
170
+ l1_latent:
171
+ type: L1Loss
172
+ space: latent
173
+ pixel_psnr_pt:
174
+ type: calculate_psnr_pt
175
+ space: pixel
176
+ crop_border: 2
177
+ test_y_channel: false
178
+ x4:
179
+ save_img: true
180
+ label: val_x4
181
+ val_sizes:
182
+ lq: 256
183
+ gt: 1024
184
+ metrics:
185
+ l1_latent:
186
+ type: L1Loss
187
+ space: latent
188
+ l2_latent:
189
+ type: MSELoss
190
+ space: latent
191
+ pixel_psnr_pt:
192
+ type: calculate_psnr_pt
193
+ space: pixel
194
+ crop_border: 2
195
+ test_y_channel: false
196
+ logger:
197
+ print_freq: 100
198
+ save_checkpoint_freq: 5000
199
+ use_tb_logger: true
200
+ wandb:
201
+ project: Swin2SR-Latent-SR
202
+ entity: kazanplova-it-more
203
+ resume_id: null
204
+ max_val_images: 10
205
+ dist_params:
206
+ backend: nccl
207
+ port: 29500
208
+ dist: true
209
+ load_networks_only: false
210
+ exp_name: '32'
211
+ name: '32'
212
+ path:
213
+ experiments_root: /data/kazanplova/latent_vae_upscale_train/runs/01_11_2025