Eviation commited on
Commit
7ccbc99
·
verified ·
1 Parent(s): 39c2024

Upload 43 files

Browse files
Files changed (44) hide show
  1. .gitattributes +14 -0
  2. 2xFullTae/fulltae.yml +207 -0
  3. 2xFullTae/models/net_g_ema_1200.safetensors +3 -0
  4. 2xFullTae/models/net_g_ema_144.safetensors +3 -0
  5. 2xFullTae/models/net_g_ema_157.safetensors +3 -0
  6. 2xFullTae/models/net_g_ema_1600.safetensors +3 -0
  7. 2xFullTae/models/net_g_ema_1708.safetensors +3 -0
  8. 2xFullTae/models/net_g_ema_1893.safetensors +3 -0
  9. 2xFullTae/models/net_g_ema_2000.safetensors +3 -0
  10. 2xFullTae/models/net_g_ema_2400.safetensors +3 -0
  11. 2xFullTae/models/net_g_ema_2800.safetensors +3 -0
  12. 2xFullTae/models/net_g_ema_3200.safetensors +3 -0
  13. 2xFullTae/models/net_g_ema_3600.safetensors +3 -0
  14. 2xFullTae/models/net_g_ema_3793.safetensors +3 -0
  15. 2xFullTae/models/net_g_ema_400.safetensors +3 -0
  16. 2xFullTae/models/net_g_ema_800.safetensors +3 -0
  17. 2xFullTae/models/resume_models/net_g_1200.safetensors +3 -0
  18. 2xFullTae/models/resume_models/net_g_144.safetensors +3 -0
  19. 2xFullTae/models/resume_models/net_g_157.safetensors +3 -0
  20. 2xFullTae/models/resume_models/net_g_1600.safetensors +3 -0
  21. 2xFullTae/models/resume_models/net_g_1708.safetensors +3 -0
  22. 2xFullTae/models/resume_models/net_g_1893.safetensors +3 -0
  23. 2xFullTae/models/resume_models/net_g_2000.safetensors +3 -0
  24. 2xFullTae/models/resume_models/net_g_2400.safetensors +3 -0
  25. 2xFullTae/models/resume_models/net_g_2800.safetensors +3 -0
  26. 2xFullTae/models/resume_models/net_g_3200.safetensors +3 -0
  27. 2xFullTae/models/resume_models/net_g_3600.safetensors +3 -0
  28. 2xFullTae/models/resume_models/net_g_3793.safetensors +3 -0
  29. 2xFullTae/models/resume_models/net_g_400.safetensors +3 -0
  30. 2xFullTae/models/resume_models/net_g_800.safetensors +3 -0
  31. 2xFullTae/training_states/1200.state +3 -0
  32. 2xFullTae/training_states/144.state +3 -0
  33. 2xFullTae/training_states/157.state +3 -0
  34. 2xFullTae/training_states/1600.state +3 -0
  35. 2xFullTae/training_states/1708.state +3 -0
  36. 2xFullTae/training_states/1893.state +3 -0
  37. 2xFullTae/training_states/2000.state +3 -0
  38. 2xFullTae/training_states/2400.state +3 -0
  39. 2xFullTae/training_states/2800.state +3 -0
  40. 2xFullTae/training_states/3200.state +3 -0
  41. 2xFullTae/training_states/3600.state +3 -0
  42. 2xFullTae/training_states/3793.state +3 -0
  43. 2xFullTae/training_states/400.state +3 -0
  44. 2xFullTae/training_states/800.state +3 -0
.gitattributes CHANGED
@@ -438,3 +438,17 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
438
  1x_NoiseMakerSoft/training_states/7000.state filter=lfs diff=lfs merge=lfs -text
439
  1x_NoiseMakerSoft/training_states/8000.state filter=lfs diff=lfs merge=lfs -text
440
  1x_NoiseMakerSoft/training_states/9000.state filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
438
  1x_NoiseMakerSoft/training_states/7000.state filter=lfs diff=lfs merge=lfs -text
439
  1x_NoiseMakerSoft/training_states/8000.state filter=lfs diff=lfs merge=lfs -text
440
  1x_NoiseMakerSoft/training_states/9000.state filter=lfs diff=lfs merge=lfs -text
441
+ 2xFullTae/training_states/1200.state filter=lfs diff=lfs merge=lfs -text
442
+ 2xFullTae/training_states/144.state filter=lfs diff=lfs merge=lfs -text
443
+ 2xFullTae/training_states/157.state filter=lfs diff=lfs merge=lfs -text
444
+ 2xFullTae/training_states/1600.state filter=lfs diff=lfs merge=lfs -text
445
+ 2xFullTae/training_states/1708.state filter=lfs diff=lfs merge=lfs -text
446
+ 2xFullTae/training_states/1893.state filter=lfs diff=lfs merge=lfs -text
447
+ 2xFullTae/training_states/2000.state filter=lfs diff=lfs merge=lfs -text
448
+ 2xFullTae/training_states/2400.state filter=lfs diff=lfs merge=lfs -text
449
+ 2xFullTae/training_states/2800.state filter=lfs diff=lfs merge=lfs -text
450
+ 2xFullTae/training_states/3200.state filter=lfs diff=lfs merge=lfs -text
451
+ 2xFullTae/training_states/3600.state filter=lfs diff=lfs merge=lfs -text
452
+ 2xFullTae/training_states/3793.state filter=lfs diff=lfs merge=lfs -text
453
+ 2xFullTae/training_states/400.state filter=lfs diff=lfs merge=lfs -text
454
+ 2xFullTae/training_states/800.state filter=lfs diff=lfs merge=lfs -text
2xFullTae/fulltae.yml ADDED
@@ -0,0 +1,207 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # GENERATE TIME: Mon Jul 21 16:37:11 2025
2
+ # CMD:
3
+ # train.py --auto_resume -opt options\fulltae.yml
4
+
5
+ # GENERATE TIME: Sat Jun 14 09:42:06 2025
6
+ # CMD:
7
+ # train.py --auto_resume -opt options/train/RealPLKSR/2xOSISRDv2_realplksr_dysample_layernorm.yml
8
+
9
+ # yaml-language-server: $schema=https://raw.githubusercontent.com/the-database/traiNNer-redux/refs/heads/master/schemas/redux-config.schema.json
10
+ #########################################################################################
11
+ # General Settings
12
+ # https://trainner-redux.readthedocs.io/en/latest/config_reference.html#top-level-options
13
+ #########################################################################################
14
+ name: 2xFullTae
15
+ scale: 2 # 1, 2, 3, 4, 8
16
+ use_amp: false # Speed up training and reduce VRAM usage. NVIDIA only.
17
+ amp_bf16: false # Use bf16 instead of fp16 for AMP, RTX 3000 series or newer only. Only recommended if fp16 doesn't work.
18
+ use_channels_last: true # Enable channels last memory format while using AMP. Reduces VRAM and speeds up training for most architectures, but some architectures are slower with channels last.
19
+ fast_matmul: false # Trade precision for performance.
20
+ num_gpu: auto
21
+ # manual_seed: 1024 # Random seed for training, useful for removing randomness when testing the effect of different settings.
22
+
23
+
24
+ ########################################################################################################################
25
+ # Dataset and Dataloader Settings
26
+ # https://trainner-redux.readthedocs.io/en/latest/config_reference.html#dataset-options-datasets-train-and-datasets-val
27
+ ########################################################################################################################
28
+ datasets:
29
+ # Settings for the training dataset.
30
+ train:
31
+ name: ABBBCDLNRRRT
32
+ type: pairedimagedataset
33
+ # Path to the HR (high res) images in your training dataset. Specify one or multiple folders, separated by commas.
34
+ dataroot_gt: [
35
+ G:\\activebigout\\2048,G:\\batch2bigout\\2048,G:\\bikinibigout\\2048,G:\\brabigout\\2048,G:\\corsetbigout\\2048,G:\\dressbigout\\2048,G:\\lingeriebigout\\2048,G:\\nudebigout\\2048,G:\\redonebigout\\2048,G:\\redtwobigout\\2048,G:\\restbigout\\2048,G:\\topsbigout\\2048
36
+ ]
37
+ # Path to the LR (low res) images in your training dataset. Specify one or multiple folders, separated by commas.
38
+ dataroot_lq: [
39
+ G:\\activebigout\\1024,G:\\batch2bigout\\1024,G:\\bikinibigout\\1024,G:\\brabigout\\1024,G:\\corsetbigout\\1024,G:\\dressbigout\\1024,G:\\lingeriebigout\\1024,G:\\nudebigout\\1024,G:\\redonebigout\\1024,G:\\redtwobigout\\1024,G:\\restbigout\\1024,G:\\topsbigout\\1024
40
+ ]
41
+ # meta_info: data/meta_info/dataset1.txt
42
+
43
+
44
+ lq_size: 96 # During training, a square of this size is cropped from LR images. Larger is usually better but uses more VRAM. Previously gt_size, use lq_size = gt_size / scale to convert. Use multiple of 8 for best performance with AMP.
45
+ use_hflip: true # Randomly flip the images horizontally.
46
+ use_rot: true # Randomly rotate the images.
47
+
48
+ num_worker_per_gpu: 1
49
+ batch_size_per_gpu: 1 # paper: 64 # Increasing stabilizes training but with diminishing returns. Use multiple of 8 for best performance with AMP.
50
+ accum_iter: 16 # Using values larger than 1 simulates higher batch size by trading performance for reduced VRAM usage. If accum_iter = 4 and batch_size_per_gpu = 6 then effective batch size = 4 * 6 = 24 but performance may be as much as 4 times as slow.
51
+ # Settings for your validation dataset (optional). These settings will
52
+ # be ignored if val_enabled is false in the Validation section below.
53
+ val:
54
+ name: Val Dataset
55
+ type: pairedimagedataset
56
+ dataroot_gt: [
57
+ G:\\valbigout\\2048
58
+ ]
59
+ dataroot_lq: [
60
+ G:\\valbigout\\1024
61
+ ]
62
+
63
+ #####################################################################
64
+ # Network Settings
65
+ # https://trainner-redux.readthedocs.io/en/latest/arch_reference.html
66
+ #####################################################################
67
+ # Generator model settings
68
+ network_g:
69
+ type: realplksr
70
+ upsampler: dysample # pixelshuffle, dysample (better quality on even number scales, but does not support dynamic onnx)
71
+ layer_norm: true # better quality, not compatible with older models
72
+
73
+ #########################################################################################
74
+ # Pretrain and Resume Paths
75
+ # https://trainner-redux.readthedocs.io/en/latest/config_reference.html#path-options-path
76
+ #########################################################################################
77
+ path:
78
+ # pretrain_network_g: # experiments/2xAllBigVal/models/net_g_ema_8400.safetensors
79
+ param_key_g: ~
80
+ strict_load_g: true # Disable strict loading to partially load a pretrain model with a different scale
81
+ resume_state: ~
82
+
83
+ ###########################################################################################
84
+ # Training Settings
85
+ # https://trainner-redux.readthedocs.io/en/latest/config_reference.html#train-options-train
86
+ ###########################################################################################
87
+ train:
88
+ ema_decay: 0.999
89
+ ema_power: 0.75 # Gradually warm up ema decay when training from scratch
90
+ grad_clip: false # Gradient clipping allows more stable training when using higher learning rates.
91
+ # Optimizer for generator model
92
+ optim_g:
93
+ type: AdamW
94
+ lr: !!float 5e-5
95
+ weight_decay: 0.01
96
+ betas: [0.9, 0.99]
97
+
98
+ scheduler:
99
+ type: MultiStepLR
100
+ milestones: [100000, 200000, 300000, 400000, 425000]
101
+ gamma: 0.5
102
+
103
+ total_iter: 50000 # Total number of iterations.
104
+ warmup_iter: -1 # Gradually ramp up learning rates until this iteration, to stabilize early training. Use -1 to disable.
105
+
106
+ # Losses - for any loss set the loss_weight to 0 to disable it.
107
+ # https://trainner-redux.readthedocs.io/en/latest/loss_reference.html
108
+ losses:
109
+ # TAESD
110
+ - type: taesd
111
+ criterion: charbonnier # charbonnier, l1, l2
112
+ loss_weight: 0.1 # was 0.5 until 1709
113
+
114
+ # Pixel-level foundation
115
+ - type: charbonnierloss
116
+ loss_weight: 0.1 # Up from 0.05 (noise suppression) 68,500 iters
117
+ eps: !!float 1e-2 # More tolerant of natural noise Prevents model from "inventing" patterns to hide noise 68,500 iters
118
+
119
+ # Structural integrity
120
+ - type: mssimloss
121
+ loss_weight: 0.3
122
+ channels: 3
123
+ downsample: false
124
+ test_y_channel: false
125
+ color_space: rgb
126
+
127
+ # Enhanced perceptual loss
128
+ - type: perceptualloss
129
+ loss_weight: 0.1 # Slightly reduced (Modify perceptual focus) 68,500 iters
130
+ layer_weights:
131
+ conv2_2: 0.3 # Lower-level features Added conv2_2 focuses on broader textures 68,500 iters
132
+ conv3_4: 0.5 # Reduced weight Reduces pressure to create high-frequency details everywhere 68,500 iters
133
+ conv4_4: 0.2
134
+ criterion: charbonnier
135
+ num_proj_fd: 128
136
+ stride_fd: 2
137
+
138
+ # Frequency-domain focus (now carries more weight)
139
+ - type: ffloss
140
+ loss_weight: 0.55 # up from 0.3 # Down from 0.55 at 68,500 iters. To fix dot pattern (over-emphasize high frequencies / Creates synthetic patterns in texture-less regions).Lower weight prevents artificial frequency amplification Stops creating "phantom" patterns in noisy regions
141
+ alpha: 1.0
142
+ patch_factor: 1
143
+ ave_spectrum: false
144
+ log_matrix: true
145
+
146
+ # Local detail preservation
147
+ - type: ldlloss
148
+ loss_weight: 0.3 # down from 0.8
149
+ criterion: charbonnier
150
+
151
+ # Gradient variance for edge enhancement
152
+ - type: gradientvarianceloss
153
+ patch_size: 24 # Larger patches for smoother textures for texture regularization 68,500 iters
154
+ criterion: charbonnier
155
+ loss_weight: 0.25 # Increased from 0.15 at 68,500 iters for texture regularization
156
+
157
+ # # GAN for realistic textures
158
+ # - type: ganloss
159
+ # gan_type: vanilla
160
+ # loss_weight: 0.05
161
+ # real_label_val: 0.9
162
+ # fake_label_val: 0.1
163
+
164
+ # Color consistency
165
+ - type: cosimloss
166
+ loss_weight: 0.3
167
+ cosim_lambda: 3
168
+
169
+ ##############################################################################################
170
+ # Validation
171
+ # https://trainner-redux.readthedocs.io/en/latest/config_reference.html#validation-options-val
172
+ ##############################################################################################
173
+ val:
174
+ val_enabled: true # Whether to enable validations. If disabled, all validation settings below are ignored.
175
+ val_freq: 290 # How often to run validations, in iterations.
176
+ save_img: true # Whether to save the validation images during validation, in the experiments/<name>/visualization folder.
177
+ tile_size: 128 # Tile size of input, reduce VRAM usage but slower inference. 0 to disable.
178
+ tile_overlap: 8 # Number of pixels to overlap tiles by, larger is slower but reduces tile seams.
179
+
180
+ metrics_enabled: true # Whether to run metrics calculations during validation.
181
+ metrics:
182
+ #topiq:
183
+ # type: calculate_topiq
184
+ psnr:
185
+ type: calculate_psnr
186
+ crop_border: 4
187
+ test_y_channel: false
188
+ ssim:
189
+ type: calculate_ssim
190
+ crop_border: 4 # Whether to crop border during validation.
191
+ test_y_channel: false # Whether to convert to Y(CbCr) for validation.
192
+ lpips:
193
+ type: calculate_lpips
194
+ better: lower
195
+ #dists:
196
+ # type: calculate_dists
197
+ # better: lower
198
+
199
+ ##############################################################################################
200
+ # Logging
201
+ # https://trainner-redux.readthedocs.io/en/latest/config_reference.html#logging-options-logger
202
+ ##############################################################################################
203
+ logger:
204
+ print_freq: 10
205
+ save_checkpoint_freq: 400
206
+ save_checkpoint_format: safetensors
207
+ use_tb_logger: true
2xFullTae/models/net_g_ema_1200.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40bc5f85e09d981481960fce51faa9149b6a50f3d430eef54a7522d0e51187bd
3
+ size 29511708
2xFullTae/models/net_g_ema_144.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:042b3b8f3ccfcbd102d11de194ce45ac43bc6016d48390223dcd5c2043f0bb47
3
+ size 29511708
2xFullTae/models/net_g_ema_157.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87ab2d3215202fe62af75ebaf6e458226a810136551cb0bd32fe1c4f93413472
3
+ size 29511708
2xFullTae/models/net_g_ema_1600.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86b9cc6534d13b8ae7b17fbccb76e68bce8e919ee0d863fda2299354c9d914e5
3
+ size 29511708
2xFullTae/models/net_g_ema_1708.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f53d1e51f97671a0276f5c745266aeb5a10085ee56decf766e33af1334dd6ae
3
+ size 29511708
2xFullTae/models/net_g_ema_1893.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49cad82e2a483e318b9a192751f9896ea580b4b752901971a0ead7988c4b2dd9
3
+ size 29511708
2xFullTae/models/net_g_ema_2000.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a0bd4b90dbc4c485234ce490ed0c50a8893c8e2be926d3166288e05da9d877f
3
+ size 29511708
2xFullTae/models/net_g_ema_2400.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:699935060f8a7f89beb2535aa08c77117618e3a3f0768d4172d00d7e6ad942da
3
+ size 29511708
2xFullTae/models/net_g_ema_2800.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6aefcb9ee258db1d27ded952368bb0e3e2d492b3ef4a840fb8d601ccf71755ad
3
+ size 29511708
2xFullTae/models/net_g_ema_3200.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2d6356d6a1e696301b23cd4654d994083be009ffef40e34d8a6b76bc7e4d03b
3
+ size 29511708
2xFullTae/models/net_g_ema_3600.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd9c2bc4c4c03c5b38dacc0039b57782e879da9592be255414d843c691763ccb
3
+ size 29511708
2xFullTae/models/net_g_ema_3793.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ef86c2fd9f9d356fcdb3ebd6cb401bde0c7a8530ad5c3c99b2154412309f9b0
3
+ size 29511708
2xFullTae/models/net_g_ema_400.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0806708c414073848fbce3e6d2082b317878c8d6fc9dd01663886991a6431dde
3
+ size 29511708
2xFullTae/models/net_g_ema_800.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a63ec178a3251159b6c0ec97e0aadb01a36e20e47b3c34929ae4561db327d75
3
+ size 29511708
2xFullTae/models/resume_models/net_g_1200.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:438f2ebe684ccdd10e1f4eef56f1cfe95e5b00f55e6462bb4d6b335e445c0744
3
+ size 29511708
2xFullTae/models/resume_models/net_g_144.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b824867f52a3457d5a7574ff7df5af4df5faa128eba6985c8ea5b734b048dfdf
3
+ size 29511708
2xFullTae/models/resume_models/net_g_157.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8cfeb50060d9aee9deb3df1d8ebca798bb53e2253a55e654a6ce6e5bd8daa6d
3
+ size 29511708
2xFullTae/models/resume_models/net_g_1600.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30ca18b5617b174fb6c69b35a4da67c203ab5670288286837011cba69995650e
3
+ size 29511708
2xFullTae/models/resume_models/net_g_1708.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66bab770720afaf0d9b19e278fe3dd6d51cee0ff96f03dbae6ad926fc9a66bc2
3
+ size 29511708
2xFullTae/models/resume_models/net_g_1893.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b303a9939480045961c1c8a34f3624d59e1e7fe7fe01bea6b4ae2ed171d7126
3
+ size 29511708
2xFullTae/models/resume_models/net_g_2000.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea72fa162b3ee8c89642af58d78618c7eea6c4a0a7a412a1e708777a46c95192
3
+ size 29511708
2xFullTae/models/resume_models/net_g_2400.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:562e31b72e8e918d4667dadd7ac6179c0939c4f2686b0feac3cbceaeac7fa727
3
+ size 29511708
2xFullTae/models/resume_models/net_g_2800.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c10565bdba55da4480ee4156e4cf9c09af51c95cac4376b3ec1584b8b046369
3
+ size 29511708
2xFullTae/models/resume_models/net_g_3200.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c100dfb5db57fdc3f2a79eb3b68fefe7b738331ae3daa5f7f6ff31e0efdcb30c
3
+ size 29511708
2xFullTae/models/resume_models/net_g_3600.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:719abb2729d3b7abb2ceb22d395f27b55fd39bcb86204f53ab6bc5e412d58bd5
3
+ size 29511708
2xFullTae/models/resume_models/net_g_3793.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:125bce05dde92708d60079622170aabb5b7408c1b3ef58378c5eb314a3150ce1
3
+ size 29511708
2xFullTae/models/resume_models/net_g_400.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1eeb2cd2093a6731cc363f129eb9d7bb3be6b930245001a91858a4e1ab7d17b
3
+ size 29511708
2xFullTae/models/resume_models/net_g_800.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44ae0b53e7129c4a069fdf379a7b587a7d522a141bc686933810dcb4e1346f1e
3
+ size 29511708
2xFullTae/training_states/1200.state ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ec08213a7069aa40a5643fd019b00bee49312a4985f537853782887f1539bd8
3
+ size 59199829
2xFullTae/training_states/144.state ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e2e18ff7b5648721bc931f05013acaee25d2a56058b13540b9f9d5f25ccfe63
3
+ size 59198595
2xFullTae/training_states/157.state ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22414e1ff6a5b2740fa3935928062399669b3fe2322bd8632e568eded06a8cfb
3
+ size 59198595
2xFullTae/training_states/1600.state ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f3df8a84f3c9358535986c6a76fd3437adaabb33a16e0bd9197fae5c8195eff
3
+ size 59199829
2xFullTae/training_states/1708.state ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b131dbed04d8c7b75bdfc5900075cfd4b487a5ecf5db98481bf7cc96e2862e75
3
+ size 59199829
2xFullTae/training_states/1893.state ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31f56dd48e8fe5f2d1e0a5b4c3b394c9dba98c013ff5be74f37bdcf6d3f6b561
3
+ size 59199829
2xFullTae/training_states/2000.state ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1eb3f11e23c1bc0637e2207333effe2f720b4f4f383d20c24177d9d6bf3afd8
3
+ size 59199829
2xFullTae/training_states/2400.state ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:850a6ce34f73aaa2e9fe95d87afc38ccbbe4c7d962b439a4ebe08fe08674d0ab
3
+ size 59199829
2xFullTae/training_states/2800.state ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64cd230a17c4252b2bcb83c826a33c74cf33d652ccd3b414e3d4fef725cc4fa6
3
+ size 59199829
2xFullTae/training_states/3200.state ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c28b154ae2bd517d4e75a68ae28c040ab69b310c80d704a7e5ef3ecc9392f8c
3
+ size 59199829
2xFullTae/training_states/3600.state ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b1892fe950577c88cb5f6ead471006bed3089c4cf47fedd1646fd2cb14b7b4a
3
+ size 59199829
2xFullTae/training_states/3793.state ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37d61c9349d65a1bad90ce79c2496248040f8166a80747e319c95bcc78d60e2d
3
+ size 59199829
2xFullTae/training_states/400.state ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2d48350410df09963da5dcb62d5e1410583fee4efa41b5675af3a3cd734ba6b
3
+ size 59198595
2xFullTae/training_states/800.state ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cabcc452ec03b7c5425f19892809e2adc35deb9b8df874a70448c2ad0c66ea57
3
+ size 59198595