rootxhacker commited on
Commit
08d2d35
·
verified ·
1 Parent(s): 6f8d066

Training in progress, step 36500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:16887918a2514cde38397627644e79ad7a7cc859123d9755cbcacbf31a8c9f7d
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f00ee81c87f16a8bdea12f78335f236b7be0e12be02042e846ac3c0d151355c
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fc4a7c0e93c6189640f847962e46079ebcf94653f638d4fdd14079010baf0e65
3
  size 1736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efdccb23bf5e306d948f48beb9315429633d26670a250b2f01f01a906be73107
3
  size 1736
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:266a495e1a502f1d39b57b5fcb15d594e7ebb88b9a37c10fcde88037b918f14e
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d4a503089b873fa58fe7ad923dc676b15a066d799ced08a885cce4c8b7d5b5a
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6c600e0bd51e53276481f2f5ca4e889540a7504baebe1c05ac1ae4c81b5b17f3
3
- size 14180
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5d0f591d999e8283f8ffbd498c9dcefdfec3c57a9cf9cddf8c2ec79814e55a9
3
+ size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:49ec2d3f3b18676d40b30b5547947dd759a32e3455d479b941fbf8c1dd3301d4
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6a361b5fce627657860521eb50e2574f27b2d4131fa370a849ae347ab3399c1
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:906c43e30b9edc8f130b3f77317e01f206c6e5f1267c22ca4899b9c5968ba8ca
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0cf91a495853427b5f841ab1be3c35766ad0edbc12ff06e2c15160a47811ed22
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 35500,
3
- "best_metric": -30.469755172729492,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-35500",
5
- "epoch": 2.7690177678640104,
6
  "eval_steps": 250,
7
- "global_step": 36000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -6200,6 +6200,92 @@
6200
  "eval_samples_per_second": 59.708,
6201
  "eval_steps_per_second": 14.927,
6202
  "step": 36000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6203
  }
6204
  ],
6205
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 36500,
3
+ "best_metric": -30.47022247314453,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-36500",
5
+ "epoch": 2.8074763479732328,
6
  "eval_steps": 250,
7
+ "global_step": 36500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
6200
  "eval_samples_per_second": 59.708,
6201
  "eval_steps_per_second": 14.927,
6202
  "step": 36000
6203
+ },
6204
+ {
6205
+ "epoch": 2.772863625874933,
6206
+ "grad_norm": 2.303079128265381,
6207
+ "learning_rate": 1.5422174895462695e-05,
6208
+ "loss": -30.4658,
6209
+ "step": 36050
6210
+ },
6211
+ {
6212
+ "epoch": 2.776709483885855,
6213
+ "grad_norm": 1.1409249305725098,
6214
+ "learning_rate": 1.516245487364621e-05,
6215
+ "loss": -30.4659,
6216
+ "step": 36100
6217
+ },
6218
+ {
6219
+ "epoch": 2.780555341896777,
6220
+ "grad_norm": 0.9111425280570984,
6221
+ "learning_rate": 1.4902734851829727e-05,
6222
+ "loss": -30.4658,
6223
+ "step": 36150
6224
+ },
6225
+ {
6226
+ "epoch": 2.784401199907699,
6227
+ "grad_norm": 1.3415731191635132,
6228
+ "learning_rate": 1.4643014830013247e-05,
6229
+ "loss": -30.4659,
6230
+ "step": 36200
6231
+ },
6232
+ {
6233
+ "epoch": 2.7882470579186216,
6234
+ "grad_norm": 0.8532393574714661,
6235
+ "learning_rate": 1.4383294808196765e-05,
6236
+ "loss": -30.4658,
6237
+ "step": 36250
6238
+ },
6239
+ {
6240
+ "epoch": 2.7882470579186216,
6241
+ "eval_loss": -30.469867706298828,
6242
+ "eval_runtime": 17.0192,
6243
+ "eval_samples_per_second": 58.757,
6244
+ "eval_steps_per_second": 14.689,
6245
+ "step": 36250
6246
+ },
6247
+ {
6248
+ "epoch": 2.792092915929544,
6249
+ "grad_norm": 1.6142336130142212,
6250
+ "learning_rate": 1.4123574786380284e-05,
6251
+ "loss": -30.4662,
6252
+ "step": 36300
6253
+ },
6254
+ {
6255
+ "epoch": 2.795938773940466,
6256
+ "grad_norm": 1.9979597330093384,
6257
+ "learning_rate": 1.3863854764563799e-05,
6258
+ "loss": -30.4652,
6259
+ "step": 36350
6260
+ },
6261
+ {
6262
+ "epoch": 2.7997846319513884,
6263
+ "grad_norm": 1.4881914854049683,
6264
+ "learning_rate": 1.3604134742747318e-05,
6265
+ "loss": -30.466,
6266
+ "step": 36400
6267
+ },
6268
+ {
6269
+ "epoch": 2.803630489962311,
6270
+ "grad_norm": 2.0614233016967773,
6271
+ "learning_rate": 1.3344414720930836e-05,
6272
+ "loss": -30.4658,
6273
+ "step": 36450
6274
+ },
6275
+ {
6276
+ "epoch": 2.8074763479732328,
6277
+ "grad_norm": 1.0104310512542725,
6278
+ "learning_rate": 1.3084694699114356e-05,
6279
+ "loss": -30.466,
6280
+ "step": 36500
6281
+ },
6282
+ {
6283
+ "epoch": 2.8074763479732328,
6284
+ "eval_loss": -30.47022247314453,
6285
+ "eval_runtime": 16.8782,
6286
+ "eval_samples_per_second": 59.248,
6287
+ "eval_steps_per_second": 14.812,
6288
+ "step": 36500
6289
  }
6290
  ],
6291
  "logging_steps": 50,