rootxhacker commited on
Commit
72cfdb6
·
verified ·
1 Parent(s): e2c9c99

Training in progress, step 7500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:52b997298d45a4fc6cd35ecfe5ccf8a749558ac28f75afb2d0909fe1d125004b
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1db6fd6ee1d14415c983da271a1a5fced7a19d546dcff5e9c52892498d1357ea
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f95ce72a1c3a7c4d5aab7200d9dec95f462be645783ddec5e51a5155488e47cc
3
  size 1736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90e5824e7895d4d885fcae3d6d27b58bf56f7ed9334ac10f6cfbd2e2106a176c
3
  size 1736
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:215f3ea94412e8b86475daca81b7346b3039dd518294dd8628b764b4fe24e130
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:636d50d708bc2858a479246fdb19d62e4927d92e0d54e0a5e51e6cc5b17b5a5c
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:517f7f125f44f7edc89c0d28a191b1cd6fb9d0b94d17a4ddcd53f5fcaa8aefcd
3
  size 14180
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:340279b529434b0154ba013e9a01bef12d3018e0cbfd933ce4ed580c0bdf70f8
3
  size 14180
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7df83cd74eb098f609a6963afc6cf853a656c30456c390df44a7c43c2165d798
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d034bc1613cdbec4742e0b9343e54d205e633fa15e81421b5d524ce6b4165764
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:93f284fcd12d6e29e5cbdf1e4a0021a53156461eb5dba25291a9ab08dfb3c1a8
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2541b57467e961f49663f187822c9553808b853819403f17ca1464ff8f14b868
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 7000,
3
- "best_metric": 1.4396251440048218,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-7000",
5
- "epoch": 0.5384201215291131,
6
  "eval_steps": 250,
7
- "global_step": 7000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -1212,6 +1212,92 @@
1212
  "eval_samples_per_second": 59.225,
1213
  "eval_steps_per_second": 14.806,
1214
  "step": 7000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1215
  }
1216
  ],
1217
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 7500,
3
+ "best_metric": 1.424470067024231,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-7500",
5
+ "epoch": 0.5768787016383355,
6
  "eval_steps": 250,
7
+ "global_step": 7500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
1212
  "eval_samples_per_second": 59.225,
1213
  "eval_steps_per_second": 14.806,
1214
  "step": 7000
1215
+ },
1216
+ {
1217
+ "epoch": 0.5422659795400354,
1218
+ "grad_norm": 1.6438534259796143,
1219
+ "learning_rate": 0.00016600264914422253,
1220
+ "loss": 1.4556,
1221
+ "step": 7050
1222
+ },
1223
+ {
1224
+ "epoch": 0.5461118375509576,
1225
+ "grad_norm": 1.0202991962432861,
1226
+ "learning_rate": 0.00016574292912240605,
1227
+ "loss": 1.4659,
1228
+ "step": 7100
1229
+ },
1230
+ {
1231
+ "epoch": 0.5499576955618799,
1232
+ "grad_norm": 1.1861631870269775,
1233
+ "learning_rate": 0.0001654884035010259,
1234
+ "loss": 1.4615,
1235
+ "step": 7150
1236
+ },
1237
+ {
1238
+ "epoch": 0.5538035535728021,
1239
+ "grad_norm": 1.3344788551330566,
1240
+ "learning_rate": 0.0001652286834792094,
1241
+ "loss": 1.4654,
1242
+ "step": 7200
1243
+ },
1244
+ {
1245
+ "epoch": 0.5576494115837243,
1246
+ "grad_norm": 1.6251221895217896,
1247
+ "learning_rate": 0.00016496896345739295,
1248
+ "loss": 1.5118,
1249
+ "step": 7250
1250
+ },
1251
+ {
1252
+ "epoch": 0.5576494115837243,
1253
+ "eval_loss": 1.4305483102798462,
1254
+ "eval_runtime": 16.9313,
1255
+ "eval_samples_per_second": 59.062,
1256
+ "eval_steps_per_second": 14.766,
1257
+ "step": 7250
1258
+ },
1259
+ {
1260
+ "epoch": 0.5614952695946466,
1261
+ "grad_norm": 1.6725965738296509,
1262
+ "learning_rate": 0.00016470924343557646,
1263
+ "loss": 1.4391,
1264
+ "step": 7300
1265
+ },
1266
+ {
1267
+ "epoch": 0.5653411276055688,
1268
+ "grad_norm": 1.2355259656906128,
1269
+ "learning_rate": 0.00016444952341375998,
1270
+ "loss": 1.4197,
1271
+ "step": 7350
1272
+ },
1273
+ {
1274
+ "epoch": 0.5691869856164911,
1275
+ "grad_norm": 2.637535572052002,
1276
+ "learning_rate": 0.0001641898033919435,
1277
+ "loss": 1.4445,
1278
+ "step": 7400
1279
+ },
1280
+ {
1281
+ "epoch": 0.5730328436274132,
1282
+ "grad_norm": 0.877162516117096,
1283
+ "learning_rate": 0.000163930083370127,
1284
+ "loss": 1.4575,
1285
+ "step": 7450
1286
+ },
1287
+ {
1288
+ "epoch": 0.5768787016383355,
1289
+ "grad_norm": 1.3708187341690063,
1290
+ "learning_rate": 0.00016367036334831052,
1291
+ "loss": 1.4965,
1292
+ "step": 7500
1293
+ },
1294
+ {
1295
+ "epoch": 0.5768787016383355,
1296
+ "eval_loss": 1.424470067024231,
1297
+ "eval_runtime": 16.8553,
1298
+ "eval_samples_per_second": 59.328,
1299
+ "eval_steps_per_second": 14.832,
1300
+ "step": 7500
1301
  }
1302
  ],
1303
  "logging_steps": 50,