rootxhacker commited on
Commit
8cac667
·
verified ·
1 Parent(s): cf06ca7

Training in progress, step 2000, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:12b232b0278243d25903d8789b4dba3a6c48116f80cf44f29b224b3a9213825f
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed248223fb076f4557238ee5eec9add808ee83dacf213c5b42e548a3180d84ba
3
  size 36730224
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f87fd12b6e3592e8e2a065da27b2e13a73a88e79824ecd70f4ced521e1d64d89
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca187211ec5ca22e22494af8fae9ec9214346dc466486151b4b1df2b8f4958fe
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:42de46b2e4c5179833d46fe28b80f469c3d40255a80ec3a104ef7ad1fd3d8210
3
- size 14180
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2512ba1bd332fd1303d75224b66504a865cf77c5c151bdcdfcda52c23dbe2db6
3
+ size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:34bac16fbc99181cf3c30f05bcb7e648ad2c92eef3f334c8cbe294300c1d92e7
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8fdbdc59fd9e02ea3e69d676110893ccd87d22af41e77bc0d26c87371fa56871
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c205be8a1beb83fa60ead2e2dc0acbbf0de06ae533775758f9d9f2396e1a4003
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e942a77a12aac95a1f43f3b73ecd979987695991d41d93e1b6557c90b404597e
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 1500,
3
- "best_metric": 4.696172714233398,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-1500",
5
- "epoch": 0.1153757403276671,
6
  "eval_steps": 250,
7
- "global_step": 1500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -266,6 +266,92 @@
266
  "eval_samples_per_second": 52.993,
267
  "eval_steps_per_second": 13.248,
268
  "step": 1500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
269
  }
270
  ],
271
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 2000,
3
+ "best_metric": 4.620576858520508,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-2000",
5
+ "epoch": 0.15383432043688947,
6
  "eval_steps": 250,
7
+ "global_step": 2000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
266
  "eval_samples_per_second": 52.993,
267
  "eval_steps_per_second": 13.248,
268
  "step": 1500
269
+ },
270
+ {
271
+ "epoch": 0.11922159833858934,
272
+ "grad_norm": 3.801748752593994,
273
+ "learning_rate": 0.0001325679202970479,
274
+ "loss": 4.7912,
275
+ "step": 1550
276
+ },
277
+ {
278
+ "epoch": 0.12306745634951158,
279
+ "grad_norm": 8.367344856262207,
280
+ "learning_rate": 0.00013231057681702878,
281
+ "loss": 4.7281,
282
+ "step": 1600
283
+ },
284
+ {
285
+ "epoch": 0.12691331436043382,
286
+ "grad_norm": 4.299734592437744,
287
+ "learning_rate": 0.00013205323333700966,
288
+ "loss": 4.7263,
289
+ "step": 1650
290
+ },
291
+ {
292
+ "epoch": 0.13075917237135604,
293
+ "grad_norm": 6.152933597564697,
294
+ "learning_rate": 0.00013179588985699054,
295
+ "loss": 4.8519,
296
+ "step": 1700
297
+ },
298
+ {
299
+ "epoch": 0.13460503038227828,
300
+ "grad_norm": 4.300355434417725,
301
+ "learning_rate": 0.00013153854637697142,
302
+ "loss": 4.8359,
303
+ "step": 1750
304
+ },
305
+ {
306
+ "epoch": 0.13460503038227828,
307
+ "eval_loss": 4.635708808898926,
308
+ "eval_runtime": 18.5455,
309
+ "eval_samples_per_second": 53.922,
310
+ "eval_steps_per_second": 13.48,
311
+ "step": 1750
312
+ },
313
+ {
314
+ "epoch": 0.13845088839320052,
315
+ "grad_norm": 2.1330080032348633,
316
+ "learning_rate": 0.0001312812028969523,
317
+ "loss": 4.807,
318
+ "step": 1800
319
+ },
320
+ {
321
+ "epoch": 0.14229674640412276,
322
+ "grad_norm": 4.667717456817627,
323
+ "learning_rate": 0.00013102385941693318,
324
+ "loss": 4.6633,
325
+ "step": 1850
326
+ },
327
+ {
328
+ "epoch": 0.146142604415045,
329
+ "grad_norm": 6.904145240783691,
330
+ "learning_rate": 0.00013076651593691408,
331
+ "loss": 4.7899,
332
+ "step": 1900
333
+ },
334
+ {
335
+ "epoch": 0.14998846242596722,
336
+ "grad_norm": 2.930926561355591,
337
+ "learning_rate": 0.00013050917245689496,
338
+ "loss": 4.6692,
339
+ "step": 1950
340
+ },
341
+ {
342
+ "epoch": 0.15383432043688947,
343
+ "grad_norm": 3.6246345043182373,
344
+ "learning_rate": 0.00013025182897687584,
345
+ "loss": 4.781,
346
+ "step": 2000
347
+ },
348
+ {
349
+ "epoch": 0.15383432043688947,
350
+ "eval_loss": 4.620576858520508,
351
+ "eval_runtime": 18.7692,
352
+ "eval_samples_per_second": 53.279,
353
+ "eval_steps_per_second": 13.32,
354
+ "step": 2000
355
  }
356
  ],
357
  "logging_steps": 50,