ljcamargo commited on
Commit
682e349
·
verified ·
1 Parent(s): 3871245

Training in progress, step 800, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1e9c89c5afa1545ed2abce6fb61b801899049ea15f9cad45a8a50d3ec0f9d3c1
3
  size 3826461296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f9a757842f36fd94aead624d11ae735bab0021a2cf1d22b12f1d19d8eb3745df
3
  size 3826461296
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8b0b1b8a92e051537f0d9657a16e5d51d9f5dddc753f3ecef5357cb38fad4fca
3
  size 2479955235
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:864f28f2db139c235def5468478555614e1208e6c7e1636a6be1a9a8a84d2903
3
  size 2479955235
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d166d430557d2569c721c8dec1c8ddf3bfe3fec272b03dceb3e3268be418ae2c
3
  size 1383
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee1ff4b6d230d52fbd75a1fdfc717e2baaa7034a01541c1dee54a5bf5dd662d6
3
  size 1383
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:55fdec1914c1bee37a4826392246117bd3062dd019b2d4c1b1f435e39b62b9ce
3
  size 1465
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87f9876af7981b4f995b217441438c53a026fb406c344a1e30a18ad2545bd292
3
  size 1465
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 0.28,
6
  "eval_steps": 500,
7
- "global_step": 700,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -498,6 +498,76 @@
498
  "learning_rate": 3.6254019292604503e-05,
499
  "loss": 0.7662,
500
  "step": 700
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
501
  }
502
  ],
503
  "logging_steps": 10,
@@ -517,7 +587,7 @@
517
  "attributes": {}
518
  }
519
  },
520
- "total_flos": 1.2633123965792256e+16,
521
  "train_batch_size": 2,
522
  "trial_name": null,
523
  "trial_params": null
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 0.32,
6
  "eval_steps": 500,
7
+ "global_step": 800,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
498
  "learning_rate": 3.6254019292604503e-05,
499
  "loss": 0.7662,
500
  "step": 700
501
+ },
502
+ {
503
+ "epoch": 0.284,
504
+ "grad_norm": 25.36922264099121,
505
+ "learning_rate": 3.605305466237942e-05,
506
+ "loss": 0.7945,
507
+ "step": 710
508
+ },
509
+ {
510
+ "epoch": 0.288,
511
+ "grad_norm": 24.251853942871094,
512
+ "learning_rate": 3.585209003215435e-05,
513
+ "loss": 0.9693,
514
+ "step": 720
515
+ },
516
+ {
517
+ "epoch": 0.292,
518
+ "grad_norm": 15.235057830810547,
519
+ "learning_rate": 3.5651125401929266e-05,
520
+ "loss": 0.8969,
521
+ "step": 730
522
+ },
523
+ {
524
+ "epoch": 0.296,
525
+ "grad_norm": 14.464040756225586,
526
+ "learning_rate": 3.5450160771704185e-05,
527
+ "loss": 0.7205,
528
+ "step": 740
529
+ },
530
+ {
531
+ "epoch": 0.3,
532
+ "grad_norm": 23.044424057006836,
533
+ "learning_rate": 3.5249196141479104e-05,
534
+ "loss": 0.839,
535
+ "step": 750
536
+ },
537
+ {
538
+ "epoch": 0.304,
539
+ "grad_norm": 25.620925903320312,
540
+ "learning_rate": 3.504823151125402e-05,
541
+ "loss": 0.8887,
542
+ "step": 760
543
+ },
544
+ {
545
+ "epoch": 0.308,
546
+ "grad_norm": 10.347396850585938,
547
+ "learning_rate": 3.484726688102894e-05,
548
+ "loss": 0.645,
549
+ "step": 770
550
+ },
551
+ {
552
+ "epoch": 0.312,
553
+ "grad_norm": 19.114471435546875,
554
+ "learning_rate": 3.464630225080386e-05,
555
+ "loss": 0.8341,
556
+ "step": 780
557
+ },
558
+ {
559
+ "epoch": 0.316,
560
+ "grad_norm": 17.528043746948242,
561
+ "learning_rate": 3.4445337620578785e-05,
562
+ "loss": 0.7108,
563
+ "step": 790
564
+ },
565
+ {
566
+ "epoch": 0.32,
567
+ "grad_norm": 13.186959266662598,
568
+ "learning_rate": 3.4244372990353704e-05,
569
+ "loss": 0.7991,
570
+ "step": 800
571
  }
572
  ],
573
  "logging_steps": 10,
 
587
  "attributes": {}
588
  }
589
  },
590
+ "total_flos": 1.442001474164736e+16,
591
  "train_batch_size": 2,
592
  "trial_name": null,
593
  "trial_params": null