ljcamargo commited on
Commit
a2ba68b
·
verified ·
1 Parent(s): b2b4757

Training in progress, step 2400, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3dfeffd5b70b240cf2e2d2863da88eee1aa83f787540436f9ab733c8f7de2bec
3
  size 3826461296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3daa573e6c713aaba8eac212b616972e27300562149c5a86cf161a20a2986ab7
3
  size 3826461296
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:40948a2d1802a19363fb766e414eb283cbde5c248e2166918a98f5c11d88b13c
3
  size 2479955235
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ecb57a7b1bc6df393a4d72695c3003da6933e5235a702cfb45b794fb4c7477bb
3
  size 2479955235
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c615193f9ae8a09e4b00ab7eadeaae7a871753d733db99177a1dcef6ff1f33b3
3
  size 1383
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db5e98494f4088139d10e50b166d461cb7e004fb4e01c00728b39a7d5b780d91
3
  size 1383
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5b73cf65e0ad93ae93850e87c8976f968fc0604f876f55ae715db103a1834c6d
3
  size 1465
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2e9f05b210aeadfc93d7bb7ed80d64988a89f8306f57d2f4dda99778443c8e5
3
  size 1465
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 0.92,
6
  "eval_steps": 500,
7
- "global_step": 2300,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -1618,6 +1618,76 @@
1618
  "learning_rate": 4.119774919614148e-06,
1619
  "loss": 0.2503,
1620
  "step": 2300
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1621
  }
1622
  ],
1623
  "logging_steps": 10,
@@ -1637,7 +1707,7 @@
1637
  "attributes": {}
1638
  }
1639
  },
1640
- "total_flos": 4.150543064863334e+16,
1641
  "train_batch_size": 2,
1642
  "trial_name": null,
1643
  "trial_params": null
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 0.96,
6
  "eval_steps": 500,
7
+ "global_step": 2400,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
1618
  "learning_rate": 4.119774919614148e-06,
1619
  "loss": 0.2503,
1620
  "step": 2300
1621
+ },
1622
+ {
1623
+ "epoch": 0.924,
1624
+ "grad_norm": 8.718388557434082,
1625
+ "learning_rate": 3.918810289389068e-06,
1626
+ "loss": 0.2793,
1627
+ "step": 2310
1628
+ },
1629
+ {
1630
+ "epoch": 0.928,
1631
+ "grad_norm": 11.070220947265625,
1632
+ "learning_rate": 3.7178456591639876e-06,
1633
+ "loss": 0.2585,
1634
+ "step": 2320
1635
+ },
1636
+ {
1637
+ "epoch": 0.932,
1638
+ "grad_norm": 7.7603759765625,
1639
+ "learning_rate": 3.516881028938907e-06,
1640
+ "loss": 0.5099,
1641
+ "step": 2330
1642
+ },
1643
+ {
1644
+ "epoch": 0.936,
1645
+ "grad_norm": 15.190449714660645,
1646
+ "learning_rate": 3.3159163987138267e-06,
1647
+ "loss": 0.2601,
1648
+ "step": 2340
1649
+ },
1650
+ {
1651
+ "epoch": 0.94,
1652
+ "grad_norm": 9.578422546386719,
1653
+ "learning_rate": 3.1149517684887463e-06,
1654
+ "loss": 0.264,
1655
+ "step": 2350
1656
+ },
1657
+ {
1658
+ "epoch": 0.944,
1659
+ "grad_norm": 10.097108840942383,
1660
+ "learning_rate": 2.913987138263666e-06,
1661
+ "loss": 0.6162,
1662
+ "step": 2360
1663
+ },
1664
+ {
1665
+ "epoch": 0.948,
1666
+ "grad_norm": 8.5523042678833,
1667
+ "learning_rate": 2.7130225080385853e-06,
1668
+ "loss": 0.2649,
1669
+ "step": 2370
1670
+ },
1671
+ {
1672
+ "epoch": 0.952,
1673
+ "grad_norm": 8.127545356750488,
1674
+ "learning_rate": 2.512057877813505e-06,
1675
+ "loss": 0.4394,
1676
+ "step": 2380
1677
+ },
1678
+ {
1679
+ "epoch": 0.956,
1680
+ "grad_norm": 8.142045021057129,
1681
+ "learning_rate": 2.311093247588425e-06,
1682
+ "loss": 0.2339,
1683
+ "step": 2390
1684
+ },
1685
+ {
1686
+ "epoch": 0.96,
1687
+ "grad_norm": 6.302300930023193,
1688
+ "learning_rate": 2.1101286173633444e-06,
1689
+ "loss": 0.2135,
1690
+ "step": 2400
1691
  }
1692
  ],
1693
  "logging_steps": 10,
 
1707
  "attributes": {}
1708
  }
1709
  },
1710
+ "total_flos": 4.334241751385702e+16,
1711
  "train_batch_size": 2,
1712
  "trial_name": null,
1713
  "trial_params": null