mgh6 commited on
Commit
2cef6c9
·
verified ·
1 Parent(s): 94f6879

Training in progress, epoch 16, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:66200acd6c20b3e5f776dd34d1dce04a462cff382acb5ae4670ee560518aaaf9
3
  size 2682482800
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85a37ab425a5bb455eb5138be42e451b4f6504a4ec63d8c4323c617cb3ebc33e
3
  size 2682482800
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d55fa5ee514060d7787608dec3a888a028a7841ddb0bd524977d24948976b24e
3
  size 5365108834
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb0e7b73e702883e0aea72dd82fc77ae713a9e1e0ea8fae495ef88376c79291d
3
  size 5365108834
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bc091a107ead0fd888fc4fea93ea268674953a0350c819218e401bfc73856e7d
3
  size 15006
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d2f665ad4a75c89f6711748909d7b7bdbe353d41461e39d9efde1e7c8c122662
3
  size 15006
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:89399495261c74d8cccbea129b463394fdf91e1c31568f5b62fa945b79a3648c
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8db07ed3922379a7896753ff18d75baed592f87392b70da4afc38e708f68e8df
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 15.997999636297509,
5
  "eval_steps": 50,
6
- "global_step": 5488,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1642,6 +1642,111 @@
1642
  "eval_samples_per_second": 41.5,
1643
  "eval_steps_per_second": 20.75,
1644
  "step": 5450
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1645
  }
1646
  ],
1647
  "logging_steps": 50,
@@ -1661,7 +1766,7 @@
1661
  "attributes": {}
1662
  }
1663
  },
1664
- "total_flos": 1.42943403473109e+18,
1665
  "train_batch_size": 2,
1666
  "trial_name": null,
1667
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 16.997999636297507,
5
  "eval_steps": 50,
6
+ "global_step": 5831,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1642
  "eval_samples_per_second": 41.5,
1643
  "eval_steps_per_second": 20.75,
1644
  "step": 5450
1645
+ },
1646
+ {
1647
+ "epoch": 16.03491543917076,
1648
+ "grad_norm": 56.59613037109375,
1649
+ "learning_rate": 1.9825072886297377e-05,
1650
+ "loss": 0.7036,
1651
+ "step": 5500
1652
+ },
1653
+ {
1654
+ "epoch": 16.03491543917076,
1655
+ "eval_loss": 0.6698302626609802,
1656
+ "eval_runtime": 118.2947,
1657
+ "eval_samples_per_second": 40.796,
1658
+ "eval_steps_per_second": 20.398,
1659
+ "step": 5500
1660
+ },
1661
+ {
1662
+ "epoch": 16.180396435715586,
1663
+ "grad_norm": 105.1061019897461,
1664
+ "learning_rate": 1.9096209912536444e-05,
1665
+ "loss": 0.6882,
1666
+ "step": 5550
1667
+ },
1668
+ {
1669
+ "epoch": 16.180396435715586,
1670
+ "eval_loss": 0.6712303161621094,
1671
+ "eval_runtime": 116.6934,
1672
+ "eval_samples_per_second": 41.356,
1673
+ "eval_steps_per_second": 20.678,
1674
+ "step": 5550
1675
+ },
1676
+ {
1677
+ "epoch": 16.325877432260413,
1678
+ "grad_norm": 48.417171478271484,
1679
+ "learning_rate": 1.836734693877551e-05,
1680
+ "loss": 0.6877,
1681
+ "step": 5600
1682
+ },
1683
+ {
1684
+ "epoch": 16.325877432260413,
1685
+ "eval_loss": 0.6700485944747925,
1686
+ "eval_runtime": 117.069,
1687
+ "eval_samples_per_second": 41.224,
1688
+ "eval_steps_per_second": 20.612,
1689
+ "step": 5600
1690
+ },
1691
+ {
1692
+ "epoch": 16.47135842880524,
1693
+ "grad_norm": 71.32566833496094,
1694
+ "learning_rate": 1.7638483965014578e-05,
1695
+ "loss": 0.6912,
1696
+ "step": 5650
1697
+ },
1698
+ {
1699
+ "epoch": 16.47135842880524,
1700
+ "eval_loss": 0.6679624915122986,
1701
+ "eval_runtime": 116.8492,
1702
+ "eval_samples_per_second": 41.301,
1703
+ "eval_steps_per_second": 20.651,
1704
+ "step": 5650
1705
+ },
1706
+ {
1707
+ "epoch": 16.616839425350065,
1708
+ "grad_norm": 50.87607955932617,
1709
+ "learning_rate": 1.6909620991253645e-05,
1710
+ "loss": 0.6874,
1711
+ "step": 5700
1712
+ },
1713
+ {
1714
+ "epoch": 16.616839425350065,
1715
+ "eval_loss": 0.6632840633392334,
1716
+ "eval_runtime": 116.778,
1717
+ "eval_samples_per_second": 41.326,
1718
+ "eval_steps_per_second": 20.663,
1719
+ "step": 5700
1720
+ },
1721
+ {
1722
+ "epoch": 16.76232042189489,
1723
+ "grad_norm": 62.6202392578125,
1724
+ "learning_rate": 1.6180758017492712e-05,
1725
+ "loss": 0.685,
1726
+ "step": 5750
1727
+ },
1728
+ {
1729
+ "epoch": 16.76232042189489,
1730
+ "eval_loss": 0.6643268465995789,
1731
+ "eval_runtime": 116.7182,
1732
+ "eval_samples_per_second": 41.347,
1733
+ "eval_steps_per_second": 20.674,
1734
+ "step": 5750
1735
+ },
1736
+ {
1737
+ "epoch": 16.907801418439718,
1738
+ "grad_norm": 57.434730529785156,
1739
+ "learning_rate": 1.545189504373178e-05,
1740
+ "loss": 0.6861,
1741
+ "step": 5800
1742
+ },
1743
+ {
1744
+ "epoch": 16.907801418439718,
1745
+ "eval_loss": 0.663483738899231,
1746
+ "eval_runtime": 116.4065,
1747
+ "eval_samples_per_second": 41.458,
1748
+ "eval_steps_per_second": 20.729,
1749
+ "step": 5800
1750
  }
1751
  ],
1752
  "logging_steps": 50,
 
1766
  "attributes": {}
1767
  }
1768
  },
1769
+ "total_flos": 1.5187665302784573e+18,
1770
  "train_batch_size": 2,
1771
  "trial_name": null,
1772
  "trial_params": null