Azrail commited on
Commit
a3e8026
·
verified ·
1 Parent(s): 51e2023

Training in progress, step 10000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:40de222a643a29f7d83ca57461447e94369db28dfe02ce1ea8dc42c4841ff5b0
3
  size 517931840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa4c5a2f220199565c612d1b585e05a7b92fff44644cb0d5fe3e2f43506e0e66
3
  size 517931840
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:17390b38b97f3c88d7498ab8a8662e59b4ff06eff339b6bdbd74ab3397b0fd3a
3
  size 1035661434
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6389d8c90e4661170f94cbc7fb36a5d3b74c5a0f6b13c2a19b9518515670df99
3
  size 1035661434
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:060f0503489879d8dfb53b047f548ca4611036feac0d8375d9686aebe8f546f0
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf18e0901d5cf90634d477e2ea56f1da923039f227c0abc08997f7c74e97f4e1
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a106eb944d9a54ad09b68ad887887e7f10d1565ba0db56d5847127b57e2c1043
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00e07de65a0221f25bbdd3fefea6366e38c04a216122618dd72af76e955ce943
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 0.1976943891861169,
6
  "eval_steps": 500,
7
- "global_step": 9000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -1610,11 +1610,189 @@
1610
  "eval_steps_per_second": 18.972,
1611
  "num_input_tokens_seen": 9437184000,
1612
  "step": 9000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1613
  }
1614
  ],
1615
  "logging_steps": 50,
1616
  "max_steps": 200000,
1617
- "num_input_tokens_seen": 9437184000,
1618
  "num_train_epochs": 5,
1619
  "save_steps": 1000,
1620
  "stateful_callbacks": {
@@ -1629,7 +1807,7 @@
1629
  "attributes": {}
1630
  }
1631
  },
1632
- "total_flos": 5.374545443684352e+18,
1633
  "train_batch_size": 64,
1634
  "trial_name": null,
1635
  "trial_params": null
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 0.21966043242901878,
6
  "eval_steps": 500,
7
+ "global_step": 10000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
1610
  "eval_steps_per_second": 18.972,
1611
  "num_input_tokens_seen": 9437184000,
1612
  "step": 9000
1613
+ },
1614
+ {
1615
+ "epoch": 0.198792691348262,
1616
+ "grad_norm": 0.17666102945804596,
1617
+ "learning_rate": 0.001,
1618
+ "loss": 3.0728,
1619
+ "num_input_tokens_seen": 9489612800,
1620
+ "step": 9050
1621
+ },
1622
+ {
1623
+ "epoch": 0.1998909935104071,
1624
+ "grad_norm": 0.202484592795372,
1625
+ "learning_rate": 0.001,
1626
+ "loss": 2.9818,
1627
+ "num_input_tokens_seen": 9542041600,
1628
+ "step": 9100
1629
+ },
1630
+ {
1631
+ "epoch": 0.2009892956725522,
1632
+ "grad_norm": 0.15095236897468567,
1633
+ "learning_rate": 0.001,
1634
+ "loss": 2.9423,
1635
+ "num_input_tokens_seen": 9594470400,
1636
+ "step": 9150
1637
+ },
1638
+ {
1639
+ "epoch": 0.2020875978346973,
1640
+ "grad_norm": 0.13089850544929504,
1641
+ "learning_rate": 0.001,
1642
+ "loss": 2.9227,
1643
+ "num_input_tokens_seen": 9646899200,
1644
+ "step": 9200
1645
+ },
1646
+ {
1647
+ "epoch": 0.2031858999968424,
1648
+ "grad_norm": 0.14022304117679596,
1649
+ "learning_rate": 0.001,
1650
+ "loss": 2.8988,
1651
+ "num_input_tokens_seen": 9699328000,
1652
+ "step": 9250
1653
+ },
1654
+ {
1655
+ "epoch": 0.2042842021589875,
1656
+ "grad_norm": 0.13116785883903503,
1657
+ "learning_rate": 0.001,
1658
+ "loss": 2.8716,
1659
+ "num_input_tokens_seen": 9751756800,
1660
+ "step": 9300
1661
+ },
1662
+ {
1663
+ "epoch": 0.20538250432113256,
1664
+ "grad_norm": 0.1395471841096878,
1665
+ "learning_rate": 0.001,
1666
+ "loss": 2.8727,
1667
+ "num_input_tokens_seen": 9804185600,
1668
+ "step": 9350
1669
+ },
1670
+ {
1671
+ "epoch": 0.20648080648327766,
1672
+ "grad_norm": 0.1271878033876419,
1673
+ "learning_rate": 0.001,
1674
+ "loss": 2.864,
1675
+ "num_input_tokens_seen": 9856614400,
1676
+ "step": 9400
1677
+ },
1678
+ {
1679
+ "epoch": 0.20757910864542276,
1680
+ "grad_norm": 0.14148685336112976,
1681
+ "learning_rate": 0.001,
1682
+ "loss": 2.8604,
1683
+ "num_input_tokens_seen": 9909043200,
1684
+ "step": 9450
1685
+ },
1686
+ {
1687
+ "epoch": 0.20867741080756785,
1688
+ "grad_norm": 0.1292584091424942,
1689
+ "learning_rate": 0.001,
1690
+ "loss": 2.8547,
1691
+ "num_input_tokens_seen": 9961472000,
1692
+ "step": 9500
1693
+ },
1694
+ {
1695
+ "epoch": 0.20867741080756785,
1696
+ "eval_loss": 2.756131649017334,
1697
+ "eval_runtime": 65.0495,
1698
+ "eval_samples_per_second": 76.865,
1699
+ "eval_steps_per_second": 19.216,
1700
+ "num_input_tokens_seen": 9961472000,
1701
+ "step": 9500
1702
+ },
1703
+ {
1704
+ "epoch": 0.20977571296971295,
1705
+ "grad_norm": 0.10929372161626816,
1706
+ "learning_rate": 0.001,
1707
+ "loss": 2.8467,
1708
+ "num_input_tokens_seen": 10013900800,
1709
+ "step": 9550
1710
+ },
1711
+ {
1712
+ "epoch": 0.21087401513185805,
1713
+ "grad_norm": 0.1180899515748024,
1714
+ "learning_rate": 0.001,
1715
+ "loss": 2.8501,
1716
+ "num_input_tokens_seen": 10066329600,
1717
+ "step": 9600
1718
+ },
1719
+ {
1720
+ "epoch": 0.21197231729400312,
1721
+ "grad_norm": 0.12041448056697845,
1722
+ "learning_rate": 0.001,
1723
+ "loss": 2.8438,
1724
+ "num_input_tokens_seen": 10118758400,
1725
+ "step": 9650
1726
+ },
1727
+ {
1728
+ "epoch": 0.21307061945614822,
1729
+ "grad_norm": 0.13195224106311798,
1730
+ "learning_rate": 0.001,
1731
+ "loss": 2.8341,
1732
+ "num_input_tokens_seen": 10171187200,
1733
+ "step": 9700
1734
+ },
1735
+ {
1736
+ "epoch": 0.21416892161829332,
1737
+ "grad_norm": 0.11887054890394211,
1738
+ "learning_rate": 0.001,
1739
+ "loss": 2.8349,
1740
+ "num_input_tokens_seen": 10223616000,
1741
+ "step": 9750
1742
+ },
1743
+ {
1744
+ "epoch": 0.21526722378043842,
1745
+ "grad_norm": 0.1044996827840805,
1746
+ "learning_rate": 0.001,
1747
+ "loss": 2.8428,
1748
+ "num_input_tokens_seen": 10276044800,
1749
+ "step": 9800
1750
+ },
1751
+ {
1752
+ "epoch": 0.21636552594258351,
1753
+ "grad_norm": 0.11951665580272675,
1754
+ "learning_rate": 0.001,
1755
+ "loss": 2.8323,
1756
+ "num_input_tokens_seen": 10328473600,
1757
+ "step": 9850
1758
+ },
1759
+ {
1760
+ "epoch": 0.2174638281047286,
1761
+ "grad_norm": 0.11673793941736221,
1762
+ "learning_rate": 0.001,
1763
+ "loss": 2.8271,
1764
+ "num_input_tokens_seen": 10380902400,
1765
+ "step": 9900
1766
+ },
1767
+ {
1768
+ "epoch": 0.21856213026687368,
1769
+ "grad_norm": 0.1178969219326973,
1770
+ "learning_rate": 0.001,
1771
+ "loss": 2.8328,
1772
+ "num_input_tokens_seen": 10433331200,
1773
+ "step": 9950
1774
+ },
1775
+ {
1776
+ "epoch": 0.21966043242901878,
1777
+ "grad_norm": 0.11995361745357513,
1778
+ "learning_rate": 0.001,
1779
+ "loss": 2.8182,
1780
+ "num_input_tokens_seen": 10485760000,
1781
+ "step": 10000
1782
+ },
1783
+ {
1784
+ "epoch": 0.21966043242901878,
1785
+ "eval_loss": 2.732673168182373,
1786
+ "eval_runtime": 66.3377,
1787
+ "eval_samples_per_second": 75.372,
1788
+ "eval_steps_per_second": 18.843,
1789
+ "num_input_tokens_seen": 10485760000,
1790
+ "step": 10000
1791
  }
1792
  ],
1793
  "logging_steps": 50,
1794
  "max_steps": 200000,
1795
+ "num_input_tokens_seen": 10485760000,
1796
  "num_train_epochs": 5,
1797
  "save_steps": 1000,
1798
  "stateful_callbacks": {
 
1807
  "attributes": {}
1808
  }
1809
  },
1810
+ "total_flos": 5.97171715964928e+18,
1811
  "train_batch_size": 64,
1812
  "trial_name": null,
1813
  "trial_params": null