Training in progress, step 10000, checkpoint
Browse files
last-checkpoint/model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 517931840
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:aa4c5a2f220199565c612d1b585e05a7b92fff44644cb0d5fe3e2f43506e0e66
|
| 3 |
size 517931840
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1035661434
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6389d8c90e4661170f94cbc7fb36a5d3b74c5a0f6b13c2a19b9518515670df99
|
| 3 |
size 1035661434
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bf18e0901d5cf90634d477e2ea56f1da923039f227c0abc08997f7c74e97f4e1
|
| 3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:00e07de65a0221f25bbdd3fefea6366e38c04a216122618dd72af76e955ce943
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -2,9 +2,9 @@
|
|
| 2 |
"best_global_step": null,
|
| 3 |
"best_metric": null,
|
| 4 |
"best_model_checkpoint": null,
|
| 5 |
-
"epoch": 0.
|
| 6 |
"eval_steps": 500,
|
| 7 |
-
"global_step":
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
@@ -1610,11 +1610,189 @@
|
|
| 1610 |
"eval_steps_per_second": 18.972,
|
| 1611 |
"num_input_tokens_seen": 9437184000,
|
| 1612 |
"step": 9000
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1613 |
}
|
| 1614 |
],
|
| 1615 |
"logging_steps": 50,
|
| 1616 |
"max_steps": 200000,
|
| 1617 |
-
"num_input_tokens_seen":
|
| 1618 |
"num_train_epochs": 5,
|
| 1619 |
"save_steps": 1000,
|
| 1620 |
"stateful_callbacks": {
|
|
@@ -1629,7 +1807,7 @@
|
|
| 1629 |
"attributes": {}
|
| 1630 |
}
|
| 1631 |
},
|
| 1632 |
-
"total_flos": 5.
|
| 1633 |
"train_batch_size": 64,
|
| 1634 |
"trial_name": null,
|
| 1635 |
"trial_params": null
|
|
|
|
| 2 |
"best_global_step": null,
|
| 3 |
"best_metric": null,
|
| 4 |
"best_model_checkpoint": null,
|
| 5 |
+
"epoch": 0.21966043242901878,
|
| 6 |
"eval_steps": 500,
|
| 7 |
+
"global_step": 10000,
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
|
|
| 1610 |
"eval_steps_per_second": 18.972,
|
| 1611 |
"num_input_tokens_seen": 9437184000,
|
| 1612 |
"step": 9000
|
| 1613 |
+
},
|
| 1614 |
+
{
|
| 1615 |
+
"epoch": 0.198792691348262,
|
| 1616 |
+
"grad_norm": 0.17666102945804596,
|
| 1617 |
+
"learning_rate": 0.001,
|
| 1618 |
+
"loss": 3.0728,
|
| 1619 |
+
"num_input_tokens_seen": 9489612800,
|
| 1620 |
+
"step": 9050
|
| 1621 |
+
},
|
| 1622 |
+
{
|
| 1623 |
+
"epoch": 0.1998909935104071,
|
| 1624 |
+
"grad_norm": 0.202484592795372,
|
| 1625 |
+
"learning_rate": 0.001,
|
| 1626 |
+
"loss": 2.9818,
|
| 1627 |
+
"num_input_tokens_seen": 9542041600,
|
| 1628 |
+
"step": 9100
|
| 1629 |
+
},
|
| 1630 |
+
{
|
| 1631 |
+
"epoch": 0.2009892956725522,
|
| 1632 |
+
"grad_norm": 0.15095236897468567,
|
| 1633 |
+
"learning_rate": 0.001,
|
| 1634 |
+
"loss": 2.9423,
|
| 1635 |
+
"num_input_tokens_seen": 9594470400,
|
| 1636 |
+
"step": 9150
|
| 1637 |
+
},
|
| 1638 |
+
{
|
| 1639 |
+
"epoch": 0.2020875978346973,
|
| 1640 |
+
"grad_norm": 0.13089850544929504,
|
| 1641 |
+
"learning_rate": 0.001,
|
| 1642 |
+
"loss": 2.9227,
|
| 1643 |
+
"num_input_tokens_seen": 9646899200,
|
| 1644 |
+
"step": 9200
|
| 1645 |
+
},
|
| 1646 |
+
{
|
| 1647 |
+
"epoch": 0.2031858999968424,
|
| 1648 |
+
"grad_norm": 0.14022304117679596,
|
| 1649 |
+
"learning_rate": 0.001,
|
| 1650 |
+
"loss": 2.8988,
|
| 1651 |
+
"num_input_tokens_seen": 9699328000,
|
| 1652 |
+
"step": 9250
|
| 1653 |
+
},
|
| 1654 |
+
{
|
| 1655 |
+
"epoch": 0.2042842021589875,
|
| 1656 |
+
"grad_norm": 0.13116785883903503,
|
| 1657 |
+
"learning_rate": 0.001,
|
| 1658 |
+
"loss": 2.8716,
|
| 1659 |
+
"num_input_tokens_seen": 9751756800,
|
| 1660 |
+
"step": 9300
|
| 1661 |
+
},
|
| 1662 |
+
{
|
| 1663 |
+
"epoch": 0.20538250432113256,
|
| 1664 |
+
"grad_norm": 0.1395471841096878,
|
| 1665 |
+
"learning_rate": 0.001,
|
| 1666 |
+
"loss": 2.8727,
|
| 1667 |
+
"num_input_tokens_seen": 9804185600,
|
| 1668 |
+
"step": 9350
|
| 1669 |
+
},
|
| 1670 |
+
{
|
| 1671 |
+
"epoch": 0.20648080648327766,
|
| 1672 |
+
"grad_norm": 0.1271878033876419,
|
| 1673 |
+
"learning_rate": 0.001,
|
| 1674 |
+
"loss": 2.864,
|
| 1675 |
+
"num_input_tokens_seen": 9856614400,
|
| 1676 |
+
"step": 9400
|
| 1677 |
+
},
|
| 1678 |
+
{
|
| 1679 |
+
"epoch": 0.20757910864542276,
|
| 1680 |
+
"grad_norm": 0.14148685336112976,
|
| 1681 |
+
"learning_rate": 0.001,
|
| 1682 |
+
"loss": 2.8604,
|
| 1683 |
+
"num_input_tokens_seen": 9909043200,
|
| 1684 |
+
"step": 9450
|
| 1685 |
+
},
|
| 1686 |
+
{
|
| 1687 |
+
"epoch": 0.20867741080756785,
|
| 1688 |
+
"grad_norm": 0.1292584091424942,
|
| 1689 |
+
"learning_rate": 0.001,
|
| 1690 |
+
"loss": 2.8547,
|
| 1691 |
+
"num_input_tokens_seen": 9961472000,
|
| 1692 |
+
"step": 9500
|
| 1693 |
+
},
|
| 1694 |
+
{
|
| 1695 |
+
"epoch": 0.20867741080756785,
|
| 1696 |
+
"eval_loss": 2.756131649017334,
|
| 1697 |
+
"eval_runtime": 65.0495,
|
| 1698 |
+
"eval_samples_per_second": 76.865,
|
| 1699 |
+
"eval_steps_per_second": 19.216,
|
| 1700 |
+
"num_input_tokens_seen": 9961472000,
|
| 1701 |
+
"step": 9500
|
| 1702 |
+
},
|
| 1703 |
+
{
|
| 1704 |
+
"epoch": 0.20977571296971295,
|
| 1705 |
+
"grad_norm": 0.10929372161626816,
|
| 1706 |
+
"learning_rate": 0.001,
|
| 1707 |
+
"loss": 2.8467,
|
| 1708 |
+
"num_input_tokens_seen": 10013900800,
|
| 1709 |
+
"step": 9550
|
| 1710 |
+
},
|
| 1711 |
+
{
|
| 1712 |
+
"epoch": 0.21087401513185805,
|
| 1713 |
+
"grad_norm": 0.1180899515748024,
|
| 1714 |
+
"learning_rate": 0.001,
|
| 1715 |
+
"loss": 2.8501,
|
| 1716 |
+
"num_input_tokens_seen": 10066329600,
|
| 1717 |
+
"step": 9600
|
| 1718 |
+
},
|
| 1719 |
+
{
|
| 1720 |
+
"epoch": 0.21197231729400312,
|
| 1721 |
+
"grad_norm": 0.12041448056697845,
|
| 1722 |
+
"learning_rate": 0.001,
|
| 1723 |
+
"loss": 2.8438,
|
| 1724 |
+
"num_input_tokens_seen": 10118758400,
|
| 1725 |
+
"step": 9650
|
| 1726 |
+
},
|
| 1727 |
+
{
|
| 1728 |
+
"epoch": 0.21307061945614822,
|
| 1729 |
+
"grad_norm": 0.13195224106311798,
|
| 1730 |
+
"learning_rate": 0.001,
|
| 1731 |
+
"loss": 2.8341,
|
| 1732 |
+
"num_input_tokens_seen": 10171187200,
|
| 1733 |
+
"step": 9700
|
| 1734 |
+
},
|
| 1735 |
+
{
|
| 1736 |
+
"epoch": 0.21416892161829332,
|
| 1737 |
+
"grad_norm": 0.11887054890394211,
|
| 1738 |
+
"learning_rate": 0.001,
|
| 1739 |
+
"loss": 2.8349,
|
| 1740 |
+
"num_input_tokens_seen": 10223616000,
|
| 1741 |
+
"step": 9750
|
| 1742 |
+
},
|
| 1743 |
+
{
|
| 1744 |
+
"epoch": 0.21526722378043842,
|
| 1745 |
+
"grad_norm": 0.1044996827840805,
|
| 1746 |
+
"learning_rate": 0.001,
|
| 1747 |
+
"loss": 2.8428,
|
| 1748 |
+
"num_input_tokens_seen": 10276044800,
|
| 1749 |
+
"step": 9800
|
| 1750 |
+
},
|
| 1751 |
+
{
|
| 1752 |
+
"epoch": 0.21636552594258351,
|
| 1753 |
+
"grad_norm": 0.11951665580272675,
|
| 1754 |
+
"learning_rate": 0.001,
|
| 1755 |
+
"loss": 2.8323,
|
| 1756 |
+
"num_input_tokens_seen": 10328473600,
|
| 1757 |
+
"step": 9850
|
| 1758 |
+
},
|
| 1759 |
+
{
|
| 1760 |
+
"epoch": 0.2174638281047286,
|
| 1761 |
+
"grad_norm": 0.11673793941736221,
|
| 1762 |
+
"learning_rate": 0.001,
|
| 1763 |
+
"loss": 2.8271,
|
| 1764 |
+
"num_input_tokens_seen": 10380902400,
|
| 1765 |
+
"step": 9900
|
| 1766 |
+
},
|
| 1767 |
+
{
|
| 1768 |
+
"epoch": 0.21856213026687368,
|
| 1769 |
+
"grad_norm": 0.1178969219326973,
|
| 1770 |
+
"learning_rate": 0.001,
|
| 1771 |
+
"loss": 2.8328,
|
| 1772 |
+
"num_input_tokens_seen": 10433331200,
|
| 1773 |
+
"step": 9950
|
| 1774 |
+
},
|
| 1775 |
+
{
|
| 1776 |
+
"epoch": 0.21966043242901878,
|
| 1777 |
+
"grad_norm": 0.11995361745357513,
|
| 1778 |
+
"learning_rate": 0.001,
|
| 1779 |
+
"loss": 2.8182,
|
| 1780 |
+
"num_input_tokens_seen": 10485760000,
|
| 1781 |
+
"step": 10000
|
| 1782 |
+
},
|
| 1783 |
+
{
|
| 1784 |
+
"epoch": 0.21966043242901878,
|
| 1785 |
+
"eval_loss": 2.732673168182373,
|
| 1786 |
+
"eval_runtime": 66.3377,
|
| 1787 |
+
"eval_samples_per_second": 75.372,
|
| 1788 |
+
"eval_steps_per_second": 18.843,
|
| 1789 |
+
"num_input_tokens_seen": 10485760000,
|
| 1790 |
+
"step": 10000
|
| 1791 |
}
|
| 1792 |
],
|
| 1793 |
"logging_steps": 50,
|
| 1794 |
"max_steps": 200000,
|
| 1795 |
+
"num_input_tokens_seen": 10485760000,
|
| 1796 |
"num_train_epochs": 5,
|
| 1797 |
"save_steps": 1000,
|
| 1798 |
"stateful_callbacks": {
|
|
|
|
| 1807 |
"attributes": {}
|
| 1808 |
}
|
| 1809 |
},
|
| 1810 |
+
"total_flos": 5.97171715964928e+18,
|
| 1811 |
"train_batch_size": 64,
|
| 1812 |
"trial_name": null,
|
| 1813 |
"trial_params": null
|