TweedleDeepLearnings commited on
Commit
f99403f
·
verified ·
1 Parent(s): 2155732

Training in progress, step 2550, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:31463c340dbc666aac7e225725fb9ad3360f9e62eea45ec33c678a92ce7f17eb
3
  size 394815560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd6be1a1e8dba63fda5837c7f72a318a4ccc16a21490df3b362f7fbc12cf3ce1
3
  size 394815560
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:11d1c8f4c06aaf25f2e2c9c7f82cb3bf81269efc82425c06532c3d05719146f7
3
  size 711495290
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9286d7bb63631c27497c1665788ea8e50b3032d5786d5d2cdc68be60a59615e0
3
  size 711495290
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ff3f41874ddf41d16b0af5bb1930c6f1c9d073c3e2c471b0a984d6a624cf5509
3
  size 14180
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40d9b2439b514e816a44784dde57cac1f97b6409754b7cfedefda3e959232c33
3
  size 14180
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:92004acc1e8872397c2b3d090ade00b417621b7e605504bae6bb3ef764136468
3
  size 1256
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33d3220dbd31e88d495721958a661ce79eb3e88c61eb9620c091aa5a27bed962
3
  size 1256
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 2.5657522678375244,
3
  "best_model_checkpoint": "./output/checkpoint-1650",
4
- "epoch": 1.288244766505636,
5
  "eval_steps": 150,
6
- "global_step": 2400,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1815,6 +1815,119 @@
1815
  "eval_samples_per_second": 13.943,
1816
  "eval_steps_per_second": 13.943,
1817
  "step": 2400
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1818
  }
1819
  ],
1820
  "logging_steps": 10,
@@ -1834,7 +1947,7 @@
1834
  "attributes": {}
1835
  }
1836
  },
1837
- "total_flos": 5427337167931392.0,
1838
  "train_batch_size": 16,
1839
  "trial_name": null,
1840
  "trial_params": null
 
1
  {
2
  "best_metric": 2.5657522678375244,
3
  "best_model_checkpoint": "./output/checkpoint-1650",
4
+ "epoch": 1.3687600644122382,
5
  "eval_steps": 150,
6
+ "global_step": 2550,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1815
  "eval_samples_per_second": 13.943,
1816
  "eval_steps_per_second": 13.943,
1817
  "step": 2400
1818
+ },
1819
+ {
1820
+ "epoch": 1.2936124530327429,
1821
+ "grad_norm": 8.595821380615234,
1822
+ "learning_rate": 4.358557235613734e-05,
1823
+ "loss": 2.1494,
1824
+ "step": 2410
1825
+ },
1826
+ {
1827
+ "epoch": 1.2989801395598497,
1828
+ "grad_norm": 8.649581909179688,
1829
+ "learning_rate": 4.333007628713158e-05,
1830
+ "loss": 2.138,
1831
+ "step": 2420
1832
+ },
1833
+ {
1834
+ "epoch": 1.3043478260869565,
1835
+ "grad_norm": 8.213851928710938,
1836
+ "learning_rate": 4.3074443331732674e-05,
1837
+ "loss": 2.1286,
1838
+ "step": 2430
1839
+ },
1840
+ {
1841
+ "epoch": 1.3097155126140634,
1842
+ "grad_norm": 8.641511917114258,
1843
+ "learning_rate": 4.281868399801016e-05,
1844
+ "loss": 2.0917,
1845
+ "step": 2440
1846
+ },
1847
+ {
1848
+ "epoch": 1.3150831991411702,
1849
+ "grad_norm": 8.759982109069824,
1850
+ "learning_rate": 4.256280879922852e-05,
1851
+ "loss": 2.0891,
1852
+ "step": 2450
1853
+ },
1854
+ {
1855
+ "epoch": 1.320450885668277,
1856
+ "grad_norm": 9.908300399780273,
1857
+ "learning_rate": 4.230682825341498e-05,
1858
+ "loss": 2.0988,
1859
+ "step": 2460
1860
+ },
1861
+ {
1862
+ "epoch": 1.325818572195384,
1863
+ "grad_norm": 9.921548843383789,
1864
+ "learning_rate": 4.205075288292717e-05,
1865
+ "loss": 2.1693,
1866
+ "step": 2470
1867
+ },
1868
+ {
1869
+ "epoch": 1.3311862587224907,
1870
+ "grad_norm": 8.90367603302002,
1871
+ "learning_rate": 4.17945932140206e-05,
1872
+ "loss": 2.0252,
1873
+ "step": 2480
1874
+ },
1875
+ {
1876
+ "epoch": 1.3365539452495974,
1877
+ "grad_norm": 8.375697135925293,
1878
+ "learning_rate": 4.1538359776415936e-05,
1879
+ "loss": 2.1367,
1880
+ "step": 2490
1881
+ },
1882
+ {
1883
+ "epoch": 1.3419216317767042,
1884
+ "grad_norm": 8.196423530578613,
1885
+ "learning_rate": 4.128206310286622e-05,
1886
+ "loss": 2.1237,
1887
+ "step": 2500
1888
+ },
1889
+ {
1890
+ "epoch": 1.347289318303811,
1891
+ "grad_norm": 9.057501792907715,
1892
+ "learning_rate": 4.102571372872382e-05,
1893
+ "loss": 2.1228,
1894
+ "step": 2510
1895
+ },
1896
+ {
1897
+ "epoch": 1.3526570048309179,
1898
+ "grad_norm": 9.608572006225586,
1899
+ "learning_rate": 4.0769322191507485e-05,
1900
+ "loss": 2.1579,
1901
+ "step": 2520
1902
+ },
1903
+ {
1904
+ "epoch": 1.3580246913580247,
1905
+ "grad_norm": 8.915650367736816,
1906
+ "learning_rate": 4.051289903046909e-05,
1907
+ "loss": 2.1421,
1908
+ "step": 2530
1909
+ },
1910
+ {
1911
+ "epoch": 1.3633923778851316,
1912
+ "grad_norm": 8.882908821105957,
1913
+ "learning_rate": 4.025645478616045e-05,
1914
+ "loss": 2.111,
1915
+ "step": 2540
1916
+ },
1917
+ {
1918
+ "epoch": 1.3687600644122382,
1919
+ "grad_norm": 8.42502498626709,
1920
+ "learning_rate": 4e-05,
1921
+ "loss": 2.1635,
1922
+ "step": 2550
1923
+ },
1924
+ {
1925
+ "epoch": 1.3687600644122382,
1926
+ "eval_loss": 2.5890932083129883,
1927
+ "eval_runtime": 36.6753,
1928
+ "eval_samples_per_second": 13.66,
1929
+ "eval_steps_per_second": 13.66,
1930
+ "step": 2550
1931
  }
1932
  ],
1933
  "logging_steps": 10,
 
1947
  "attributes": {}
1948
  }
1949
  },
1950
+ "total_flos": 5767973348014080.0,
1951
  "train_batch_size": 16,
1952
  "trial_name": null,
1953
  "trial_params": null