mgh6 commited on
Commit
b0daf72
·
verified ·
1 Parent(s): ffda4fd

Training in progress, epoch 18, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:659e1a6a1ff1a9fbec759df1d78ad027f1d391213c1a217bf5b92f48ccb3b2a7
3
  size 2682482800
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa112d7382802b71b92578187eb839ff8484215e39cd2eebf633f2cdf07840b3
3
  size 2682482800
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b5d5e4710df7ed104bcaa0a9051033613a80b4363c6f76bae6c58a4863fa30ad
3
  size 5365108834
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36e264601b26cf4f05d8d9b8bf2be8d4c0837531e47c0482bd376cbd5b8441a5
3
  size 5365108834
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:acb524c16816d50e3851812f08ed31af0f04d2253498b52b5121962571f22c75
3
  size 15006
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00a1f49e04dfd6249428b9586875a339bc93164dac6605e82474c6ccf8f127eb
3
  size 15006
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ee6db23f5e321edc29e923809ce0b3dff2a73c49bba17a4de22c710250ea7d6e
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fbcba6a24d9c72c080feefc697485672e2b8856f539e65accd7175f3e99dc162
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 17.997999636297507,
5
  "eval_steps": 50,
6
- "global_step": 6174,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1852,6 +1852,111 @@
1852
  "eval_samples_per_second": 41.452,
1853
  "eval_steps_per_second": 20.726,
1854
  "step": 6150
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1855
  }
1856
  ],
1857
  "logging_steps": 50,
@@ -1871,7 +1976,7 @@
1871
  "attributes": {}
1872
  }
1873
  },
1874
- "total_flos": 1.6080269119827476e+18,
1875
  "train_batch_size": 2,
1876
  "trial_name": null,
1877
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 18.997999636297507,
5
  "eval_steps": 50,
6
+ "global_step": 6517,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1852
  "eval_samples_per_second": 41.452,
1853
  "eval_steps_per_second": 20.726,
1854
  "step": 6150
1855
+ },
1856
+ {
1857
+ "epoch": 18.07565011820331,
1858
+ "grad_norm": 74.32230377197266,
1859
+ "learning_rate": 9.620991253644314e-06,
1860
+ "loss": 0.6763,
1861
+ "step": 6200
1862
+ },
1863
+ {
1864
+ "epoch": 18.07565011820331,
1865
+ "eval_loss": 0.6560296416282654,
1866
+ "eval_runtime": 116.3284,
1867
+ "eval_samples_per_second": 41.486,
1868
+ "eval_steps_per_second": 20.743,
1869
+ "step": 6200
1870
+ },
1871
+ {
1872
+ "epoch": 18.221131114748136,
1873
+ "grad_norm": 62.084503173828125,
1874
+ "learning_rate": 8.892128279883383e-06,
1875
+ "loss": 0.6667,
1876
+ "step": 6250
1877
+ },
1878
+ {
1879
+ "epoch": 18.221131114748136,
1880
+ "eval_loss": 0.6540623903274536,
1881
+ "eval_runtime": 116.3245,
1882
+ "eval_samples_per_second": 41.487,
1883
+ "eval_steps_per_second": 20.744,
1884
+ "step": 6250
1885
+ },
1886
+ {
1887
+ "epoch": 18.366612111292962,
1888
+ "grad_norm": 43.38812255859375,
1889
+ "learning_rate": 8.163265306122448e-06,
1890
+ "loss": 0.6761,
1891
+ "step": 6300
1892
+ },
1893
+ {
1894
+ "epoch": 18.366612111292962,
1895
+ "eval_loss": 0.6527947187423706,
1896
+ "eval_runtime": 116.5368,
1897
+ "eval_samples_per_second": 41.412,
1898
+ "eval_steps_per_second": 20.706,
1899
+ "step": 6300
1900
+ },
1901
+ {
1902
+ "epoch": 18.51209310783779,
1903
+ "grad_norm": 81.97037506103516,
1904
+ "learning_rate": 7.434402332361516e-06,
1905
+ "loss": 0.6708,
1906
+ "step": 6350
1907
+ },
1908
+ {
1909
+ "epoch": 18.51209310783779,
1910
+ "eval_loss": 0.656480073928833,
1911
+ "eval_runtime": 116.5086,
1912
+ "eval_samples_per_second": 41.422,
1913
+ "eval_steps_per_second": 20.711,
1914
+ "step": 6350
1915
+ },
1916
+ {
1917
+ "epoch": 18.657574104382615,
1918
+ "grad_norm": 60.97893524169922,
1919
+ "learning_rate": 6.705539358600584e-06,
1920
+ "loss": 0.6704,
1921
+ "step": 6400
1922
+ },
1923
+ {
1924
+ "epoch": 18.657574104382615,
1925
+ "eval_loss": 0.6558669209480286,
1926
+ "eval_runtime": 116.735,
1927
+ "eval_samples_per_second": 41.342,
1928
+ "eval_steps_per_second": 20.671,
1929
+ "step": 6400
1930
+ },
1931
+ {
1932
+ "epoch": 18.80305510092744,
1933
+ "grad_norm": 75.70346069335938,
1934
+ "learning_rate": 5.97667638483965e-06,
1935
+ "loss": 0.6695,
1936
+ "step": 6450
1937
+ },
1938
+ {
1939
+ "epoch": 18.80305510092744,
1940
+ "eval_loss": 0.6515429615974426,
1941
+ "eval_runtime": 116.3122,
1942
+ "eval_samples_per_second": 41.492,
1943
+ "eval_steps_per_second": 20.746,
1944
+ "step": 6450
1945
+ },
1946
+ {
1947
+ "epoch": 18.948536097472267,
1948
+ "grad_norm": 49.48427963256836,
1949
+ "learning_rate": 5.247813411078718e-06,
1950
+ "loss": 0.6716,
1951
+ "step": 6500
1952
+ },
1953
+ {
1954
+ "epoch": 18.948536097472267,
1955
+ "eval_loss": 0.6520217061042786,
1956
+ "eval_runtime": 116.8293,
1957
+ "eval_samples_per_second": 41.308,
1958
+ "eval_steps_per_second": 20.654,
1959
+ "step": 6500
1960
  }
1961
  ],
1962
  "logging_steps": 50,
 
1976
  "attributes": {}
1977
  }
1978
  },
1979
+ "total_flos": 1.697366505768878e+18,
1980
  "train_batch_size": 2,
1981
  "trial_name": null,
1982
  "trial_params": null