irishprancer commited on
Commit
3ca540f
·
verified ·
1 Parent(s): d5e4250

Training in progress, step 2100, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:558bd7240ae36bf407fa59d99bc87b4f2a83238e3eefe9e7d80228b0d675c5f2
3
  size 527048968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:023bdf4fb37105c005af155a89b3617ec253742740b1c443640c3cfd919b153c
3
  size 527048968
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c9c44e87a84da7f09944098b8b24a46d1e1784343e14aa00f21290dad1d35d03
3
  size 1054135994
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee34169c145fc9db54236abebff2e7bbd35b77780ea424fc21c3c1a2442d1d2d
3
  size 1054135994
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b99872e409bc674b8b81f4e59047c6dd181202eceec2bd8ec5e0879c5adbc8fc
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0245389a7bc82d29e971a63bfd36a481643b7f85d29704591ececfc67d58955b
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b4b6379912da098573e94d032fbbd0f173cdb17e7336f428aac213f60f9a5145
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1bce2a8653bdd97c984ac0cb8f8fc01c58c1e2c98ab50a771eb0dc001a601b0
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.7177689671516418,
3
  "best_model_checkpoint": "./output/checkpoint-450",
4
- "epoch": 84.78260869565217,
5
  "eval_steps": 150,
6
- "global_step": 1950,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1892,6 +1892,151 @@
1892
  "EMA_steps_per_second": 25.224,
1893
  "epoch": 84.78260869565217,
1894
  "step": 1950
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1895
  }
1896
  ],
1897
  "logging_steps": 10,
@@ -1911,7 +2056,7 @@
1911
  "attributes": {}
1912
  }
1913
  },
1914
- "total_flos": 5.02053291067392e+16,
1915
  "train_batch_size": 4,
1916
  "trial_name": null,
1917
  "trial_params": null
 
1
  {
2
  "best_metric": 0.7177689671516418,
3
  "best_model_checkpoint": "./output/checkpoint-450",
4
+ "epoch": 91.30434782608695,
5
  "eval_steps": 150,
6
+ "global_step": 2100,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1892
  "EMA_steps_per_second": 25.224,
1893
  "epoch": 84.78260869565217,
1894
  "step": 1950
1895
+ },
1896
+ {
1897
+ "epoch": 85.21739130434783,
1898
+ "grad_norm": 2.3564674854278564,
1899
+ "learning_rate": 7.487875215855521e-07,
1900
+ "loss": 0.265,
1901
+ "step": 1960
1902
+ },
1903
+ {
1904
+ "epoch": 85.65217391304348,
1905
+ "grad_norm": 1.8858447074890137,
1906
+ "learning_rate": 1.4975750431711041e-06,
1907
+ "loss": 0.2704,
1908
+ "step": 1970
1909
+ },
1910
+ {
1911
+ "epoch": 86.08695652173913,
1912
+ "grad_norm": 2.1835811138153076,
1913
+ "learning_rate": 2.2463625647566557e-06,
1914
+ "loss": 0.2531,
1915
+ "step": 1980
1916
+ },
1917
+ {
1918
+ "epoch": 86.52173913043478,
1919
+ "grad_norm": 1.674813985824585,
1920
+ "learning_rate": 2.9951500863422082e-06,
1921
+ "loss": 0.2767,
1922
+ "step": 1990
1923
+ },
1924
+ {
1925
+ "epoch": 86.95652173913044,
1926
+ "grad_norm": 2.0097134113311768,
1927
+ "learning_rate": 3.74393760792776e-06,
1928
+ "loss": 0.2766,
1929
+ "step": 2000
1930
+ },
1931
+ {
1932
+ "epoch": 87.3913043478261,
1933
+ "grad_norm": 1.4214787483215332,
1934
+ "learning_rate": 4.4927251295133115e-06,
1935
+ "loss": 0.2779,
1936
+ "step": 2010
1937
+ },
1938
+ {
1939
+ "epoch": 87.82608695652173,
1940
+ "grad_norm": 2.0007896423339844,
1941
+ "learning_rate": 5.241512651098863e-06,
1942
+ "loss": 0.2588,
1943
+ "step": 2020
1944
+ },
1945
+ {
1946
+ "epoch": 88.26086956521739,
1947
+ "grad_norm": 2.0449113845825195,
1948
+ "learning_rate": 5.9903001726844164e-06,
1949
+ "loss": 0.2614,
1950
+ "step": 2030
1951
+ },
1952
+ {
1953
+ "epoch": 88.69565217391305,
1954
+ "grad_norm": 1.7983092069625854,
1955
+ "learning_rate": 6.739087694269968e-06,
1956
+ "loss": 0.2852,
1957
+ "step": 2040
1958
+ },
1959
+ {
1960
+ "epoch": 89.1304347826087,
1961
+ "grad_norm": 1.8373875617980957,
1962
+ "learning_rate": 7.48787521585552e-06,
1963
+ "loss": 0.2914,
1964
+ "step": 2050
1965
+ },
1966
+ {
1967
+ "epoch": 89.56521739130434,
1968
+ "grad_norm": 1.543720006942749,
1969
+ "learning_rate": 7.487874473866896e-06,
1970
+ "loss": 0.2467,
1971
+ "step": 2060
1972
+ },
1973
+ {
1974
+ "epoch": 90.0,
1975
+ "grad_norm": 1.6378145217895508,
1976
+ "learning_rate": 7.487872247901318e-06,
1977
+ "loss": 0.2524,
1978
+ "step": 2070
1979
+ },
1980
+ {
1981
+ "epoch": 90.43478260869566,
1982
+ "grad_norm": 1.8025075197219849,
1983
+ "learning_rate": 7.4878685379596685e-06,
1984
+ "loss": 0.2572,
1985
+ "step": 2080
1986
+ },
1987
+ {
1988
+ "epoch": 90.8695652173913,
1989
+ "grad_norm": 1.7167291641235352,
1990
+ "learning_rate": 7.487863344043418e-06,
1991
+ "loss": 0.283,
1992
+ "step": 2090
1993
+ },
1994
+ {
1995
+ "epoch": 91.30434782608695,
1996
+ "grad_norm": 1.7985183000564575,
1997
+ "learning_rate": 7.487856666154626e-06,
1998
+ "loss": 0.2721,
1999
+ "step": 2100
2000
+ },
2001
+ {
2002
+ "epoch": 91.30434782608695,
2003
+ "eval_loss": 0.8964352607727051,
2004
+ "eval_runtime": 0.4971,
2005
+ "eval_samples_per_second": 20.118,
2006
+ "eval_steps_per_second": 20.118,
2007
+ "step": 2100
2008
+ },
2009
+ {
2010
+ "Start_State_loss": 0.861186683177948,
2011
+ "Start_State_runtime": 0.5202,
2012
+ "Start_State_samples_per_second": 19.223,
2013
+ "Start_State_steps_per_second": 19.223,
2014
+ "epoch": 91.30434782608695,
2015
+ "step": 2100
2016
+ },
2017
+ {
2018
+ "Raw_Model_loss": 0.8964352607727051,
2019
+ "Raw_Model_runtime": 0.4702,
2020
+ "Raw_Model_samples_per_second": 21.269,
2021
+ "Raw_Model_steps_per_second": 21.269,
2022
+ "epoch": 91.30434782608695,
2023
+ "step": 2100
2024
+ },
2025
+ {
2026
+ "SWA_loss": 0.7461259365081787,
2027
+ "SWA_runtime": 0.3981,
2028
+ "SWA_samples_per_second": 25.122,
2029
+ "SWA_steps_per_second": 25.122,
2030
+ "epoch": 91.30434782608695,
2031
+ "step": 2100
2032
+ },
2033
+ {
2034
+ "EMA_loss": 0.8607404828071594,
2035
+ "EMA_runtime": 0.3991,
2036
+ "EMA_samples_per_second": 25.058,
2037
+ "EMA_steps_per_second": 25.058,
2038
+ "epoch": 91.30434782608695,
2039
+ "step": 2100
2040
  }
2041
  ],
2042
  "logging_steps": 10,
 
2056
  "attributes": {}
2057
  }
2058
  },
2059
+ "total_flos": 5.411670961736909e+16,
2060
  "train_batch_size": 4,
2061
  "trial_name": null,
2062
  "trial_params": null