NairaRahim commited on
Commit
22eeae6
·
verified ·
1 Parent(s): c899a9f

Training in progress, epoch 21, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6f875ff0bfa8c3f03718200317018a9c1320ef659ed8be49eb8d1545f90dca2b
3
  size 1227009528
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:759e11911622505bdb5a77d511b901f277ce0bbc06c416da623a53f9c0f0b663
3
  size 1227009528
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4edde3d63fc51cb76d2b8798e35123bee17f10b37ac6770074f24fbb1849dc32
3
  size 2454133690
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c251d2bf032c0284e58db70391eefabbbff2c2fb3bc7e50e9d6ba42d7c0345d5
3
  size 2454133690
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5b0066a3b21610aa70bfcf0b5c4ca5da7f43ab12c9e601ab15813e745474a36d
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4db8bac039de345d3f184975bc589d15c0b93c070db1287ed5147dba9f5ba405
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d8c5b80067b711daea816f97793263fb6b6d08534034a3999a4ce7590fa85de8
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7adf3363afe258d5bb46d306ca21d84331e969bff032715bb85737194722fbc
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 34.54485321044922,
3
  "best_model_checkpoint": "/kaggle/working/output/checkpoint-20880",
4
- "epoch": 20.0,
5
  "eval_steps": 500,
6
- "global_step": 26100,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1994,6 +1994,105 @@
1994
  "eval_samples_per_second": 26.478,
1995
  "eval_steps_per_second": 3.328,
1996
  "step": 26100
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1997
  }
1998
  ],
1999
  "logging_steps": 100,
@@ -2008,7 +2107,7 @@
2008
  "early_stopping_threshold": 0.0
2009
  },
2010
  "attributes": {
2011
- "early_stopping_patience_counter": 4
2012
  }
2013
  },
2014
  "TrainerControl": {
@@ -2022,7 +2121,7 @@
2022
  "attributes": {}
2023
  }
2024
  },
2025
- "total_flos": 2.814621806094336e+16,
2026
  "train_batch_size": 8,
2027
  "trial_name": null,
2028
  "trial_params": null
 
1
  {
2
  "best_metric": 34.54485321044922,
3
  "best_model_checkpoint": "/kaggle/working/output/checkpoint-20880",
4
+ "epoch": 21.0,
5
  "eval_steps": 500,
6
+ "global_step": 27405,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1994
  "eval_samples_per_second": 26.478,
1995
  "eval_steps_per_second": 3.328,
1996
  "step": 26100
1997
+ },
1998
+ {
1999
+ "epoch": 20.07662835249042,
2000
+ "grad_norm": 4.07724142074585,
2001
+ "learning_rate": 3.745785440613027e-05,
2002
+ "loss": 33.562,
2003
+ "step": 26200
2004
+ },
2005
+ {
2006
+ "epoch": 20.153256704980844,
2007
+ "grad_norm": 4.335379600524902,
2008
+ "learning_rate": 3.7409961685823756e-05,
2009
+ "loss": 33.166,
2010
+ "step": 26300
2011
+ },
2012
+ {
2013
+ "epoch": 20.229885057471265,
2014
+ "grad_norm": 5.472820281982422,
2015
+ "learning_rate": 3.736206896551724e-05,
2016
+ "loss": 33.8918,
2017
+ "step": 26400
2018
+ },
2019
+ {
2020
+ "epoch": 20.306513409961685,
2021
+ "grad_norm": 3.011789321899414,
2022
+ "learning_rate": 3.731417624521073e-05,
2023
+ "loss": 33.395,
2024
+ "step": 26500
2025
+ },
2026
+ {
2027
+ "epoch": 20.38314176245211,
2028
+ "grad_norm": 3.251089096069336,
2029
+ "learning_rate": 3.7266283524904216e-05,
2030
+ "loss": 32.9072,
2031
+ "step": 26600
2032
+ },
2033
+ {
2034
+ "epoch": 20.45977011494253,
2035
+ "grad_norm": 2.7508978843688965,
2036
+ "learning_rate": 3.72183908045977e-05,
2037
+ "loss": 33.92,
2038
+ "step": 26700
2039
+ },
2040
+ {
2041
+ "epoch": 20.53639846743295,
2042
+ "grad_norm": 2.8051536083221436,
2043
+ "learning_rate": 3.717049808429119e-05,
2044
+ "loss": 33.9392,
2045
+ "step": 26800
2046
+ },
2047
+ {
2048
+ "epoch": 20.613026819923373,
2049
+ "grad_norm": 7.377379417419434,
2050
+ "learning_rate": 3.712260536398468e-05,
2051
+ "loss": 33.0382,
2052
+ "step": 26900
2053
+ },
2054
+ {
2055
+ "epoch": 20.689655172413794,
2056
+ "grad_norm": 3.7770464420318604,
2057
+ "learning_rate": 3.7074712643678164e-05,
2058
+ "loss": 32.6836,
2059
+ "step": 27000
2060
+ },
2061
+ {
2062
+ "epoch": 20.766283524904214,
2063
+ "grad_norm": 4.923346996307373,
2064
+ "learning_rate": 3.702681992337165e-05,
2065
+ "loss": 33.2129,
2066
+ "step": 27100
2067
+ },
2068
+ {
2069
+ "epoch": 20.842911877394634,
2070
+ "grad_norm": 4.790703773498535,
2071
+ "learning_rate": 3.697892720306513e-05,
2072
+ "loss": 33.5413,
2073
+ "step": 27200
2074
+ },
2075
+ {
2076
+ "epoch": 20.919540229885058,
2077
+ "grad_norm": 4.592926025390625,
2078
+ "learning_rate": 3.6931034482758624e-05,
2079
+ "loss": 33.2436,
2080
+ "step": 27300
2081
+ },
2082
+ {
2083
+ "epoch": 20.99616858237548,
2084
+ "grad_norm": 3.0529520511627197,
2085
+ "learning_rate": 3.688314176245211e-05,
2086
+ "loss": 33.2415,
2087
+ "step": 27400
2088
+ },
2089
+ {
2090
+ "epoch": 21.0,
2091
+ "eval_loss": 34.59661865234375,
2092
+ "eval_runtime": 49.3345,
2093
+ "eval_samples_per_second": 26.452,
2094
+ "eval_steps_per_second": 3.324,
2095
+ "step": 27405
2096
  }
2097
  ],
2098
  "logging_steps": 100,
 
2107
  "early_stopping_threshold": 0.0
2108
  },
2109
  "attributes": {
2110
+ "early_stopping_patience_counter": 5
2111
  }
2112
  },
2113
  "TrainerControl": {
 
2121
  "attributes": {}
2122
  }
2123
  },
2124
+ "total_flos": 2.955352896399053e+16,
2125
  "train_batch_size": 8,
2126
  "trial_name": null,
2127
  "trial_params": null