mgh6 commited on
Commit
16a6469
·
verified ·
1 Parent(s): d4da27e

Training in progress, epoch 11, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0c30855507821eae60b9a8981ec29a51e2006c4710fe70f6eddd878cdf727bc5
3
  size 2682482800
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fae6d18c31da2f3123d26fe82f253230f081a7f04d6096c55edf8a79d57c0bff
3
  size 2682482800
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bd9acbde3767dd0193d5328dd03bad7ab2d4f2ea5faf5f9f997bed12d35d2d5f
3
  size 5365108834
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57eff680a79bd13191e5837ccb2bf3e205b17f06697567bf05c733d21f919c9e
3
  size 5365108834
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9e18b7a226bd09a1c0b40f0d1408dd50fa3221531aba57476aa3bd09f3bf91da
3
  size 15006
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7620d0d74aa4af76afd5408d773ed4748a8acb91b0e964a8cddab8e88f040c64
3
  size 15006
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:14ff11420ab6e05bda6bd86fde16e8ed4af7e505a292540a3a802d06de28458c
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b2b83ebf1ec1af378519da07bfb09fb79fd0733745840a41e1fc2e6668cb73a
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 10.997999636297509,
5
  "eval_steps": 50,
6
- "global_step": 3773,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1132,6 +1132,111 @@
1132
  "eval_samples_per_second": 41.449,
1133
  "eval_steps_per_second": 20.725,
1134
  "step": 3750
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1135
  }
1136
  ],
1137
  "logging_steps": 50,
@@ -1151,7 +1256,7 @@
1151
  "attributes": {}
1152
  }
1153
  },
1154
- "total_flos": 9.827787294554194e+17,
1155
  "train_batch_size": 2,
1156
  "trial_name": null,
1157
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 11.997999636297509,
5
  "eval_steps": 50,
6
+ "global_step": 4116,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1132
  "eval_samples_per_second": 41.449,
1133
  "eval_steps_per_second": 20.725,
1134
  "step": 3750
1135
+ },
1136
+ {
1137
+ "epoch": 11.078559738134206,
1138
+ "grad_norm": 61.34590530395508,
1139
+ "learning_rate": 4.46064139941691e-05,
1140
+ "loss": 0.7639,
1141
+ "step": 3800
1142
+ },
1143
+ {
1144
+ "epoch": 11.078559738134206,
1145
+ "eval_loss": 0.731596052646637,
1146
+ "eval_runtime": 116.4502,
1147
+ "eval_samples_per_second": 41.443,
1148
+ "eval_steps_per_second": 20.721,
1149
+ "step": 3800
1150
+ },
1151
+ {
1152
+ "epoch": 11.224040734679033,
1153
+ "grad_norm": 74.86040496826172,
1154
+ "learning_rate": 4.387755102040816e-05,
1155
+ "loss": 0.7515,
1156
+ "step": 3850
1157
+ },
1158
+ {
1159
+ "epoch": 11.224040734679033,
1160
+ "eval_loss": 0.7297601699829102,
1161
+ "eval_runtime": 116.5358,
1162
+ "eval_samples_per_second": 41.412,
1163
+ "eval_steps_per_second": 20.706,
1164
+ "step": 3850
1165
+ },
1166
+ {
1167
+ "epoch": 11.369521731223859,
1168
+ "grad_norm": 53.91130828857422,
1169
+ "learning_rate": 4.314868804664723e-05,
1170
+ "loss": 0.7567,
1171
+ "step": 3900
1172
+ },
1173
+ {
1174
+ "epoch": 11.369521731223859,
1175
+ "eval_loss": 0.7294139266014099,
1176
+ "eval_runtime": 116.7175,
1177
+ "eval_samples_per_second": 41.348,
1178
+ "eval_steps_per_second": 20.674,
1179
+ "step": 3900
1180
+ },
1181
+ {
1182
+ "epoch": 11.515002727768685,
1183
+ "grad_norm": 63.51275634765625,
1184
+ "learning_rate": 4.2419825072886296e-05,
1185
+ "loss": 0.7519,
1186
+ "step": 3950
1187
+ },
1188
+ {
1189
+ "epoch": 11.515002727768685,
1190
+ "eval_loss": 0.7229241728782654,
1191
+ "eval_runtime": 116.7452,
1192
+ "eval_samples_per_second": 41.338,
1193
+ "eval_steps_per_second": 20.669,
1194
+ "step": 3950
1195
+ },
1196
+ {
1197
+ "epoch": 11.660483724313512,
1198
+ "grad_norm": 44.026119232177734,
1199
+ "learning_rate": 4.1690962099125366e-05,
1200
+ "loss": 0.7494,
1201
+ "step": 4000
1202
+ },
1203
+ {
1204
+ "epoch": 11.660483724313512,
1205
+ "eval_loss": 0.7239974141120911,
1206
+ "eval_runtime": 116.4117,
1207
+ "eval_samples_per_second": 41.456,
1208
+ "eval_steps_per_second": 20.728,
1209
+ "step": 4000
1210
+ },
1211
+ {
1212
+ "epoch": 11.805964720858338,
1213
+ "grad_norm": 54.68580627441406,
1214
+ "learning_rate": 4.0962099125364436e-05,
1215
+ "loss": 0.7452,
1216
+ "step": 4050
1217
+ },
1218
+ {
1219
+ "epoch": 11.805964720858338,
1220
+ "eval_loss": 0.7212454080581665,
1221
+ "eval_runtime": 116.4854,
1222
+ "eval_samples_per_second": 41.43,
1223
+ "eval_steps_per_second": 20.715,
1224
+ "step": 4050
1225
+ },
1226
+ {
1227
+ "epoch": 11.951445717403164,
1228
+ "grad_norm": 68.30030822753906,
1229
+ "learning_rate": 4.02332361516035e-05,
1230
+ "loss": 0.7498,
1231
+ "step": 4100
1232
+ },
1233
+ {
1234
+ "epoch": 11.951445717403164,
1235
+ "eval_loss": 0.7175871133804321,
1236
+ "eval_runtime": 116.3899,
1237
+ "eval_samples_per_second": 41.464,
1238
+ "eval_steps_per_second": 20.732,
1239
+ "step": 4100
1240
  }
1241
  ],
1242
  "logging_steps": 50,
 
1256
  "attributes": {}
1257
  }
1258
  },
1259
+ "total_flos": 1.072157800433451e+18,
1260
  "train_batch_size": 2,
1261
  "trial_name": null,
1262
  "trial_params": null