ljcamargo commited on
Commit
442a536
·
verified ·
1 Parent(s): ed41bf9

Training in progress, step 1800, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:19f020f57d4a61360034ec006ae66facade0f2d21653389e6f14602a4142a050
3
  size 2558403928
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e3ce014c4ffdeb3ed2d1bb4f167ab2f28f0e06615589c0b4b8532fa2bda557a
3
  size 2558403928
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bc606d36977ccbf3d4b122958eb1f61737f84fd18ecaaffc4930cc8fd07e9364
3
  size 1313638993
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6bbd954c177e56d9e20277125d9cf6754fff55e598bb9de9bdec4e2cbe1afd8
3
  size 1313638993
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b62ca0429ec7faf1375b45507b4b6556f7e8ec80e94acc260c0727d9c01b414c
3
  size 14645
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c51602986df284fcdbb1c3432b905ff86ef24e41fe8fb42533918ae0e10ec6f
3
  size 14645
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ca372268f4fa9335030c0cb7aedb6cdba75f457da50e7a4034abb1a2d0843689
3
  size 1383
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa0408efb69cab96d5bab9a1aaf44cedbc9fc8d34f4cef378d81605e5c026d5c
3
  size 1383
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fa39772c5e0acad2bda0112b852159f96e3b408cefd615906b49c8dfa5e48e61
3
  size 1465
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:020caa0fd59e7adabae53f0d39b914916a456345c07ea74ca27ee8ec408257e7
3
  size 1465
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 1.2,
6
  "eval_steps": 500,
7
- "global_step": 1500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -1065,6 +1065,216 @@
1065
  "learning_rate": 7.445598762092537e-05,
1066
  "loss": 5.4473,
1067
  "step": 1500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1068
  }
1069
  ],
1070
  "logging_steps": 10,
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 1.44,
6
  "eval_steps": 500,
7
+ "global_step": 1800,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
1065
  "learning_rate": 7.445598762092537e-05,
1066
  "loss": 5.4473,
1067
  "step": 1500
1068
+ },
1069
+ {
1070
+ "epoch": 1.208,
1071
+ "grad_norm": 7.027050495147705,
1072
+ "learning_rate": 7.319053606355061e-05,
1073
+ "loss": 5.1231,
1074
+ "step": 1510
1075
+ },
1076
+ {
1077
+ "epoch": 1.216,
1078
+ "grad_norm": 6.161197662353516,
1079
+ "learning_rate": 7.192969352016383e-05,
1080
+ "loss": 5.1066,
1081
+ "step": 1520
1082
+ },
1083
+ {
1084
+ "epoch": 1.224,
1085
+ "grad_norm": 5.8758745193481445,
1086
+ "learning_rate": 7.067367675156758e-05,
1087
+ "loss": 5.1743,
1088
+ "step": 1530
1089
+ },
1090
+ {
1091
+ "epoch": 1.232,
1092
+ "grad_norm": 8.138237953186035,
1093
+ "learning_rate": 6.942270168892959e-05,
1094
+ "loss": 5.2096,
1095
+ "step": 1540
1096
+ },
1097
+ {
1098
+ "epoch": 1.24,
1099
+ "grad_norm": 7.498131275177002,
1100
+ "learning_rate": 6.817698339666066e-05,
1101
+ "loss": 5.1718,
1102
+ "step": 1550
1103
+ },
1104
+ {
1105
+ "epoch": 1.248,
1106
+ "grad_norm": 5.885447025299072,
1107
+ "learning_rate": 6.693673603544097e-05,
1108
+ "loss": 5.4034,
1109
+ "step": 1560
1110
+ },
1111
+ {
1112
+ "epoch": 1.256,
1113
+ "grad_norm": 5.564108848571777,
1114
+ "learning_rate": 6.570217282540247e-05,
1115
+ "loss": 5.0762,
1116
+ "step": 1570
1117
+ },
1118
+ {
1119
+ "epoch": 1.264,
1120
+ "grad_norm": 5.904695510864258,
1121
+ "learning_rate": 6.447350600947236e-05,
1122
+ "loss": 4.9994,
1123
+ "step": 1580
1124
+ },
1125
+ {
1126
+ "epoch": 1.272,
1127
+ "grad_norm": 6.91570520401001,
1128
+ "learning_rate": 6.325094681688503e-05,
1129
+ "loss": 5.1286,
1130
+ "step": 1590
1131
+ },
1132
+ {
1133
+ "epoch": 1.28,
1134
+ "grad_norm": 5.768303394317627,
1135
+ "learning_rate": 6.203470542686798e-05,
1136
+ "loss": 4.975,
1137
+ "step": 1600
1138
+ },
1139
+ {
1140
+ "epoch": 1.288,
1141
+ "grad_norm": 9.351409912109375,
1142
+ "learning_rate": 6.082499093250831e-05,
1143
+ "loss": 4.9613,
1144
+ "step": 1610
1145
+ },
1146
+ {
1147
+ "epoch": 1.296,
1148
+ "grad_norm": 9.572948455810547,
1149
+ "learning_rate": 5.962201130480618e-05,
1150
+ "loss": 4.7479,
1151
+ "step": 1620
1152
+ },
1153
+ {
1154
+ "epoch": 1.304,
1155
+ "grad_norm": 7.169508457183838,
1156
+ "learning_rate": 5.842597335692067e-05,
1157
+ "loss": 5.2066,
1158
+ "step": 1630
1159
+ },
1160
+ {
1161
+ "epoch": 1.312,
1162
+ "grad_norm": 7.029344081878662,
1163
+ "learning_rate": 5.7237082708615186e-05,
1164
+ "loss": 4.7997,
1165
+ "step": 1640
1166
+ },
1167
+ {
1168
+ "epoch": 1.32,
1169
+ "grad_norm": 6.824025630950928,
1170
+ "learning_rate": 5.605554375090784e-05,
1171
+ "loss": 4.6027,
1172
+ "step": 1650
1173
+ },
1174
+ {
1175
+ "epoch": 1.328,
1176
+ "grad_norm": 12.187256813049316,
1177
+ "learning_rate": 5.488155961093298e-05,
1178
+ "loss": 5.03,
1179
+ "step": 1660
1180
+ },
1181
+ {
1182
+ "epoch": 1.336,
1183
+ "grad_norm": 5.913212776184082,
1184
+ "learning_rate": 5.3715332117020154e-05,
1185
+ "loss": 4.7387,
1186
+ "step": 1670
1187
+ },
1188
+ {
1189
+ "epoch": 1.3439999999999999,
1190
+ "grad_norm": 6.245068550109863,
1191
+ "learning_rate": 5.2557061763996354e-05,
1192
+ "loss": 5.0907,
1193
+ "step": 1680
1194
+ },
1195
+ {
1196
+ "epoch": 1.3519999999999999,
1197
+ "grad_norm": 9.76169490814209,
1198
+ "learning_rate": 5.140694767871733e-05,
1199
+ "loss": 5.1469,
1200
+ "step": 1690
1201
+ },
1202
+ {
1203
+ "epoch": 1.3599999999999999,
1204
+ "grad_norm": 10.653762817382812,
1205
+ "learning_rate": 5.026518758583434e-05,
1206
+ "loss": 4.9712,
1207
+ "step": 1700
1208
+ },
1209
+ {
1210
+ "epoch": 1.3679999999999999,
1211
+ "grad_norm": 7.157083511352539,
1212
+ "learning_rate": 4.9131977773801765e-05,
1213
+ "loss": 4.6749,
1214
+ "step": 1710
1215
+ },
1216
+ {
1217
+ "epoch": 1.376,
1218
+ "grad_norm": 8.09749698638916,
1219
+ "learning_rate": 4.8007513061131736e-05,
1220
+ "loss": 4.5003,
1221
+ "step": 1720
1222
+ },
1223
+ {
1224
+ "epoch": 1.384,
1225
+ "grad_norm": 6.536433219909668,
1226
+ "learning_rate": 4.6891986762901406e-05,
1227
+ "loss": 4.2644,
1228
+ "step": 1730
1229
+ },
1230
+ {
1231
+ "epoch": 1.392,
1232
+ "grad_norm": 7.866100311279297,
1233
+ "learning_rate": 4.578559065751873e-05,
1234
+ "loss": 4.721,
1235
+ "step": 1740
1236
+ },
1237
+ {
1238
+ "epoch": 1.4,
1239
+ "grad_norm": 5.806945323944092,
1240
+ "learning_rate": 4.4688514953752515e-05,
1241
+ "loss": 4.6537,
1242
+ "step": 1750
1243
+ },
1244
+ {
1245
+ "epoch": 1.408,
1246
+ "grad_norm": 8.229887962341309,
1247
+ "learning_rate": 4.360094825803203e-05,
1248
+ "loss": 4.7068,
1249
+ "step": 1760
1250
+ },
1251
+ {
1252
+ "epoch": 1.416,
1253
+ "grad_norm": 10.401880264282227,
1254
+ "learning_rate": 4.25230775420224e-05,
1255
+ "loss": 4.5295,
1256
+ "step": 1770
1257
+ },
1258
+ {
1259
+ "epoch": 1.424,
1260
+ "grad_norm": 7.166602611541748,
1261
+ "learning_rate": 4.145508811048089e-05,
1262
+ "loss": 4.5715,
1263
+ "step": 1780
1264
+ },
1265
+ {
1266
+ "epoch": 1.432,
1267
+ "grad_norm": 8.82656478881836,
1268
+ "learning_rate": 4.039716356939981e-05,
1269
+ "loss": 4.8924,
1270
+ "step": 1790
1271
+ },
1272
+ {
1273
+ "epoch": 1.44,
1274
+ "grad_norm": 8.747090339660645,
1275
+ "learning_rate": 3.9349485794441395e-05,
1276
+ "loss": 4.5401,
1277
+ "step": 1800
1278
  }
1279
  ],
1280
  "logging_steps": 10,