Training in progress, step 40, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 8668296
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:69250899be317f17eb6150755da5a0a35d49a63d155d515e801a858c946226c1
|
| 3 |
size 8668296
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 17405562
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cc80bb61a5290e89e718be215e85f6cf4a9fd78360d6ee75b2bc094d230cb93e
|
| 3 |
size 17405562
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 13990
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:905434e45334ec4cba7c0ef0c068c1bd1f793066d9f7aa304f9fefb141203eb2
|
| 3 |
size 13990
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:76ae49a69ba5bc0ce251dde9168db7d92fe29ee0c2ea6e44dcdcace2cd0e9412
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 37.79527559055118,
|
| 3 |
"best_model_checkpoint": "./exp/wft-test-model/checkpoint-30",
|
| 4 |
-
"epoch":
|
| 5 |
"eval_steps": 10,
|
| 6 |
-
"global_step":
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
@@ -198,6 +198,7 @@
|
|
| 198 |
},
|
| 199 |
{
|
| 200 |
"epoch": 1.02,
|
|
|
|
| 201 |
"grad_norm": 4.428985118865967,
|
| 202 |
"learning_rate": 0.00021,
|
| 203 |
"loss": 1.2975,
|
|
@@ -205,6 +206,7 @@
|
|
| 205 |
},
|
| 206 |
{
|
| 207 |
"epoch": 1.03,
|
|
|
|
| 208 |
"grad_norm": 4.232193470001221,
|
| 209 |
"learning_rate": 0.00022,
|
| 210 |
"loss": 1.3636,
|
|
@@ -212,6 +214,7 @@
|
|
| 212 |
},
|
| 213 |
{
|
| 214 |
"epoch": 1.04,
|
|
|
|
| 215 |
"grad_norm": 5.8515167236328125,
|
| 216 |
"learning_rate": 0.00023,
|
| 217 |
"loss": 1.3856,
|
|
@@ -219,6 +222,7 @@
|
|
| 219 |
},
|
| 220 |
{
|
| 221 |
"epoch": 1.05,
|
|
|
|
| 222 |
"grad_norm": 3.324622869491577,
|
| 223 |
"learning_rate": 0.00024,
|
| 224 |
"loss": 0.9259,
|
|
@@ -226,6 +230,7 @@
|
|
| 226 |
},
|
| 227 |
{
|
| 228 |
"epoch": 1.06,
|
|
|
|
| 229 |
"grad_norm": 5.164034366607666,
|
| 230 |
"learning_rate": 0.00025,
|
| 231 |
"loss": 1.1038,
|
|
@@ -233,6 +238,7 @@
|
|
| 233 |
},
|
| 234 |
{
|
| 235 |
"epoch": 1.07,
|
|
|
|
| 236 |
"grad_norm": 8.681169509887695,
|
| 237 |
"learning_rate": 0.00026000000000000003,
|
| 238 |
"loss": 1.4779,
|
|
@@ -240,6 +246,7 @@
|
|
| 240 |
},
|
| 241 |
{
|
| 242 |
"epoch": 1.08,
|
|
|
|
| 243 |
"grad_norm": 3.7555525302886963,
|
| 244 |
"learning_rate": 0.00027,
|
| 245 |
"loss": 1.0174,
|
|
@@ -247,6 +254,7 @@
|
|
| 247 |
},
|
| 248 |
{
|
| 249 |
"epoch": 1.09,
|
|
|
|
| 250 |
"grad_norm": 4.5491042137146,
|
| 251 |
"learning_rate": 0.00028000000000000003,
|
| 252 |
"loss": 1.1516,
|
|
@@ -254,6 +262,7 @@
|
|
| 254 |
},
|
| 255 |
{
|
| 256 |
"epoch": 1.1,
|
|
|
|
| 257 |
"grad_norm": 3.3533918857574463,
|
| 258 |
"learning_rate": 0.00029,
|
| 259 |
"loss": 0.7413,
|
|
@@ -261,6 +270,7 @@
|
|
| 261 |
},
|
| 262 |
{
|
| 263 |
"epoch": 1.11,
|
|
|
|
| 264 |
"grad_norm": 3.5604212284088135,
|
| 265 |
"learning_rate": 0.0003,
|
| 266 |
"loss": 0.8667,
|
|
@@ -279,6 +289,90 @@
|
|
| 279 |
"eval_wer": 37.79527559055118,
|
| 280 |
"eval_wer_time": 0.03960013389587402,
|
| 281 |
"step": 30
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 282 |
}
|
| 283 |
],
|
| 284 |
"logging_steps": 1,
|
|
@@ -298,7 +392,7 @@
|
|
| 298 |
"attributes": {}
|
| 299 |
}
|
| 300 |
},
|
| 301 |
-
"total_flos":
|
| 302 |
"train_batch_size": 4,
|
| 303 |
"trial_name": null,
|
| 304 |
"trial_params": null
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 37.79527559055118,
|
| 3 |
"best_model_checkpoint": "./exp/wft-test-model/checkpoint-30",
|
| 4 |
+
"epoch": 2.02,
|
| 5 |
"eval_steps": 10,
|
| 6 |
+
"global_step": 40,
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
|
|
| 198 |
},
|
| 199 |
{
|
| 200 |
"epoch": 1.02,
|
| 201 |
+
"eval_pred": null,
|
| 202 |
"grad_norm": 4.428985118865967,
|
| 203 |
"learning_rate": 0.00021,
|
| 204 |
"loss": 1.2975,
|
|
|
|
| 206 |
},
|
| 207 |
{
|
| 208 |
"epoch": 1.03,
|
| 209 |
+
"eval_pred": null,
|
| 210 |
"grad_norm": 4.232193470001221,
|
| 211 |
"learning_rate": 0.00022,
|
| 212 |
"loss": 1.3636,
|
|
|
|
| 214 |
},
|
| 215 |
{
|
| 216 |
"epoch": 1.04,
|
| 217 |
+
"eval_pred": null,
|
| 218 |
"grad_norm": 5.8515167236328125,
|
| 219 |
"learning_rate": 0.00023,
|
| 220 |
"loss": 1.3856,
|
|
|
|
| 222 |
},
|
| 223 |
{
|
| 224 |
"epoch": 1.05,
|
| 225 |
+
"eval_pred": null,
|
| 226 |
"grad_norm": 3.324622869491577,
|
| 227 |
"learning_rate": 0.00024,
|
| 228 |
"loss": 0.9259,
|
|
|
|
| 230 |
},
|
| 231 |
{
|
| 232 |
"epoch": 1.06,
|
| 233 |
+
"eval_pred": null,
|
| 234 |
"grad_norm": 5.164034366607666,
|
| 235 |
"learning_rate": 0.00025,
|
| 236 |
"loss": 1.1038,
|
|
|
|
| 238 |
},
|
| 239 |
{
|
| 240 |
"epoch": 1.07,
|
| 241 |
+
"eval_pred": null,
|
| 242 |
"grad_norm": 8.681169509887695,
|
| 243 |
"learning_rate": 0.00026000000000000003,
|
| 244 |
"loss": 1.4779,
|
|
|
|
| 246 |
},
|
| 247 |
{
|
| 248 |
"epoch": 1.08,
|
| 249 |
+
"eval_pred": null,
|
| 250 |
"grad_norm": 3.7555525302886963,
|
| 251 |
"learning_rate": 0.00027,
|
| 252 |
"loss": 1.0174,
|
|
|
|
| 254 |
},
|
| 255 |
{
|
| 256 |
"epoch": 1.09,
|
| 257 |
+
"eval_pred": null,
|
| 258 |
"grad_norm": 4.5491042137146,
|
| 259 |
"learning_rate": 0.00028000000000000003,
|
| 260 |
"loss": 1.1516,
|
|
|
|
| 262 |
},
|
| 263 |
{
|
| 264 |
"epoch": 1.1,
|
| 265 |
+
"eval_pred": null,
|
| 266 |
"grad_norm": 3.3533918857574463,
|
| 267 |
"learning_rate": 0.00029,
|
| 268 |
"loss": 0.7413,
|
|
|
|
| 270 |
},
|
| 271 |
{
|
| 272 |
"epoch": 1.11,
|
| 273 |
+
"eval_pred": null,
|
| 274 |
"grad_norm": 3.5604212284088135,
|
| 275 |
"learning_rate": 0.0003,
|
| 276 |
"loss": 0.8667,
|
|
|
|
| 289 |
"eval_wer": 37.79527559055118,
|
| 290 |
"eval_wer_time": 0.03960013389587402,
|
| 291 |
"step": 30
|
| 292 |
+
},
|
| 293 |
+
{
|
| 294 |
+
"epoch": 1.12,
|
| 295 |
+
"grad_norm": 3.4594056606292725,
|
| 296 |
+
"learning_rate": 0.00031,
|
| 297 |
+
"loss": 0.8506,
|
| 298 |
+
"step": 31
|
| 299 |
+
},
|
| 300 |
+
{
|
| 301 |
+
"epoch": 1.13,
|
| 302 |
+
"grad_norm": 2.7810869216918945,
|
| 303 |
+
"learning_rate": 0.00032,
|
| 304 |
+
"loss": 0.6984,
|
| 305 |
+
"step": 32
|
| 306 |
+
},
|
| 307 |
+
{
|
| 308 |
+
"epoch": 1.1400000000000001,
|
| 309 |
+
"grad_norm": 3.656834840774536,
|
| 310 |
+
"learning_rate": 0.00033,
|
| 311 |
+
"loss": 0.7591,
|
| 312 |
+
"step": 33
|
| 313 |
+
},
|
| 314 |
+
{
|
| 315 |
+
"epoch": 1.15,
|
| 316 |
+
"grad_norm": 3.26741886138916,
|
| 317 |
+
"learning_rate": 0.00034,
|
| 318 |
+
"loss": 0.7747,
|
| 319 |
+
"step": 34
|
| 320 |
+
},
|
| 321 |
+
{
|
| 322 |
+
"epoch": 1.16,
|
| 323 |
+
"grad_norm": 2.717848777770996,
|
| 324 |
+
"learning_rate": 0.00035,
|
| 325 |
+
"loss": 0.7748,
|
| 326 |
+
"step": 35
|
| 327 |
+
},
|
| 328 |
+
{
|
| 329 |
+
"epoch": 1.17,
|
| 330 |
+
"grad_norm": 2.2561049461364746,
|
| 331 |
+
"learning_rate": 0.00035999999999999997,
|
| 332 |
+
"loss": 0.8973,
|
| 333 |
+
"step": 36
|
| 334 |
+
},
|
| 335 |
+
{
|
| 336 |
+
"epoch": 1.18,
|
| 337 |
+
"grad_norm": 3.305250406265259,
|
| 338 |
+
"learning_rate": 0.00037,
|
| 339 |
+
"loss": 0.8124,
|
| 340 |
+
"step": 37
|
| 341 |
+
},
|
| 342 |
+
{
|
| 343 |
+
"epoch": 1.19,
|
| 344 |
+
"grad_norm": 5.550938129425049,
|
| 345 |
+
"learning_rate": 0.00038,
|
| 346 |
+
"loss": 0.7307,
|
| 347 |
+
"step": 38
|
| 348 |
+
},
|
| 349 |
+
{
|
| 350 |
+
"epoch": 2.01,
|
| 351 |
+
"grad_norm": 2.3965165615081787,
|
| 352 |
+
"learning_rate": 0.00039000000000000005,
|
| 353 |
+
"loss": 0.5959,
|
| 354 |
+
"step": 39
|
| 355 |
+
},
|
| 356 |
+
{
|
| 357 |
+
"epoch": 2.02,
|
| 358 |
+
"grad_norm": 1.919407844543457,
|
| 359 |
+
"learning_rate": 0.0004,
|
| 360 |
+
"loss": 0.4517,
|
| 361 |
+
"step": 40
|
| 362 |
+
},
|
| 363 |
+
{
|
| 364 |
+
"epoch": 2.02,
|
| 365 |
+
"eval_cer": 73.79261363636364,
|
| 366 |
+
"eval_cer_time": 0.002995014190673828,
|
| 367 |
+
"eval_decode_time": 0.5137369632720947,
|
| 368 |
+
"eval_loss": 0.6337034106254578,
|
| 369 |
+
"eval_pred": "| i | Label | Prediction |\n| --- | --- | --- |\n| 0 | MISTER QUILTER IS THE APOSTLE OF THE MIDDLE CLASSES AND WE ARE GLAD TO WELCOME HIS GOSPEL | MISTER QUILTER IS THE APOSLE OF THE MDLE CLASSES AND WE ARE GLAD TO WELCOME HIS GOSPLETHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTH#THTHTHTH# \u266a# -####TH# - - -THTHTH#\"- \u266aTHTHTH#THTH#\"-\"-##++++++\"-\"-\"-++\"-\"-\"-\"-\"-\"-\"-\"-++\"-\"-\"-\"-\"-\"-\"-\"-+\"-\"-\"-\"-\"-\"-\"-\"-+\"- |\n| 1 | NOR IS MISTER QUILTER'S MANNER LESS INTERESTING THAN HIS MATTER | NOR IS MrISTER QUILTERSS MANNER LESS INTERESTING THAN HIS MATTERMM)) - -)))) - - - - -### - - - - - -\"-\"- - - -\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-++\"-\"-\"-\"-\"-\"- |\n| 2 | HE TELLS US THAT AT THIS FESTIVE SEASON OF THE YEAR WITH CHRISTMAS AND ROAST BEEF LOOMING BEFORE US SIMILES DRAWN FROM EATING AND ITS RESULTS OCCUR MOST READILY TO THE MIND | HE TELLS US THAT AT THIS FESTTIVE SEASON OF THE YEAR WITH CHRISTMAS AND ROAST BEEEF LUMING BEFORE US SIMILES DRAWN FROM EATING AND ITS RESULTS OCRUR MOST READDILY TO THE MINDGTHTHG%%%###%####%####%####%###%%###%####%### |\n| 3 | HE HAS GRAVE DOUBTS WHETHER SIR FREDERICK LEIGHTON'S WORK IS REALLY GREEK AFTER ALL AND CAN DISCOVER IN IT BUT LITTLE OF ROCKY ITHACA | HE HAS GRAVE DOTSTS WHETHER SERIR FRDIC LITTON'SS WORK IS REALLY GREK AFTER ALL AND CAN DISCOVER IN IT BUT LITTLE OF ROCKY ITHAKAGTHTHTHGTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTH##GTHTHTH#TH####THTHTH##TH##THTHTHTH |\n| 4 | LINNELL'S PICTURES ARE A SORT OF UP GUARDS AND AT EM PAINTINGS AND MASON'S EXQUISITE IDYLLS ARE AS NATIONAL AS A JINGO POEM MISTER BIRKET FOSTER'S LANDSCAPES SMILE AT ONE MUCH IN THE SAME WAY THAT MISTER CARKER USED TO FLASH HIS TEETH AND MISTER JOHN COLLIER GIVES HIS SITTER A CHEERFUL SLAP ON THE BACK BEFORE HE SAYS LIKE A SHAMPOOER IN A TURKISH BATH NEXT MAN | LINELEL'S PICTURES ARE A SORT OF UPGARDS AND ADT PINTINGS AND MASSS EXQUISIT ITOLLEIS ARE AS NATIONAL AS A GINGO PEM MrISTER BRIRKET FOSTER'S LANDSAPES SMILE AT ONE MUCH IN THE SAME WAY THAT MISTER CARCER USED TO FASH HIS TEETH AND MISTER John COLLIER GIVES HIS CITTER A CHURFUL SLAP IN THE BACK BEFORE HE SAYS LIKE A SHAMPURER AND A TURKISH BATH NEXT MAN |\n| 5 | IT IS OBVIOUSLY UNNECESSARY FOR US TO POINT OUT HOW LUMINOUS THESE CRITICISMS ARE HOW DELICATE IN EXPRESSION | IT IS OBVIOUSLY UNNESSESSARY FOR US TO POINT OUT HOW LUMINUS THESE CRITICISMS ARE HOW DEELICATE IN EXPRESSIONTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTH-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"\"-\"--\"-\"-\"-\"\"--\"\"-\"-\"-\"--\" |\n| 6 | ON THE GENERAL PRINCIPLES OF ART MISTER QUILTER WRITES WITH EQUAL LUCIDITY | ON THE GENERAL PRINCEIPLES OF ART MISTER QUILTER RITES WITH EQUIL LUCIDITYEEE - -EEEEEEE+ -++++++++++++\"-++++++\"-\"-\"-+++\"-\"-\"-\"-+\"-+\"-++\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-e\"-\"-\"-\"-\"-\"-e\"-\"-ee\"-\"-\"-\"-\"-\"-\"-ee\"-\"- |\n| 7 | PAINTING HE TELLS US IS OF A DIFFERENT QUALITY TO MATHEMATICS AND FINISH IN ART IS ADDING MORE FACT | PINING HE TELLS US IS OF A DIFFERENT QUALITY TO MHEMATICS AND FINISH IN ART IS ADDING MORE FACT%%%%%%%%%%%%%%%%#%%%%%%++%%+++++++++++++++++++++++++++++++++++++++++++++\"-++++++++\"-+++++++ |\n| 8 | AS FOR ETCHINGS THEY ARE OF TWO KINDS BRITISH AND FOREIGN | AS FOR ECHINGS THEY ARE OF TWO KINDS BRITISH AND FOREIGN \u266a \u266a) \u266a \u266a \u266a \u266a \u266a \u266a \u266a \u266a \u266a \u266a \u266a \u266a \u266a \u266a \u266aTH |\n| 9 | HE LAMENTS MOST BITTERLY THE DIVORCE THAT HAS BEEN MADE BETWEEN DECORATIVE ART AND WHAT WE USUALLY CALL PICTURES MAKES THE CUSTOMARY APPEAL TO THE LAST JUDGMENT AND REMINDS US THAT IN THE GREAT DAYS OF ART MICHAEL ANGELO WAS THE FURNISHING UPHOLSTERER | HE LAMENTS MOST BITTERLY THE DEIVORCE THAT HAS BEEN MADE BEWEEN DEECRATIVE ART AND WHAT WE USELALLY CALL PICTURES MAKES A CUSTOMER APPEAL TO THE LAST JGENT AND REMINES US THAT IN THE GREAT DAYS OF ART M ANGELO WAS THE FURNISHING APHOLSTTERER |\n",
|
| 370 |
+
"eval_runtime": 1.3876,
|
| 371 |
+
"eval_samples_per_second": 7.207,
|
| 372 |
+
"eval_steps_per_second": 1.441,
|
| 373 |
+
"eval_wer": 40.94488188976378,
|
| 374 |
+
"eval_wer_time": 0.02174687385559082,
|
| 375 |
+
"step": 40
|
| 376 |
}
|
| 377 |
],
|
| 378 |
"logging_steps": 1,
|
|
|
|
| 392 |
"attributes": {}
|
| 393 |
}
|
| 394 |
},
|
| 395 |
+
"total_flos": 4270905999360000.0,
|
| 396 |
"train_batch_size": 4,
|
| 397 |
"trial_name": null,
|
| 398 |
"trial_params": null
|