JacobLinCool commited on
Commit
e33ddfe
·
verified ·
1 Parent(s): bb7595a

Training in progress, step 50, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:69250899be317f17eb6150755da5a0a35d49a63d155d515e801a858c946226c1
3
  size 8668296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24b449598bad47929b0ce7ce9703863717369e58b5e614b756d22f3d1ac8a9ba
3
  size 8668296
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cc80bb61a5290e89e718be215e85f6cf4a9fd78360d6ee75b2bc094d230cb93e
3
  size 17405562
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:accaa4c683aa5e1fe5f340f3b9cbc03bfdc27beb6f4c087a84121f607359471d
3
  size 17405562
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:905434e45334ec4cba7c0ef0c068c1bd1f793066d9f7aa304f9fefb141203eb2
3
  size 13990
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7974626dae7c2fcda8d413a7af00ed25e0f770e9e83e99fdad74474c21304fba
3
  size 13990
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:76ae49a69ba5bc0ce251dde9168db7d92fe29ee0c2ea6e44dcdcace2cd0e9412
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a225c467e7b8ca846ccba612e68b44c91016d285329de640f01561d0afab82dc
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 37.79527559055118,
3
- "best_model_checkpoint": "./exp/wft-test-model/checkpoint-30",
4
- "epoch": 2.02,
5
  "eval_steps": 10,
6
- "global_step": 40,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -292,6 +292,7 @@
292
  },
293
  {
294
  "epoch": 1.12,
 
295
  "grad_norm": 3.4594056606292725,
296
  "learning_rate": 0.00031,
297
  "loss": 0.8506,
@@ -299,6 +300,7 @@
299
  },
300
  {
301
  "epoch": 1.13,
 
302
  "grad_norm": 2.7810869216918945,
303
  "learning_rate": 0.00032,
304
  "loss": 0.6984,
@@ -306,6 +308,7 @@
306
  },
307
  {
308
  "epoch": 1.1400000000000001,
 
309
  "grad_norm": 3.656834840774536,
310
  "learning_rate": 0.00033,
311
  "loss": 0.7591,
@@ -313,6 +316,7 @@
313
  },
314
  {
315
  "epoch": 1.15,
 
316
  "grad_norm": 3.26741886138916,
317
  "learning_rate": 0.00034,
318
  "loss": 0.7747,
@@ -320,6 +324,7 @@
320
  },
321
  {
322
  "epoch": 1.16,
 
323
  "grad_norm": 2.717848777770996,
324
  "learning_rate": 0.00035,
325
  "loss": 0.7748,
@@ -327,6 +332,7 @@
327
  },
328
  {
329
  "epoch": 1.17,
 
330
  "grad_norm": 2.2561049461364746,
331
  "learning_rate": 0.00035999999999999997,
332
  "loss": 0.8973,
@@ -334,6 +340,7 @@
334
  },
335
  {
336
  "epoch": 1.18,
 
337
  "grad_norm": 3.305250406265259,
338
  "learning_rate": 0.00037,
339
  "loss": 0.8124,
@@ -341,6 +348,7 @@
341
  },
342
  {
343
  "epoch": 1.19,
 
344
  "grad_norm": 5.550938129425049,
345
  "learning_rate": 0.00038,
346
  "loss": 0.7307,
@@ -348,6 +356,7 @@
348
  },
349
  {
350
  "epoch": 2.01,
 
351
  "grad_norm": 2.3965165615081787,
352
  "learning_rate": 0.00039000000000000005,
353
  "loss": 0.5959,
@@ -355,6 +364,7 @@
355
  },
356
  {
357
  "epoch": 2.02,
 
358
  "grad_norm": 1.919407844543457,
359
  "learning_rate": 0.0004,
360
  "loss": 0.4517,
@@ -373,6 +383,90 @@
373
  "eval_wer": 40.94488188976378,
374
  "eval_wer_time": 0.02174687385559082,
375
  "step": 40
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
376
  }
377
  ],
378
  "logging_steps": 1,
@@ -392,7 +486,7 @@
392
  "attributes": {}
393
  }
394
  },
395
- "total_flos": 4270905999360000.0,
396
  "train_batch_size": 4,
397
  "trial_name": null,
398
  "trial_params": null
 
1
  {
2
+ "best_metric": 16.535433070866144,
3
+ "best_model_checkpoint": "./exp/wft-test-model/checkpoint-50",
4
+ "epoch": 2.12,
5
  "eval_steps": 10,
6
+ "global_step": 50,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
292
  },
293
  {
294
  "epoch": 1.12,
295
+ "eval_pred": null,
296
  "grad_norm": 3.4594056606292725,
297
  "learning_rate": 0.00031,
298
  "loss": 0.8506,
 
300
  },
301
  {
302
  "epoch": 1.13,
303
+ "eval_pred": null,
304
  "grad_norm": 2.7810869216918945,
305
  "learning_rate": 0.00032,
306
  "loss": 0.6984,
 
308
  },
309
  {
310
  "epoch": 1.1400000000000001,
311
+ "eval_pred": null,
312
  "grad_norm": 3.656834840774536,
313
  "learning_rate": 0.00033,
314
  "loss": 0.7591,
 
316
  },
317
  {
318
  "epoch": 1.15,
319
+ "eval_pred": null,
320
  "grad_norm": 3.26741886138916,
321
  "learning_rate": 0.00034,
322
  "loss": 0.7747,
 
324
  },
325
  {
326
  "epoch": 1.16,
327
+ "eval_pred": null,
328
  "grad_norm": 2.717848777770996,
329
  "learning_rate": 0.00035,
330
  "loss": 0.7748,
 
332
  },
333
  {
334
  "epoch": 1.17,
335
+ "eval_pred": null,
336
  "grad_norm": 2.2561049461364746,
337
  "learning_rate": 0.00035999999999999997,
338
  "loss": 0.8973,
 
340
  },
341
  {
342
  "epoch": 1.18,
343
+ "eval_pred": null,
344
  "grad_norm": 3.305250406265259,
345
  "learning_rate": 0.00037,
346
  "loss": 0.8124,
 
348
  },
349
  {
350
  "epoch": 1.19,
351
+ "eval_pred": null,
352
  "grad_norm": 5.550938129425049,
353
  "learning_rate": 0.00038,
354
  "loss": 0.7307,
 
356
  },
357
  {
358
  "epoch": 2.01,
359
+ "eval_pred": null,
360
  "grad_norm": 2.3965165615081787,
361
  "learning_rate": 0.00039000000000000005,
362
  "loss": 0.5959,
 
364
  },
365
  {
366
  "epoch": 2.02,
367
+ "eval_pred": null,
368
  "grad_norm": 1.919407844543457,
369
  "learning_rate": 0.0004,
370
  "loss": 0.4517,
 
383
  "eval_wer": 40.94488188976378,
384
  "eval_wer_time": 0.02174687385559082,
385
  "step": 40
386
+ },
387
+ {
388
+ "epoch": 2.03,
389
+ "grad_norm": 2.023576021194458,
390
+ "learning_rate": 0.00041,
391
+ "loss": 0.6744,
392
+ "step": 41
393
+ },
394
+ {
395
+ "epoch": 2.04,
396
+ "grad_norm": 1.814568281173706,
397
+ "learning_rate": 0.00042,
398
+ "loss": 0.4547,
399
+ "step": 42
400
+ },
401
+ {
402
+ "epoch": 2.05,
403
+ "grad_norm": 2.640611171722412,
404
+ "learning_rate": 0.00043,
405
+ "loss": 0.5251,
406
+ "step": 43
407
+ },
408
+ {
409
+ "epoch": 2.06,
410
+ "grad_norm": 2.0394105911254883,
411
+ "learning_rate": 0.00044,
412
+ "loss": 0.561,
413
+ "step": 44
414
+ },
415
+ {
416
+ "epoch": 2.07,
417
+ "grad_norm": 2.003967046737671,
418
+ "learning_rate": 0.00045000000000000004,
419
+ "loss": 0.4994,
420
+ "step": 45
421
+ },
422
+ {
423
+ "epoch": 2.08,
424
+ "grad_norm": 2.668008804321289,
425
+ "learning_rate": 0.00046,
426
+ "loss": 0.51,
427
+ "step": 46
428
+ },
429
+ {
430
+ "epoch": 2.09,
431
+ "grad_norm": 2.123786687850952,
432
+ "learning_rate": 0.00047,
433
+ "loss": 0.6328,
434
+ "step": 47
435
+ },
436
+ {
437
+ "epoch": 2.1,
438
+ "grad_norm": 2.4257912635803223,
439
+ "learning_rate": 0.00048,
440
+ "loss": 0.3986,
441
+ "step": 48
442
+ },
443
+ {
444
+ "epoch": 2.11,
445
+ "grad_norm": 2.094670057296753,
446
+ "learning_rate": 0.00049,
447
+ "loss": 0.4531,
448
+ "step": 49
449
+ },
450
+ {
451
+ "epoch": 2.12,
452
+ "grad_norm": 2.141101837158203,
453
+ "learning_rate": 0.0005,
454
+ "loss": 0.4352,
455
+ "step": 50
456
+ },
457
+ {
458
+ "epoch": 2.12,
459
+ "eval_cer": 88.13920454545455,
460
+ "eval_cer_time": 0.0032167434692382812,
461
+ "eval_decode_time": 0.5202732086181641,
462
+ "eval_loss": 0.4492875337600708,
463
+ "eval_pred": "| i | Label | Prediction |\n| --- | --- | --- |\n| 0 | MISTER QUILTER IS THE APOSTLE OF THE MIDDLE CLASSES AND WE ARE GLAD TO WELCOME HIS GOSPEL | MISTER QUILTER IS THE APOSTLE OF THE MIDDLE CLASSES AND WE ARE GLAD TO WELCOME HIS GOSPELBBBBBBTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTH\"-\"-\"-TH\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"- |\n| 1 | NOR IS MISTER QUILTER'S MANNER LESS INTERESTING THAN HIS MATTER | NOR IS MISTER QUILTER'S MANNER LESS INTERESTING THAN HIS MATTERSSMMMTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTH\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"- |\n| 2 | HE TELLS US THAT AT THIS FESTIVE SEASON OF THE YEAR WITH CHRISTMAS AND ROAST BEEF LOOMING BEFORE US SIMILES DRAWN FROM EATING AND ITS RESULTS OCCUR MOST READILY TO THE MIND | HE TELLS US THAT AT THIS FESTIVE SEASON OF THE YEAR WITH CHRISTMAS AND ROAST BEEF LUMING BEFORE US SIMILES DRAWN FROM EATING AND ITS RESULTS OCCUR MOST READDILY TO THE MINDSSSSSSSSSTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTH##THTHTHTHTH#TH###THTHTHTH |\n| 3 | HE HAS GRAVE DOUBTS WHETHER SIR FREDERICK LEIGHTON'S WORK IS REALLY GREEK AFTER ALL AND CAN DISCOVER IN IT BUT LITTLE OF ROCKY ITHACA | HE HAS GRAVE DOTSTS WHETHER SIR FREDIC LIGHTTON'S WORK IS REALLY GREEK AFTER ALL AND CAN DISCOVER IN IT BUT LITTLE OF ROCKY ITHAKAAAAEEEEEEEEETHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTH |\n| 4 | LINNELL'S PICTURES ARE A SORT OF UP GUARDS AND AT EM PAINTINGS AND MASON'S EXQUISITE IDYLLS ARE AS NATIONAL AS A JINGO POEM MISTER BIRKET FOSTER'S LANDSCAPES SMILE AT ONE MUCH IN THE SAME WAY THAT MISTER CARKER USED TO FLASH HIS TEETH AND MISTER JOHN COLLIER GIVES HIS SITTER A CHEERFUL SLAP ON THE BACK BEFORE HE SAYS LIKE A SHAMPOOER IN A TURKISH BATH NEXT MAN | LINELELL'S PICTURES ARE A SORT OF UP GUARDS AND ATEM PAINTINGS AND MAS'S EXQUISATE ITOLLESS ARE AS NATIONAL AS A JINGO POEM MISTER BIRKET FOSTER'S LANDSAPES SMILE AT ONE MUCH IN THE SAME WAY THAT MISTER CARCER USED TO FLASH HIS TEETH AND MISTER JOHN COLLIER GIVES HIS SITTER A CHIRFUL SLAP ON THE BACK BEFORE HE SAYS LIKE A SHAMPOOER IN A TURKISH BATH NEXT MAN |\n| 5 | IT IS OBVIOUSLY UNNECESSARY FOR US TO POINT OUT HOW LUMINOUS THESE CRITICISMS ARE HOW DELICATE IN EXPRESSION | HE IS OVIOUSLY UNNESSESSARY FOR US TO POINT OUT HOW LUMENUS THESE CRITICISMS ARE HOW DEELICATE IN EXPRESSIONENENENENEN))THTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHCOMTH-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"\"--\"-\" |\n| 6 | ON THE GENERAL PRINCIPLES OF ART MISTER QUILTER WRITES WITH EQUAL LUCIDITY | ON THE GENERAL PRINCEIPLES OF ART MISTER QUILTER RIGHTITES WITH EQUIL LUCIDITYSEEEEEEEEEEEDEEEEEEEDDDED\"-D\"-\"-D++\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"-\"- |\n| 7 | PAINTING HE TELLS US IS OF A DIFFERENT QUALITY TO MATHEMATICS AND FINISH IN ART IS ADDING MORE FACT | PINTING HE TELLS US IS OF A DIFFERENT QUALITY TO MATHEMATICS AND FINISH IN ART IS ADDING MORE FACTTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTHTH++%++++++++++++++++++++++++++++++++++++++\"-+++++\"-\"-+\"-+\"-\"-++\"-\"-\"-\"-\"-++++ |\n| 8 | AS FOR ETCHINGS THEY ARE OF TWO KINDS BRITISH AND FOREIGN | AS FOR ATCHINGS THEY ARE OF TWO KINDS BRITISH AND FOREIGNTHTHTHTHTH |\n| 9 | HE LAMENTS MOST BITTERLY THE DIVORCE THAT HAS BEEN MADE BETWEEN DECORATIVE ART AND WHAT WE USUALLY CALL PICTURES MAKES THE CUSTOMARY APPEAL TO THE LAST JUDGMENT AND REMINDS US THAT IN THE GREAT DAYS OF ART MICHAEL ANGELO WAS THE FURNISHING UPHOLSTERER | HE LAMENTS MOST BITTERLY THE DEIVORCE THAT HAS BEEN MADE BEWEEN DEECRATIVE ART AND WHAT WE USEALLY CALL PICTURES MAKEES A CUSTOMARY APPEAL TO THE LAST JGENT AND REMINES US THAT IN THE GREAT DAYS OF ART M ANGELO WAS THE FURNISHING APHOLSTTERER |\n",
464
+ "eval_runtime": 1.3178,
465
+ "eval_samples_per_second": 7.588,
466
+ "eval_steps_per_second": 1.518,
467
+ "eval_wer": 16.535433070866144,
468
+ "eval_wer_time": 0.005387067794799805,
469
+ "step": 50
470
  }
471
  ],
472
  "logging_steps": 1,
 
486
  "attributes": {}
487
  }
488
  },
489
+ "total_flos": 5380232232960000.0,
490
  "train_batch_size": 4,
491
  "trial_name": null,
492
  "trial_params": null