SystemAdmin123 commited on
Commit
bb219e9
·
verified ·
1 Parent(s): e317f0f

Training in progress, step 800, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:19465df8c765c100b1ba5242ef17b58f9b637c549a87ac2747076fa42c39e0a6
3
  size 723674912
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f20ee9d85e078def0a1968e25a4afa65844c8a597217ce9d41e6cc154906c92
3
  size 723674912
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9552643304156bc0ff61539cee518c025cee2d6eea5aac4a16b479330d4d36b2
3
  size 735625626
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8b113fad7926aed3da6821423f32ab4f6eae5ec06013cff7b635f914358c996
3
  size 735625626
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9feae33b2fec0a6229240e7adaee6ecc8f5cfdf1a8bd0e827b1d8a241424e3c0
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c431bcafebc4c8ee346d130e382b11c81be579ca0bfd3918fae07b16e10b92f
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a673aaf85c0fe6b6c29cb8f3e7dbd829eef637110e4ad9a775f3fcf001c92591
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40b6b717644e21f80a22ec98694b3a2fd9d62a6467e549d64314725dba905d52
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.11837821840781296,
5
  "eval_steps": 200,
6
- "global_step": 400,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -311,6 +311,302 @@
311
  "eval_samples_per_second": 39.591,
312
  "eval_steps_per_second": 9.911,
313
  "step": 400
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
314
  }
315
  ],
316
  "logging_steps": 10,
@@ -330,7 +626,7 @@
330
  "attributes": {}
331
  }
332
  },
333
- "total_flos": 6216909638860800.0,
334
  "train_batch_size": 4,
335
  "trial_name": null,
336
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.23675643681562591,
5
  "eval_steps": 200,
6
+ "global_step": 800,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
311
  "eval_samples_per_second": 39.591,
312
  "eval_steps_per_second": 9.911,
313
  "step": 400
314
+ },
315
+ {
316
+ "epoch": 0.12133767386800828,
317
+ "grad_norm": 2.15625,
318
+ "learning_rate": 0.00019297764858882514,
319
+ "loss": 2.3662,
320
+ "step": 410
321
+ },
322
+ {
323
+ "epoch": 0.12429712932820361,
324
+ "grad_norm": 3.171875,
325
+ "learning_rate": 0.00019248258232139388,
326
+ "loss": 2.1696,
327
+ "step": 420
328
+ },
329
+ {
330
+ "epoch": 0.12725658478839894,
331
+ "grad_norm": 4.3125,
332
+ "learning_rate": 0.00019197133427991436,
333
+ "loss": 2.0936,
334
+ "step": 430
335
+ },
336
+ {
337
+ "epoch": 0.13021604024859426,
338
+ "grad_norm": 4.15625,
339
+ "learning_rate": 0.00019144399391799043,
340
+ "loss": 1.9139,
341
+ "step": 440
342
+ },
343
+ {
344
+ "epoch": 0.1331754957087896,
345
+ "grad_norm": 17.75,
346
+ "learning_rate": 0.00019090065350491626,
347
+ "loss": 2.2588,
348
+ "step": 450
349
+ },
350
+ {
351
+ "epoch": 0.1361349511689849,
352
+ "grad_norm": 2.21875,
353
+ "learning_rate": 0.0001903414081095315,
354
+ "loss": 2.3196,
355
+ "step": 460
356
+ },
357
+ {
358
+ "epoch": 0.13909440662918024,
359
+ "grad_norm": 2.46875,
360
+ "learning_rate": 0.00018976635558358722,
361
+ "loss": 2.0432,
362
+ "step": 470
363
+ },
364
+ {
365
+ "epoch": 0.14205386208937557,
366
+ "grad_norm": 4.40625,
367
+ "learning_rate": 0.00018917559654462474,
368
+ "loss": 2.1155,
369
+ "step": 480
370
+ },
371
+ {
372
+ "epoch": 0.1450133175495709,
373
+ "grad_norm": 5.21875,
374
+ "learning_rate": 0.00018856923435837022,
375
+ "loss": 2.0727,
376
+ "step": 490
377
+ },
378
+ {
379
+ "epoch": 0.1479727730097662,
380
+ "grad_norm": 12.25,
381
+ "learning_rate": 0.0001879473751206489,
382
+ "loss": 1.5163,
383
+ "step": 500
384
+ },
385
+ {
386
+ "epoch": 0.15093222846996152,
387
+ "grad_norm": 2.171875,
388
+ "learning_rate": 0.00018731012763882133,
389
+ "loss": 2.3348,
390
+ "step": 510
391
+ },
392
+ {
393
+ "epoch": 0.15389168393015684,
394
+ "grad_norm": 3.109375,
395
+ "learning_rate": 0.00018665760341274505,
396
+ "loss": 2.1121,
397
+ "step": 520
398
+ },
399
+ {
400
+ "epoch": 0.15685113939035217,
401
+ "grad_norm": 3.75,
402
+ "learning_rate": 0.00018598991661526572,
403
+ "loss": 2.201,
404
+ "step": 530
405
+ },
406
+ {
407
+ "epoch": 0.1598105948505475,
408
+ "grad_norm": 3.859375,
409
+ "learning_rate": 0.00018530718407223974,
410
+ "loss": 2.1016,
411
+ "step": 540
412
+ },
413
+ {
414
+ "epoch": 0.16277005031074282,
415
+ "grad_norm": 16.125,
416
+ "learning_rate": 0.00018460952524209355,
417
+ "loss": 1.8499,
418
+ "step": 550
419
+ },
420
+ {
421
+ "epoch": 0.16572950577093815,
422
+ "grad_norm": 2.109375,
423
+ "learning_rate": 0.00018389706219492147,
424
+ "loss": 2.1731,
425
+ "step": 560
426
+ },
427
+ {
428
+ "epoch": 0.16868896123113347,
429
+ "grad_norm": 2.515625,
430
+ "learning_rate": 0.00018316991959112716,
431
+ "loss": 2.3097,
432
+ "step": 570
433
+ },
434
+ {
435
+ "epoch": 0.1716484166913288,
436
+ "grad_norm": 3.296875,
437
+ "learning_rate": 0.00018242822465961176,
438
+ "loss": 1.7914,
439
+ "step": 580
440
+ },
441
+ {
442
+ "epoch": 0.17460787215152412,
443
+ "grad_norm": 6.90625,
444
+ "learning_rate": 0.00018167210717551224,
445
+ "loss": 1.9812,
446
+ "step": 590
447
+ },
448
+ {
449
+ "epoch": 0.17756732761171945,
450
+ "grad_norm": 15.125,
451
+ "learning_rate": 0.00018090169943749476,
452
+ "loss": 1.7489,
453
+ "step": 600
454
+ },
455
+ {
456
+ "epoch": 0.17756732761171945,
457
+ "eval_loss": 2.0746848583221436,
458
+ "eval_runtime": 37.5843,
459
+ "eval_samples_per_second": 39.964,
460
+ "eval_steps_per_second": 10.004,
461
+ "step": 600
462
+ },
463
+ {
464
+ "epoch": 0.18052678307191478,
465
+ "grad_norm": 2.859375,
466
+ "learning_rate": 0.00018011713624460608,
467
+ "loss": 1.9754,
468
+ "step": 610
469
+ },
470
+ {
471
+ "epoch": 0.1834862385321101,
472
+ "grad_norm": 2.453125,
473
+ "learning_rate": 0.00017931855487268782,
474
+ "loss": 1.9933,
475
+ "step": 620
476
+ },
477
+ {
478
+ "epoch": 0.18644569399230543,
479
+ "grad_norm": 3.140625,
480
+ "learning_rate": 0.0001785060950503568,
481
+ "loss": 2.3726,
482
+ "step": 630
483
+ },
484
+ {
485
+ "epoch": 0.18940514945250073,
486
+ "grad_norm": 10.125,
487
+ "learning_rate": 0.00017767989893455698,
488
+ "loss": 2.0989,
489
+ "step": 640
490
+ },
491
+ {
492
+ "epoch": 0.19236460491269605,
493
+ "grad_norm": 16.125,
494
+ "learning_rate": 0.00017684011108568592,
495
+ "loss": 2.0883,
496
+ "step": 650
497
+ },
498
+ {
499
+ "epoch": 0.19532406037289138,
500
+ "grad_norm": 2.0625,
501
+ "learning_rate": 0.00017598687844230088,
502
+ "loss": 2.3784,
503
+ "step": 660
504
+ },
505
+ {
506
+ "epoch": 0.1982835158330867,
507
+ "grad_norm": 2.640625,
508
+ "learning_rate": 0.00017512035029540885,
509
+ "loss": 2.1031,
510
+ "step": 670
511
+ },
512
+ {
513
+ "epoch": 0.20124297129328203,
514
+ "grad_norm": 4.0625,
515
+ "learning_rate": 0.000174240678262345,
516
+ "loss": 2.1489,
517
+ "step": 680
518
+ },
519
+ {
520
+ "epoch": 0.20420242675347736,
521
+ "grad_norm": 6.0,
522
+ "learning_rate": 0.000173348016260244,
523
+ "loss": 1.8543,
524
+ "step": 690
525
+ },
526
+ {
527
+ "epoch": 0.20716188221367268,
528
+ "grad_norm": 18.375,
529
+ "learning_rate": 0.00017244252047910892,
530
+ "loss": 1.7568,
531
+ "step": 700
532
+ },
533
+ {
534
+ "epoch": 0.210121337673868,
535
+ "grad_norm": 2.296875,
536
+ "learning_rate": 0.00017152434935448256,
537
+ "loss": 2.1479,
538
+ "step": 710
539
+ },
540
+ {
541
+ "epoch": 0.21308079313406333,
542
+ "grad_norm": 2.796875,
543
+ "learning_rate": 0.0001705936635397259,
544
+ "loss": 2.2194,
545
+ "step": 720
546
+ },
547
+ {
548
+ "epoch": 0.21604024859425866,
549
+ "grad_norm": 3.578125,
550
+ "learning_rate": 0.00016965062587790823,
551
+ "loss": 2.0353,
552
+ "step": 730
553
+ },
554
+ {
555
+ "epoch": 0.218999704054454,
556
+ "grad_norm": 3.203125,
557
+ "learning_rate": 0.00016869540137331445,
558
+ "loss": 1.8538,
559
+ "step": 740
560
+ },
561
+ {
562
+ "epoch": 0.2219591595146493,
563
+ "grad_norm": 10.4375,
564
+ "learning_rate": 0.00016772815716257412,
565
+ "loss": 1.8197,
566
+ "step": 750
567
+ },
568
+ {
569
+ "epoch": 0.22491861497484464,
570
+ "grad_norm": 2.21875,
571
+ "learning_rate": 0.00016674906248541726,
572
+ "loss": 2.2368,
573
+ "step": 760
574
+ },
575
+ {
576
+ "epoch": 0.22787807043503996,
577
+ "grad_norm": 3.890625,
578
+ "learning_rate": 0.00016575828865506245,
579
+ "loss": 2.0362,
580
+ "step": 770
581
+ },
582
+ {
583
+ "epoch": 0.2308375258952353,
584
+ "grad_norm": 4.0625,
585
+ "learning_rate": 0.0001647560090282419,
586
+ "loss": 1.9445,
587
+ "step": 780
588
+ },
589
+ {
590
+ "epoch": 0.2337969813554306,
591
+ "grad_norm": 6.75,
592
+ "learning_rate": 0.000163742398974869,
593
+ "loss": 1.819,
594
+ "step": 790
595
+ },
596
+ {
597
+ "epoch": 0.23675643681562591,
598
+ "grad_norm": 28.25,
599
+ "learning_rate": 0.0001627176358473537,
600
+ "loss": 1.913,
601
+ "step": 800
602
+ },
603
+ {
604
+ "epoch": 0.23675643681562591,
605
+ "eval_loss": 2.06573748588562,
606
+ "eval_runtime": 37.9697,
607
+ "eval_samples_per_second": 39.558,
608
+ "eval_steps_per_second": 9.903,
609
+ "step": 800
610
  }
611
  ],
612
  "logging_steps": 10,
 
626
  "attributes": {}
627
  }
628
  },
629
+ "total_flos": 1.24028893790208e+16,
630
  "train_batch_size": 4,
631
  "trial_name": null,
632
  "trial_params": null