irishprancer commited on
Commit
ff0d141
·
verified ·
1 Parent(s): d426085

Training in progress, step 600, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f3c3d59965295ce7450641ca00b68f5a92f415b6fac45dff70940bd51e95cf96
3
  size 527048968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba7511c2605cd169f9d4925ef16d0aab1204009d184f031c3e7da4d154628430
3
  size 527048968
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3548425916972973cfad6bfd0e5b8015f30a53f0657f6ef6e17a1ab45db45c4c
3
  size 1054135994
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4450e064cd4e1c2b71c64918ea5ef37084002b46be2aef8983814714f6dfcb31
3
  size 1054135994
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f76c4c2dca19bfb0a463693a0e409b44510488650d816e566bccd2a2851e9524
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3db7d037567f9907c679104cc622f6046bdf2f33143ca3a29365b2fa4072ce8
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:da0b9a1e2fa3da24e9af8b74787d70ff4c95d9bc9b74eeab135df0350a00b462
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5ff073b2d9654300593116ff77249d188bd5752ecbe3d0cb800964ea5d21c80
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.7166430950164795,
3
  "best_model_checkpoint": "./output/checkpoint-450",
4
- "epoch": 19.565217391304348,
5
  "eval_steps": 150,
6
- "global_step": 450,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -442,6 +442,151 @@
442
  "EMA_steps_per_second": 23.424,
443
  "epoch": 19.565217391304348,
444
  "step": 450
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
445
  }
446
  ],
447
  "logging_steps": 10,
@@ -461,7 +606,7 @@
461
  "attributes": {}
462
  }
463
  },
464
- "total_flos": 1.1591800814075904e+16,
465
  "train_batch_size": 4,
466
  "trial_name": null,
467
  "trial_params": null
 
1
  {
2
  "best_metric": 0.7166430950164795,
3
  "best_model_checkpoint": "./output/checkpoint-450",
4
+ "epoch": 26.08695652173913,
5
  "eval_steps": 150,
6
+ "global_step": 600,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
442
  "EMA_steps_per_second": 23.424,
443
  "epoch": 19.565217391304348,
444
  "step": 450
445
+ },
446
+ {
447
+ "epoch": 20.0,
448
+ "grad_norm": 2.1295206546783447,
449
+ "learning_rate": 2.9996147467351856e-05,
450
+ "loss": 0.5146,
451
+ "step": 460
452
+ },
453
+ {
454
+ "epoch": 20.434782608695652,
455
+ "grad_norm": 1.2770106792449951,
456
+ "learning_rate": 2.9995930474939773e-05,
457
+ "loss": 0.4784,
458
+ "step": 470
459
+ },
460
+ {
461
+ "epoch": 20.869565217391305,
462
+ "grad_norm": 1.4706833362579346,
463
+ "learning_rate": 2.9995707538619975e-05,
464
+ "loss": 0.5705,
465
+ "step": 480
466
+ },
467
+ {
468
+ "epoch": 21.304347826086957,
469
+ "grad_norm": 1.3234336376190186,
470
+ "learning_rate": 2.9995478658480822e-05,
471
+ "loss": 0.5167,
472
+ "step": 490
473
+ },
474
+ {
475
+ "epoch": 21.73913043478261,
476
+ "grad_norm": 1.242873191833496,
477
+ "learning_rate": 2.9995243834613043e-05,
478
+ "loss": 0.5207,
479
+ "step": 500
480
+ },
481
+ {
482
+ "epoch": 22.17391304347826,
483
+ "grad_norm": 1.77828049659729,
484
+ "learning_rate": 2.9995003067109707e-05,
485
+ "loss": 0.4832,
486
+ "step": 510
487
+ },
488
+ {
489
+ "epoch": 22.608695652173914,
490
+ "grad_norm": 1.5282888412475586,
491
+ "learning_rate": 2.9994756356066246e-05,
492
+ "loss": 0.5615,
493
+ "step": 520
494
+ },
495
+ {
496
+ "epoch": 23.043478260869566,
497
+ "grad_norm": 1.7345402240753174,
498
+ "learning_rate": 2.999450370158046e-05,
499
+ "loss": 0.4929,
500
+ "step": 530
501
+ },
502
+ {
503
+ "epoch": 23.47826086956522,
504
+ "grad_norm": 1.3091520071029663,
505
+ "learning_rate": 2.9994245103752478e-05,
506
+ "loss": 0.4383,
507
+ "step": 540
508
+ },
509
+ {
510
+ "epoch": 23.91304347826087,
511
+ "grad_norm": 1.2344285249710083,
512
+ "learning_rate": 2.999398056268481e-05,
513
+ "loss": 0.5264,
514
+ "step": 550
515
+ },
516
+ {
517
+ "epoch": 24.347826086956523,
518
+ "grad_norm": 1.409712791442871,
519
+ "learning_rate": 2.9993710078482306e-05,
520
+ "loss": 0.5206,
521
+ "step": 560
522
+ },
523
+ {
524
+ "epoch": 24.782608695652176,
525
+ "grad_norm": 0.9513388872146606,
526
+ "learning_rate": 2.9993433651252185e-05,
527
+ "loss": 0.443,
528
+ "step": 570
529
+ },
530
+ {
531
+ "epoch": 25.217391304347824,
532
+ "grad_norm": 1.7172473669052124,
533
+ "learning_rate": 2.9993151281104006e-05,
534
+ "loss": 0.5329,
535
+ "step": 580
536
+ },
537
+ {
538
+ "epoch": 25.652173913043477,
539
+ "grad_norm": 1.1298749446868896,
540
+ "learning_rate": 2.9992862968149695e-05,
541
+ "loss": 0.4737,
542
+ "step": 590
543
+ },
544
+ {
545
+ "epoch": 26.08695652173913,
546
+ "grad_norm": 1.1705174446105957,
547
+ "learning_rate": 2.9992568712503533e-05,
548
+ "loss": 0.4611,
549
+ "step": 600
550
+ },
551
+ {
552
+ "epoch": 26.08695652173913,
553
+ "eval_loss": 0.7199234366416931,
554
+ "eval_runtime": 0.4236,
555
+ "eval_samples_per_second": 23.605,
556
+ "eval_steps_per_second": 23.605,
557
+ "step": 600
558
+ },
559
+ {
560
+ "Start_State_loss": 0.8601926565170288,
561
+ "Start_State_runtime": 0.4141,
562
+ "Start_State_samples_per_second": 24.146,
563
+ "Start_State_steps_per_second": 24.146,
564
+ "epoch": 26.08695652173913,
565
+ "step": 600
566
+ },
567
+ {
568
+ "Raw_Model_loss": 0.7199234366416931,
569
+ "Raw_Model_runtime": 0.4101,
570
+ "Raw_Model_samples_per_second": 24.387,
571
+ "Raw_Model_steps_per_second": 24.387,
572
+ "epoch": 26.08695652173913,
573
+ "step": 600
574
+ },
575
+ {
576
+ "SWA_loss": 0.7341524362564087,
577
+ "SWA_runtime": 0.4412,
578
+ "SWA_samples_per_second": 22.668,
579
+ "SWA_steps_per_second": 22.668,
580
+ "epoch": 26.08695652173913,
581
+ "step": 600
582
+ },
583
+ {
584
+ "EMA_loss": 0.8609703183174133,
585
+ "EMA_runtime": 0.424,
586
+ "EMA_samples_per_second": 23.585,
587
+ "EMA_steps_per_second": 23.585,
588
+ "epoch": 26.08695652173913,
589
+ "step": 600
590
  }
591
  ],
592
  "logging_steps": 10,
 
606
  "attributes": {}
607
  }
608
  },
609
+ "total_flos": 1.5363065839583232e+16,
610
  "train_batch_size": 4,
611
  "trial_name": null,
612
  "trial_params": null