CocoRoF commited on
Commit
41ab23c
·
verified ·
1 Parent(s): 8b3f33a

Training in progress, step 8000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9633a850a059cf491302b372793014c9f51d624cecee02d810f1ffb4f948cb5a
3
  size 613004648
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1964f2d5c44ba4b4c7469db7fbf8939a22f3ba4ba4fc4a9f4f93ad2d0dc89b1
3
  size 613004648
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0ecb623df2903c40cf6dd05e8ef7a7d544e2ade6ea023dc6571c5439d0804ef0
3
  size 1226096954
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd9ace385e6a3aedc5db9e21d5d93eaf44570601ca89516573fdb084906c785a
3
  size 1226096954
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9819055317e0aa1215ad120239bc4cecc175225c0dc18c98ca0bffe9f465133f
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff2a98b7b58dd272a86869334fd0edf641ab47ceb102b634b242c3ff92151a26
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e9e4824f828cb336ae95eabed6c26dfd8b564a017ccd96bc4ddd4d495508006c
3
  size 1000
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f41a5cf6a87e1088bf38fb8d7948db1a1b81802306d68edaf56a922e0f3b1c95
3
  size 1000
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 3.280224929709466,
5
  "eval_steps": 250,
6
- "global_step": 7000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -5355,6 +5355,770 @@
5355
  "eval_spearman_manhattan": 0.767145329588772,
5356
  "eval_steps_per_second": 37.861,
5357
  "step": 7000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5358
  }
5359
  ],
5360
  "logging_steps": 10,
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 3.7488284910965324,
5
  "eval_steps": 250,
6
+ "global_step": 8000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
5355
  "eval_spearman_manhattan": 0.767145329588772,
5356
  "eval_steps_per_second": 37.861,
5357
  "step": 7000
5358
+ },
5359
+ {
5360
+ "epoch": 3.2849109653233364,
5361
+ "grad_norm": 2.2868146896362305,
5362
+ "learning_rate": 9.589386129334584e-06,
5363
+ "loss": 0.1909,
5364
+ "step": 7010
5365
+ },
5366
+ {
5367
+ "epoch": 3.2895970009372073,
5368
+ "grad_norm": 2.2928576469421387,
5369
+ "learning_rate": 9.58880037488285e-06,
5370
+ "loss": 0.1867,
5371
+ "step": 7020
5372
+ },
5373
+ {
5374
+ "epoch": 3.294283036551078,
5375
+ "grad_norm": 2.36120343208313,
5376
+ "learning_rate": 9.588214620431116e-06,
5377
+ "loss": 0.2003,
5378
+ "step": 7030
5379
+ },
5380
+ {
5381
+ "epoch": 3.2989690721649483,
5382
+ "grad_norm": 2.88966965675354,
5383
+ "learning_rate": 9.587628865979383e-06,
5384
+ "loss": 0.1916,
5385
+ "step": 7040
5386
+ },
5387
+ {
5388
+ "epoch": 3.303655107778819,
5389
+ "grad_norm": 3.188649892807007,
5390
+ "learning_rate": 9.587043111527648e-06,
5391
+ "loss": 0.1839,
5392
+ "step": 7050
5393
+ },
5394
+ {
5395
+ "epoch": 3.3083411433926897,
5396
+ "grad_norm": 2.4474985599517822,
5397
+ "learning_rate": 9.586457357075915e-06,
5398
+ "loss": 0.2007,
5399
+ "step": 7060
5400
+ },
5401
+ {
5402
+ "epoch": 3.3130271790065606,
5403
+ "grad_norm": 3.1308703422546387,
5404
+ "learning_rate": 9.58587160262418e-06,
5405
+ "loss": 0.2049,
5406
+ "step": 7070
5407
+ },
5408
+ {
5409
+ "epoch": 3.317713214620431,
5410
+ "grad_norm": 2.200005292892456,
5411
+ "learning_rate": 9.585285848172447e-06,
5412
+ "loss": 0.1926,
5413
+ "step": 7080
5414
+ },
5415
+ {
5416
+ "epoch": 3.3223992502343016,
5417
+ "grad_norm": 3.012788772583008,
5418
+ "learning_rate": 9.584700093720714e-06,
5419
+ "loss": 0.1753,
5420
+ "step": 7090
5421
+ },
5422
+ {
5423
+ "epoch": 3.3270852858481725,
5424
+ "grad_norm": 2.8640036582946777,
5425
+ "learning_rate": 9.58411433926898e-06,
5426
+ "loss": 0.2004,
5427
+ "step": 7100
5428
+ },
5429
+ {
5430
+ "epoch": 3.331771321462043,
5431
+ "grad_norm": 2.8414249420166016,
5432
+ "learning_rate": 9.583528584817245e-06,
5433
+ "loss": 0.22,
5434
+ "step": 7110
5435
+ },
5436
+ {
5437
+ "epoch": 3.336457357075914,
5438
+ "grad_norm": 3.137575149536133,
5439
+ "learning_rate": 9.582942830365512e-06,
5440
+ "loss": 0.1985,
5441
+ "step": 7120
5442
+ },
5443
+ {
5444
+ "epoch": 3.3411433926897844,
5445
+ "grad_norm": 3.0193729400634766,
5446
+ "learning_rate": 9.582357075913777e-06,
5447
+ "loss": 0.1918,
5448
+ "step": 7130
5449
+ },
5450
+ {
5451
+ "epoch": 3.345829428303655,
5452
+ "grad_norm": 2.8278656005859375,
5453
+ "learning_rate": 9.581771321462044e-06,
5454
+ "loss": 0.2042,
5455
+ "step": 7140
5456
+ },
5457
+ {
5458
+ "epoch": 3.350515463917526,
5459
+ "grad_norm": 3.6021783351898193,
5460
+ "learning_rate": 9.58118556701031e-06,
5461
+ "loss": 0.2013,
5462
+ "step": 7150
5463
+ },
5464
+ {
5465
+ "epoch": 3.3552014995313963,
5466
+ "grad_norm": 2.698171615600586,
5467
+ "learning_rate": 9.580599812558576e-06,
5468
+ "loss": 0.1776,
5469
+ "step": 7160
5470
+ },
5471
+ {
5472
+ "epoch": 3.3598875351452673,
5473
+ "grad_norm": 2.7866642475128174,
5474
+ "learning_rate": 9.580014058106843e-06,
5475
+ "loss": 0.2219,
5476
+ "step": 7170
5477
+ },
5478
+ {
5479
+ "epoch": 3.3645735707591378,
5480
+ "grad_norm": 2.216646671295166,
5481
+ "learning_rate": 9.579428303655108e-06,
5482
+ "loss": 0.1971,
5483
+ "step": 7180
5484
+ },
5485
+ {
5486
+ "epoch": 3.3692596063730083,
5487
+ "grad_norm": 2.300733804702759,
5488
+ "learning_rate": 9.578842549203375e-06,
5489
+ "loss": 0.1866,
5490
+ "step": 7190
5491
+ },
5492
+ {
5493
+ "epoch": 3.373945641986879,
5494
+ "grad_norm": 2.985243320465088,
5495
+ "learning_rate": 9.578256794751642e-06,
5496
+ "loss": 0.1884,
5497
+ "step": 7200
5498
+ },
5499
+ {
5500
+ "epoch": 3.3786316776007497,
5501
+ "grad_norm": 2.4142348766326904,
5502
+ "learning_rate": 9.577671040299907e-06,
5503
+ "loss": 0.2,
5504
+ "step": 7210
5505
+ },
5506
+ {
5507
+ "epoch": 3.3833177132146206,
5508
+ "grad_norm": 3.3411073684692383,
5509
+ "learning_rate": 9.577085285848174e-06,
5510
+ "loss": 0.1961,
5511
+ "step": 7220
5512
+ },
5513
+ {
5514
+ "epoch": 3.388003748828491,
5515
+ "grad_norm": 2.240095376968384,
5516
+ "learning_rate": 9.57649953139644e-06,
5517
+ "loss": 0.1759,
5518
+ "step": 7230
5519
+ },
5520
+ {
5521
+ "epoch": 3.3926897844423616,
5522
+ "grad_norm": 2.9961328506469727,
5523
+ "learning_rate": 9.575913776944706e-06,
5524
+ "loss": 0.2184,
5525
+ "step": 7240
5526
+ },
5527
+ {
5528
+ "epoch": 3.3973758200562325,
5529
+ "grad_norm": 3.188340663909912,
5530
+ "learning_rate": 9.575328022492971e-06,
5531
+ "loss": 0.225,
5532
+ "step": 7250
5533
+ },
5534
+ {
5535
+ "epoch": 3.3973758200562325,
5536
+ "eval_loss": 0.05125829204916954,
5537
+ "eval_pearson_cosine": 0.7683713936958014,
5538
+ "eval_pearson_dot": 0.5867459849479424,
5539
+ "eval_pearson_euclidean": 0.7513948436830518,
5540
+ "eval_pearson_manhattan": 0.7514791526214619,
5541
+ "eval_runtime": 39.3707,
5542
+ "eval_samples_per_second": 38.099,
5543
+ "eval_spearman_cosine": 0.7749219401036014,
5544
+ "eval_spearman_dot": 0.589399728241704,
5545
+ "eval_spearman_euclidean": 0.7692196211001016,
5546
+ "eval_spearman_manhattan": 0.7689230301417627,
5547
+ "eval_steps_per_second": 38.099,
5548
+ "step": 7250
5549
+ },
5550
+ {
5551
+ "epoch": 3.402061855670103,
5552
+ "grad_norm": 2.632519006729126,
5553
+ "learning_rate": 9.574742268041238e-06,
5554
+ "loss": 0.2081,
5555
+ "step": 7260
5556
+ },
5557
+ {
5558
+ "epoch": 3.406747891283974,
5559
+ "grad_norm": 2.775071620941162,
5560
+ "learning_rate": 9.574156513589503e-06,
5561
+ "loss": 0.2054,
5562
+ "step": 7270
5563
+ },
5564
+ {
5565
+ "epoch": 3.4114339268978444,
5566
+ "grad_norm": 2.4009063243865967,
5567
+ "learning_rate": 9.57357075913777e-06,
5568
+ "loss": 0.188,
5569
+ "step": 7280
5570
+ },
5571
+ {
5572
+ "epoch": 3.416119962511715,
5573
+ "grad_norm": 3.2851004600524902,
5574
+ "learning_rate": 9.572985004686036e-06,
5575
+ "loss": 0.1901,
5576
+ "step": 7290
5577
+ },
5578
+ {
5579
+ "epoch": 3.420805998125586,
5580
+ "grad_norm": 2.978980541229248,
5581
+ "learning_rate": 9.572399250234303e-06,
5582
+ "loss": 0.1906,
5583
+ "step": 7300
5584
+ },
5585
+ {
5586
+ "epoch": 3.4254920337394563,
5587
+ "grad_norm": 3.4765729904174805,
5588
+ "learning_rate": 9.57181349578257e-06,
5589
+ "loss": 0.163,
5590
+ "step": 7310
5591
+ },
5592
+ {
5593
+ "epoch": 3.4301780693533273,
5594
+ "grad_norm": 2.3053479194641113,
5595
+ "learning_rate": 9.571227741330835e-06,
5596
+ "loss": 0.2051,
5597
+ "step": 7320
5598
+ },
5599
+ {
5600
+ "epoch": 3.4348641049671977,
5601
+ "grad_norm": 2.520573139190674,
5602
+ "learning_rate": 9.570641986879102e-06,
5603
+ "loss": 0.2171,
5604
+ "step": 7330
5605
+ },
5606
+ {
5607
+ "epoch": 3.4395501405810682,
5608
+ "grad_norm": 2.7145869731903076,
5609
+ "learning_rate": 9.570056232427367e-06,
5610
+ "loss": 0.1946,
5611
+ "step": 7340
5612
+ },
5613
+ {
5614
+ "epoch": 3.444236176194939,
5615
+ "grad_norm": 2.9998998641967773,
5616
+ "learning_rate": 9.569470477975634e-06,
5617
+ "loss": 0.1942,
5618
+ "step": 7350
5619
+ },
5620
+ {
5621
+ "epoch": 3.4489222118088096,
5622
+ "grad_norm": 3.165696620941162,
5623
+ "learning_rate": 9.568884723523899e-06,
5624
+ "loss": 0.2204,
5625
+ "step": 7360
5626
+ },
5627
+ {
5628
+ "epoch": 3.4536082474226806,
5629
+ "grad_norm": 2.271026849746704,
5630
+ "learning_rate": 9.568298969072166e-06,
5631
+ "loss": 0.1974,
5632
+ "step": 7370
5633
+ },
5634
+ {
5635
+ "epoch": 3.458294283036551,
5636
+ "grad_norm": 2.50876522064209,
5637
+ "learning_rate": 9.567713214620433e-06,
5638
+ "loss": 0.1998,
5639
+ "step": 7380
5640
+ },
5641
+ {
5642
+ "epoch": 3.4629803186504216,
5643
+ "grad_norm": 2.734961986541748,
5644
+ "learning_rate": 9.567127460168698e-06,
5645
+ "loss": 0.1918,
5646
+ "step": 7390
5647
+ },
5648
+ {
5649
+ "epoch": 3.4676663542642925,
5650
+ "grad_norm": 2.7971320152282715,
5651
+ "learning_rate": 9.566541705716965e-06,
5652
+ "loss": 0.2187,
5653
+ "step": 7400
5654
+ },
5655
+ {
5656
+ "epoch": 3.472352389878163,
5657
+ "grad_norm": 2.5854151248931885,
5658
+ "learning_rate": 9.56595595126523e-06,
5659
+ "loss": 0.1825,
5660
+ "step": 7410
5661
+ },
5662
+ {
5663
+ "epoch": 3.477038425492034,
5664
+ "grad_norm": 3.7421298027038574,
5665
+ "learning_rate": 9.565370196813497e-06,
5666
+ "loss": 0.2332,
5667
+ "step": 7420
5668
+ },
5669
+ {
5670
+ "epoch": 3.4817244611059044,
5671
+ "grad_norm": 3.0620522499084473,
5672
+ "learning_rate": 9.564784442361762e-06,
5673
+ "loss": 0.1976,
5674
+ "step": 7430
5675
+ },
5676
+ {
5677
+ "epoch": 3.486410496719775,
5678
+ "grad_norm": 2.0397024154663086,
5679
+ "learning_rate": 9.56419868791003e-06,
5680
+ "loss": 0.1765,
5681
+ "step": 7440
5682
+ },
5683
+ {
5684
+ "epoch": 3.491096532333646,
5685
+ "grad_norm": 3.286588191986084,
5686
+ "learning_rate": 9.563612933458294e-06,
5687
+ "loss": 0.2395,
5688
+ "step": 7450
5689
+ },
5690
+ {
5691
+ "epoch": 3.4957825679475163,
5692
+ "grad_norm": 2.812507390975952,
5693
+ "learning_rate": 9.563027179006561e-06,
5694
+ "loss": 0.1873,
5695
+ "step": 7460
5696
+ },
5697
+ {
5698
+ "epoch": 3.5004686035613872,
5699
+ "grad_norm": 3.213799476623535,
5700
+ "learning_rate": 9.562441424554827e-06,
5701
+ "loss": 0.2309,
5702
+ "step": 7470
5703
+ },
5704
+ {
5705
+ "epoch": 3.5051546391752577,
5706
+ "grad_norm": 2.9520084857940674,
5707
+ "learning_rate": 9.561855670103093e-06,
5708
+ "loss": 0.1993,
5709
+ "step": 7480
5710
+ },
5711
+ {
5712
+ "epoch": 3.509840674789128,
5713
+ "grad_norm": 2.5035924911499023,
5714
+ "learning_rate": 9.56126991565136e-06,
5715
+ "loss": 0.1745,
5716
+ "step": 7490
5717
+ },
5718
+ {
5719
+ "epoch": 3.514526710402999,
5720
+ "grad_norm": 1.8526318073272705,
5721
+ "learning_rate": 9.560684161199626e-06,
5722
+ "loss": 0.1748,
5723
+ "step": 7500
5724
+ },
5725
+ {
5726
+ "epoch": 3.514526710402999,
5727
+ "eval_loss": 0.050157614052295685,
5728
+ "eval_pearson_cosine": 0.7752481977984473,
5729
+ "eval_pearson_dot": 0.5877283656585206,
5730
+ "eval_pearson_euclidean": 0.7460582363929404,
5731
+ "eval_pearson_manhattan": 0.7459100147594686,
5732
+ "eval_runtime": 42.2146,
5733
+ "eval_samples_per_second": 35.533,
5734
+ "eval_spearman_cosine": 0.7800582604837842,
5735
+ "eval_spearman_dot": 0.5948918375623936,
5736
+ "eval_spearman_euclidean": 0.7635782236371852,
5737
+ "eval_spearman_manhattan": 0.7629943017245161,
5738
+ "eval_steps_per_second": 35.533,
5739
+ "step": 7500
5740
+ },
5741
+ {
5742
+ "epoch": 3.5192127460168696,
5743
+ "grad_norm": 2.638265371322632,
5744
+ "learning_rate": 9.560098406747893e-06,
5745
+ "loss": 0.1919,
5746
+ "step": 7510
5747
+ },
5748
+ {
5749
+ "epoch": 3.5238987816307406,
5750
+ "grad_norm": 2.9258508682250977,
5751
+ "learning_rate": 9.559512652296158e-06,
5752
+ "loss": 0.2036,
5753
+ "step": 7520
5754
+ },
5755
+ {
5756
+ "epoch": 3.528584817244611,
5757
+ "grad_norm": 3.6239101886749268,
5758
+ "learning_rate": 9.558926897844425e-06,
5759
+ "loss": 0.1855,
5760
+ "step": 7530
5761
+ },
5762
+ {
5763
+ "epoch": 3.5332708528584815,
5764
+ "grad_norm": 3.1037397384643555,
5765
+ "learning_rate": 9.558341143392692e-06,
5766
+ "loss": 0.2081,
5767
+ "step": 7540
5768
+ },
5769
+ {
5770
+ "epoch": 3.5379568884723525,
5771
+ "grad_norm": 2.8824028968811035,
5772
+ "learning_rate": 9.557755388940957e-06,
5773
+ "loss": 0.2028,
5774
+ "step": 7550
5775
+ },
5776
+ {
5777
+ "epoch": 3.542642924086223,
5778
+ "grad_norm": 2.8066060543060303,
5779
+ "learning_rate": 9.557169634489224e-06,
5780
+ "loss": 0.2031,
5781
+ "step": 7560
5782
+ },
5783
+ {
5784
+ "epoch": 3.547328959700094,
5785
+ "grad_norm": 3.691577911376953,
5786
+ "learning_rate": 9.556583880037489e-06,
5787
+ "loss": 0.2095,
5788
+ "step": 7570
5789
+ },
5790
+ {
5791
+ "epoch": 3.5520149953139644,
5792
+ "grad_norm": 1.769102692604065,
5793
+ "learning_rate": 9.555998125585754e-06,
5794
+ "loss": 0.1927,
5795
+ "step": 7580
5796
+ },
5797
+ {
5798
+ "epoch": 3.556701030927835,
5799
+ "grad_norm": 3.4198482036590576,
5800
+ "learning_rate": 9.555412371134021e-06,
5801
+ "loss": 0.217,
5802
+ "step": 7590
5803
+ },
5804
+ {
5805
+ "epoch": 3.561387066541706,
5806
+ "grad_norm": 2.0379226207733154,
5807
+ "learning_rate": 9.554826616682288e-06,
5808
+ "loss": 0.1977,
5809
+ "step": 7600
5810
+ },
5811
+ {
5812
+ "epoch": 3.5660731021555763,
5813
+ "grad_norm": 2.7821481227874756,
5814
+ "learning_rate": 9.554240862230553e-06,
5815
+ "loss": 0.2084,
5816
+ "step": 7610
5817
+ },
5818
+ {
5819
+ "epoch": 3.570759137769447,
5820
+ "grad_norm": 2.9597930908203125,
5821
+ "learning_rate": 9.55365510777882e-06,
5822
+ "loss": 0.1858,
5823
+ "step": 7620
5824
+ },
5825
+ {
5826
+ "epoch": 3.5754451733833177,
5827
+ "grad_norm": 3.193235397338867,
5828
+ "learning_rate": 9.553069353327085e-06,
5829
+ "loss": 0.1936,
5830
+ "step": 7630
5831
+ },
5832
+ {
5833
+ "epoch": 3.580131208997188,
5834
+ "grad_norm": 2.8011372089385986,
5835
+ "learning_rate": 9.552483598875352e-06,
5836
+ "loss": 0.1958,
5837
+ "step": 7640
5838
+ },
5839
+ {
5840
+ "epoch": 3.584817244611059,
5841
+ "grad_norm": 2.2420766353607178,
5842
+ "learning_rate": 9.55189784442362e-06,
5843
+ "loss": 0.1988,
5844
+ "step": 7650
5845
+ },
5846
+ {
5847
+ "epoch": 3.5895032802249296,
5848
+ "grad_norm": 2.6297037601470947,
5849
+ "learning_rate": 9.551312089971884e-06,
5850
+ "loss": 0.1809,
5851
+ "step": 7660
5852
+ },
5853
+ {
5854
+ "epoch": 3.5941893158388005,
5855
+ "grad_norm": 2.1531927585601807,
5856
+ "learning_rate": 9.550726335520151e-06,
5857
+ "loss": 0.197,
5858
+ "step": 7670
5859
+ },
5860
+ {
5861
+ "epoch": 3.598875351452671,
5862
+ "grad_norm": 3.1594254970550537,
5863
+ "learning_rate": 9.550140581068417e-06,
5864
+ "loss": 0.2049,
5865
+ "step": 7680
5866
+ },
5867
+ {
5868
+ "epoch": 3.6035613870665415,
5869
+ "grad_norm": 3.0111396312713623,
5870
+ "learning_rate": 9.549554826616683e-06,
5871
+ "loss": 0.2213,
5872
+ "step": 7690
5873
+ },
5874
+ {
5875
+ "epoch": 3.6082474226804124,
5876
+ "grad_norm": 2.9145541191101074,
5877
+ "learning_rate": 9.54896907216495e-06,
5878
+ "loss": 0.2027,
5879
+ "step": 7700
5880
+ },
5881
+ {
5882
+ "epoch": 3.612933458294283,
5883
+ "grad_norm": 2.465318441390991,
5884
+ "learning_rate": 9.548383317713216e-06,
5885
+ "loss": 0.1937,
5886
+ "step": 7710
5887
+ },
5888
+ {
5889
+ "epoch": 3.617619493908154,
5890
+ "grad_norm": 2.6440296173095703,
5891
+ "learning_rate": 9.547797563261482e-06,
5892
+ "loss": 0.2379,
5893
+ "step": 7720
5894
+ },
5895
+ {
5896
+ "epoch": 3.6223055295220243,
5897
+ "grad_norm": 3.0767295360565186,
5898
+ "learning_rate": 9.547211808809748e-06,
5899
+ "loss": 0.2245,
5900
+ "step": 7730
5901
+ },
5902
+ {
5903
+ "epoch": 3.626991565135895,
5904
+ "grad_norm": 2.5648956298828125,
5905
+ "learning_rate": 9.546626054358013e-06,
5906
+ "loss": 0.2085,
5907
+ "step": 7740
5908
+ },
5909
+ {
5910
+ "epoch": 3.6316776007497658,
5911
+ "grad_norm": 2.750375509262085,
5912
+ "learning_rate": 9.54604029990628e-06,
5913
+ "loss": 0.2045,
5914
+ "step": 7750
5915
+ },
5916
+ {
5917
+ "epoch": 3.6316776007497658,
5918
+ "eval_loss": 0.0512368269264698,
5919
+ "eval_pearson_cosine": 0.7786877204785156,
5920
+ "eval_pearson_dot": 0.6112588074907244,
5921
+ "eval_pearson_euclidean": 0.7460156578487087,
5922
+ "eval_pearson_manhattan": 0.7457383323528859,
5923
+ "eval_runtime": 40.2911,
5924
+ "eval_samples_per_second": 37.229,
5925
+ "eval_spearman_cosine": 0.7855704741784334,
5926
+ "eval_spearman_dot": 0.615574294440821,
5927
+ "eval_spearman_euclidean": 0.764218948503257,
5928
+ "eval_spearman_manhattan": 0.7636335157024128,
5929
+ "eval_steps_per_second": 37.229,
5930
+ "step": 7750
5931
+ },
5932
+ {
5933
+ "epoch": 3.6363636363636362,
5934
+ "grad_norm": 2.703327178955078,
5935
+ "learning_rate": 9.545454545454547e-06,
5936
+ "loss": 0.2241,
5937
+ "step": 7760
5938
+ },
5939
+ {
5940
+ "epoch": 3.641049671977507,
5941
+ "grad_norm": 2.124490976333618,
5942
+ "learning_rate": 9.544868791002812e-06,
5943
+ "loss": 0.1919,
5944
+ "step": 7770
5945
+ },
5946
+ {
5947
+ "epoch": 3.6457357075913777,
5948
+ "grad_norm": 2.9296751022338867,
5949
+ "learning_rate": 9.544283036551079e-06,
5950
+ "loss": 0.2316,
5951
+ "step": 7780
5952
+ },
5953
+ {
5954
+ "epoch": 3.650421743205248,
5955
+ "grad_norm": 2.1380982398986816,
5956
+ "learning_rate": 9.543697282099344e-06,
5957
+ "loss": 0.2119,
5958
+ "step": 7790
5959
+ },
5960
+ {
5961
+ "epoch": 3.655107778819119,
5962
+ "grad_norm": 2.6718966960906982,
5963
+ "learning_rate": 9.543111527647611e-06,
5964
+ "loss": 0.1828,
5965
+ "step": 7800
5966
+ },
5967
+ {
5968
+ "epoch": 3.6597938144329896,
5969
+ "grad_norm": 2.4571399688720703,
5970
+ "learning_rate": 9.542525773195878e-06,
5971
+ "loss": 0.2349,
5972
+ "step": 7810
5973
+ },
5974
+ {
5975
+ "epoch": 3.6644798500468605,
5976
+ "grad_norm": 2.056894540786743,
5977
+ "learning_rate": 9.541940018744143e-06,
5978
+ "loss": 0.1866,
5979
+ "step": 7820
5980
+ },
5981
+ {
5982
+ "epoch": 3.669165885660731,
5983
+ "grad_norm": 2.6094396114349365,
5984
+ "learning_rate": 9.54135426429241e-06,
5985
+ "loss": 0.1974,
5986
+ "step": 7830
5987
+ },
5988
+ {
5989
+ "epoch": 3.6738519212746015,
5990
+ "grad_norm": 2.887286901473999,
5991
+ "learning_rate": 9.540768509840675e-06,
5992
+ "loss": 0.2095,
5993
+ "step": 7840
5994
+ },
5995
+ {
5996
+ "epoch": 3.6785379568884724,
5997
+ "grad_norm": 2.494635581970215,
5998
+ "learning_rate": 9.540182755388942e-06,
5999
+ "loss": 0.2114,
6000
+ "step": 7850
6001
+ },
6002
+ {
6003
+ "epoch": 3.683223992502343,
6004
+ "grad_norm": 3.544060707092285,
6005
+ "learning_rate": 9.539597000937207e-06,
6006
+ "loss": 0.2046,
6007
+ "step": 7860
6008
+ },
6009
+ {
6010
+ "epoch": 3.687910028116214,
6011
+ "grad_norm": 2.7348105907440186,
6012
+ "learning_rate": 9.539011246485474e-06,
6013
+ "loss": 0.1959,
6014
+ "step": 7870
6015
+ },
6016
+ {
6017
+ "epoch": 3.6925960637300843,
6018
+ "grad_norm": 2.1513376235961914,
6019
+ "learning_rate": 9.538425492033741e-06,
6020
+ "loss": 0.1886,
6021
+ "step": 7880
6022
+ },
6023
+ {
6024
+ "epoch": 3.697282099343955,
6025
+ "grad_norm": 3.102924346923828,
6026
+ "learning_rate": 9.537839737582007e-06,
6027
+ "loss": 0.1918,
6028
+ "step": 7890
6029
+ },
6030
+ {
6031
+ "epoch": 3.7019681349578257,
6032
+ "grad_norm": 3.166285276412964,
6033
+ "learning_rate": 9.537253983130272e-06,
6034
+ "loss": 0.241,
6035
+ "step": 7900
6036
+ },
6037
+ {
6038
+ "epoch": 3.706654170571696,
6039
+ "grad_norm": 2.2208495140075684,
6040
+ "learning_rate": 9.536668228678539e-06,
6041
+ "loss": 0.2227,
6042
+ "step": 7910
6043
+ },
6044
+ {
6045
+ "epoch": 3.711340206185567,
6046
+ "grad_norm": 2.3546998500823975,
6047
+ "learning_rate": 9.536082474226806e-06,
6048
+ "loss": 0.1939,
6049
+ "step": 7920
6050
+ },
6051
+ {
6052
+ "epoch": 3.7160262417994376,
6053
+ "grad_norm": 2.840576171875,
6054
+ "learning_rate": 9.53549671977507e-06,
6055
+ "loss": 0.1999,
6056
+ "step": 7930
6057
+ },
6058
+ {
6059
+ "epoch": 3.720712277413308,
6060
+ "grad_norm": 2.5227537155151367,
6061
+ "learning_rate": 9.534910965323338e-06,
6062
+ "loss": 0.1939,
6063
+ "step": 7940
6064
+ },
6065
+ {
6066
+ "epoch": 3.725398313027179,
6067
+ "grad_norm": 2.9870660305023193,
6068
+ "learning_rate": 9.534325210871603e-06,
6069
+ "loss": 0.1905,
6070
+ "step": 7950
6071
+ },
6072
+ {
6073
+ "epoch": 3.7300843486410495,
6074
+ "grad_norm": 2.755246877670288,
6075
+ "learning_rate": 9.53373945641987e-06,
6076
+ "loss": 0.2267,
6077
+ "step": 7960
6078
+ },
6079
+ {
6080
+ "epoch": 3.7347703842549205,
6081
+ "grad_norm": 2.6030502319335938,
6082
+ "learning_rate": 9.533153701968135e-06,
6083
+ "loss": 0.2079,
6084
+ "step": 7970
6085
+ },
6086
+ {
6087
+ "epoch": 3.739456419868791,
6088
+ "grad_norm": 3.440239906311035,
6089
+ "learning_rate": 9.532567947516402e-06,
6090
+ "loss": 0.2159,
6091
+ "step": 7980
6092
+ },
6093
+ {
6094
+ "epoch": 3.7441424554826614,
6095
+ "grad_norm": 2.320068359375,
6096
+ "learning_rate": 9.531982193064669e-06,
6097
+ "loss": 0.204,
6098
+ "step": 7990
6099
+ },
6100
+ {
6101
+ "epoch": 3.7488284910965324,
6102
+ "grad_norm": 1.8982993364334106,
6103
+ "learning_rate": 9.531396438612934e-06,
6104
+ "loss": 0.1821,
6105
+ "step": 8000
6106
+ },
6107
+ {
6108
+ "epoch": 3.7488284910965324,
6109
+ "eval_loss": 0.05020766332745552,
6110
+ "eval_pearson_cosine": 0.7782144788553325,
6111
+ "eval_pearson_dot": 0.6044765942520343,
6112
+ "eval_pearson_euclidean": 0.7544531777073331,
6113
+ "eval_pearson_manhattan": 0.7542794078830823,
6114
+ "eval_runtime": 40.4619,
6115
+ "eval_samples_per_second": 37.072,
6116
+ "eval_spearman_cosine": 0.784196959157774,
6117
+ "eval_spearman_dot": 0.6068889930424554,
6118
+ "eval_spearman_euclidean": 0.7710302682957489,
6119
+ "eval_spearman_manhattan": 0.7707235385401614,
6120
+ "eval_steps_per_second": 37.072,
6121
+ "step": 8000
6122
  }
6123
  ],
6124
  "logging_steps": 10,