KMB_SimCSE_test / last-checkpoint /trainer_state.json
CocoRoF's picture
Training in progress, step 8536, checkpoint
d2589ec verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 4.0,
"eval_steps": 250,
"global_step": 8536,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.004686035613870665,
"grad_norm": 0.5978041887283325,
"learning_rate": 1.999707122774133e-05,
"loss": 3.4332,
"step": 10
},
{
"epoch": 0.00937207122774133,
"grad_norm": 1.5130422115325928,
"learning_rate": 1.9994142455482664e-05,
"loss": 3.2454,
"step": 20
},
{
"epoch": 0.014058106841611996,
"grad_norm": 1.607094168663025,
"learning_rate": 1.9991213683223994e-05,
"loss": 2.9102,
"step": 30
},
{
"epoch": 0.01874414245548266,
"grad_norm": 2.6665945053100586,
"learning_rate": 1.9988284910965326e-05,
"loss": 2.7066,
"step": 40
},
{
"epoch": 0.023430178069353328,
"grad_norm": 3.212167978286743,
"learning_rate": 1.9985356138706656e-05,
"loss": 2.5895,
"step": 50
},
{
"epoch": 0.028116213683223992,
"grad_norm": 3.8364155292510986,
"learning_rate": 1.998242736644799e-05,
"loss": 2.4215,
"step": 60
},
{
"epoch": 0.03280224929709466,
"grad_norm": 3.560389280319214,
"learning_rate": 1.9979498594189318e-05,
"loss": 2.0972,
"step": 70
},
{
"epoch": 0.03748828491096532,
"grad_norm": 4.126266956329346,
"learning_rate": 1.9976569821930648e-05,
"loss": 1.8428,
"step": 80
},
{
"epoch": 0.04217432052483599,
"grad_norm": 4.506312370300293,
"learning_rate": 1.997364104967198e-05,
"loss": 1.5853,
"step": 90
},
{
"epoch": 0.046860356138706656,
"grad_norm": 4.499500751495361,
"learning_rate": 1.997071227741331e-05,
"loss": 1.4429,
"step": 100
},
{
"epoch": 0.05154639175257732,
"grad_norm": 4.746804237365723,
"learning_rate": 1.996778350515464e-05,
"loss": 1.3275,
"step": 110
},
{
"epoch": 0.056232427366447985,
"grad_norm": 4.36592435836792,
"learning_rate": 1.996485473289597e-05,
"loss": 1.0965,
"step": 120
},
{
"epoch": 0.06091846298031865,
"grad_norm": 4.177589416503906,
"learning_rate": 1.9961925960637302e-05,
"loss": 1.0654,
"step": 130
},
{
"epoch": 0.06560449859418932,
"grad_norm": 4.377724647521973,
"learning_rate": 1.995899718837863e-05,
"loss": 0.9687,
"step": 140
},
{
"epoch": 0.07029053420805999,
"grad_norm": 3.9814295768737793,
"learning_rate": 1.9956068416119964e-05,
"loss": 1.043,
"step": 150
},
{
"epoch": 0.07497656982193064,
"grad_norm": 3.938041925430298,
"learning_rate": 1.9953139643861294e-05,
"loss": 0.9639,
"step": 160
},
{
"epoch": 0.07966260543580131,
"grad_norm": 4.323759078979492,
"learning_rate": 1.9950210871602627e-05,
"loss": 0.9344,
"step": 170
},
{
"epoch": 0.08434864104967198,
"grad_norm": 4.037902355194092,
"learning_rate": 1.9947282099343956e-05,
"loss": 0.8302,
"step": 180
},
{
"epoch": 0.08903467666354264,
"grad_norm": 3.9691343307495117,
"learning_rate": 1.994435332708529e-05,
"loss": 0.8573,
"step": 190
},
{
"epoch": 0.09372071227741331,
"grad_norm": 3.8372201919555664,
"learning_rate": 1.994142455482662e-05,
"loss": 0.808,
"step": 200
},
{
"epoch": 0.09840674789128398,
"grad_norm": 4.224605083465576,
"learning_rate": 1.9938495782567948e-05,
"loss": 0.8942,
"step": 210
},
{
"epoch": 0.10309278350515463,
"grad_norm": 3.7245891094207764,
"learning_rate": 1.993556701030928e-05,
"loss": 0.8321,
"step": 220
},
{
"epoch": 0.1077788191190253,
"grad_norm": 4.344135761260986,
"learning_rate": 1.993263823805061e-05,
"loss": 0.8035,
"step": 230
},
{
"epoch": 0.11246485473289597,
"grad_norm": 3.6320858001708984,
"learning_rate": 1.9929709465791943e-05,
"loss": 0.723,
"step": 240
},
{
"epoch": 0.11715089034676664,
"grad_norm": 2.934122085571289,
"learning_rate": 1.9926780693533273e-05,
"loss": 0.761,
"step": 250
},
{
"epoch": 0.11715089034676664,
"eval_loss": 0.13966944813728333,
"eval_pearson_cosine": 0.7191493566389795,
"eval_pearson_dot": 0.43422507697410495,
"eval_pearson_euclidean": 0.7135360350736899,
"eval_pearson_manhattan": 0.7129270627309872,
"eval_runtime": 3.3956,
"eval_samples_per_second": 441.749,
"eval_spearman_cosine": 0.7365500082397669,
"eval_spearman_dot": 0.43017280395518565,
"eval_spearman_euclidean": 0.7209662281529704,
"eval_spearman_manhattan": 0.720539010106817,
"eval_steps_per_second": 27.683,
"step": 250
},
{
"epoch": 0.1218369259606373,
"grad_norm": 3.1349332332611084,
"learning_rate": 1.9923851921274602e-05,
"loss": 0.7916,
"step": 260
},
{
"epoch": 0.12652296157450796,
"grad_norm": 4.0211286544799805,
"learning_rate": 1.9920923149015935e-05,
"loss": 0.7147,
"step": 270
},
{
"epoch": 0.13120899718837864,
"grad_norm": 3.262589693069458,
"learning_rate": 1.9917994376757265e-05,
"loss": 0.7689,
"step": 280
},
{
"epoch": 0.1358950328022493,
"grad_norm": 4.0925397872924805,
"learning_rate": 1.9915065604498594e-05,
"loss": 0.7556,
"step": 290
},
{
"epoch": 0.14058106841611998,
"grad_norm": 3.7041923999786377,
"learning_rate": 1.9912136832239924e-05,
"loss": 0.7239,
"step": 300
},
{
"epoch": 0.14526710402999063,
"grad_norm": 3.712703227996826,
"learning_rate": 1.9909208059981257e-05,
"loss": 0.6685,
"step": 310
},
{
"epoch": 0.14995313964386128,
"grad_norm": 3.6240854263305664,
"learning_rate": 1.9906279287722586e-05,
"loss": 0.7263,
"step": 320
},
{
"epoch": 0.15463917525773196,
"grad_norm": 4.013660907745361,
"learning_rate": 1.990335051546392e-05,
"loss": 0.7287,
"step": 330
},
{
"epoch": 0.15932521087160262,
"grad_norm": 2.8780667781829834,
"learning_rate": 1.990042174320525e-05,
"loss": 0.6709,
"step": 340
},
{
"epoch": 0.1640112464854733,
"grad_norm": 3.035186290740967,
"learning_rate": 1.989749297094658e-05,
"loss": 0.6977,
"step": 350
},
{
"epoch": 0.16869728209934395,
"grad_norm": 3.433403491973877,
"learning_rate": 1.989456419868791e-05,
"loss": 0.6656,
"step": 360
},
{
"epoch": 0.1733833177132146,
"grad_norm": 3.8357644081115723,
"learning_rate": 1.9891635426429244e-05,
"loss": 0.6192,
"step": 370
},
{
"epoch": 0.1780693533270853,
"grad_norm": 3.506263494491577,
"learning_rate": 1.9888706654170573e-05,
"loss": 0.6516,
"step": 380
},
{
"epoch": 0.18275538894095594,
"grad_norm": 2.839836835861206,
"learning_rate": 1.9885777881911906e-05,
"loss": 0.6056,
"step": 390
},
{
"epoch": 0.18744142455482662,
"grad_norm": 3.2694239616394043,
"learning_rate": 1.9882849109653236e-05,
"loss": 0.6753,
"step": 400
},
{
"epoch": 0.19212746016869728,
"grad_norm": 3.5731680393218994,
"learning_rate": 1.9879920337394565e-05,
"loss": 0.6848,
"step": 410
},
{
"epoch": 0.19681349578256796,
"grad_norm": 2.8836023807525635,
"learning_rate": 1.9876991565135898e-05,
"loss": 0.659,
"step": 420
},
{
"epoch": 0.2014995313964386,
"grad_norm": 3.5604217052459717,
"learning_rate": 1.9874062792877228e-05,
"loss": 0.6745,
"step": 430
},
{
"epoch": 0.20618556701030927,
"grad_norm": 2.7834393978118896,
"learning_rate": 1.9871134020618557e-05,
"loss": 0.623,
"step": 440
},
{
"epoch": 0.21087160262417995,
"grad_norm": 3.5130224227905273,
"learning_rate": 1.986820524835989e-05,
"loss": 0.579,
"step": 450
},
{
"epoch": 0.2155576382380506,
"grad_norm": 2.9537723064422607,
"learning_rate": 1.986527647610122e-05,
"loss": 0.6042,
"step": 460
},
{
"epoch": 0.22024367385192128,
"grad_norm": 2.926415205001831,
"learning_rate": 1.986234770384255e-05,
"loss": 0.5785,
"step": 470
},
{
"epoch": 0.22492970946579194,
"grad_norm": 4.53303861618042,
"learning_rate": 1.9859418931583882e-05,
"loss": 0.5957,
"step": 480
},
{
"epoch": 0.2296157450796626,
"grad_norm": 2.5960335731506348,
"learning_rate": 1.985649015932521e-05,
"loss": 0.6223,
"step": 490
},
{
"epoch": 0.23430178069353327,
"grad_norm": 2.509232997894287,
"learning_rate": 1.985356138706654e-05,
"loss": 0.6275,
"step": 500
},
{
"epoch": 0.23430178069353327,
"eval_loss": 0.12402218580245972,
"eval_pearson_cosine": 0.7535085617332911,
"eval_pearson_dot": 0.4526570980382658,
"eval_pearson_euclidean": 0.7442016683051946,
"eval_pearson_manhattan": 0.7442320131099294,
"eval_runtime": 3.5327,
"eval_samples_per_second": 424.602,
"eval_spearman_cosine": 0.7638158285362802,
"eval_spearman_dot": 0.4533334110281352,
"eval_spearman_euclidean": 0.7505934110060137,
"eval_spearman_manhattan": 0.7505444222182343,
"eval_steps_per_second": 26.608,
"step": 500
},
{
"epoch": 0.23898781630740393,
"grad_norm": 3.2064850330352783,
"learning_rate": 1.9850632614807874e-05,
"loss": 0.6257,
"step": 510
},
{
"epoch": 0.2436738519212746,
"grad_norm": 3.235630512237549,
"learning_rate": 1.9847703842549203e-05,
"loss": 0.6069,
"step": 520
},
{
"epoch": 0.24835988753514526,
"grad_norm": 2.8067846298217773,
"learning_rate": 1.9844775070290536e-05,
"loss": 0.6137,
"step": 530
},
{
"epoch": 0.2530459231490159,
"grad_norm": 3.6333608627319336,
"learning_rate": 1.9841846298031866e-05,
"loss": 0.6111,
"step": 540
},
{
"epoch": 0.25773195876288657,
"grad_norm": 2.410837411880493,
"learning_rate": 1.98389175257732e-05,
"loss": 0.5463,
"step": 550
},
{
"epoch": 0.2624179943767573,
"grad_norm": 3.0617260932922363,
"learning_rate": 1.9835988753514528e-05,
"loss": 0.5571,
"step": 560
},
{
"epoch": 0.26710402999062793,
"grad_norm": 2.576580762863159,
"learning_rate": 1.983305998125586e-05,
"loss": 0.5428,
"step": 570
},
{
"epoch": 0.2717900656044986,
"grad_norm": 3.412797451019287,
"learning_rate": 1.983013120899719e-05,
"loss": 0.5541,
"step": 580
},
{
"epoch": 0.27647610121836924,
"grad_norm": 2.929755210876465,
"learning_rate": 1.9827202436738523e-05,
"loss": 0.5941,
"step": 590
},
{
"epoch": 0.28116213683223995,
"grad_norm": 2.80749773979187,
"learning_rate": 1.9824273664479853e-05,
"loss": 0.5702,
"step": 600
},
{
"epoch": 0.2858481724461106,
"grad_norm": 2.91896390914917,
"learning_rate": 1.9821344892221182e-05,
"loss": 0.5971,
"step": 610
},
{
"epoch": 0.29053420805998126,
"grad_norm": 2.754328727722168,
"learning_rate": 1.9818416119962512e-05,
"loss": 0.5544,
"step": 620
},
{
"epoch": 0.2952202436738519,
"grad_norm": 3.4075613021850586,
"learning_rate": 1.9815487347703845e-05,
"loss": 0.5981,
"step": 630
},
{
"epoch": 0.29990627928772257,
"grad_norm": 2.6344642639160156,
"learning_rate": 1.9812558575445174e-05,
"loss": 0.5539,
"step": 640
},
{
"epoch": 0.3045923149015933,
"grad_norm": 2.417914867401123,
"learning_rate": 1.9809629803186504e-05,
"loss": 0.5078,
"step": 650
},
{
"epoch": 0.30927835051546393,
"grad_norm": 2.52506685256958,
"learning_rate": 1.9806701030927837e-05,
"loss": 0.532,
"step": 660
},
{
"epoch": 0.3139643861293346,
"grad_norm": 2.830923557281494,
"learning_rate": 1.9803772258669166e-05,
"loss": 0.5537,
"step": 670
},
{
"epoch": 0.31865042174320524,
"grad_norm": 3.1638379096984863,
"learning_rate": 1.98008434864105e-05,
"loss": 0.5375,
"step": 680
},
{
"epoch": 0.3233364573570759,
"grad_norm": 3.447112798690796,
"learning_rate": 1.979791471415183e-05,
"loss": 0.5755,
"step": 690
},
{
"epoch": 0.3280224929709466,
"grad_norm": 2.3407933712005615,
"learning_rate": 1.9794985941893158e-05,
"loss": 0.514,
"step": 700
},
{
"epoch": 0.33270852858481725,
"grad_norm": 3.226551055908203,
"learning_rate": 1.979205716963449e-05,
"loss": 0.5167,
"step": 710
},
{
"epoch": 0.3373945641986879,
"grad_norm": 3.1462602615356445,
"learning_rate": 1.978912839737582e-05,
"loss": 0.5431,
"step": 720
},
{
"epoch": 0.34208059981255856,
"grad_norm": 3.051692008972168,
"learning_rate": 1.9786199625117153e-05,
"loss": 0.5129,
"step": 730
},
{
"epoch": 0.3467666354264292,
"grad_norm": 2.5202085971832275,
"learning_rate": 1.9783270852858483e-05,
"loss": 0.5752,
"step": 740
},
{
"epoch": 0.3514526710402999,
"grad_norm": 2.6891393661499023,
"learning_rate": 1.9780342080599816e-05,
"loss": 0.5326,
"step": 750
},
{
"epoch": 0.3514526710402999,
"eval_loss": 0.11494184285402298,
"eval_pearson_cosine": 0.7539853173929743,
"eval_pearson_dot": 0.47863358738842976,
"eval_pearson_euclidean": 0.7326912597509363,
"eval_pearson_manhattan": 0.7320283236539638,
"eval_runtime": 3.5674,
"eval_samples_per_second": 420.478,
"eval_spearman_cosine": 0.7697587773995457,
"eval_spearman_dot": 0.47367550293168154,
"eval_spearman_euclidean": 0.7466430705786936,
"eval_spearman_manhattan": 0.7460682347775467,
"eval_steps_per_second": 26.35,
"step": 750
},
{
"epoch": 0.3561387066541706,
"grad_norm": 2.9754297733306885,
"learning_rate": 1.9777413308341145e-05,
"loss": 0.4959,
"step": 760
},
{
"epoch": 0.36082474226804123,
"grad_norm": 3.2427656650543213,
"learning_rate": 1.9774484536082478e-05,
"loss": 0.539,
"step": 770
},
{
"epoch": 0.3655107778819119,
"grad_norm": 2.6402335166931152,
"learning_rate": 1.9771555763823808e-05,
"loss": 0.5328,
"step": 780
},
{
"epoch": 0.3701968134957826,
"grad_norm": 2.9217915534973145,
"learning_rate": 1.9768626991565137e-05,
"loss": 0.4895,
"step": 790
},
{
"epoch": 0.37488284910965325,
"grad_norm": 2.363999128341675,
"learning_rate": 1.976569821930647e-05,
"loss": 0.4805,
"step": 800
},
{
"epoch": 0.3795688847235239,
"grad_norm": 2.491255044937134,
"learning_rate": 1.97627694470478e-05,
"loss": 0.5281,
"step": 810
},
{
"epoch": 0.38425492033739456,
"grad_norm": 2.595560312271118,
"learning_rate": 1.975984067478913e-05,
"loss": 0.5104,
"step": 820
},
{
"epoch": 0.3889409559512652,
"grad_norm": 2.991326332092285,
"learning_rate": 1.975691190253046e-05,
"loss": 0.508,
"step": 830
},
{
"epoch": 0.3936269915651359,
"grad_norm": 2.6440255641937256,
"learning_rate": 1.975398313027179e-05,
"loss": 0.5643,
"step": 840
},
{
"epoch": 0.3983130271790066,
"grad_norm": 3.06538987159729,
"learning_rate": 1.975105435801312e-05,
"loss": 0.4944,
"step": 850
},
{
"epoch": 0.4029990627928772,
"grad_norm": 2.751331329345703,
"learning_rate": 1.9748125585754454e-05,
"loss": 0.5215,
"step": 860
},
{
"epoch": 0.4076850984067479,
"grad_norm": 3.2387125492095947,
"learning_rate": 1.9745196813495783e-05,
"loss": 0.5925,
"step": 870
},
{
"epoch": 0.41237113402061853,
"grad_norm": 2.5455057621002197,
"learning_rate": 1.9742268041237116e-05,
"loss": 0.4956,
"step": 880
},
{
"epoch": 0.41705716963448924,
"grad_norm": 2.864497423171997,
"learning_rate": 1.9739339268978446e-05,
"loss": 0.4855,
"step": 890
},
{
"epoch": 0.4217432052483599,
"grad_norm": 2.6306967735290527,
"learning_rate": 1.9736410496719775e-05,
"loss": 0.4542,
"step": 900
},
{
"epoch": 0.42642924086223055,
"grad_norm": 2.294090747833252,
"learning_rate": 1.9733481724461108e-05,
"loss": 0.4247,
"step": 910
},
{
"epoch": 0.4311152764761012,
"grad_norm": 2.9401843547821045,
"learning_rate": 1.9730552952202438e-05,
"loss": 0.5219,
"step": 920
},
{
"epoch": 0.43580131208997186,
"grad_norm": 2.5331194400787354,
"learning_rate": 1.972762417994377e-05,
"loss": 0.5193,
"step": 930
},
{
"epoch": 0.44048734770384257,
"grad_norm": 3.265944719314575,
"learning_rate": 1.97246954076851e-05,
"loss": 0.4733,
"step": 940
},
{
"epoch": 0.4451733833177132,
"grad_norm": 2.6586129665374756,
"learning_rate": 1.9721766635426433e-05,
"loss": 0.4939,
"step": 950
},
{
"epoch": 0.4498594189315839,
"grad_norm": 2.126664638519287,
"learning_rate": 1.9718837863167762e-05,
"loss": 0.5149,
"step": 960
},
{
"epoch": 0.45454545454545453,
"grad_norm": 2.8597347736358643,
"learning_rate": 1.9715909090909092e-05,
"loss": 0.5089,
"step": 970
},
{
"epoch": 0.4592314901593252,
"grad_norm": 2.5802338123321533,
"learning_rate": 1.9712980318650425e-05,
"loss": 0.5264,
"step": 980
},
{
"epoch": 0.4639175257731959,
"grad_norm": 2.4247639179229736,
"learning_rate": 1.9710051546391754e-05,
"loss": 0.5416,
"step": 990
},
{
"epoch": 0.46860356138706655,
"grad_norm": 2.6614298820495605,
"learning_rate": 1.9707122774133084e-05,
"loss": 0.4917,
"step": 1000
},
{
"epoch": 0.46860356138706655,
"eval_loss": 0.1028011292219162,
"eval_pearson_cosine": 0.7630335129215418,
"eval_pearson_dot": 0.5427640329363896,
"eval_pearson_euclidean": 0.7394698860945113,
"eval_pearson_manhattan": 0.7395035409540256,
"eval_runtime": 3.2727,
"eval_samples_per_second": 458.333,
"eval_spearman_cosine": 0.7777749785163052,
"eval_spearman_dot": 0.5403903388556195,
"eval_spearman_euclidean": 0.7530551416243708,
"eval_spearman_manhattan": 0.753178994867466,
"eval_steps_per_second": 28.722,
"step": 1000
},
{
"epoch": 0.4732895970009372,
"grad_norm": 2.8377857208251953,
"learning_rate": 1.9704194001874417e-05,
"loss": 0.4806,
"step": 1010
},
{
"epoch": 0.47797563261480785,
"grad_norm": 2.3730361461639404,
"learning_rate": 1.9701265229615746e-05,
"loss": 0.4946,
"step": 1020
},
{
"epoch": 0.48266166822867856,
"grad_norm": 3.101628541946411,
"learning_rate": 1.9698336457357076e-05,
"loss": 0.4958,
"step": 1030
},
{
"epoch": 0.4873477038425492,
"grad_norm": 2.4200384616851807,
"learning_rate": 1.969540768509841e-05,
"loss": 0.5231,
"step": 1040
},
{
"epoch": 0.49203373945641987,
"grad_norm": 2.9535088539123535,
"learning_rate": 1.9692478912839738e-05,
"loss": 0.4952,
"step": 1050
},
{
"epoch": 0.4967197750702905,
"grad_norm": 2.3175110816955566,
"learning_rate": 1.968955014058107e-05,
"loss": 0.4845,
"step": 1060
},
{
"epoch": 0.5014058106841612,
"grad_norm": 2.934502124786377,
"learning_rate": 1.96866213683224e-05,
"loss": 0.5229,
"step": 1070
},
{
"epoch": 0.5060918462980318,
"grad_norm": 2.5961878299713135,
"learning_rate": 1.9683692596063733e-05,
"loss": 0.5111,
"step": 1080
},
{
"epoch": 0.5107778819119025,
"grad_norm": 2.9093525409698486,
"learning_rate": 1.9680763823805063e-05,
"loss": 0.545,
"step": 1090
},
{
"epoch": 0.5154639175257731,
"grad_norm": 2.733900547027588,
"learning_rate": 1.9677835051546392e-05,
"loss": 0.4589,
"step": 1100
},
{
"epoch": 0.5201499531396439,
"grad_norm": 2.3905482292175293,
"learning_rate": 1.9674906279287725e-05,
"loss": 0.5197,
"step": 1110
},
{
"epoch": 0.5248359887535146,
"grad_norm": 2.5557029247283936,
"learning_rate": 1.9671977507029055e-05,
"loss": 0.5089,
"step": 1120
},
{
"epoch": 0.5295220243673852,
"grad_norm": 2.3789846897125244,
"learning_rate": 1.9669048734770388e-05,
"loss": 0.5373,
"step": 1130
},
{
"epoch": 0.5342080599812559,
"grad_norm": 2.848998785018921,
"learning_rate": 1.9666119962511717e-05,
"loss": 0.5321,
"step": 1140
},
{
"epoch": 0.5388940955951266,
"grad_norm": 2.2063546180725098,
"learning_rate": 1.9663191190253047e-05,
"loss": 0.4809,
"step": 1150
},
{
"epoch": 0.5435801312089972,
"grad_norm": 2.7333438396453857,
"learning_rate": 1.966026241799438e-05,
"loss": 0.4615,
"step": 1160
},
{
"epoch": 0.5482661668228679,
"grad_norm": 3.124687671661377,
"learning_rate": 1.965733364573571e-05,
"loss": 0.5177,
"step": 1170
},
{
"epoch": 0.5529522024367385,
"grad_norm": 2.726681709289551,
"learning_rate": 1.965440487347704e-05,
"loss": 0.4891,
"step": 1180
},
{
"epoch": 0.5576382380506092,
"grad_norm": 2.9802463054656982,
"learning_rate": 1.965147610121837e-05,
"loss": 0.5018,
"step": 1190
},
{
"epoch": 0.5623242736644799,
"grad_norm": 2.844006299972534,
"learning_rate": 1.96485473289597e-05,
"loss": 0.5149,
"step": 1200
},
{
"epoch": 0.5670103092783505,
"grad_norm": 2.442472219467163,
"learning_rate": 1.964561855670103e-05,
"loss": 0.4808,
"step": 1210
},
{
"epoch": 0.5716963448922212,
"grad_norm": 1.9972388744354248,
"learning_rate": 1.9642689784442363e-05,
"loss": 0.4743,
"step": 1220
},
{
"epoch": 0.5763823805060918,
"grad_norm": 2.6430776119232178,
"learning_rate": 1.9639761012183693e-05,
"loss": 0.5057,
"step": 1230
},
{
"epoch": 0.5810684161199625,
"grad_norm": 2.37652850151062,
"learning_rate": 1.9636832239925026e-05,
"loss": 0.4851,
"step": 1240
},
{
"epoch": 0.5857544517338332,
"grad_norm": 2.755315065383911,
"learning_rate": 1.9633903467666355e-05,
"loss": 0.4451,
"step": 1250
},
{
"epoch": 0.5857544517338332,
"eval_loss": 0.09590280055999756,
"eval_pearson_cosine": 0.7634019154373535,
"eval_pearson_dot": 0.5909367397822605,
"eval_pearson_euclidean": 0.7507597414427352,
"eval_pearson_manhattan": 0.7505103943331619,
"eval_runtime": 3.3916,
"eval_samples_per_second": 442.267,
"eval_spearman_cosine": 0.7803407693262222,
"eval_spearman_dot": 0.592884901593322,
"eval_spearman_euclidean": 0.7652017734526011,
"eval_spearman_manhattan": 0.7648713625602643,
"eval_steps_per_second": 27.715,
"step": 1250
},
{
"epoch": 0.5904404873477038,
"grad_norm": 3.0767931938171387,
"learning_rate": 1.9630974695407688e-05,
"loss": 0.5093,
"step": 1260
},
{
"epoch": 0.5951265229615745,
"grad_norm": 3.0984067916870117,
"learning_rate": 1.9628045923149017e-05,
"loss": 0.4888,
"step": 1270
},
{
"epoch": 0.5998125585754451,
"grad_norm": 2.496957778930664,
"learning_rate": 1.962511715089035e-05,
"loss": 0.5005,
"step": 1280
},
{
"epoch": 0.6044985941893158,
"grad_norm": 2.1109137535095215,
"learning_rate": 1.962218837863168e-05,
"loss": 0.4729,
"step": 1290
},
{
"epoch": 0.6091846298031866,
"grad_norm": 2.663804054260254,
"learning_rate": 1.961925960637301e-05,
"loss": 0.4454,
"step": 1300
},
{
"epoch": 0.6138706654170571,
"grad_norm": 1.8708107471466064,
"learning_rate": 1.9616330834114342e-05,
"loss": 0.4949,
"step": 1310
},
{
"epoch": 0.6185567010309279,
"grad_norm": 2.9178926944732666,
"learning_rate": 1.9613402061855672e-05,
"loss": 0.4438,
"step": 1320
},
{
"epoch": 0.6232427366447985,
"grad_norm": 2.1928579807281494,
"learning_rate": 1.9610473289597005e-05,
"loss": 0.4483,
"step": 1330
},
{
"epoch": 0.6279287722586692,
"grad_norm": 2.3122403621673584,
"learning_rate": 1.9607544517338334e-05,
"loss": 0.4677,
"step": 1340
},
{
"epoch": 0.6326148078725399,
"grad_norm": 2.5242831707000732,
"learning_rate": 1.9604615745079664e-05,
"loss": 0.4822,
"step": 1350
},
{
"epoch": 0.6373008434864105,
"grad_norm": 2.572446346282959,
"learning_rate": 1.9601686972820993e-05,
"loss": 0.46,
"step": 1360
},
{
"epoch": 0.6419868791002812,
"grad_norm": 2.6114673614501953,
"learning_rate": 1.9598758200562326e-05,
"loss": 0.4606,
"step": 1370
},
{
"epoch": 0.6466729147141518,
"grad_norm": 2.6151556968688965,
"learning_rate": 1.9595829428303655e-05,
"loss": 0.459,
"step": 1380
},
{
"epoch": 0.6513589503280225,
"grad_norm": 2.9040744304656982,
"learning_rate": 1.959290065604499e-05,
"loss": 0.4994,
"step": 1390
},
{
"epoch": 0.6560449859418932,
"grad_norm": 2.9549591541290283,
"learning_rate": 1.9589971883786318e-05,
"loss": 0.4335,
"step": 1400
},
{
"epoch": 0.6607310215557638,
"grad_norm": 2.2790753841400146,
"learning_rate": 1.9587043111527647e-05,
"loss": 0.4642,
"step": 1410
},
{
"epoch": 0.6654170571696345,
"grad_norm": 2.265308141708374,
"learning_rate": 1.958411433926898e-05,
"loss": 0.4654,
"step": 1420
},
{
"epoch": 0.6701030927835051,
"grad_norm": 2.1528425216674805,
"learning_rate": 1.958118556701031e-05,
"loss": 0.4983,
"step": 1430
},
{
"epoch": 0.6747891283973758,
"grad_norm": 2.840533494949341,
"learning_rate": 1.9578256794751643e-05,
"loss": 0.4989,
"step": 1440
},
{
"epoch": 0.6794751640112465,
"grad_norm": 2.4118032455444336,
"learning_rate": 1.9575328022492972e-05,
"loss": 0.4207,
"step": 1450
},
{
"epoch": 0.6841611996251171,
"grad_norm": 2.845712900161743,
"learning_rate": 1.9572399250234305e-05,
"loss": 0.4219,
"step": 1460
},
{
"epoch": 0.6888472352389878,
"grad_norm": 2.6068544387817383,
"learning_rate": 1.9569470477975635e-05,
"loss": 0.4618,
"step": 1470
},
{
"epoch": 0.6935332708528584,
"grad_norm": 2.6239466667175293,
"learning_rate": 1.9566541705716967e-05,
"loss": 0.4613,
"step": 1480
},
{
"epoch": 0.6982193064667291,
"grad_norm": 2.417541742324829,
"learning_rate": 1.9563612933458297e-05,
"loss": 0.4457,
"step": 1490
},
{
"epoch": 0.7029053420805998,
"grad_norm": 2.444243907928467,
"learning_rate": 1.9560684161199626e-05,
"loss": 0.4682,
"step": 1500
},
{
"epoch": 0.7029053420805998,
"eval_loss": 0.10573244839906693,
"eval_pearson_cosine": 0.7687240391174335,
"eval_pearson_dot": 0.52706539661569,
"eval_pearson_euclidean": 0.7544794171981408,
"eval_pearson_manhattan": 0.7540749762273435,
"eval_runtime": 3.5044,
"eval_samples_per_second": 428.028,
"eval_spearman_cosine": 0.7854868415020917,
"eval_spearman_dot": 0.5189687789236848,
"eval_spearman_euclidean": 0.7685180176151998,
"eval_spearman_manhattan": 0.7680518081256867,
"eval_steps_per_second": 26.823,
"step": 1500
},
{
"epoch": 0.7075913776944704,
"grad_norm": 2.4107391834259033,
"learning_rate": 1.955775538894096e-05,
"loss": 0.4231,
"step": 1510
},
{
"epoch": 0.7122774133083412,
"grad_norm": 2.2897465229034424,
"learning_rate": 1.955482661668229e-05,
"loss": 0.495,
"step": 1520
},
{
"epoch": 0.7169634489222118,
"grad_norm": 2.1320059299468994,
"learning_rate": 1.9551897844423618e-05,
"loss": 0.4362,
"step": 1530
},
{
"epoch": 0.7216494845360825,
"grad_norm": 2.1766531467437744,
"learning_rate": 1.9548969072164948e-05,
"loss": 0.4324,
"step": 1540
},
{
"epoch": 0.7263355201499532,
"grad_norm": 2.1343445777893066,
"learning_rate": 1.954604029990628e-05,
"loss": 0.4356,
"step": 1550
},
{
"epoch": 0.7310215557638238,
"grad_norm": 2.6357688903808594,
"learning_rate": 1.954311152764761e-05,
"loss": 0.4617,
"step": 1560
},
{
"epoch": 0.7357075913776945,
"grad_norm": 2.2863566875457764,
"learning_rate": 1.9540182755388943e-05,
"loss": 0.4444,
"step": 1570
},
{
"epoch": 0.7403936269915652,
"grad_norm": 1.9845598936080933,
"learning_rate": 1.9537253983130273e-05,
"loss": 0.4459,
"step": 1580
},
{
"epoch": 0.7450796626054358,
"grad_norm": 2.3808443546295166,
"learning_rate": 1.9534325210871605e-05,
"loss": 0.4664,
"step": 1590
},
{
"epoch": 0.7497656982193065,
"grad_norm": 2.976896286010742,
"learning_rate": 1.9531396438612935e-05,
"loss": 0.4317,
"step": 1600
},
{
"epoch": 0.7544517338331771,
"grad_norm": 3.114450693130493,
"learning_rate": 1.9528467666354264e-05,
"loss": 0.4365,
"step": 1610
},
{
"epoch": 0.7591377694470478,
"grad_norm": 2.208798408508301,
"learning_rate": 1.9525538894095597e-05,
"loss": 0.4357,
"step": 1620
},
{
"epoch": 0.7638238050609185,
"grad_norm": 3.005563259124756,
"learning_rate": 1.9522610121836927e-05,
"loss": 0.4184,
"step": 1630
},
{
"epoch": 0.7685098406747891,
"grad_norm": 2.711338758468628,
"learning_rate": 1.951968134957826e-05,
"loss": 0.4479,
"step": 1640
},
{
"epoch": 0.7731958762886598,
"grad_norm": 1.9399685859680176,
"learning_rate": 1.951675257731959e-05,
"loss": 0.4642,
"step": 1650
},
{
"epoch": 0.7778819119025304,
"grad_norm": 2.5553548336029053,
"learning_rate": 1.9513823805060922e-05,
"loss": 0.4851,
"step": 1660
},
{
"epoch": 0.7825679475164011,
"grad_norm": 2.7015600204467773,
"learning_rate": 1.951089503280225e-05,
"loss": 0.4294,
"step": 1670
},
{
"epoch": 0.7872539831302718,
"grad_norm": 3.027247428894043,
"learning_rate": 1.950796626054358e-05,
"loss": 0.4359,
"step": 1680
},
{
"epoch": 0.7919400187441424,
"grad_norm": 2.603200674057007,
"learning_rate": 1.9505037488284914e-05,
"loss": 0.4833,
"step": 1690
},
{
"epoch": 0.7966260543580131,
"grad_norm": 2.2245004177093506,
"learning_rate": 1.9502108716026243e-05,
"loss": 0.4633,
"step": 1700
},
{
"epoch": 0.8013120899718837,
"grad_norm": 2.1866180896759033,
"learning_rate": 1.9499179943767573e-05,
"loss": 0.4465,
"step": 1710
},
{
"epoch": 0.8059981255857545,
"grad_norm": 2.464083671569824,
"learning_rate": 1.9496251171508906e-05,
"loss": 0.4282,
"step": 1720
},
{
"epoch": 0.8106841611996252,
"grad_norm": 2.5016162395477295,
"learning_rate": 1.9493322399250235e-05,
"loss": 0.4566,
"step": 1730
},
{
"epoch": 0.8153701968134958,
"grad_norm": 2.9877169132232666,
"learning_rate": 1.9490393626991565e-05,
"loss": 0.462,
"step": 1740
},
{
"epoch": 0.8200562324273665,
"grad_norm": 2.2195188999176025,
"learning_rate": 1.9487464854732898e-05,
"loss": 0.4489,
"step": 1750
},
{
"epoch": 0.8200562324273665,
"eval_loss": 0.09936515241861343,
"eval_pearson_cosine": 0.7657869832789288,
"eval_pearson_dot": 0.576463122226734,
"eval_pearson_euclidean": 0.7513952467595857,
"eval_pearson_manhattan": 0.7505485236956133,
"eval_runtime": 3.3329,
"eval_samples_per_second": 450.052,
"eval_spearman_cosine": 0.7800032031654975,
"eval_spearman_dot": 0.576004160544721,
"eval_spearman_euclidean": 0.7627214415950069,
"eval_spearman_manhattan": 0.7623633881575729,
"eval_steps_per_second": 28.203,
"step": 1750
},
{
"epoch": 0.8247422680412371,
"grad_norm": 2.793635606765747,
"learning_rate": 1.9484536082474227e-05,
"loss": 0.4434,
"step": 1760
},
{
"epoch": 0.8294283036551078,
"grad_norm": 2.403629779815674,
"learning_rate": 1.948160731021556e-05,
"loss": 0.4034,
"step": 1770
},
{
"epoch": 0.8341143392689785,
"grad_norm": 2.6248741149902344,
"learning_rate": 1.947867853795689e-05,
"loss": 0.4328,
"step": 1780
},
{
"epoch": 0.8388003748828491,
"grad_norm": 1.8423967361450195,
"learning_rate": 1.9475749765698223e-05,
"loss": 0.458,
"step": 1790
},
{
"epoch": 0.8434864104967198,
"grad_norm": 2.8771181106567383,
"learning_rate": 1.9472820993439552e-05,
"loss": 0.3796,
"step": 1800
},
{
"epoch": 0.8481724461105904,
"grad_norm": 2.6302261352539062,
"learning_rate": 1.946989222118088e-05,
"loss": 0.4494,
"step": 1810
},
{
"epoch": 0.8528584817244611,
"grad_norm": 1.8728386163711548,
"learning_rate": 1.9466963448922214e-05,
"loss": 0.4624,
"step": 1820
},
{
"epoch": 0.8575445173383318,
"grad_norm": 1.9615451097488403,
"learning_rate": 1.9464034676663544e-05,
"loss": 0.4666,
"step": 1830
},
{
"epoch": 0.8622305529522024,
"grad_norm": 2.423076629638672,
"learning_rate": 1.9461105904404877e-05,
"loss": 0.4659,
"step": 1840
},
{
"epoch": 0.8669165885660731,
"grad_norm": 2.1707210540771484,
"learning_rate": 1.9458177132146206e-05,
"loss": 0.4536,
"step": 1850
},
{
"epoch": 0.8716026241799437,
"grad_norm": 2.3241922855377197,
"learning_rate": 1.9455248359887536e-05,
"loss": 0.4589,
"step": 1860
},
{
"epoch": 0.8762886597938144,
"grad_norm": 2.628664255142212,
"learning_rate": 1.945231958762887e-05,
"loss": 0.4801,
"step": 1870
},
{
"epoch": 0.8809746954076851,
"grad_norm": 2.25126314163208,
"learning_rate": 1.9449390815370198e-05,
"loss": 0.4127,
"step": 1880
},
{
"epoch": 0.8856607310215557,
"grad_norm": 2.988956928253174,
"learning_rate": 1.9446462043111528e-05,
"loss": 0.4431,
"step": 1890
},
{
"epoch": 0.8903467666354264,
"grad_norm": 2.3071420192718506,
"learning_rate": 1.944353327085286e-05,
"loss": 0.4606,
"step": 1900
},
{
"epoch": 0.895032802249297,
"grad_norm": 2.712460517883301,
"learning_rate": 1.944060449859419e-05,
"loss": 0.4853,
"step": 1910
},
{
"epoch": 0.8997188378631678,
"grad_norm": 2.555626153945923,
"learning_rate": 1.943767572633552e-05,
"loss": 0.4187,
"step": 1920
},
{
"epoch": 0.9044048734770385,
"grad_norm": 2.3509678840637207,
"learning_rate": 1.9434746954076852e-05,
"loss": 0.4144,
"step": 1930
},
{
"epoch": 0.9090909090909091,
"grad_norm": 2.63004994392395,
"learning_rate": 1.9431818181818182e-05,
"loss": 0.399,
"step": 1940
},
{
"epoch": 0.9137769447047798,
"grad_norm": 2.8148534297943115,
"learning_rate": 1.9428889409559515e-05,
"loss": 0.4309,
"step": 1950
},
{
"epoch": 0.9184629803186504,
"grad_norm": 2.594778537750244,
"learning_rate": 1.9425960637300844e-05,
"loss": 0.4753,
"step": 1960
},
{
"epoch": 0.9231490159325211,
"grad_norm": 2.4502975940704346,
"learning_rate": 1.9423031865042177e-05,
"loss": 0.4317,
"step": 1970
},
{
"epoch": 0.9278350515463918,
"grad_norm": 2.335707664489746,
"learning_rate": 1.9420103092783507e-05,
"loss": 0.405,
"step": 1980
},
{
"epoch": 0.9325210871602624,
"grad_norm": 2.0419764518737793,
"learning_rate": 1.941717432052484e-05,
"loss": 0.4669,
"step": 1990
},
{
"epoch": 0.9372071227741331,
"grad_norm": 2.2041420936584473,
"learning_rate": 1.941424554826617e-05,
"loss": 0.4696,
"step": 2000
},
{
"epoch": 0.9372071227741331,
"eval_loss": 0.10549145191907883,
"eval_pearson_cosine": 0.7617707450696116,
"eval_pearson_dot": 0.5909647410786931,
"eval_pearson_euclidean": 0.7526201710813609,
"eval_pearson_manhattan": 0.7514050722082715,
"eval_runtime": 3.4394,
"eval_samples_per_second": 436.124,
"eval_spearman_cosine": 0.7834635716420068,
"eval_spearman_dot": 0.5835376546734291,
"eval_spearman_euclidean": 0.7675113486625056,
"eval_spearman_manhattan": 0.7668717701506194,
"eval_steps_per_second": 27.33,
"step": 2000
},
{
"epoch": 0.9418931583880038,
"grad_norm": 1.7690527439117432,
"learning_rate": 1.94113167760075e-05,
"loss": 0.39,
"step": 2010
},
{
"epoch": 0.9465791940018744,
"grad_norm": 2.247870683670044,
"learning_rate": 1.940838800374883e-05,
"loss": 0.4283,
"step": 2020
},
{
"epoch": 0.9512652296157451,
"grad_norm": 2.7865118980407715,
"learning_rate": 1.940545923149016e-05,
"loss": 0.4041,
"step": 2030
},
{
"epoch": 0.9559512652296157,
"grad_norm": 2.7173750400543213,
"learning_rate": 1.9402530459231494e-05,
"loss": 0.428,
"step": 2040
},
{
"epoch": 0.9606373008434864,
"grad_norm": 2.4290664196014404,
"learning_rate": 1.9399601686972823e-05,
"loss": 0.4363,
"step": 2050
},
{
"epoch": 0.9653233364573571,
"grad_norm": 2.9478437900543213,
"learning_rate": 1.9396672914714153e-05,
"loss": 0.4341,
"step": 2060
},
{
"epoch": 0.9700093720712277,
"grad_norm": 2.5642406940460205,
"learning_rate": 1.9393744142455482e-05,
"loss": 0.4402,
"step": 2070
},
{
"epoch": 0.9746954076850984,
"grad_norm": 2.622969388961792,
"learning_rate": 1.9390815370196815e-05,
"loss": 0.4415,
"step": 2080
},
{
"epoch": 0.979381443298969,
"grad_norm": 1.8864187002182007,
"learning_rate": 1.9387886597938145e-05,
"loss": 0.421,
"step": 2090
},
{
"epoch": 0.9840674789128397,
"grad_norm": 2.145385503768921,
"learning_rate": 1.9384957825679474e-05,
"loss": 0.4142,
"step": 2100
},
{
"epoch": 0.9887535145267105,
"grad_norm": 2.5870542526245117,
"learning_rate": 1.9382029053420807e-05,
"loss": 0.4701,
"step": 2110
},
{
"epoch": 0.993439550140581,
"grad_norm": 2.5176923274993896,
"learning_rate": 1.9379100281162137e-05,
"loss": 0.3776,
"step": 2120
},
{
"epoch": 0.9981255857544518,
"grad_norm": 2.8260464668273926,
"learning_rate": 1.937617150890347e-05,
"loss": 0.4453,
"step": 2130
},
{
"epoch": 1.0028116213683225,
"grad_norm": 2.2735579013824463,
"learning_rate": 1.93732427366448e-05,
"loss": 0.3617,
"step": 2140
},
{
"epoch": 1.007497656982193,
"grad_norm": 1.9802652597427368,
"learning_rate": 1.9370313964386132e-05,
"loss": 0.3103,
"step": 2150
},
{
"epoch": 1.0121836925960637,
"grad_norm": 2.4149436950683594,
"learning_rate": 1.936738519212746e-05,
"loss": 0.3183,
"step": 2160
},
{
"epoch": 1.0168697282099344,
"grad_norm": 2.9166879653930664,
"learning_rate": 1.9364456419868794e-05,
"loss": 0.2927,
"step": 2170
},
{
"epoch": 1.021555763823805,
"grad_norm": 1.8774973154067993,
"learning_rate": 1.9361527647610124e-05,
"loss": 0.3217,
"step": 2180
},
{
"epoch": 1.0262417994376758,
"grad_norm": 2.0291717052459717,
"learning_rate": 1.9358598875351457e-05,
"loss": 0.3236,
"step": 2190
},
{
"epoch": 1.0309278350515463,
"grad_norm": 2.500614643096924,
"learning_rate": 1.9355670103092786e-05,
"loss": 0.3187,
"step": 2200
},
{
"epoch": 1.035613870665417,
"grad_norm": 2.4826183319091797,
"learning_rate": 1.9352741330834116e-05,
"loss": 0.3284,
"step": 2210
},
{
"epoch": 1.0402999062792877,
"grad_norm": 2.647724151611328,
"learning_rate": 1.934981255857545e-05,
"loss": 0.3432,
"step": 2220
},
{
"epoch": 1.0449859418931584,
"grad_norm": 2.356365203857422,
"learning_rate": 1.9346883786316778e-05,
"loss": 0.3321,
"step": 2230
},
{
"epoch": 1.0496719775070291,
"grad_norm": 2.514002561569214,
"learning_rate": 1.9343955014058108e-05,
"loss": 0.3115,
"step": 2240
},
{
"epoch": 1.0543580131208996,
"grad_norm": 2.049299716949463,
"learning_rate": 1.9341026241799437e-05,
"loss": 0.3474,
"step": 2250
},
{
"epoch": 1.0543580131208996,
"eval_loss": 0.0818348377943039,
"eval_pearson_cosine": 0.7662947284133068,
"eval_pearson_dot": 0.5774155023095773,
"eval_pearson_euclidean": 0.7536499802604126,
"eval_pearson_manhattan": 0.7526750534069606,
"eval_runtime": 3.4518,
"eval_samples_per_second": 434.553,
"eval_spearman_cosine": 0.7776893657594519,
"eval_spearman_dot": 0.5747770364870878,
"eval_spearman_euclidean": 0.7641629894500831,
"eval_spearman_manhattan": 0.7635836296791149,
"eval_steps_per_second": 27.232,
"step": 2250
},
{
"epoch": 1.0590440487347703,
"grad_norm": 1.9242613315582275,
"learning_rate": 1.933809746954077e-05,
"loss": 0.3362,
"step": 2260
},
{
"epoch": 1.063730084348641,
"grad_norm": 2.6948235034942627,
"learning_rate": 1.93351686972821e-05,
"loss": 0.3524,
"step": 2270
},
{
"epoch": 1.0684161199625117,
"grad_norm": 2.162079334259033,
"learning_rate": 1.9332239925023432e-05,
"loss": 0.3292,
"step": 2280
},
{
"epoch": 1.0731021555763824,
"grad_norm": 2.1351771354675293,
"learning_rate": 1.9329311152764762e-05,
"loss": 0.3409,
"step": 2290
},
{
"epoch": 1.077788191190253,
"grad_norm": 2.527005434036255,
"learning_rate": 1.932638238050609e-05,
"loss": 0.3027,
"step": 2300
},
{
"epoch": 1.0824742268041236,
"grad_norm": 2.446911334991455,
"learning_rate": 1.9323453608247424e-05,
"loss": 0.3376,
"step": 2310
},
{
"epoch": 1.0871602624179943,
"grad_norm": 2.478301763534546,
"learning_rate": 1.9320524835988754e-05,
"loss": 0.298,
"step": 2320
},
{
"epoch": 1.091846298031865,
"grad_norm": 2.537336587905884,
"learning_rate": 1.9317596063730087e-05,
"loss": 0.3312,
"step": 2330
},
{
"epoch": 1.0965323336457358,
"grad_norm": 2.4405462741851807,
"learning_rate": 1.9314667291471416e-05,
"loss": 0.2998,
"step": 2340
},
{
"epoch": 1.1012183692596063,
"grad_norm": 2.2255520820617676,
"learning_rate": 1.931173851921275e-05,
"loss": 0.2987,
"step": 2350
},
{
"epoch": 1.105904404873477,
"grad_norm": 2.541879177093506,
"learning_rate": 1.930880974695408e-05,
"loss": 0.3286,
"step": 2360
},
{
"epoch": 1.1105904404873477,
"grad_norm": 2.1010048389434814,
"learning_rate": 1.930588097469541e-05,
"loss": 0.2989,
"step": 2370
},
{
"epoch": 1.1152764761012184,
"grad_norm": 1.9881564378738403,
"learning_rate": 1.930295220243674e-05,
"loss": 0.3537,
"step": 2380
},
{
"epoch": 1.119962511715089,
"grad_norm": 2.503845691680908,
"learning_rate": 1.930002343017807e-05,
"loss": 0.3242,
"step": 2390
},
{
"epoch": 1.1246485473289598,
"grad_norm": 2.773836612701416,
"learning_rate": 1.9297094657919403e-05,
"loss": 0.3144,
"step": 2400
},
{
"epoch": 1.1293345829428303,
"grad_norm": 2.7757749557495117,
"learning_rate": 1.9294165885660733e-05,
"loss": 0.3114,
"step": 2410
},
{
"epoch": 1.134020618556701,
"grad_norm": 2.707315444946289,
"learning_rate": 1.9291237113402062e-05,
"loss": 0.2872,
"step": 2420
},
{
"epoch": 1.1387066541705717,
"grad_norm": 2.1421258449554443,
"learning_rate": 1.9288308341143395e-05,
"loss": 0.3122,
"step": 2430
},
{
"epoch": 1.1433926897844424,
"grad_norm": 2.2792320251464844,
"learning_rate": 1.9285379568884725e-05,
"loss": 0.304,
"step": 2440
},
{
"epoch": 1.148078725398313,
"grad_norm": 2.1971569061279297,
"learning_rate": 1.9282450796626054e-05,
"loss": 0.3571,
"step": 2450
},
{
"epoch": 1.1527647610121836,
"grad_norm": 2.101923704147339,
"learning_rate": 1.9279522024367387e-05,
"loss": 0.3069,
"step": 2460
},
{
"epoch": 1.1574507966260543,
"grad_norm": 2.395801305770874,
"learning_rate": 1.9276593252108717e-05,
"loss": 0.3145,
"step": 2470
},
{
"epoch": 1.162136832239925,
"grad_norm": 1.9261037111282349,
"learning_rate": 1.927366447985005e-05,
"loss": 0.3229,
"step": 2480
},
{
"epoch": 1.1668228678537957,
"grad_norm": 2.185063362121582,
"learning_rate": 1.927073570759138e-05,
"loss": 0.3289,
"step": 2490
},
{
"epoch": 1.1715089034676662,
"grad_norm": 2.2372336387634277,
"learning_rate": 1.926780693533271e-05,
"loss": 0.319,
"step": 2500
},
{
"epoch": 1.1715089034676662,
"eval_loss": 0.07521134614944458,
"eval_pearson_cosine": 0.7753374743054451,
"eval_pearson_dot": 0.5929128805152892,
"eval_pearson_euclidean": 0.7592332760910097,
"eval_pearson_manhattan": 0.7589109467518398,
"eval_runtime": 3.6686,
"eval_samples_per_second": 408.877,
"eval_spearman_cosine": 0.785757446712241,
"eval_spearman_dot": 0.5918750155733082,
"eval_spearman_euclidean": 0.7692376569048597,
"eval_spearman_manhattan": 0.7692457770843739,
"eval_steps_per_second": 25.623,
"step": 2500
},
{
"epoch": 1.176194939081537,
"grad_norm": 2.2003061771392822,
"learning_rate": 1.926487816307404e-05,
"loss": 0.2979,
"step": 2510
},
{
"epoch": 1.1808809746954076,
"grad_norm": 1.9160058498382568,
"learning_rate": 1.926194939081537e-05,
"loss": 0.3119,
"step": 2520
},
{
"epoch": 1.1855670103092784,
"grad_norm": 2.5396270751953125,
"learning_rate": 1.9259020618556704e-05,
"loss": 0.2882,
"step": 2530
},
{
"epoch": 1.190253045923149,
"grad_norm": 2.184056520462036,
"learning_rate": 1.9256091846298033e-05,
"loss": 0.3012,
"step": 2540
},
{
"epoch": 1.1949390815370198,
"grad_norm": 2.4930028915405273,
"learning_rate": 1.9253163074039366e-05,
"loss": 0.3399,
"step": 2550
},
{
"epoch": 1.1996251171508903,
"grad_norm": 2.0577898025512695,
"learning_rate": 1.9250234301780696e-05,
"loss": 0.3013,
"step": 2560
},
{
"epoch": 1.204311152764761,
"grad_norm": 1.9735450744628906,
"learning_rate": 1.9247305529522025e-05,
"loss": 0.315,
"step": 2570
},
{
"epoch": 1.2089971883786317,
"grad_norm": 2.815910577774048,
"learning_rate": 1.9244376757263358e-05,
"loss": 0.3134,
"step": 2580
},
{
"epoch": 1.2136832239925024,
"grad_norm": 2.4166650772094727,
"learning_rate": 1.9241447985004687e-05,
"loss": 0.2924,
"step": 2590
},
{
"epoch": 1.218369259606373,
"grad_norm": 2.65451979637146,
"learning_rate": 1.9238519212746017e-05,
"loss": 0.313,
"step": 2600
},
{
"epoch": 1.2230552952202436,
"grad_norm": 3.048393726348877,
"learning_rate": 1.923559044048735e-05,
"loss": 0.3396,
"step": 2610
},
{
"epoch": 1.2277413308341143,
"grad_norm": 2.1255080699920654,
"learning_rate": 1.923266166822868e-05,
"loss": 0.2918,
"step": 2620
},
{
"epoch": 1.232427366447985,
"grad_norm": 2.4257168769836426,
"learning_rate": 1.922973289597001e-05,
"loss": 0.3221,
"step": 2630
},
{
"epoch": 1.2371134020618557,
"grad_norm": 2.3977224826812744,
"learning_rate": 1.9226804123711342e-05,
"loss": 0.3777,
"step": 2640
},
{
"epoch": 1.2417994376757264,
"grad_norm": 2.0743911266326904,
"learning_rate": 1.922387535145267e-05,
"loss": 0.2981,
"step": 2650
},
{
"epoch": 1.246485473289597,
"grad_norm": 2.740622043609619,
"learning_rate": 1.9220946579194004e-05,
"loss": 0.3239,
"step": 2660
},
{
"epoch": 1.2511715089034676,
"grad_norm": 2.5174942016601562,
"learning_rate": 1.9218017806935334e-05,
"loss": 0.2981,
"step": 2670
},
{
"epoch": 1.2558575445173383,
"grad_norm": 2.3390488624572754,
"learning_rate": 1.9215089034676667e-05,
"loss": 0.3509,
"step": 2680
},
{
"epoch": 1.260543580131209,
"grad_norm": 2.495445966720581,
"learning_rate": 1.9212160262417996e-05,
"loss": 0.3153,
"step": 2690
},
{
"epoch": 1.2652296157450795,
"grad_norm": 2.384673595428467,
"learning_rate": 1.9209231490159326e-05,
"loss": 0.3541,
"step": 2700
},
{
"epoch": 1.2699156513589505,
"grad_norm": 1.9707324504852295,
"learning_rate": 1.920630271790066e-05,
"loss": 0.3053,
"step": 2710
},
{
"epoch": 1.274601686972821,
"grad_norm": 2.0584781169891357,
"learning_rate": 1.9203373945641988e-05,
"loss": 0.2716,
"step": 2720
},
{
"epoch": 1.2792877225866917,
"grad_norm": 1.9361029863357544,
"learning_rate": 1.920044517338332e-05,
"loss": 0.2988,
"step": 2730
},
{
"epoch": 1.2839737582005624,
"grad_norm": 2.3929100036621094,
"learning_rate": 1.919751640112465e-05,
"loss": 0.3356,
"step": 2740
},
{
"epoch": 1.2886597938144329,
"grad_norm": 2.7867496013641357,
"learning_rate": 1.9194587628865983e-05,
"loss": 0.3682,
"step": 2750
},
{
"epoch": 1.2886597938144329,
"eval_loss": 0.07667941600084305,
"eval_pearson_cosine": 0.7735852714798206,
"eval_pearson_dot": 0.578414803552846,
"eval_pearson_euclidean": 0.7564370361137946,
"eval_pearson_manhattan": 0.7556229015098381,
"eval_runtime": 3.3395,
"eval_samples_per_second": 449.165,
"eval_spearman_cosine": 0.7851408708188664,
"eval_spearman_dot": 0.5785312770170487,
"eval_spearman_euclidean": 0.7670951414033645,
"eval_spearman_manhattan": 0.7667469763884154,
"eval_steps_per_second": 28.148,
"step": 2750
},
{
"epoch": 1.2933458294283038,
"grad_norm": 2.0457334518432617,
"learning_rate": 1.9191658856607313e-05,
"loss": 0.3118,
"step": 2760
},
{
"epoch": 1.2980318650421743,
"grad_norm": 2.1723532676696777,
"learning_rate": 1.9188730084348642e-05,
"loss": 0.2918,
"step": 2770
},
{
"epoch": 1.302717900656045,
"grad_norm": 2.041738748550415,
"learning_rate": 1.9185801312089972e-05,
"loss": 0.3084,
"step": 2780
},
{
"epoch": 1.3074039362699157,
"grad_norm": 2.774698257446289,
"learning_rate": 1.9182872539831305e-05,
"loss": 0.3337,
"step": 2790
},
{
"epoch": 1.3120899718837864,
"grad_norm": 1.9762837886810303,
"learning_rate": 1.9179943767572634e-05,
"loss": 0.3029,
"step": 2800
},
{
"epoch": 1.316776007497657,
"grad_norm": 2.1785385608673096,
"learning_rate": 1.9177014995313964e-05,
"loss": 0.3337,
"step": 2810
},
{
"epoch": 1.3214620431115276,
"grad_norm": 2.0208494663238525,
"learning_rate": 1.9174086223055296e-05,
"loss": 0.2931,
"step": 2820
},
{
"epoch": 1.3261480787253983,
"grad_norm": 2.3976972103118896,
"learning_rate": 1.9171157450796626e-05,
"loss": 0.3316,
"step": 2830
},
{
"epoch": 1.330834114339269,
"grad_norm": 1.9782285690307617,
"learning_rate": 1.916822867853796e-05,
"loss": 0.3173,
"step": 2840
},
{
"epoch": 1.3355201499531397,
"grad_norm": 2.0938830375671387,
"learning_rate": 1.916529990627929e-05,
"loss": 0.2858,
"step": 2850
},
{
"epoch": 1.3402061855670104,
"grad_norm": 2.079247236251831,
"learning_rate": 1.916237113402062e-05,
"loss": 0.2911,
"step": 2860
},
{
"epoch": 1.344892221180881,
"grad_norm": 2.5487163066864014,
"learning_rate": 1.915944236176195e-05,
"loss": 0.2782,
"step": 2870
},
{
"epoch": 1.3495782567947516,
"grad_norm": 2.4930672645568848,
"learning_rate": 1.9156513589503284e-05,
"loss": 0.309,
"step": 2880
},
{
"epoch": 1.3542642924086223,
"grad_norm": 2.208987236022949,
"learning_rate": 1.9153584817244613e-05,
"loss": 0.302,
"step": 2890
},
{
"epoch": 1.358950328022493,
"grad_norm": 2.2777154445648193,
"learning_rate": 1.9150656044985943e-05,
"loss": 0.2932,
"step": 2900
},
{
"epoch": 1.3636363636363638,
"grad_norm": 2.0704426765441895,
"learning_rate": 1.9147727272727276e-05,
"loss": 0.3326,
"step": 2910
},
{
"epoch": 1.3683223992502342,
"grad_norm": 2.550323486328125,
"learning_rate": 1.9144798500468605e-05,
"loss": 0.2945,
"step": 2920
},
{
"epoch": 1.373008434864105,
"grad_norm": 2.1471896171569824,
"learning_rate": 1.9141869728209938e-05,
"loss": 0.3258,
"step": 2930
},
{
"epoch": 1.3776944704779757,
"grad_norm": 2.4434566497802734,
"learning_rate": 1.9138940955951267e-05,
"loss": 0.3,
"step": 2940
},
{
"epoch": 1.3823805060918464,
"grad_norm": 2.2125484943389893,
"learning_rate": 1.9136012183692597e-05,
"loss": 0.3618,
"step": 2950
},
{
"epoch": 1.387066541705717,
"grad_norm": 2.5156352519989014,
"learning_rate": 1.913308341143393e-05,
"loss": 0.3102,
"step": 2960
},
{
"epoch": 1.3917525773195876,
"grad_norm": 2.1281540393829346,
"learning_rate": 1.913015463917526e-05,
"loss": 0.31,
"step": 2970
},
{
"epoch": 1.3964386129334583,
"grad_norm": 2.1260693073272705,
"learning_rate": 1.912722586691659e-05,
"loss": 0.2946,
"step": 2980
},
{
"epoch": 1.401124648547329,
"grad_norm": 2.591770887374878,
"learning_rate": 1.912429709465792e-05,
"loss": 0.2906,
"step": 2990
},
{
"epoch": 1.4058106841611997,
"grad_norm": 1.832505226135254,
"learning_rate": 1.912136832239925e-05,
"loss": 0.3033,
"step": 3000
},
{
"epoch": 1.4058106841611997,
"eval_loss": 0.07158821821212769,
"eval_pearson_cosine": 0.7835729367892306,
"eval_pearson_dot": 0.5987220055863389,
"eval_pearson_euclidean": 0.7599913918565986,
"eval_pearson_manhattan": 0.7589542766532986,
"eval_runtime": 3.4365,
"eval_samples_per_second": 436.486,
"eval_spearman_cosine": 0.7961581681624044,
"eval_spearman_dot": 0.5975812797826083,
"eval_spearman_euclidean": 0.7727385023345612,
"eval_spearman_manhattan": 0.7723228512152759,
"eval_steps_per_second": 27.353,
"step": 3000
},
{
"epoch": 1.4104967197750704,
"grad_norm": 2.3309245109558105,
"learning_rate": 1.911843955014058e-05,
"loss": 0.3238,
"step": 3010
},
{
"epoch": 1.415182755388941,
"grad_norm": 2.3359642028808594,
"learning_rate": 1.9115510777881914e-05,
"loss": 0.2933,
"step": 3020
},
{
"epoch": 1.4198687910028116,
"grad_norm": 2.363147497177124,
"learning_rate": 1.9112582005623243e-05,
"loss": 0.3204,
"step": 3030
},
{
"epoch": 1.4245548266166823,
"grad_norm": 2.353273391723633,
"learning_rate": 1.9109653233364576e-05,
"loss": 0.295,
"step": 3040
},
{
"epoch": 1.429240862230553,
"grad_norm": 2.418267011642456,
"learning_rate": 1.9106724461105905e-05,
"loss": 0.3397,
"step": 3050
},
{
"epoch": 1.4339268978444237,
"grad_norm": 2.612607955932617,
"learning_rate": 1.910379568884724e-05,
"loss": 0.3001,
"step": 3060
},
{
"epoch": 1.4386129334582942,
"grad_norm": 2.506488084793091,
"learning_rate": 1.9100866916588568e-05,
"loss": 0.3101,
"step": 3070
},
{
"epoch": 1.443298969072165,
"grad_norm": 2.437333345413208,
"learning_rate": 1.90979381443299e-05,
"loss": 0.3017,
"step": 3080
},
{
"epoch": 1.4479850046860356,
"grad_norm": 2.3391661643981934,
"learning_rate": 1.909500937207123e-05,
"loss": 0.2943,
"step": 3090
},
{
"epoch": 1.4526710402999063,
"grad_norm": 2.410454511642456,
"learning_rate": 1.909208059981256e-05,
"loss": 0.3242,
"step": 3100
},
{
"epoch": 1.457357075913777,
"grad_norm": 2.6086418628692627,
"learning_rate": 1.9089151827553893e-05,
"loss": 0.3442,
"step": 3110
},
{
"epoch": 1.4620431115276475,
"grad_norm": 2.1769497394561768,
"learning_rate": 1.9086223055295222e-05,
"loss": 0.3114,
"step": 3120
},
{
"epoch": 1.4667291471415183,
"grad_norm": 1.9582467079162598,
"learning_rate": 1.908329428303655e-05,
"loss": 0.3261,
"step": 3130
},
{
"epoch": 1.471415182755389,
"grad_norm": 2.577211380004883,
"learning_rate": 1.9080365510777884e-05,
"loss": 0.31,
"step": 3140
},
{
"epoch": 1.4761012183692597,
"grad_norm": 2.725773811340332,
"learning_rate": 1.9077436738519214e-05,
"loss": 0.2859,
"step": 3150
},
{
"epoch": 1.4807872539831304,
"grad_norm": 2.779578685760498,
"learning_rate": 1.9074507966260543e-05,
"loss": 0.3304,
"step": 3160
},
{
"epoch": 1.4854732895970009,
"grad_norm": 1.9493879079818726,
"learning_rate": 1.9071579194001876e-05,
"loss": 0.2992,
"step": 3170
},
{
"epoch": 1.4901593252108716,
"grad_norm": 3.1011288166046143,
"learning_rate": 1.9068650421743206e-05,
"loss": 0.3406,
"step": 3180
},
{
"epoch": 1.4948453608247423,
"grad_norm": 2.466146230697632,
"learning_rate": 1.906572164948454e-05,
"loss": 0.357,
"step": 3190
},
{
"epoch": 1.499531396438613,
"grad_norm": 2.0382285118103027,
"learning_rate": 1.9062792877225868e-05,
"loss": 0.2968,
"step": 3200
},
{
"epoch": 1.5042174320524837,
"grad_norm": 2.331540584564209,
"learning_rate": 1.9059864104967198e-05,
"loss": 0.3125,
"step": 3210
},
{
"epoch": 1.5089034676663542,
"grad_norm": 2.590193271636963,
"learning_rate": 1.905693533270853e-05,
"loss": 0.2762,
"step": 3220
},
{
"epoch": 1.513589503280225,
"grad_norm": 2.878938913345337,
"learning_rate": 1.905400656044986e-05,
"loss": 0.2868,
"step": 3230
},
{
"epoch": 1.5182755388940956,
"grad_norm": 2.4432168006896973,
"learning_rate": 1.9051077788191193e-05,
"loss": 0.3021,
"step": 3240
},
{
"epoch": 1.522961574507966,
"grad_norm": 2.849559783935547,
"learning_rate": 1.9048149015932523e-05,
"loss": 0.3247,
"step": 3250
},
{
"epoch": 1.522961574507966,
"eval_loss": 0.07676094025373459,
"eval_pearson_cosine": 0.7779113850636818,
"eval_pearson_dot": 0.5637961042064976,
"eval_pearson_euclidean": 0.7621398908224677,
"eval_pearson_manhattan": 0.7612994290780648,
"eval_runtime": 3.2908,
"eval_samples_per_second": 455.817,
"eval_spearman_cosine": 0.7910676412952927,
"eval_spearman_dot": 0.562324725935799,
"eval_spearman_euclidean": 0.7734903614875008,
"eval_spearman_manhattan": 0.7731474604573987,
"eval_steps_per_second": 28.565,
"step": 3250
},
{
"epoch": 1.527647610121837,
"grad_norm": 2.372697591781616,
"learning_rate": 1.9045220243673855e-05,
"loss": 0.3423,
"step": 3260
},
{
"epoch": 1.5323336457357075,
"grad_norm": 2.9068384170532227,
"learning_rate": 1.9042291471415185e-05,
"loss": 0.3133,
"step": 3270
},
{
"epoch": 1.5370196813495782,
"grad_norm": 2.738391160964966,
"learning_rate": 1.9039362699156518e-05,
"loss": 0.3018,
"step": 3280
},
{
"epoch": 1.541705716963449,
"grad_norm": 2.121777296066284,
"learning_rate": 1.9036433926897847e-05,
"loss": 0.3203,
"step": 3290
},
{
"epoch": 1.5463917525773194,
"grad_norm": 2.1887810230255127,
"learning_rate": 1.9033505154639177e-05,
"loss": 0.3639,
"step": 3300
},
{
"epoch": 1.5510777881911904,
"grad_norm": 2.4203438758850098,
"learning_rate": 1.9030576382380506e-05,
"loss": 0.3116,
"step": 3310
},
{
"epoch": 1.5557638238050608,
"grad_norm": 2.00972318649292,
"learning_rate": 1.902764761012184e-05,
"loss": 0.3139,
"step": 3320
},
{
"epoch": 1.5604498594189316,
"grad_norm": 2.093622922897339,
"learning_rate": 1.902471883786317e-05,
"loss": 0.3248,
"step": 3330
},
{
"epoch": 1.5651358950328023,
"grad_norm": 2.5851595401763916,
"learning_rate": 1.9021790065604498e-05,
"loss": 0.3386,
"step": 3340
},
{
"epoch": 1.569821930646673,
"grad_norm": 2.208651304244995,
"learning_rate": 1.901886129334583e-05,
"loss": 0.3138,
"step": 3350
},
{
"epoch": 1.5745079662605437,
"grad_norm": 2.5663788318634033,
"learning_rate": 1.901593252108716e-05,
"loss": 0.3165,
"step": 3360
},
{
"epoch": 1.5791940018744142,
"grad_norm": 2.443042039871216,
"learning_rate": 1.9013003748828493e-05,
"loss": 0.3178,
"step": 3370
},
{
"epoch": 1.5838800374882849,
"grad_norm": 2.385019540786743,
"learning_rate": 1.9010074976569823e-05,
"loss": 0.2779,
"step": 3380
},
{
"epoch": 1.5885660731021556,
"grad_norm": 2.5440433025360107,
"learning_rate": 1.9007146204311156e-05,
"loss": 0.3182,
"step": 3390
},
{
"epoch": 1.5932521087160263,
"grad_norm": 2.1016016006469727,
"learning_rate": 1.9004217432052485e-05,
"loss": 0.3224,
"step": 3400
},
{
"epoch": 1.597938144329897,
"grad_norm": 1.8776601552963257,
"learning_rate": 1.9001288659793815e-05,
"loss": 0.2926,
"step": 3410
},
{
"epoch": 1.6026241799437675,
"grad_norm": 1.7855247259140015,
"learning_rate": 1.8998359887535148e-05,
"loss": 0.3087,
"step": 3420
},
{
"epoch": 1.6073102155576382,
"grad_norm": 2.239816188812256,
"learning_rate": 1.8995431115276477e-05,
"loss": 0.3254,
"step": 3430
},
{
"epoch": 1.611996251171509,
"grad_norm": 2.309274673461914,
"learning_rate": 1.899250234301781e-05,
"loss": 0.2741,
"step": 3440
},
{
"epoch": 1.6166822867853796,
"grad_norm": 2.672464370727539,
"learning_rate": 1.898957357075914e-05,
"loss": 0.3137,
"step": 3450
},
{
"epoch": 1.6213683223992503,
"grad_norm": 2.3761508464813232,
"learning_rate": 1.8986644798500473e-05,
"loss": 0.322,
"step": 3460
},
{
"epoch": 1.6260543580131208,
"grad_norm": 2.497772216796875,
"learning_rate": 1.8983716026241802e-05,
"loss": 0.2832,
"step": 3470
},
{
"epoch": 1.6307403936269915,
"grad_norm": 2.2633016109466553,
"learning_rate": 1.898078725398313e-05,
"loss": 0.2659,
"step": 3480
},
{
"epoch": 1.6354264292408622,
"grad_norm": 2.2554588317871094,
"learning_rate": 1.897785848172446e-05,
"loss": 0.3039,
"step": 3490
},
{
"epoch": 1.640112464854733,
"grad_norm": 2.0419113636016846,
"learning_rate": 1.8974929709465794e-05,
"loss": 0.26,
"step": 3500
},
{
"epoch": 1.640112464854733,
"eval_loss": 0.06864778697490692,
"eval_pearson_cosine": 0.7792468184875077,
"eval_pearson_dot": 0.6003709070725307,
"eval_pearson_euclidean": 0.762291216964627,
"eval_pearson_manhattan": 0.7615198846806042,
"eval_runtime": 3.2918,
"eval_samples_per_second": 455.681,
"eval_spearman_cosine": 0.7902077133733839,
"eval_spearman_dot": 0.5998471986095437,
"eval_spearman_euclidean": 0.7733862684662854,
"eval_spearman_manhattan": 0.7732713082323596,
"eval_steps_per_second": 28.556,
"step": 3500
},
{
"epoch": 1.6447985004686037,
"grad_norm": 2.1808488368988037,
"learning_rate": 1.8972000937207123e-05,
"loss": 0.2977,
"step": 3510
},
{
"epoch": 1.6494845360824741,
"grad_norm": 2.5023481845855713,
"learning_rate": 1.8969072164948453e-05,
"loss": 0.3275,
"step": 3520
},
{
"epoch": 1.6541705716963448,
"grad_norm": 1.9074852466583252,
"learning_rate": 1.8966143392689786e-05,
"loss": 0.3084,
"step": 3530
},
{
"epoch": 1.6588566073102156,
"grad_norm": 2.4666028022766113,
"learning_rate": 1.8963214620431115e-05,
"loss": 0.3202,
"step": 3540
},
{
"epoch": 1.6635426429240863,
"grad_norm": 2.2712929248809814,
"learning_rate": 1.8960285848172448e-05,
"loss": 0.3023,
"step": 3550
},
{
"epoch": 1.668228678537957,
"grad_norm": 2.274965286254883,
"learning_rate": 1.8957357075913778e-05,
"loss": 0.3134,
"step": 3560
},
{
"epoch": 1.6729147141518275,
"grad_norm": 2.2933590412139893,
"learning_rate": 1.895442830365511e-05,
"loss": 0.2935,
"step": 3570
},
{
"epoch": 1.6776007497656982,
"grad_norm": 2.492459774017334,
"learning_rate": 1.895149953139644e-05,
"loss": 0.2891,
"step": 3580
},
{
"epoch": 1.6822867853795689,
"grad_norm": 2.4401142597198486,
"learning_rate": 1.894857075913777e-05,
"loss": 0.329,
"step": 3590
},
{
"epoch": 1.6869728209934396,
"grad_norm": 2.0816047191619873,
"learning_rate": 1.8945641986879102e-05,
"loss": 0.3042,
"step": 3600
},
{
"epoch": 1.6916588566073103,
"grad_norm": 2.2334861755371094,
"learning_rate": 1.8942713214620432e-05,
"loss": 0.2909,
"step": 3610
},
{
"epoch": 1.6963448922211808,
"grad_norm": 2.3555166721343994,
"learning_rate": 1.8939784442361765e-05,
"loss": 0.3284,
"step": 3620
},
{
"epoch": 1.7010309278350515,
"grad_norm": 1.751710295677185,
"learning_rate": 1.8936855670103094e-05,
"loss": 0.2848,
"step": 3630
},
{
"epoch": 1.7057169634489222,
"grad_norm": 1.920424461364746,
"learning_rate": 1.8933926897844427e-05,
"loss": 0.279,
"step": 3640
},
{
"epoch": 1.710402999062793,
"grad_norm": 2.840911865234375,
"learning_rate": 1.8930998125585757e-05,
"loss": 0.3116,
"step": 3650
},
{
"epoch": 1.7150890346766636,
"grad_norm": 2.429884672164917,
"learning_rate": 1.8928069353327086e-05,
"loss": 0.264,
"step": 3660
},
{
"epoch": 1.7197750702905341,
"grad_norm": 2.2484772205352783,
"learning_rate": 1.892514058106842e-05,
"loss": 0.3201,
"step": 3670
},
{
"epoch": 1.7244611059044048,
"grad_norm": 2.342460870742798,
"learning_rate": 1.892221180880975e-05,
"loss": 0.3068,
"step": 3680
},
{
"epoch": 1.7291471415182755,
"grad_norm": 2.332886219024658,
"learning_rate": 1.8919283036551078e-05,
"loss": 0.3238,
"step": 3690
},
{
"epoch": 1.7338331771321462,
"grad_norm": 1.7878814935684204,
"learning_rate": 1.8916354264292408e-05,
"loss": 0.2866,
"step": 3700
},
{
"epoch": 1.738519212746017,
"grad_norm": 2.455904722213745,
"learning_rate": 1.891342549203374e-05,
"loss": 0.2887,
"step": 3710
},
{
"epoch": 1.7432052483598874,
"grad_norm": 2.273310899734497,
"learning_rate": 1.891049671977507e-05,
"loss": 0.3377,
"step": 3720
},
{
"epoch": 1.7478912839737581,
"grad_norm": 2.77280592918396,
"learning_rate": 1.8907567947516403e-05,
"loss": 0.2928,
"step": 3730
},
{
"epoch": 1.7525773195876289,
"grad_norm": 2.2917518615722656,
"learning_rate": 1.8904639175257732e-05,
"loss": 0.3154,
"step": 3740
},
{
"epoch": 1.7572633552014996,
"grad_norm": 3.0059680938720703,
"learning_rate": 1.8901710402999065e-05,
"loss": 0.3216,
"step": 3750
},
{
"epoch": 1.7572633552014996,
"eval_loss": 0.07067140191793442,
"eval_pearson_cosine": 0.7850573661327331,
"eval_pearson_dot": 0.6098455249231733,
"eval_pearson_euclidean": 0.7677167356750871,
"eval_pearson_manhattan": 0.7668202443827852,
"eval_runtime": 3.5184,
"eval_samples_per_second": 426.329,
"eval_spearman_cosine": 0.7949681180333369,
"eval_spearman_dot": 0.6135646781276878,
"eval_spearman_euclidean": 0.7791366533328872,
"eval_spearman_manhattan": 0.7786680488005274,
"eval_steps_per_second": 26.717,
"step": 3750
},
{
"epoch": 1.7619493908153703,
"grad_norm": 2.2580602169036865,
"learning_rate": 1.8898781630740395e-05,
"loss": 0.2838,
"step": 3760
},
{
"epoch": 1.7666354264292408,
"grad_norm": 2.6639926433563232,
"learning_rate": 1.8895852858481728e-05,
"loss": 0.3388,
"step": 3770
},
{
"epoch": 1.7713214620431117,
"grad_norm": 2.4491138458251953,
"learning_rate": 1.8892924086223057e-05,
"loss": 0.2905,
"step": 3780
},
{
"epoch": 1.7760074976569822,
"grad_norm": 2.4412620067596436,
"learning_rate": 1.8889995313964387e-05,
"loss": 0.2982,
"step": 3790
},
{
"epoch": 1.780693533270853,
"grad_norm": 2.356416702270508,
"learning_rate": 1.888706654170572e-05,
"loss": 0.3302,
"step": 3800
},
{
"epoch": 1.7853795688847236,
"grad_norm": 2.3429980278015137,
"learning_rate": 1.888413776944705e-05,
"loss": 0.3069,
"step": 3810
},
{
"epoch": 1.790065604498594,
"grad_norm": 2.228518486022949,
"learning_rate": 1.8881208997188382e-05,
"loss": 0.3016,
"step": 3820
},
{
"epoch": 1.794751640112465,
"grad_norm": 2.8037281036376953,
"learning_rate": 1.887828022492971e-05,
"loss": 0.3324,
"step": 3830
},
{
"epoch": 1.7994376757263355,
"grad_norm": 1.893836498260498,
"learning_rate": 1.887535145267104e-05,
"loss": 0.2592,
"step": 3840
},
{
"epoch": 1.8041237113402062,
"grad_norm": 2.6253368854522705,
"learning_rate": 1.8872422680412374e-05,
"loss": 0.295,
"step": 3850
},
{
"epoch": 1.808809746954077,
"grad_norm": 2.2410082817077637,
"learning_rate": 1.8869493908153703e-05,
"loss": 0.2495,
"step": 3860
},
{
"epoch": 1.8134957825679474,
"grad_norm": 2.3530426025390625,
"learning_rate": 1.8866565135895033e-05,
"loss": 0.3004,
"step": 3870
},
{
"epoch": 1.8181818181818183,
"grad_norm": 2.178670644760132,
"learning_rate": 1.8863636363636366e-05,
"loss": 0.3386,
"step": 3880
},
{
"epoch": 1.8228678537956888,
"grad_norm": 2.0660688877105713,
"learning_rate": 1.8860707591377695e-05,
"loss": 0.3117,
"step": 3890
},
{
"epoch": 1.8275538894095595,
"grad_norm": 2.5521416664123535,
"learning_rate": 1.8857778819119025e-05,
"loss": 0.3011,
"step": 3900
},
{
"epoch": 1.8322399250234302,
"grad_norm": 2.3752758502960205,
"learning_rate": 1.8854850046860358e-05,
"loss": 0.3203,
"step": 3910
},
{
"epoch": 1.8369259606373007,
"grad_norm": 2.0836801528930664,
"learning_rate": 1.8851921274601687e-05,
"loss": 0.2706,
"step": 3920
},
{
"epoch": 1.8416119962511717,
"grad_norm": 2.767982006072998,
"learning_rate": 1.884899250234302e-05,
"loss": 0.3077,
"step": 3930
},
{
"epoch": 1.8462980318650422,
"grad_norm": 2.7820098400115967,
"learning_rate": 1.884606373008435e-05,
"loss": 0.2991,
"step": 3940
},
{
"epoch": 1.8509840674789129,
"grad_norm": 2.6888327598571777,
"learning_rate": 1.8843134957825682e-05,
"loss": 0.2862,
"step": 3950
},
{
"epoch": 1.8556701030927836,
"grad_norm": 2.4160315990448,
"learning_rate": 1.8840206185567012e-05,
"loss": 0.2725,
"step": 3960
},
{
"epoch": 1.860356138706654,
"grad_norm": 2.778198719024658,
"learning_rate": 1.8837277413308345e-05,
"loss": 0.3509,
"step": 3970
},
{
"epoch": 1.865042174320525,
"grad_norm": 1.8606864213943481,
"learning_rate": 1.8834348641049674e-05,
"loss": 0.2971,
"step": 3980
},
{
"epoch": 1.8697282099343955,
"grad_norm": 1.98170804977417,
"learning_rate": 1.8831419868791004e-05,
"loss": 0.3131,
"step": 3990
},
{
"epoch": 1.8744142455482662,
"grad_norm": 1.828123927116394,
"learning_rate": 1.8828491096532337e-05,
"loss": 0.3166,
"step": 4000
},
{
"epoch": 1.8744142455482662,
"eval_loss": 0.07189416140317917,
"eval_pearson_cosine": 0.7798801652938181,
"eval_pearson_dot": 0.5736722449381091,
"eval_pearson_euclidean": 0.7563095863468678,
"eval_pearson_manhattan": 0.7550479478827015,
"eval_runtime": 3.2996,
"eval_samples_per_second": 454.607,
"eval_spearman_cosine": 0.7910837741143455,
"eval_spearman_dot": 0.5753555851409241,
"eval_spearman_euclidean": 0.7700926689739851,
"eval_spearman_manhattan": 0.7692694796240825,
"eval_steps_per_second": 28.489,
"step": 4000
},
{
"epoch": 1.879100281162137,
"grad_norm": 1.8813470602035522,
"learning_rate": 1.8825562324273666e-05,
"loss": 0.2484,
"step": 4010
},
{
"epoch": 1.8837863167760074,
"grad_norm": 2.005892038345337,
"learning_rate": 1.8822633552014996e-05,
"loss": 0.3137,
"step": 4020
},
{
"epoch": 1.8884723523898783,
"grad_norm": 2.067732572555542,
"learning_rate": 1.881970477975633e-05,
"loss": 0.3143,
"step": 4030
},
{
"epoch": 1.8931583880037488,
"grad_norm": 2.33829665184021,
"learning_rate": 1.8816776007497658e-05,
"loss": 0.3252,
"step": 4040
},
{
"epoch": 1.8978444236176195,
"grad_norm": 1.655529260635376,
"learning_rate": 1.8813847235238987e-05,
"loss": 0.2889,
"step": 4050
},
{
"epoch": 1.9025304592314902,
"grad_norm": 2.849501609802246,
"learning_rate": 1.881091846298032e-05,
"loss": 0.323,
"step": 4060
},
{
"epoch": 1.9072164948453607,
"grad_norm": 2.8345351219177246,
"learning_rate": 1.880798969072165e-05,
"loss": 0.316,
"step": 4070
},
{
"epoch": 1.9119025304592316,
"grad_norm": 2.061378240585327,
"learning_rate": 1.8805060918462983e-05,
"loss": 0.3112,
"step": 4080
},
{
"epoch": 1.9165885660731021,
"grad_norm": 1.883039116859436,
"learning_rate": 1.8802132146204312e-05,
"loss": 0.2625,
"step": 4090
},
{
"epoch": 1.9212746016869728,
"grad_norm": 2.2609167098999023,
"learning_rate": 1.8799203373945642e-05,
"loss": 0.3043,
"step": 4100
},
{
"epoch": 1.9259606373008435,
"grad_norm": 2.27116060256958,
"learning_rate": 1.8796274601686975e-05,
"loss": 0.3022,
"step": 4110
},
{
"epoch": 1.930646672914714,
"grad_norm": 1.9829683303833008,
"learning_rate": 1.8793345829428304e-05,
"loss": 0.3147,
"step": 4120
},
{
"epoch": 1.935332708528585,
"grad_norm": 2.073610305786133,
"learning_rate": 1.8790417057169637e-05,
"loss": 0.2779,
"step": 4130
},
{
"epoch": 1.9400187441424555,
"grad_norm": 2.183823585510254,
"learning_rate": 1.8787488284910967e-05,
"loss": 0.3041,
"step": 4140
},
{
"epoch": 1.9447047797563262,
"grad_norm": 2.373502016067505,
"learning_rate": 1.87845595126523e-05,
"loss": 0.3037,
"step": 4150
},
{
"epoch": 1.9493908153701969,
"grad_norm": 3.190950393676758,
"learning_rate": 1.878163074039363e-05,
"loss": 0.2876,
"step": 4160
},
{
"epoch": 1.9540768509840674,
"grad_norm": 2.2904293537139893,
"learning_rate": 1.8778701968134962e-05,
"loss": 0.2957,
"step": 4170
},
{
"epoch": 1.9587628865979383,
"grad_norm": 2.5620875358581543,
"learning_rate": 1.877577319587629e-05,
"loss": 0.2865,
"step": 4180
},
{
"epoch": 1.9634489222118088,
"grad_norm": 1.8882808685302734,
"learning_rate": 1.877284442361762e-05,
"loss": 0.302,
"step": 4190
},
{
"epoch": 1.9681349578256795,
"grad_norm": 2.736853837966919,
"learning_rate": 1.876991565135895e-05,
"loss": 0.2783,
"step": 4200
},
{
"epoch": 1.9728209934395502,
"grad_norm": 2.6126515865325928,
"learning_rate": 1.8766986879100283e-05,
"loss": 0.3337,
"step": 4210
},
{
"epoch": 1.9775070290534207,
"grad_norm": 1.9608129262924194,
"learning_rate": 1.8764058106841613e-05,
"loss": 0.2817,
"step": 4220
},
{
"epoch": 1.9821930646672916,
"grad_norm": 2.7429044246673584,
"learning_rate": 1.8761129334582942e-05,
"loss": 0.2827,
"step": 4230
},
{
"epoch": 1.986879100281162,
"grad_norm": 2.322411298751831,
"learning_rate": 1.8758200562324275e-05,
"loss": 0.2967,
"step": 4240
},
{
"epoch": 1.9915651358950328,
"grad_norm": 1.80522882938385,
"learning_rate": 1.8755271790065605e-05,
"loss": 0.315,
"step": 4250
},
{
"epoch": 1.9915651358950328,
"eval_loss": 0.0709637850522995,
"eval_pearson_cosine": 0.7818087092412611,
"eval_pearson_dot": 0.5917647670204929,
"eval_pearson_euclidean": 0.7671767812906438,
"eval_pearson_manhattan": 0.7657443892799449,
"eval_runtime": 3.3617,
"eval_samples_per_second": 446.198,
"eval_spearman_cosine": 0.7925261968709182,
"eval_spearman_dot": 0.5930341342679433,
"eval_spearman_euclidean": 0.7790026457642986,
"eval_spearman_manhattan": 0.7779600937996545,
"eval_steps_per_second": 27.962,
"step": 4250
},
{
"epoch": 1.9962511715089035,
"grad_norm": 2.3154516220092773,
"learning_rate": 1.8752343017806937e-05,
"loss": 0.3019,
"step": 4260
},
{
"epoch": 2.000937207122774,
"grad_norm": 1.769853949546814,
"learning_rate": 1.8749414245548267e-05,
"loss": 0.2639,
"step": 4270
},
{
"epoch": 2.005623242736645,
"grad_norm": 2.7089197635650635,
"learning_rate": 1.87464854732896e-05,
"loss": 0.1872,
"step": 4280
},
{
"epoch": 2.0103092783505154,
"grad_norm": 2.1212291717529297,
"learning_rate": 1.874355670103093e-05,
"loss": 0.189,
"step": 4290
},
{
"epoch": 2.014995313964386,
"grad_norm": 1.8745218515396118,
"learning_rate": 1.874062792877226e-05,
"loss": 0.1749,
"step": 4300
},
{
"epoch": 2.019681349578257,
"grad_norm": 2.0145514011383057,
"learning_rate": 1.8737699156513592e-05,
"loss": 0.1809,
"step": 4310
},
{
"epoch": 2.0243673851921273,
"grad_norm": 2.370055913925171,
"learning_rate": 1.873477038425492e-05,
"loss": 0.2056,
"step": 4320
},
{
"epoch": 2.0290534208059983,
"grad_norm": 1.7560367584228516,
"learning_rate": 1.8731841611996254e-05,
"loss": 0.1742,
"step": 4330
},
{
"epoch": 2.0337394564198688,
"grad_norm": 2.693603277206421,
"learning_rate": 1.8728912839737584e-05,
"loss": 0.2017,
"step": 4340
},
{
"epoch": 2.0384254920337392,
"grad_norm": 2.1456475257873535,
"learning_rate": 1.8725984067478917e-05,
"loss": 0.1967,
"step": 4350
},
{
"epoch": 2.04311152764761,
"grad_norm": 1.8832343816757202,
"learning_rate": 1.8723055295220246e-05,
"loss": 0.182,
"step": 4360
},
{
"epoch": 2.0477975632614807,
"grad_norm": 2.2446205615997314,
"learning_rate": 1.8720126522961576e-05,
"loss": 0.1895,
"step": 4370
},
{
"epoch": 2.0524835988753516,
"grad_norm": 1.8662675619125366,
"learning_rate": 1.871719775070291e-05,
"loss": 0.2091,
"step": 4380
},
{
"epoch": 2.057169634489222,
"grad_norm": 2.3427059650421143,
"learning_rate": 1.8714268978444238e-05,
"loss": 0.2046,
"step": 4390
},
{
"epoch": 2.0618556701030926,
"grad_norm": 1.8921583890914917,
"learning_rate": 1.8711340206185567e-05,
"loss": 0.21,
"step": 4400
},
{
"epoch": 2.0665417057169635,
"grad_norm": 2.3345112800598145,
"learning_rate": 1.8708411433926897e-05,
"loss": 0.211,
"step": 4410
},
{
"epoch": 2.071227741330834,
"grad_norm": 2.8094024658203125,
"learning_rate": 1.870548266166823e-05,
"loss": 0.185,
"step": 4420
},
{
"epoch": 2.075913776944705,
"grad_norm": 2.182375192642212,
"learning_rate": 1.870255388940956e-05,
"loss": 0.1743,
"step": 4430
},
{
"epoch": 2.0805998125585754,
"grad_norm": 1.9309943914413452,
"learning_rate": 1.8699625117150892e-05,
"loss": 0.2002,
"step": 4440
},
{
"epoch": 2.085285848172446,
"grad_norm": 2.410891532897949,
"learning_rate": 1.869669634489222e-05,
"loss": 0.2034,
"step": 4450
},
{
"epoch": 2.089971883786317,
"grad_norm": 1.9030256271362305,
"learning_rate": 1.8693767572633555e-05,
"loss": 0.1985,
"step": 4460
},
{
"epoch": 2.0946579194001873,
"grad_norm": 2.387491226196289,
"learning_rate": 1.8690838800374884e-05,
"loss": 0.1657,
"step": 4470
},
{
"epoch": 2.0993439550140582,
"grad_norm": 2.448557138442993,
"learning_rate": 1.8687910028116217e-05,
"loss": 0.1889,
"step": 4480
},
{
"epoch": 2.1040299906279287,
"grad_norm": 1.8442336320877075,
"learning_rate": 1.8684981255857546e-05,
"loss": 0.1638,
"step": 4490
},
{
"epoch": 2.108716026241799,
"grad_norm": 2.192814588546753,
"learning_rate": 1.8682052483598876e-05,
"loss": 0.2117,
"step": 4500
},
{
"epoch": 2.108716026241799,
"eval_loss": 0.0544867068529129,
"eval_pearson_cosine": 0.7772149225219422,
"eval_pearson_dot": 0.6058722441499071,
"eval_pearson_euclidean": 0.7566861769282127,
"eval_pearson_manhattan": 0.7550653692259459,
"eval_runtime": 3.562,
"eval_samples_per_second": 421.112,
"eval_spearman_cosine": 0.7889882574404774,
"eval_spearman_dot": 0.609585139530218,
"eval_spearman_euclidean": 0.7711664957300913,
"eval_spearman_manhattan": 0.7702480077346849,
"eval_steps_per_second": 26.39,
"step": 4500
},
{
"epoch": 2.11340206185567,
"grad_norm": 1.964051604270935,
"learning_rate": 1.867912371134021e-05,
"loss": 0.1727,
"step": 4510
},
{
"epoch": 2.1180880974695406,
"grad_norm": 2.180359363555908,
"learning_rate": 1.867619493908154e-05,
"loss": 0.1929,
"step": 4520
},
{
"epoch": 2.1227741330834116,
"grad_norm": 1.5920389890670776,
"learning_rate": 1.867326616682287e-05,
"loss": 0.215,
"step": 4530
},
{
"epoch": 2.127460168697282,
"grad_norm": 2.0330374240875244,
"learning_rate": 1.86703373945642e-05,
"loss": 0.1782,
"step": 4540
},
{
"epoch": 2.1321462043111525,
"grad_norm": 3.1396799087524414,
"learning_rate": 1.866740862230553e-05,
"loss": 0.1843,
"step": 4550
},
{
"epoch": 2.1368322399250235,
"grad_norm": 2.507819652557373,
"learning_rate": 1.8664479850046863e-05,
"loss": 0.2272,
"step": 4560
},
{
"epoch": 2.141518275538894,
"grad_norm": 1.9377321004867554,
"learning_rate": 1.8661551077788193e-05,
"loss": 0.2003,
"step": 4570
},
{
"epoch": 2.146204311152765,
"grad_norm": 2.868741512298584,
"learning_rate": 1.8658622305529522e-05,
"loss": 0.2266,
"step": 4580
},
{
"epoch": 2.1508903467666354,
"grad_norm": 1.8727847337722778,
"learning_rate": 1.8655693533270855e-05,
"loss": 0.1965,
"step": 4590
},
{
"epoch": 2.155576382380506,
"grad_norm": 1.937414526939392,
"learning_rate": 1.8652764761012184e-05,
"loss": 0.1957,
"step": 4600
},
{
"epoch": 2.160262417994377,
"grad_norm": 2.084174156188965,
"learning_rate": 1.8649835988753514e-05,
"loss": 0.1668,
"step": 4610
},
{
"epoch": 2.1649484536082473,
"grad_norm": 1.846552848815918,
"learning_rate": 1.8646907216494847e-05,
"loss": 0.184,
"step": 4620
},
{
"epoch": 2.169634489222118,
"grad_norm": 1.9087766408920288,
"learning_rate": 1.8643978444236176e-05,
"loss": 0.1961,
"step": 4630
},
{
"epoch": 2.1743205248359887,
"grad_norm": 2.011108875274658,
"learning_rate": 1.864104967197751e-05,
"loss": 0.1848,
"step": 4640
},
{
"epoch": 2.179006560449859,
"grad_norm": 2.8976380825042725,
"learning_rate": 1.863812089971884e-05,
"loss": 0.2082,
"step": 4650
},
{
"epoch": 2.18369259606373,
"grad_norm": 2.1731081008911133,
"learning_rate": 1.863519212746017e-05,
"loss": 0.1852,
"step": 4660
},
{
"epoch": 2.1883786316776006,
"grad_norm": 1.7672990560531616,
"learning_rate": 1.86322633552015e-05,
"loss": 0.2023,
"step": 4670
},
{
"epoch": 2.1930646672914715,
"grad_norm": 1.6886131763458252,
"learning_rate": 1.8629334582942834e-05,
"loss": 0.2126,
"step": 4680
},
{
"epoch": 2.197750702905342,
"grad_norm": 1.4913769960403442,
"learning_rate": 1.8626405810684164e-05,
"loss": 0.1836,
"step": 4690
},
{
"epoch": 2.2024367385192125,
"grad_norm": 2.0632176399230957,
"learning_rate": 1.8623477038425493e-05,
"loss": 0.2048,
"step": 4700
},
{
"epoch": 2.2071227741330834,
"grad_norm": 1.9084508419036865,
"learning_rate": 1.8620548266166826e-05,
"loss": 0.1836,
"step": 4710
},
{
"epoch": 2.211808809746954,
"grad_norm": 1.9352916479110718,
"learning_rate": 1.8617619493908155e-05,
"loss": 0.1653,
"step": 4720
},
{
"epoch": 2.216494845360825,
"grad_norm": 1.974461317062378,
"learning_rate": 1.8614690721649485e-05,
"loss": 0.2053,
"step": 4730
},
{
"epoch": 2.2211808809746953,
"grad_norm": 2.1307456493377686,
"learning_rate": 1.8611761949390818e-05,
"loss": 0.1835,
"step": 4740
},
{
"epoch": 2.2258669165885663,
"grad_norm": 1.413037896156311,
"learning_rate": 1.8608833177132147e-05,
"loss": 0.1725,
"step": 4750
},
{
"epoch": 2.2258669165885663,
"eval_loss": 0.05439988151192665,
"eval_pearson_cosine": 0.7779719017710534,
"eval_pearson_dot": 0.6064920766481805,
"eval_pearson_euclidean": 0.7604967084059453,
"eval_pearson_manhattan": 0.7592805758538006,
"eval_runtime": 3.2739,
"eval_samples_per_second": 458.175,
"eval_spearman_cosine": 0.786759628149343,
"eval_spearman_dot": 0.6127818290418755,
"eval_spearman_euclidean": 0.7721037448944629,
"eval_spearman_manhattan": 0.7714373415381524,
"eval_steps_per_second": 28.712,
"step": 4750
},
{
"epoch": 2.2305529522024368,
"grad_norm": 2.033710241317749,
"learning_rate": 1.8605904404873477e-05,
"loss": 0.1769,
"step": 4760
},
{
"epoch": 2.2352389878163073,
"grad_norm": 2.065793752670288,
"learning_rate": 1.860297563261481e-05,
"loss": 0.1896,
"step": 4770
},
{
"epoch": 2.239925023430178,
"grad_norm": 2.0393176078796387,
"learning_rate": 1.860004686035614e-05,
"loss": 0.1944,
"step": 4780
},
{
"epoch": 2.2446110590440487,
"grad_norm": 1.967092514038086,
"learning_rate": 1.859711808809747e-05,
"loss": 0.218,
"step": 4790
},
{
"epoch": 2.2492970946579196,
"grad_norm": 2.614811420440674,
"learning_rate": 1.85941893158388e-05,
"loss": 0.2158,
"step": 4800
},
{
"epoch": 2.25398313027179,
"grad_norm": 1.9915626049041748,
"learning_rate": 1.859126054358013e-05,
"loss": 0.199,
"step": 4810
},
{
"epoch": 2.2586691658856606,
"grad_norm": 2.308933973312378,
"learning_rate": 1.8588331771321464e-05,
"loss": 0.1746,
"step": 4820
},
{
"epoch": 2.2633552014995315,
"grad_norm": 2.2984583377838135,
"learning_rate": 1.8585402999062793e-05,
"loss": 0.1905,
"step": 4830
},
{
"epoch": 2.268041237113402,
"grad_norm": 2.0273101329803467,
"learning_rate": 1.8582474226804126e-05,
"loss": 0.1655,
"step": 4840
},
{
"epoch": 2.2727272727272725,
"grad_norm": 1.745671272277832,
"learning_rate": 1.8579545454545456e-05,
"loss": 0.1745,
"step": 4850
},
{
"epoch": 2.2774133083411434,
"grad_norm": 1.9413400888442993,
"learning_rate": 1.857661668228679e-05,
"loss": 0.1899,
"step": 4860
},
{
"epoch": 2.282099343955014,
"grad_norm": 2.113260269165039,
"learning_rate": 1.8573687910028118e-05,
"loss": 0.1826,
"step": 4870
},
{
"epoch": 2.286785379568885,
"grad_norm": 1.9468377828598022,
"learning_rate": 1.857075913776945e-05,
"loss": 0.1596,
"step": 4880
},
{
"epoch": 2.2914714151827553,
"grad_norm": 2.1250803470611572,
"learning_rate": 1.856783036551078e-05,
"loss": 0.1969,
"step": 4890
},
{
"epoch": 2.296157450796626,
"grad_norm": 1.5394418239593506,
"learning_rate": 1.856490159325211e-05,
"loss": 0.1731,
"step": 4900
},
{
"epoch": 2.3008434864104967,
"grad_norm": 2.2445003986358643,
"learning_rate": 1.8561972820993443e-05,
"loss": 0.1817,
"step": 4910
},
{
"epoch": 2.3055295220243672,
"grad_norm": 2.201733350753784,
"learning_rate": 1.8559044048734773e-05,
"loss": 0.175,
"step": 4920
},
{
"epoch": 2.310215557638238,
"grad_norm": 2.077092409133911,
"learning_rate": 1.8556115276476102e-05,
"loss": 0.1755,
"step": 4930
},
{
"epoch": 2.3149015932521086,
"grad_norm": 2.0630977153778076,
"learning_rate": 1.855318650421743e-05,
"loss": 0.1967,
"step": 4940
},
{
"epoch": 2.319587628865979,
"grad_norm": 2.737861156463623,
"learning_rate": 1.8550257731958764e-05,
"loss": 0.2175,
"step": 4950
},
{
"epoch": 2.32427366447985,
"grad_norm": 2.0301871299743652,
"learning_rate": 1.8547328959700094e-05,
"loss": 0.1938,
"step": 4960
},
{
"epoch": 2.3289597000937206,
"grad_norm": 1.6436868906021118,
"learning_rate": 1.8544400187441427e-05,
"loss": 0.2215,
"step": 4970
},
{
"epoch": 2.3336457357075915,
"grad_norm": 2.0046422481536865,
"learning_rate": 1.8541471415182756e-05,
"loss": 0.189,
"step": 4980
},
{
"epoch": 2.338331771321462,
"grad_norm": 2.598639488220215,
"learning_rate": 1.8538542642924086e-05,
"loss": 0.2075,
"step": 4990
},
{
"epoch": 2.3430178069353325,
"grad_norm": 1.8523390293121338,
"learning_rate": 1.853561387066542e-05,
"loss": 0.1985,
"step": 5000
},
{
"epoch": 2.3430178069353325,
"eval_loss": 0.05401456356048584,
"eval_pearson_cosine": 0.781767118943602,
"eval_pearson_dot": 0.6017329721420595,
"eval_pearson_euclidean": 0.7626066877478763,
"eval_pearson_manhattan": 0.7620550203728129,
"eval_runtime": 3.2064,
"eval_samples_per_second": 467.808,
"eval_spearman_cosine": 0.7915644905642973,
"eval_spearman_dot": 0.6077979618715047,
"eval_spearman_euclidean": 0.773425356513352,
"eval_spearman_manhattan": 0.7732819218811388,
"eval_steps_per_second": 29.316,
"step": 5000
},
{
"epoch": 2.3477038425492034,
"grad_norm": 2.0773608684539795,
"learning_rate": 1.8532685098406748e-05,
"loss": 0.1749,
"step": 5010
},
{
"epoch": 2.352389878163074,
"grad_norm": 1.6482021808624268,
"learning_rate": 1.852975632614808e-05,
"loss": 0.1719,
"step": 5020
},
{
"epoch": 2.357075913776945,
"grad_norm": 2.1038990020751953,
"learning_rate": 1.852682755388941e-05,
"loss": 0.2014,
"step": 5030
},
{
"epoch": 2.3617619493908153,
"grad_norm": 2.396871566772461,
"learning_rate": 1.8523898781630743e-05,
"loss": 0.1855,
"step": 5040
},
{
"epoch": 2.3664479850046862,
"grad_norm": 2.3098223209381104,
"learning_rate": 1.8520970009372073e-05,
"loss": 0.1819,
"step": 5050
},
{
"epoch": 2.3711340206185567,
"grad_norm": 1.9691858291625977,
"learning_rate": 1.8518041237113406e-05,
"loss": 0.1786,
"step": 5060
},
{
"epoch": 2.375820056232427,
"grad_norm": 2.8055005073547363,
"learning_rate": 1.8515112464854735e-05,
"loss": 0.1864,
"step": 5070
},
{
"epoch": 2.380506091846298,
"grad_norm": 2.2753820419311523,
"learning_rate": 1.8512183692596065e-05,
"loss": 0.1918,
"step": 5080
},
{
"epoch": 2.3851921274601686,
"grad_norm": 1.632279634475708,
"learning_rate": 1.8509254920337398e-05,
"loss": 0.1767,
"step": 5090
},
{
"epoch": 2.3898781630740396,
"grad_norm": 2.177457571029663,
"learning_rate": 1.8506326148078727e-05,
"loss": 0.1866,
"step": 5100
},
{
"epoch": 2.39456419868791,
"grad_norm": 1.8641494512557983,
"learning_rate": 1.8503397375820057e-05,
"loss": 0.1833,
"step": 5110
},
{
"epoch": 2.3992502343017805,
"grad_norm": 2.0116262435913086,
"learning_rate": 1.8500468603561386e-05,
"loss": 0.1975,
"step": 5120
},
{
"epoch": 2.4039362699156515,
"grad_norm": 1.130846381187439,
"learning_rate": 1.849753983130272e-05,
"loss": 0.1992,
"step": 5130
},
{
"epoch": 2.408622305529522,
"grad_norm": 2.743990421295166,
"learning_rate": 1.849461105904405e-05,
"loss": 0.1996,
"step": 5140
},
{
"epoch": 2.413308341143393,
"grad_norm": 2.4368762969970703,
"learning_rate": 1.849168228678538e-05,
"loss": 0.1722,
"step": 5150
},
{
"epoch": 2.4179943767572634,
"grad_norm": 2.659679889678955,
"learning_rate": 1.848875351452671e-05,
"loss": 0.1834,
"step": 5160
},
{
"epoch": 2.422680412371134,
"grad_norm": 1.9601231813430786,
"learning_rate": 1.8485824742268044e-05,
"loss": 0.1943,
"step": 5170
},
{
"epoch": 2.427366447985005,
"grad_norm": 1.521337628364563,
"learning_rate": 1.8482895970009373e-05,
"loss": 0.2009,
"step": 5180
},
{
"epoch": 2.4320524835988753,
"grad_norm": 1.4766628742218018,
"learning_rate": 1.8479967197750703e-05,
"loss": 0.1848,
"step": 5190
},
{
"epoch": 2.436738519212746,
"grad_norm": 2.3502743244171143,
"learning_rate": 1.8477038425492036e-05,
"loss": 0.165,
"step": 5200
},
{
"epoch": 2.4414245548266167,
"grad_norm": 1.5100184679031372,
"learning_rate": 1.8474109653233365e-05,
"loss": 0.1983,
"step": 5210
},
{
"epoch": 2.446110590440487,
"grad_norm": 2.336106061935425,
"learning_rate": 1.8471180880974698e-05,
"loss": 0.1871,
"step": 5220
},
{
"epoch": 2.450796626054358,
"grad_norm": 2.1695876121520996,
"learning_rate": 1.8468252108716028e-05,
"loss": 0.1996,
"step": 5230
},
{
"epoch": 2.4554826616682286,
"grad_norm": 2.047522783279419,
"learning_rate": 1.846532333645736e-05,
"loss": 0.2088,
"step": 5240
},
{
"epoch": 2.4601686972820995,
"grad_norm": 1.9469248056411743,
"learning_rate": 1.846239456419869e-05,
"loss": 0.1871,
"step": 5250
},
{
"epoch": 2.4601686972820995,
"eval_loss": 0.05267513543367386,
"eval_pearson_cosine": 0.7830377509621513,
"eval_pearson_dot": 0.5843463467596735,
"eval_pearson_euclidean": 0.7586592990276699,
"eval_pearson_manhattan": 0.7576312827364555,
"eval_runtime": 3.3095,
"eval_samples_per_second": 453.24,
"eval_spearman_cosine": 0.7898144287463136,
"eval_spearman_dot": 0.5893690662486756,
"eval_spearman_euclidean": 0.7723820863953171,
"eval_spearman_manhattan": 0.7718409919252442,
"eval_steps_per_second": 28.403,
"step": 5250
},
{
"epoch": 2.46485473289597,
"grad_norm": 2.1970055103302,
"learning_rate": 1.845946579194002e-05,
"loss": 0.2059,
"step": 5260
},
{
"epoch": 2.4695407685098405,
"grad_norm": 1.7079726457595825,
"learning_rate": 1.8456537019681352e-05,
"loss": 0.1677,
"step": 5270
},
{
"epoch": 2.4742268041237114,
"grad_norm": 2.1129045486450195,
"learning_rate": 1.8453608247422682e-05,
"loss": 0.1774,
"step": 5280
},
{
"epoch": 2.478912839737582,
"grad_norm": 2.0570807456970215,
"learning_rate": 1.845067947516401e-05,
"loss": 0.1863,
"step": 5290
},
{
"epoch": 2.483598875351453,
"grad_norm": 2.095052719116211,
"learning_rate": 1.8447750702905344e-05,
"loss": 0.1675,
"step": 5300
},
{
"epoch": 2.4882849109653233,
"grad_norm": 2.5408618450164795,
"learning_rate": 1.8444821930646674e-05,
"loss": 0.2033,
"step": 5310
},
{
"epoch": 2.492970946579194,
"grad_norm": 2.0885772705078125,
"learning_rate": 1.8441893158388003e-05,
"loss": 0.178,
"step": 5320
},
{
"epoch": 2.4976569821930648,
"grad_norm": 2.907564640045166,
"learning_rate": 1.8438964386129336e-05,
"loss": 0.2121,
"step": 5330
},
{
"epoch": 2.5023430178069352,
"grad_norm": 1.3953640460968018,
"learning_rate": 1.8436035613870666e-05,
"loss": 0.1838,
"step": 5340
},
{
"epoch": 2.5070290534208057,
"grad_norm": 1.825296401977539,
"learning_rate": 1.8433106841612e-05,
"loss": 0.1688,
"step": 5350
},
{
"epoch": 2.5117150890346767,
"grad_norm": 2.5563881397247314,
"learning_rate": 1.8430178069353328e-05,
"loss": 0.1798,
"step": 5360
},
{
"epoch": 2.5164011246485476,
"grad_norm": 2.355055570602417,
"learning_rate": 1.842724929709466e-05,
"loss": 0.1858,
"step": 5370
},
{
"epoch": 2.521087160262418,
"grad_norm": 2.4546165466308594,
"learning_rate": 1.842432052483599e-05,
"loss": 0.2149,
"step": 5380
},
{
"epoch": 2.5257731958762886,
"grad_norm": 2.0621817111968994,
"learning_rate": 1.842139175257732e-05,
"loss": 0.2215,
"step": 5390
},
{
"epoch": 2.530459231490159,
"grad_norm": 1.7873655557632446,
"learning_rate": 1.8418462980318653e-05,
"loss": 0.1872,
"step": 5400
},
{
"epoch": 2.53514526710403,
"grad_norm": 1.8408267498016357,
"learning_rate": 1.8415534208059982e-05,
"loss": 0.181,
"step": 5410
},
{
"epoch": 2.539831302717901,
"grad_norm": 2.56502628326416,
"learning_rate": 1.8412605435801315e-05,
"loss": 0.1946,
"step": 5420
},
{
"epoch": 2.5445173383317714,
"grad_norm": 1.9909722805023193,
"learning_rate": 1.8409676663542645e-05,
"loss": 0.1912,
"step": 5430
},
{
"epoch": 2.549203373945642,
"grad_norm": 2.530014753341675,
"learning_rate": 1.8406747891283974e-05,
"loss": 0.1898,
"step": 5440
},
{
"epoch": 2.5538894095595124,
"grad_norm": 2.452223300933838,
"learning_rate": 1.8403819119025307e-05,
"loss": 0.2114,
"step": 5450
},
{
"epoch": 2.5585754451733833,
"grad_norm": 2.0805137157440186,
"learning_rate": 1.8400890346766637e-05,
"loss": 0.2237,
"step": 5460
},
{
"epoch": 2.5632614807872542,
"grad_norm": 1.8469469547271729,
"learning_rate": 1.8397961574507966e-05,
"loss": 0.1895,
"step": 5470
},
{
"epoch": 2.5679475164011247,
"grad_norm": 2.0643043518066406,
"learning_rate": 1.83950328022493e-05,
"loss": 0.1943,
"step": 5480
},
{
"epoch": 2.572633552014995,
"grad_norm": 2.1368064880371094,
"learning_rate": 1.839210402999063e-05,
"loss": 0.1851,
"step": 5490
},
{
"epoch": 2.5773195876288657,
"grad_norm": 1.7404872179031372,
"learning_rate": 1.8389175257731958e-05,
"loss": 0.17,
"step": 5500
},
{
"epoch": 2.5773195876288657,
"eval_loss": 0.052087847143411636,
"eval_pearson_cosine": 0.7876996404495316,
"eval_pearson_dot": 0.6239791899062368,
"eval_pearson_euclidean": 0.7632876525798107,
"eval_pearson_manhattan": 0.7621358820486535,
"eval_runtime": 3.2114,
"eval_samples_per_second": 467.081,
"eval_spearman_cosine": 0.7959069730064968,
"eval_spearman_dot": 0.6246328836004849,
"eval_spearman_euclidean": 0.775279722994119,
"eval_spearman_manhattan": 0.7746119015682993,
"eval_steps_per_second": 29.27,
"step": 5500
},
{
"epoch": 2.5820056232427366,
"grad_norm": 1.7516288757324219,
"learning_rate": 1.838624648547329e-05,
"loss": 0.2103,
"step": 5510
},
{
"epoch": 2.5866916588566076,
"grad_norm": 2.0095157623291016,
"learning_rate": 1.838331771321462e-05,
"loss": 0.2149,
"step": 5520
},
{
"epoch": 2.591377694470478,
"grad_norm": 2.3408539295196533,
"learning_rate": 1.8380388940955953e-05,
"loss": 0.1737,
"step": 5530
},
{
"epoch": 2.5960637300843485,
"grad_norm": 2.1319546699523926,
"learning_rate": 1.8377460168697283e-05,
"loss": 0.1963,
"step": 5540
},
{
"epoch": 2.600749765698219,
"grad_norm": 1.6769325733184814,
"learning_rate": 1.8374531396438616e-05,
"loss": 0.1953,
"step": 5550
},
{
"epoch": 2.60543580131209,
"grad_norm": 2.503340244293213,
"learning_rate": 1.8371602624179945e-05,
"loss": 0.1872,
"step": 5560
},
{
"epoch": 2.610121836925961,
"grad_norm": 1.9402929544448853,
"learning_rate": 1.8368673851921278e-05,
"loss": 0.2101,
"step": 5570
},
{
"epoch": 2.6148078725398314,
"grad_norm": 1.980652928352356,
"learning_rate": 1.8365745079662608e-05,
"loss": 0.1847,
"step": 5580
},
{
"epoch": 2.619493908153702,
"grad_norm": 2.6117632389068604,
"learning_rate": 1.8362816307403937e-05,
"loss": 0.1934,
"step": 5590
},
{
"epoch": 2.624179943767573,
"grad_norm": 1.8052171468734741,
"learning_rate": 1.835988753514527e-05,
"loss": 0.1827,
"step": 5600
},
{
"epoch": 2.6288659793814433,
"grad_norm": 2.384441375732422,
"learning_rate": 1.83569587628866e-05,
"loss": 0.1765,
"step": 5610
},
{
"epoch": 2.633552014995314,
"grad_norm": 1.8288124799728394,
"learning_rate": 1.8354029990627932e-05,
"loss": 0.189,
"step": 5620
},
{
"epoch": 2.6382380506091847,
"grad_norm": 1.8904608488082886,
"learning_rate": 1.8351101218369262e-05,
"loss": 0.1858,
"step": 5630
},
{
"epoch": 2.642924086223055,
"grad_norm": 2.4750778675079346,
"learning_rate": 1.834817244611059e-05,
"loss": 0.2139,
"step": 5640
},
{
"epoch": 2.647610121836926,
"grad_norm": 1.642340898513794,
"learning_rate": 1.834524367385192e-05,
"loss": 0.2069,
"step": 5650
},
{
"epoch": 2.6522961574507966,
"grad_norm": 2.656212091445923,
"learning_rate": 1.8342314901593254e-05,
"loss": 0.1975,
"step": 5660
},
{
"epoch": 2.6569821930646675,
"grad_norm": 1.8371754884719849,
"learning_rate": 1.8339386129334583e-05,
"loss": 0.2044,
"step": 5670
},
{
"epoch": 2.661668228678538,
"grad_norm": 1.4980151653289795,
"learning_rate": 1.8336457357075916e-05,
"loss": 0.1641,
"step": 5680
},
{
"epoch": 2.6663542642924085,
"grad_norm": 2.300854206085205,
"learning_rate": 1.8333528584817246e-05,
"loss": 0.22,
"step": 5690
},
{
"epoch": 2.6710402999062794,
"grad_norm": 1.6950479745864868,
"learning_rate": 1.8330599812558575e-05,
"loss": 0.194,
"step": 5700
},
{
"epoch": 2.67572633552015,
"grad_norm": 2.4695844650268555,
"learning_rate": 1.8327671040299908e-05,
"loss": 0.2099,
"step": 5710
},
{
"epoch": 2.680412371134021,
"grad_norm": 1.6376041173934937,
"learning_rate": 1.8324742268041237e-05,
"loss": 0.1844,
"step": 5720
},
{
"epoch": 2.6850984067478914,
"grad_norm": 2.3736367225646973,
"learning_rate": 1.832181349578257e-05,
"loss": 0.1886,
"step": 5730
},
{
"epoch": 2.689784442361762,
"grad_norm": 1.9710888862609863,
"learning_rate": 1.83188847235239e-05,
"loss": 0.1881,
"step": 5740
},
{
"epoch": 2.6944704779756328,
"grad_norm": 2.38665771484375,
"learning_rate": 1.8315955951265233e-05,
"loss": 0.174,
"step": 5750
},
{
"epoch": 2.6944704779756328,
"eval_loss": 0.052811067551374435,
"eval_pearson_cosine": 0.7875700428566148,
"eval_pearson_dot": 0.6195988091254279,
"eval_pearson_euclidean": 0.7602836109774316,
"eval_pearson_manhattan": 0.7593741208437699,
"eval_runtime": 3.3995,
"eval_samples_per_second": 441.245,
"eval_spearman_cosine": 0.7949301691809272,
"eval_spearman_dot": 0.6234473775126543,
"eval_spearman_euclidean": 0.771585595419442,
"eval_spearman_manhattan": 0.7712513119620406,
"eval_steps_per_second": 27.651,
"step": 5750
},
{
"epoch": 2.6991565135895033,
"grad_norm": 1.9556349515914917,
"learning_rate": 1.8313027179006562e-05,
"loss": 0.1984,
"step": 5760
},
{
"epoch": 2.703842549203374,
"grad_norm": 1.8807729482650757,
"learning_rate": 1.8310098406747895e-05,
"loss": 0.2057,
"step": 5770
},
{
"epoch": 2.7085285848172447,
"grad_norm": 1.74807870388031,
"learning_rate": 1.8307169634489225e-05,
"loss": 0.2158,
"step": 5780
},
{
"epoch": 2.713214620431115,
"grad_norm": 2.3334672451019287,
"learning_rate": 1.8304240862230554e-05,
"loss": 0.1676,
"step": 5790
},
{
"epoch": 2.717900656044986,
"grad_norm": 2.0100314617156982,
"learning_rate": 1.8301312089971887e-05,
"loss": 0.1785,
"step": 5800
},
{
"epoch": 2.7225866916588566,
"grad_norm": 1.9850749969482422,
"learning_rate": 1.8298383317713217e-05,
"loss": 0.191,
"step": 5810
},
{
"epoch": 2.7272727272727275,
"grad_norm": 2.1841342449188232,
"learning_rate": 1.8295454545454546e-05,
"loss": 0.1814,
"step": 5820
},
{
"epoch": 2.731958762886598,
"grad_norm": 2.6331992149353027,
"learning_rate": 1.8292525773195876e-05,
"loss": 0.2218,
"step": 5830
},
{
"epoch": 2.7366447985004685,
"grad_norm": 2.161170482635498,
"learning_rate": 1.828959700093721e-05,
"loss": 0.1834,
"step": 5840
},
{
"epoch": 2.7413308341143394,
"grad_norm": 2.3174726963043213,
"learning_rate": 1.8286668228678538e-05,
"loss": 0.1863,
"step": 5850
},
{
"epoch": 2.74601686972821,
"grad_norm": 1.9110337495803833,
"learning_rate": 1.828373945641987e-05,
"loss": 0.179,
"step": 5860
},
{
"epoch": 2.750702905342081,
"grad_norm": 2.59794545173645,
"learning_rate": 1.82808106841612e-05,
"loss": 0.1891,
"step": 5870
},
{
"epoch": 2.7553889409559513,
"grad_norm": 2.1763062477111816,
"learning_rate": 1.8277881911902533e-05,
"loss": 0.2094,
"step": 5880
},
{
"epoch": 2.760074976569822,
"grad_norm": 2.3349921703338623,
"learning_rate": 1.8274953139643863e-05,
"loss": 0.1848,
"step": 5890
},
{
"epoch": 2.7647610121836927,
"grad_norm": 2.5409889221191406,
"learning_rate": 1.8272024367385192e-05,
"loss": 0.2392,
"step": 5900
},
{
"epoch": 2.7694470477975632,
"grad_norm": 2.0501179695129395,
"learning_rate": 1.8269095595126525e-05,
"loss": 0.2116,
"step": 5910
},
{
"epoch": 2.774133083411434,
"grad_norm": 2.486163854598999,
"learning_rate": 1.8266166822867855e-05,
"loss": 0.1899,
"step": 5920
},
{
"epoch": 2.7788191190253047,
"grad_norm": 2.43495774269104,
"learning_rate": 1.8263238050609187e-05,
"loss": 0.212,
"step": 5930
},
{
"epoch": 2.783505154639175,
"grad_norm": 1.9861983060836792,
"learning_rate": 1.8260309278350517e-05,
"loss": 0.2135,
"step": 5940
},
{
"epoch": 2.788191190253046,
"grad_norm": 2.46321964263916,
"learning_rate": 1.825738050609185e-05,
"loss": 0.2033,
"step": 5950
},
{
"epoch": 2.7928772258669166,
"grad_norm": 2.2375664710998535,
"learning_rate": 1.825445173383318e-05,
"loss": 0.2483,
"step": 5960
},
{
"epoch": 2.7975632614807875,
"grad_norm": 2.2775328159332275,
"learning_rate": 1.825152296157451e-05,
"loss": 0.1921,
"step": 5970
},
{
"epoch": 2.802249297094658,
"grad_norm": 2.1675665378570557,
"learning_rate": 1.8248594189315842e-05,
"loss": 0.1666,
"step": 5980
},
{
"epoch": 2.8069353327085285,
"grad_norm": 1.969119668006897,
"learning_rate": 1.824566541705717e-05,
"loss": 0.1997,
"step": 5990
},
{
"epoch": 2.8116213683223994,
"grad_norm": 3.1404953002929688,
"learning_rate": 1.82427366447985e-05,
"loss": 0.1896,
"step": 6000
},
{
"epoch": 2.8116213683223994,
"eval_loss": 0.0506160594522953,
"eval_pearson_cosine": 0.7847906228898438,
"eval_pearson_dot": 0.6051671435295134,
"eval_pearson_euclidean": 0.7605907930754796,
"eval_pearson_manhattan": 0.7595079560474858,
"eval_runtime": 3.1727,
"eval_samples_per_second": 472.783,
"eval_spearman_cosine": 0.7891227097506602,
"eval_spearman_dot": 0.6083246161085648,
"eval_spearman_euclidean": 0.7718421311199012,
"eval_spearman_manhattan": 0.7711946620901435,
"eval_steps_per_second": 29.628,
"step": 6000
},
{
"epoch": 2.81630740393627,
"grad_norm": 2.5197956562042236,
"learning_rate": 1.8239807872539834e-05,
"loss": 0.19,
"step": 6010
},
{
"epoch": 2.820993439550141,
"grad_norm": 2.342001438140869,
"learning_rate": 1.8236879100281163e-05,
"loss": 0.1902,
"step": 6020
},
{
"epoch": 2.8256794751640113,
"grad_norm": 2.081195592880249,
"learning_rate": 1.8233950328022493e-05,
"loss": 0.1861,
"step": 6030
},
{
"epoch": 2.830365510777882,
"grad_norm": 2.2222466468811035,
"learning_rate": 1.8231021555763825e-05,
"loss": 0.1956,
"step": 6040
},
{
"epoch": 2.8350515463917527,
"grad_norm": 2.531655788421631,
"learning_rate": 1.8228092783505155e-05,
"loss": 0.2001,
"step": 6050
},
{
"epoch": 2.839737582005623,
"grad_norm": 1.8432163000106812,
"learning_rate": 1.8225164011246488e-05,
"loss": 0.1917,
"step": 6060
},
{
"epoch": 2.844423617619494,
"grad_norm": 1.8071563243865967,
"learning_rate": 1.8222235238987817e-05,
"loss": 0.2292,
"step": 6070
},
{
"epoch": 2.8491096532333646,
"grad_norm": 2.596858501434326,
"learning_rate": 1.821930646672915e-05,
"loss": 0.2025,
"step": 6080
},
{
"epoch": 2.853795688847235,
"grad_norm": 2.516810655593872,
"learning_rate": 1.821637769447048e-05,
"loss": 0.2008,
"step": 6090
},
{
"epoch": 2.858481724461106,
"grad_norm": 2.685643196105957,
"learning_rate": 1.821344892221181e-05,
"loss": 0.2181,
"step": 6100
},
{
"epoch": 2.8631677600749765,
"grad_norm": 2.3013880252838135,
"learning_rate": 1.8210520149953142e-05,
"loss": 0.2089,
"step": 6110
},
{
"epoch": 2.8678537956888475,
"grad_norm": 2.1551058292388916,
"learning_rate": 1.820759137769447e-05,
"loss": 0.2096,
"step": 6120
},
{
"epoch": 2.872539831302718,
"grad_norm": 2.2167651653289795,
"learning_rate": 1.8204662605435805e-05,
"loss": 0.1928,
"step": 6130
},
{
"epoch": 2.8772258669165884,
"grad_norm": 2.3424344062805176,
"learning_rate": 1.8201733833177134e-05,
"loss": 0.1843,
"step": 6140
},
{
"epoch": 2.8819119025304594,
"grad_norm": 2.8250808715820312,
"learning_rate": 1.8198805060918464e-05,
"loss": 0.2221,
"step": 6150
},
{
"epoch": 2.88659793814433,
"grad_norm": 2.3614280223846436,
"learning_rate": 1.8195876288659796e-05,
"loss": 0.1862,
"step": 6160
},
{
"epoch": 2.891283973758201,
"grad_norm": 1.918118953704834,
"learning_rate": 1.8192947516401126e-05,
"loss": 0.1867,
"step": 6170
},
{
"epoch": 2.8959700093720713,
"grad_norm": 1.8377277851104736,
"learning_rate": 1.8190018744142455e-05,
"loss": 0.1826,
"step": 6180
},
{
"epoch": 2.9006560449859418,
"grad_norm": 2.527493715286255,
"learning_rate": 1.8187089971883788e-05,
"loss": 0.1986,
"step": 6190
},
{
"epoch": 2.9053420805998127,
"grad_norm": 2.125455379486084,
"learning_rate": 1.8184161199625118e-05,
"loss": 0.1936,
"step": 6200
},
{
"epoch": 2.910028116213683,
"grad_norm": 2.2447023391723633,
"learning_rate": 1.8181232427366447e-05,
"loss": 0.1887,
"step": 6210
},
{
"epoch": 2.914714151827554,
"grad_norm": 2.3427064418792725,
"learning_rate": 1.817830365510778e-05,
"loss": 0.197,
"step": 6220
},
{
"epoch": 2.9194001874414246,
"grad_norm": 2.480989933013916,
"learning_rate": 1.817537488284911e-05,
"loss": 0.2145,
"step": 6230
},
{
"epoch": 2.924086223055295,
"grad_norm": 2.2610814571380615,
"learning_rate": 1.8172446110590443e-05,
"loss": 0.1854,
"step": 6240
},
{
"epoch": 2.928772258669166,
"grad_norm": 2.220090866088867,
"learning_rate": 1.8169517338331772e-05,
"loss": 0.1897,
"step": 6250
},
{
"epoch": 2.928772258669166,
"eval_loss": 0.05493560805916786,
"eval_pearson_cosine": 0.7819311073703084,
"eval_pearson_dot": 0.5956676516410369,
"eval_pearson_euclidean": 0.7532898404318757,
"eval_pearson_manhattan": 0.7521288710513865,
"eval_runtime": 3.273,
"eval_samples_per_second": 458.299,
"eval_spearman_cosine": 0.7902389528249426,
"eval_spearman_dot": 0.5980515739599185,
"eval_spearman_euclidean": 0.7666646754982684,
"eval_spearman_manhattan": 0.7664381416282928,
"eval_steps_per_second": 28.72,
"step": 6250
},
{
"epoch": 2.9334582942830365,
"grad_norm": 2.517014741897583,
"learning_rate": 1.8166588566073105e-05,
"loss": 0.2132,
"step": 6260
},
{
"epoch": 2.9381443298969074,
"grad_norm": 1.9690264463424683,
"learning_rate": 1.8163659793814434e-05,
"loss": 0.1756,
"step": 6270
},
{
"epoch": 2.942830365510778,
"grad_norm": 1.812945008277893,
"learning_rate": 1.8160731021555767e-05,
"loss": 0.1848,
"step": 6280
},
{
"epoch": 2.9475164011246484,
"grad_norm": 2.003582000732422,
"learning_rate": 1.8157802249297097e-05,
"loss": 0.2071,
"step": 6290
},
{
"epoch": 2.9522024367385193,
"grad_norm": 3.1531078815460205,
"learning_rate": 1.8154873477038426e-05,
"loss": 0.2146,
"step": 6300
},
{
"epoch": 2.95688847235239,
"grad_norm": 2.5869102478027344,
"learning_rate": 1.815194470477976e-05,
"loss": 0.1854,
"step": 6310
},
{
"epoch": 2.9615745079662608,
"grad_norm": 2.2422142028808594,
"learning_rate": 1.814901593252109e-05,
"loss": 0.179,
"step": 6320
},
{
"epoch": 2.9662605435801312,
"grad_norm": 2.346750497817993,
"learning_rate": 1.814608716026242e-05,
"loss": 0.1937,
"step": 6330
},
{
"epoch": 2.9709465791940017,
"grad_norm": 1.6241106986999512,
"learning_rate": 1.814315838800375e-05,
"loss": 0.1945,
"step": 6340
},
{
"epoch": 2.9756326148078727,
"grad_norm": 2.8439619541168213,
"learning_rate": 1.814022961574508e-05,
"loss": 0.2265,
"step": 6350
},
{
"epoch": 2.980318650421743,
"grad_norm": 1.69356107711792,
"learning_rate": 1.813730084348641e-05,
"loss": 0.1869,
"step": 6360
},
{
"epoch": 2.985004686035614,
"grad_norm": 2.3386266231536865,
"learning_rate": 1.8134372071227743e-05,
"loss": 0.1752,
"step": 6370
},
{
"epoch": 2.9896907216494846,
"grad_norm": 1.873734712600708,
"learning_rate": 1.8131443298969072e-05,
"loss": 0.1972,
"step": 6380
},
{
"epoch": 2.994376757263355,
"grad_norm": 2.3732969760894775,
"learning_rate": 1.8128514526710402e-05,
"loss": 0.2229,
"step": 6390
},
{
"epoch": 2.999062792877226,
"grad_norm": 2.2805166244506836,
"learning_rate": 1.8125585754451735e-05,
"loss": 0.2022,
"step": 6400
},
{
"epoch": 3.0037488284910965,
"grad_norm": 1.8053455352783203,
"learning_rate": 1.8122656982193064e-05,
"loss": 0.132,
"step": 6410
},
{
"epoch": 3.0084348641049674,
"grad_norm": 2.322193145751953,
"learning_rate": 1.8119728209934397e-05,
"loss": 0.1341,
"step": 6420
},
{
"epoch": 3.013120899718838,
"grad_norm": 2.2462315559387207,
"learning_rate": 1.8116799437675727e-05,
"loss": 0.1294,
"step": 6430
},
{
"epoch": 3.0178069353327084,
"grad_norm": 1.5314648151397705,
"learning_rate": 1.811387066541706e-05,
"loss": 0.0974,
"step": 6440
},
{
"epoch": 3.0224929709465793,
"grad_norm": 1.4587002992630005,
"learning_rate": 1.811094189315839e-05,
"loss": 0.1019,
"step": 6450
},
{
"epoch": 3.02717900656045,
"grad_norm": 0.887192964553833,
"learning_rate": 1.8108013120899722e-05,
"loss": 0.0945,
"step": 6460
},
{
"epoch": 3.0318650421743207,
"grad_norm": 1.8087823390960693,
"learning_rate": 1.810508434864105e-05,
"loss": 0.1103,
"step": 6470
},
{
"epoch": 3.036551077788191,
"grad_norm": 2.4967281818389893,
"learning_rate": 1.8102155576382384e-05,
"loss": 0.1261,
"step": 6480
},
{
"epoch": 3.0412371134020617,
"grad_norm": 1.4955415725708008,
"learning_rate": 1.8099226804123714e-05,
"loss": 0.1255,
"step": 6490
},
{
"epoch": 3.0459231490159326,
"grad_norm": 1.0758466720581055,
"learning_rate": 1.8096298031865043e-05,
"loss": 0.105,
"step": 6500
},
{
"epoch": 3.0459231490159326,
"eval_loss": 0.04503355920314789,
"eval_pearson_cosine": 0.7887213700331159,
"eval_pearson_dot": 0.638523586851079,
"eval_pearson_euclidean": 0.7527277991826509,
"eval_pearson_manhattan": 0.7516127412435569,
"eval_runtime": 3.2262,
"eval_samples_per_second": 464.937,
"eval_spearman_cosine": 0.7931480987271526,
"eval_spearman_dot": 0.6449590450062397,
"eval_spearman_euclidean": 0.7674534939783395,
"eval_spearman_manhattan": 0.7668609796988295,
"eval_steps_per_second": 29.136,
"step": 6500
},
{
"epoch": 3.050609184629803,
"grad_norm": 2.3318862915039062,
"learning_rate": 1.8093369259606376e-05,
"loss": 0.1364,
"step": 6510
},
{
"epoch": 3.055295220243674,
"grad_norm": 1.8145966529846191,
"learning_rate": 1.8090440487347706e-05,
"loss": 0.1236,
"step": 6520
},
{
"epoch": 3.0599812558575445,
"grad_norm": 1.584845781326294,
"learning_rate": 1.8087511715089035e-05,
"loss": 0.128,
"step": 6530
},
{
"epoch": 3.064667291471415,
"grad_norm": 1.2164894342422485,
"learning_rate": 1.8084582942830368e-05,
"loss": 0.1233,
"step": 6540
},
{
"epoch": 3.069353327085286,
"grad_norm": 1.761679768562317,
"learning_rate": 1.8081654170571698e-05,
"loss": 0.1107,
"step": 6550
},
{
"epoch": 3.0740393626991565,
"grad_norm": 1.9818497896194458,
"learning_rate": 1.8078725398313027e-05,
"loss": 0.1373,
"step": 6560
},
{
"epoch": 3.0787253983130274,
"grad_norm": 2.124472141265869,
"learning_rate": 1.807579662605436e-05,
"loss": 0.1409,
"step": 6570
},
{
"epoch": 3.083411433926898,
"grad_norm": 1.6288917064666748,
"learning_rate": 1.807286785379569e-05,
"loss": 0.1022,
"step": 6580
},
{
"epoch": 3.0880974695407684,
"grad_norm": 1.5154041051864624,
"learning_rate": 1.806993908153702e-05,
"loss": 0.1167,
"step": 6590
},
{
"epoch": 3.0927835051546393,
"grad_norm": 1.736624836921692,
"learning_rate": 1.8067010309278352e-05,
"loss": 0.0932,
"step": 6600
},
{
"epoch": 3.0974695407685098,
"grad_norm": 2.2292051315307617,
"learning_rate": 1.806408153701968e-05,
"loss": 0.1458,
"step": 6610
},
{
"epoch": 3.1021555763823807,
"grad_norm": 1.8836901187896729,
"learning_rate": 1.8061152764761014e-05,
"loss": 0.1145,
"step": 6620
},
{
"epoch": 3.106841611996251,
"grad_norm": 2.0602309703826904,
"learning_rate": 1.8058223992502344e-05,
"loss": 0.1091,
"step": 6630
},
{
"epoch": 3.1115276476101217,
"grad_norm": 1.7996163368225098,
"learning_rate": 1.8055295220243677e-05,
"loss": 0.1155,
"step": 6640
},
{
"epoch": 3.1162136832239926,
"grad_norm": 1.695610761642456,
"learning_rate": 1.8052366447985006e-05,
"loss": 0.1268,
"step": 6650
},
{
"epoch": 3.120899718837863,
"grad_norm": 1.6327197551727295,
"learning_rate": 1.804943767572634e-05,
"loss": 0.1157,
"step": 6660
},
{
"epoch": 3.125585754451734,
"grad_norm": 2.457167387008667,
"learning_rate": 1.804650890346767e-05,
"loss": 0.1348,
"step": 6670
},
{
"epoch": 3.1302717900656045,
"grad_norm": 1.1226837635040283,
"learning_rate": 1.8043580131208998e-05,
"loss": 0.1112,
"step": 6680
},
{
"epoch": 3.134957825679475,
"grad_norm": 1.904137134552002,
"learning_rate": 1.804065135895033e-05,
"loss": 0.1288,
"step": 6690
},
{
"epoch": 3.139643861293346,
"grad_norm": 1.8711892366409302,
"learning_rate": 1.803772258669166e-05,
"loss": 0.1341,
"step": 6700
},
{
"epoch": 3.1443298969072164,
"grad_norm": 1.8008291721343994,
"learning_rate": 1.803479381443299e-05,
"loss": 0.1198,
"step": 6710
},
{
"epoch": 3.1490159325210874,
"grad_norm": 1.8362482786178589,
"learning_rate": 1.8031865042174323e-05,
"loss": 0.1479,
"step": 6720
},
{
"epoch": 3.153701968134958,
"grad_norm": 1.6150809526443481,
"learning_rate": 1.8028936269915652e-05,
"loss": 0.1138,
"step": 6730
},
{
"epoch": 3.1583880037488283,
"grad_norm": 2.2239766120910645,
"learning_rate": 1.8026007497656982e-05,
"loss": 0.1309,
"step": 6740
},
{
"epoch": 3.1630740393626993,
"grad_norm": 1.669233798980713,
"learning_rate": 1.8023078725398315e-05,
"loss": 0.1055,
"step": 6750
},
{
"epoch": 3.1630740393626993,
"eval_loss": 0.04599650204181671,
"eval_pearson_cosine": 0.7875010466521104,
"eval_pearson_dot": 0.6256249285068609,
"eval_pearson_euclidean": 0.752459338375715,
"eval_pearson_manhattan": 0.7515050600000066,
"eval_runtime": 3.3759,
"eval_samples_per_second": 444.332,
"eval_spearman_cosine": 0.792665874705252,
"eval_spearman_dot": 0.6331527343675145,
"eval_spearman_euclidean": 0.7657151594858318,
"eval_spearman_manhattan": 0.7651890994431135,
"eval_steps_per_second": 27.845,
"step": 6750
},
{
"epoch": 3.1677600749765698,
"grad_norm": 2.1133861541748047,
"learning_rate": 1.8020149953139644e-05,
"loss": 0.1128,
"step": 6760
},
{
"epoch": 3.1724461105904407,
"grad_norm": 1.5470513105392456,
"learning_rate": 1.8017221180880977e-05,
"loss": 0.1246,
"step": 6770
},
{
"epoch": 3.177132146204311,
"grad_norm": 1.9087189435958862,
"learning_rate": 1.8014292408622307e-05,
"loss": 0.1137,
"step": 6780
},
{
"epoch": 3.1818181818181817,
"grad_norm": 1.9185723066329956,
"learning_rate": 1.8011363636363636e-05,
"loss": 0.1196,
"step": 6790
},
{
"epoch": 3.1865042174320526,
"grad_norm": 1.7757681608200073,
"learning_rate": 1.800843486410497e-05,
"loss": 0.0978,
"step": 6800
},
{
"epoch": 3.191190253045923,
"grad_norm": 1.6667931079864502,
"learning_rate": 1.80055060918463e-05,
"loss": 0.113,
"step": 6810
},
{
"epoch": 3.195876288659794,
"grad_norm": 2.101773500442505,
"learning_rate": 1.800257731958763e-05,
"loss": 0.1034,
"step": 6820
},
{
"epoch": 3.2005623242736645,
"grad_norm": 1.6078181266784668,
"learning_rate": 1.799964854732896e-05,
"loss": 0.1276,
"step": 6830
},
{
"epoch": 3.205248359887535,
"grad_norm": 1.3552807569503784,
"learning_rate": 1.7996719775070294e-05,
"loss": 0.1184,
"step": 6840
},
{
"epoch": 3.209934395501406,
"grad_norm": 1.5063031911849976,
"learning_rate": 1.7993791002811623e-05,
"loss": 0.1367,
"step": 6850
},
{
"epoch": 3.2146204311152764,
"grad_norm": 1.9200366735458374,
"learning_rate": 1.7990862230552956e-05,
"loss": 0.1222,
"step": 6860
},
{
"epoch": 3.2193064667291473,
"grad_norm": 1.4902859926223755,
"learning_rate": 1.7987933458294286e-05,
"loss": 0.1414,
"step": 6870
},
{
"epoch": 3.223992502343018,
"grad_norm": 1.7316768169403076,
"learning_rate": 1.7985004686035615e-05,
"loss": 0.1451,
"step": 6880
},
{
"epoch": 3.2286785379568883,
"grad_norm": 1.8446965217590332,
"learning_rate": 1.7982075913776945e-05,
"loss": 0.1227,
"step": 6890
},
{
"epoch": 3.2333645735707592,
"grad_norm": 1.5571191310882568,
"learning_rate": 1.7979147141518278e-05,
"loss": 0.1193,
"step": 6900
},
{
"epoch": 3.2380506091846297,
"grad_norm": 2.3433127403259277,
"learning_rate": 1.7976218369259607e-05,
"loss": 0.1175,
"step": 6910
},
{
"epoch": 3.2427366447985007,
"grad_norm": 1.8916155099868774,
"learning_rate": 1.7973289597000937e-05,
"loss": 0.1352,
"step": 6920
},
{
"epoch": 3.247422680412371,
"grad_norm": 1.3708059787750244,
"learning_rate": 1.797036082474227e-05,
"loss": 0.1174,
"step": 6930
},
{
"epoch": 3.2521087160262416,
"grad_norm": 1.5119550228118896,
"learning_rate": 1.79674320524836e-05,
"loss": 0.1331,
"step": 6940
},
{
"epoch": 3.2567947516401126,
"grad_norm": 1.8884799480438232,
"learning_rate": 1.7964503280224932e-05,
"loss": 0.125,
"step": 6950
},
{
"epoch": 3.261480787253983,
"grad_norm": 1.6222647428512573,
"learning_rate": 1.796157450796626e-05,
"loss": 0.1191,
"step": 6960
},
{
"epoch": 3.266166822867854,
"grad_norm": 1.2746548652648926,
"learning_rate": 1.7958645735707594e-05,
"loss": 0.1256,
"step": 6970
},
{
"epoch": 3.2708528584817245,
"grad_norm": 1.883239984512329,
"learning_rate": 1.7955716963448924e-05,
"loss": 0.1283,
"step": 6980
},
{
"epoch": 3.275538894095595,
"grad_norm": 2.0994622707366943,
"learning_rate": 1.7952788191190253e-05,
"loss": 0.1316,
"step": 6990
},
{
"epoch": 3.280224929709466,
"grad_norm": 1.844297170639038,
"learning_rate": 1.7949859418931586e-05,
"loss": 0.1145,
"step": 7000
},
{
"epoch": 3.280224929709466,
"eval_loss": 0.045286137610673904,
"eval_pearson_cosine": 0.7924534055849843,
"eval_pearson_dot": 0.6315508963322145,
"eval_pearson_euclidean": 0.7559423098506386,
"eval_pearson_manhattan": 0.7547566595358202,
"eval_runtime": 3.1537,
"eval_samples_per_second": 475.63,
"eval_spearman_cosine": 0.7977173793324885,
"eval_spearman_dot": 0.6408281685024612,
"eval_spearman_euclidean": 0.7678066060450297,
"eval_spearman_manhattan": 0.7671464837523794,
"eval_steps_per_second": 29.806,
"step": 7000
},
{
"epoch": 3.2849109653233364,
"grad_norm": 1.3316779136657715,
"learning_rate": 1.7946930646672916e-05,
"loss": 0.1139,
"step": 7010
},
{
"epoch": 3.2895970009372073,
"grad_norm": 1.3782374858856201,
"learning_rate": 1.794400187441425e-05,
"loss": 0.1238,
"step": 7020
},
{
"epoch": 3.294283036551078,
"grad_norm": 1.4582282304763794,
"learning_rate": 1.7941073102155578e-05,
"loss": 0.1279,
"step": 7030
},
{
"epoch": 3.2989690721649483,
"grad_norm": 2.3581652641296387,
"learning_rate": 1.793814432989691e-05,
"loss": 0.118,
"step": 7040
},
{
"epoch": 3.303655107778819,
"grad_norm": 2.1047656536102295,
"learning_rate": 1.793521555763824e-05,
"loss": 0.1086,
"step": 7050
},
{
"epoch": 3.3083411433926897,
"grad_norm": 1.748353123664856,
"learning_rate": 1.793228678537957e-05,
"loss": 0.1357,
"step": 7060
},
{
"epoch": 3.3130271790065606,
"grad_norm": 1.905609130859375,
"learning_rate": 1.79293580131209e-05,
"loss": 0.1239,
"step": 7070
},
{
"epoch": 3.317713214620431,
"grad_norm": 1.5777533054351807,
"learning_rate": 1.7926429240862232e-05,
"loss": 0.1179,
"step": 7080
},
{
"epoch": 3.3223992502343016,
"grad_norm": 1.8429454565048218,
"learning_rate": 1.7923500468603562e-05,
"loss": 0.1103,
"step": 7090
},
{
"epoch": 3.3270852858481725,
"grad_norm": 1.601769208908081,
"learning_rate": 1.792057169634489e-05,
"loss": 0.1282,
"step": 7100
},
{
"epoch": 3.331771321462043,
"grad_norm": 1.948339581489563,
"learning_rate": 1.7917642924086224e-05,
"loss": 0.1314,
"step": 7110
},
{
"epoch": 3.336457357075914,
"grad_norm": 2.035210609436035,
"learning_rate": 1.7914714151827554e-05,
"loss": 0.1232,
"step": 7120
},
{
"epoch": 3.3411433926897844,
"grad_norm": 1.3871088027954102,
"learning_rate": 1.7911785379568887e-05,
"loss": 0.1225,
"step": 7130
},
{
"epoch": 3.345829428303655,
"grad_norm": 1.9721599817276,
"learning_rate": 1.7908856607310216e-05,
"loss": 0.1364,
"step": 7140
},
{
"epoch": 3.350515463917526,
"grad_norm": 2.2410221099853516,
"learning_rate": 1.790592783505155e-05,
"loss": 0.1277,
"step": 7150
},
{
"epoch": 3.3552014995313963,
"grad_norm": 1.8568767309188843,
"learning_rate": 1.790299906279288e-05,
"loss": 0.1114,
"step": 7160
},
{
"epoch": 3.3598875351452673,
"grad_norm": 2.367185115814209,
"learning_rate": 1.790007029053421e-05,
"loss": 0.1334,
"step": 7170
},
{
"epoch": 3.3645735707591378,
"grad_norm": 1.6528865098953247,
"learning_rate": 1.789714151827554e-05,
"loss": 0.1223,
"step": 7180
},
{
"epoch": 3.3692596063730083,
"grad_norm": 1.9136816263198853,
"learning_rate": 1.789421274601687e-05,
"loss": 0.1066,
"step": 7190
},
{
"epoch": 3.373945641986879,
"grad_norm": 2.2870426177978516,
"learning_rate": 1.7891283973758203e-05,
"loss": 0.1058,
"step": 7200
},
{
"epoch": 3.3786316776007497,
"grad_norm": 1.5177735090255737,
"learning_rate": 1.7888355201499533e-05,
"loss": 0.1179,
"step": 7210
},
{
"epoch": 3.3833177132146206,
"grad_norm": 2.215899705886841,
"learning_rate": 1.7885426429240866e-05,
"loss": 0.1321,
"step": 7220
},
{
"epoch": 3.388003748828491,
"grad_norm": 1.7164250612258911,
"learning_rate": 1.7882497656982195e-05,
"loss": 0.1111,
"step": 7230
},
{
"epoch": 3.3926897844423616,
"grad_norm": 2.3435146808624268,
"learning_rate": 1.7879568884723525e-05,
"loss": 0.1513,
"step": 7240
},
{
"epoch": 3.3973758200562325,
"grad_norm": 1.6385475397109985,
"learning_rate": 1.7876640112464858e-05,
"loss": 0.1252,
"step": 7250
},
{
"epoch": 3.3973758200562325,
"eval_loss": 0.047000955790281296,
"eval_pearson_cosine": 0.7888804465638501,
"eval_pearson_dot": 0.6256929197655836,
"eval_pearson_euclidean": 0.7570957556748965,
"eval_pearson_manhattan": 0.7561323835241964,
"eval_runtime": 3.4083,
"eval_samples_per_second": 440.106,
"eval_spearman_cosine": 0.7947042550328448,
"eval_spearman_dot": 0.6282838912238756,
"eval_spearman_euclidean": 0.7692784528325559,
"eval_spearman_manhattan": 0.7683456146298958,
"eval_steps_per_second": 27.58,
"step": 7250
},
{
"epoch": 3.402061855670103,
"grad_norm": 1.5348442792892456,
"learning_rate": 1.7873711340206187e-05,
"loss": 0.1304,
"step": 7260
},
{
"epoch": 3.406747891283974,
"grad_norm": 1.854861855506897,
"learning_rate": 1.7870782567947517e-05,
"loss": 0.1296,
"step": 7270
},
{
"epoch": 3.4114339268978444,
"grad_norm": 1.5103728771209717,
"learning_rate": 1.7867853795688846e-05,
"loss": 0.1182,
"step": 7280
},
{
"epoch": 3.416119962511715,
"grad_norm": 1.4604400396347046,
"learning_rate": 1.786492502343018e-05,
"loss": 0.1155,
"step": 7290
},
{
"epoch": 3.420805998125586,
"grad_norm": 1.9087272882461548,
"learning_rate": 1.786199625117151e-05,
"loss": 0.1265,
"step": 7300
},
{
"epoch": 3.4254920337394563,
"grad_norm": 2.3286941051483154,
"learning_rate": 1.785906747891284e-05,
"loss": 0.0926,
"step": 7310
},
{
"epoch": 3.4301780693533273,
"grad_norm": 1.8551883697509766,
"learning_rate": 1.785613870665417e-05,
"loss": 0.122,
"step": 7320
},
{
"epoch": 3.4348641049671977,
"grad_norm": 1.7142447233200073,
"learning_rate": 1.7853209934395504e-05,
"loss": 0.138,
"step": 7330
},
{
"epoch": 3.4395501405810682,
"grad_norm": 1.5548756122589111,
"learning_rate": 1.7850281162136833e-05,
"loss": 0.1123,
"step": 7340
},
{
"epoch": 3.444236176194939,
"grad_norm": 1.3273271322250366,
"learning_rate": 1.7847352389878166e-05,
"loss": 0.129,
"step": 7350
},
{
"epoch": 3.4489222118088096,
"grad_norm": 2.410999298095703,
"learning_rate": 1.7844423617619496e-05,
"loss": 0.1329,
"step": 7360
},
{
"epoch": 3.4536082474226806,
"grad_norm": 2.0272533893585205,
"learning_rate": 1.784149484536083e-05,
"loss": 0.1251,
"step": 7370
},
{
"epoch": 3.458294283036551,
"grad_norm": 2.4070749282836914,
"learning_rate": 1.7838566073102158e-05,
"loss": 0.1384,
"step": 7380
},
{
"epoch": 3.4629803186504216,
"grad_norm": 2.0983331203460693,
"learning_rate": 1.7835637300843487e-05,
"loss": 0.1268,
"step": 7390
},
{
"epoch": 3.4676663542642925,
"grad_norm": 1.5650460720062256,
"learning_rate": 1.783270852858482e-05,
"loss": 0.1218,
"step": 7400
},
{
"epoch": 3.472352389878163,
"grad_norm": 1.4505574703216553,
"learning_rate": 1.782977975632615e-05,
"loss": 0.1278,
"step": 7410
},
{
"epoch": 3.477038425492034,
"grad_norm": 2.4104747772216797,
"learning_rate": 1.782685098406748e-05,
"loss": 0.1506,
"step": 7420
},
{
"epoch": 3.4817244611059044,
"grad_norm": 2.2749085426330566,
"learning_rate": 1.7823922211808812e-05,
"loss": 0.1382,
"step": 7430
},
{
"epoch": 3.486410496719775,
"grad_norm": 1.619983434677124,
"learning_rate": 1.7820993439550142e-05,
"loss": 0.1145,
"step": 7440
},
{
"epoch": 3.491096532333646,
"grad_norm": 1.797600269317627,
"learning_rate": 1.781806466729147e-05,
"loss": 0.1381,
"step": 7450
},
{
"epoch": 3.4957825679475163,
"grad_norm": 1.2865104675292969,
"learning_rate": 1.7815135895032804e-05,
"loss": 0.1088,
"step": 7460
},
{
"epoch": 3.5004686035613872,
"grad_norm": 2.2955312728881836,
"learning_rate": 1.7812207122774134e-05,
"loss": 0.1389,
"step": 7470
},
{
"epoch": 3.5051546391752577,
"grad_norm": 1.884758472442627,
"learning_rate": 1.7809278350515463e-05,
"loss": 0.1399,
"step": 7480
},
{
"epoch": 3.509840674789128,
"grad_norm": 2.0076119899749756,
"learning_rate": 1.7806349578256796e-05,
"loss": 0.1028,
"step": 7490
},
{
"epoch": 3.514526710402999,
"grad_norm": 1.5606528520584106,
"learning_rate": 1.7803420805998125e-05,
"loss": 0.1058,
"step": 7500
},
{
"epoch": 3.514526710402999,
"eval_loss": 0.044608332216739655,
"eval_pearson_cosine": 0.7913143880230251,
"eval_pearson_dot": 0.6221286850907433,
"eval_pearson_euclidean": 0.7578325537672654,
"eval_pearson_manhattan": 0.7571791901529252,
"eval_runtime": 3.3012,
"eval_samples_per_second": 454.374,
"eval_spearman_cosine": 0.7958246101441605,
"eval_spearman_dot": 0.6338421877202784,
"eval_spearman_euclidean": 0.7715109007041803,
"eval_spearman_manhattan": 0.7714185275105551,
"eval_steps_per_second": 28.474,
"step": 7500
},
{
"epoch": 3.5192127460168696,
"grad_norm": 1.9054310321807861,
"learning_rate": 1.780049203373946e-05,
"loss": 0.1149,
"step": 7510
},
{
"epoch": 3.5238987816307406,
"grad_norm": 2.2473042011260986,
"learning_rate": 1.7797563261480788e-05,
"loss": 0.1241,
"step": 7520
},
{
"epoch": 3.528584817244611,
"grad_norm": 2.9301676750183105,
"learning_rate": 1.779463448922212e-05,
"loss": 0.1155,
"step": 7530
},
{
"epoch": 3.5332708528584815,
"grad_norm": 1.8250157833099365,
"learning_rate": 1.779170571696345e-05,
"loss": 0.118,
"step": 7540
},
{
"epoch": 3.5379568884723525,
"grad_norm": 2.101794958114624,
"learning_rate": 1.7788776944704783e-05,
"loss": 0.1338,
"step": 7550
},
{
"epoch": 3.542642924086223,
"grad_norm": 1.9058297872543335,
"learning_rate": 1.7785848172446113e-05,
"loss": 0.1172,
"step": 7560
},
{
"epoch": 3.547328959700094,
"grad_norm": 2.1106228828430176,
"learning_rate": 1.7782919400187446e-05,
"loss": 0.119,
"step": 7570
},
{
"epoch": 3.5520149953139644,
"grad_norm": 1.572771430015564,
"learning_rate": 1.7779990627928775e-05,
"loss": 0.1303,
"step": 7580
},
{
"epoch": 3.556701030927835,
"grad_norm": 2.670691967010498,
"learning_rate": 1.7777061855670105e-05,
"loss": 0.1367,
"step": 7590
},
{
"epoch": 3.561387066541706,
"grad_norm": 1.0899403095245361,
"learning_rate": 1.7774133083411434e-05,
"loss": 0.131,
"step": 7600
},
{
"epoch": 3.5660731021555763,
"grad_norm": 1.66856050491333,
"learning_rate": 1.7771204311152767e-05,
"loss": 0.1282,
"step": 7610
},
{
"epoch": 3.570759137769447,
"grad_norm": 1.9693597555160522,
"learning_rate": 1.7768275538894096e-05,
"loss": 0.1141,
"step": 7620
},
{
"epoch": 3.5754451733833177,
"grad_norm": 2.5107643604278564,
"learning_rate": 1.7765346766635426e-05,
"loss": 0.1302,
"step": 7630
},
{
"epoch": 3.580131208997188,
"grad_norm": 1.843782663345337,
"learning_rate": 1.776241799437676e-05,
"loss": 0.1295,
"step": 7640
},
{
"epoch": 3.584817244611059,
"grad_norm": 1.5396898984909058,
"learning_rate": 1.7759489222118088e-05,
"loss": 0.1146,
"step": 7650
},
{
"epoch": 3.5895032802249296,
"grad_norm": 1.250181794166565,
"learning_rate": 1.775656044985942e-05,
"loss": 0.0997,
"step": 7660
},
{
"epoch": 3.5941893158388005,
"grad_norm": 1.0038021802902222,
"learning_rate": 1.775363167760075e-05,
"loss": 0.131,
"step": 7670
},
{
"epoch": 3.598875351452671,
"grad_norm": 1.7689967155456543,
"learning_rate": 1.775070290534208e-05,
"loss": 0.1323,
"step": 7680
},
{
"epoch": 3.6035613870665415,
"grad_norm": 1.6515671014785767,
"learning_rate": 1.7747774133083413e-05,
"loss": 0.1228,
"step": 7690
},
{
"epoch": 3.6082474226804124,
"grad_norm": 1.6234405040740967,
"learning_rate": 1.7744845360824743e-05,
"loss": 0.1177,
"step": 7700
},
{
"epoch": 3.612933458294283,
"grad_norm": 1.1468828916549683,
"learning_rate": 1.7741916588566075e-05,
"loss": 0.121,
"step": 7710
},
{
"epoch": 3.617619493908154,
"grad_norm": 2.2512290477752686,
"learning_rate": 1.7738987816307405e-05,
"loss": 0.1506,
"step": 7720
},
{
"epoch": 3.6223055295220243,
"grad_norm": 1.967869758605957,
"learning_rate": 1.7736059044048738e-05,
"loss": 0.1278,
"step": 7730
},
{
"epoch": 3.626991565135895,
"grad_norm": 1.6843619346618652,
"learning_rate": 1.7733130271790067e-05,
"loss": 0.1292,
"step": 7740
},
{
"epoch": 3.6316776007497658,
"grad_norm": 1.7128106355667114,
"learning_rate": 1.77302014995314e-05,
"loss": 0.1144,
"step": 7750
},
{
"epoch": 3.6316776007497658,
"eval_loss": 0.043338797986507416,
"eval_pearson_cosine": 0.7938926107691149,
"eval_pearson_dot": 0.6518687502689318,
"eval_pearson_euclidean": 0.7541829192023144,
"eval_pearson_manhattan": 0.7533667893162033,
"eval_runtime": 3.5292,
"eval_samples_per_second": 425.028,
"eval_spearman_cosine": 0.7989181032598132,
"eval_spearman_dot": 0.658312239789335,
"eval_spearman_euclidean": 0.7677205029787961,
"eval_spearman_manhattan": 0.7673451483642694,
"eval_steps_per_second": 26.635,
"step": 7750
},
{
"epoch": 3.6363636363636362,
"grad_norm": 2.2347500324249268,
"learning_rate": 1.772727272727273e-05,
"loss": 0.145,
"step": 7760
},
{
"epoch": 3.641049671977507,
"grad_norm": 1.1075607538223267,
"learning_rate": 1.772434395501406e-05,
"loss": 0.122,
"step": 7770
},
{
"epoch": 3.6457357075913777,
"grad_norm": 1.9819576740264893,
"learning_rate": 1.772141518275539e-05,
"loss": 0.1542,
"step": 7780
},
{
"epoch": 3.650421743205248,
"grad_norm": 1.3385450839996338,
"learning_rate": 1.771848641049672e-05,
"loss": 0.1269,
"step": 7790
},
{
"epoch": 3.655107778819119,
"grad_norm": 1.2146624326705933,
"learning_rate": 1.771555763823805e-05,
"loss": 0.1124,
"step": 7800
},
{
"epoch": 3.6597938144329896,
"grad_norm": 1.8543137311935425,
"learning_rate": 1.771262886597938e-05,
"loss": 0.1513,
"step": 7810
},
{
"epoch": 3.6644798500468605,
"grad_norm": 1.147740125656128,
"learning_rate": 1.7709700093720713e-05,
"loss": 0.1207,
"step": 7820
},
{
"epoch": 3.669165885660731,
"grad_norm": 2.1737101078033447,
"learning_rate": 1.7706771321462043e-05,
"loss": 0.1345,
"step": 7830
},
{
"epoch": 3.6738519212746015,
"grad_norm": 2.481095790863037,
"learning_rate": 1.7703842549203376e-05,
"loss": 0.1304,
"step": 7840
},
{
"epoch": 3.6785379568884724,
"grad_norm": 1.9203423261642456,
"learning_rate": 1.7700913776944705e-05,
"loss": 0.1257,
"step": 7850
},
{
"epoch": 3.683223992502343,
"grad_norm": 2.324158191680908,
"learning_rate": 1.7697985004686038e-05,
"loss": 0.126,
"step": 7860
},
{
"epoch": 3.687910028116214,
"grad_norm": 1.6233323812484741,
"learning_rate": 1.7695056232427368e-05,
"loss": 0.1228,
"step": 7870
},
{
"epoch": 3.6925960637300843,
"grad_norm": 1.0567041635513306,
"learning_rate": 1.7692127460168697e-05,
"loss": 0.1129,
"step": 7880
},
{
"epoch": 3.697282099343955,
"grad_norm": 1.4684457778930664,
"learning_rate": 1.768919868791003e-05,
"loss": 0.1164,
"step": 7890
},
{
"epoch": 3.7019681349578257,
"grad_norm": 2.070171594619751,
"learning_rate": 1.768626991565136e-05,
"loss": 0.1488,
"step": 7900
},
{
"epoch": 3.706654170571696,
"grad_norm": 1.93710458278656,
"learning_rate": 1.7683341143392693e-05,
"loss": 0.1423,
"step": 7910
},
{
"epoch": 3.711340206185567,
"grad_norm": 2.109379529953003,
"learning_rate": 1.7680412371134022e-05,
"loss": 0.1247,
"step": 7920
},
{
"epoch": 3.7160262417994376,
"grad_norm": 1.9402395486831665,
"learning_rate": 1.7677483598875355e-05,
"loss": 0.1258,
"step": 7930
},
{
"epoch": 3.720712277413308,
"grad_norm": 1.3770403861999512,
"learning_rate": 1.7674554826616684e-05,
"loss": 0.1152,
"step": 7940
},
{
"epoch": 3.725398313027179,
"grad_norm": 1.8845633268356323,
"learning_rate": 1.7671626054358014e-05,
"loss": 0.1128,
"step": 7950
},
{
"epoch": 3.7300843486410495,
"grad_norm": 1.80978524684906,
"learning_rate": 1.7668697282099347e-05,
"loss": 0.1506,
"step": 7960
},
{
"epoch": 3.7347703842549205,
"grad_norm": 1.714689016342163,
"learning_rate": 1.7665768509840676e-05,
"loss": 0.1295,
"step": 7970
},
{
"epoch": 3.739456419868791,
"grad_norm": 2.2683634757995605,
"learning_rate": 1.7662839737582006e-05,
"loss": 0.153,
"step": 7980
},
{
"epoch": 3.7441424554826614,
"grad_norm": 1.1487226486206055,
"learning_rate": 1.7659910965323335e-05,
"loss": 0.1305,
"step": 7990
},
{
"epoch": 3.7488284910965324,
"grad_norm": 1.4769738912582397,
"learning_rate": 1.7656982193064668e-05,
"loss": 0.0971,
"step": 8000
},
{
"epoch": 3.7488284910965324,
"eval_loss": 0.04381699115037918,
"eval_pearson_cosine": 0.7952038190054296,
"eval_pearson_dot": 0.6344663529067152,
"eval_pearson_euclidean": 0.7547335910463318,
"eval_pearson_manhattan": 0.7536726944037966,
"eval_runtime": 3.6935,
"eval_samples_per_second": 406.123,
"eval_spearman_cosine": 0.7993143516157027,
"eval_spearman_dot": 0.6382938851288388,
"eval_spearman_euclidean": 0.7679077610624258,
"eval_spearman_manhattan": 0.767544325158077,
"eval_steps_per_second": 25.45,
"step": 8000
},
{
"epoch": 3.753514526710403,
"grad_norm": 2.859675645828247,
"learning_rate": 1.7654053420805998e-05,
"loss": 0.1442,
"step": 8010
},
{
"epoch": 3.758200562324274,
"grad_norm": 1.54740571975708,
"learning_rate": 1.765112464854733e-05,
"loss": 0.1194,
"step": 8020
},
{
"epoch": 3.7628865979381443,
"grad_norm": 1.5696630477905273,
"learning_rate": 1.764819587628866e-05,
"loss": 0.1542,
"step": 8030
},
{
"epoch": 3.7675726335520148,
"grad_norm": 1.8833867311477661,
"learning_rate": 1.7645267104029993e-05,
"loss": 0.1222,
"step": 8040
},
{
"epoch": 3.7722586691658857,
"grad_norm": 2.2312562465667725,
"learning_rate": 1.7642338331771322e-05,
"loss": 0.1378,
"step": 8050
},
{
"epoch": 3.776944704779756,
"grad_norm": 1.9470106363296509,
"learning_rate": 1.7639409559512655e-05,
"loss": 0.136,
"step": 8060
},
{
"epoch": 3.781630740393627,
"grad_norm": 1.736140489578247,
"learning_rate": 1.7636480787253985e-05,
"loss": 0.144,
"step": 8070
},
{
"epoch": 3.7863167760074976,
"grad_norm": 2.0280628204345703,
"learning_rate": 1.7633552014995314e-05,
"loss": 0.1331,
"step": 8080
},
{
"epoch": 3.791002811621368,
"grad_norm": 1.1331037282943726,
"learning_rate": 1.7630623242736647e-05,
"loss": 0.1327,
"step": 8090
},
{
"epoch": 3.795688847235239,
"grad_norm": 2.00110125541687,
"learning_rate": 1.7627694470477977e-05,
"loss": 0.1335,
"step": 8100
},
{
"epoch": 3.8003748828491095,
"grad_norm": 1.589747428894043,
"learning_rate": 1.762476569821931e-05,
"loss": 0.1222,
"step": 8110
},
{
"epoch": 3.8050609184629804,
"grad_norm": 1.8979542255401611,
"learning_rate": 1.762183692596064e-05,
"loss": 0.1534,
"step": 8120
},
{
"epoch": 3.809746954076851,
"grad_norm": 2.384608268737793,
"learning_rate": 1.761890815370197e-05,
"loss": 0.1395,
"step": 8130
},
{
"epoch": 3.8144329896907214,
"grad_norm": 1.8555763959884644,
"learning_rate": 1.76159793814433e-05,
"loss": 0.1298,
"step": 8140
},
{
"epoch": 3.8191190253045924,
"grad_norm": 2.0810048580169678,
"learning_rate": 1.761305060918463e-05,
"loss": 0.1165,
"step": 8150
},
{
"epoch": 3.823805060918463,
"grad_norm": 1.9116188287734985,
"learning_rate": 1.761012183692596e-05,
"loss": 0.1316,
"step": 8160
},
{
"epoch": 3.8284910965323338,
"grad_norm": 2.3420250415802,
"learning_rate": 1.7607193064667293e-05,
"loss": 0.1176,
"step": 8170
},
{
"epoch": 3.8331771321462043,
"grad_norm": 2.0588436126708984,
"learning_rate": 1.7604264292408623e-05,
"loss": 0.1365,
"step": 8180
},
{
"epoch": 3.8378631677600747,
"grad_norm": 1.89832603931427,
"learning_rate": 1.7601335520149952e-05,
"loss": 0.1459,
"step": 8190
},
{
"epoch": 3.8425492033739457,
"grad_norm": 2.0133140087127686,
"learning_rate": 1.7598406747891285e-05,
"loss": 0.1075,
"step": 8200
},
{
"epoch": 3.847235238987816,
"grad_norm": 1.7235685586929321,
"learning_rate": 1.7595477975632615e-05,
"loss": 0.1143,
"step": 8210
},
{
"epoch": 3.851921274601687,
"grad_norm": 1.6717524528503418,
"learning_rate": 1.7592549203373948e-05,
"loss": 0.1269,
"step": 8220
},
{
"epoch": 3.8566073102155576,
"grad_norm": 1.5314955711364746,
"learning_rate": 1.7589620431115277e-05,
"loss": 0.1264,
"step": 8230
},
{
"epoch": 3.861293345829428,
"grad_norm": 2.4005510807037354,
"learning_rate": 1.758669165885661e-05,
"loss": 0.1299,
"step": 8240
},
{
"epoch": 3.865979381443299,
"grad_norm": 1.6033201217651367,
"learning_rate": 1.758376288659794e-05,
"loss": 0.1107,
"step": 8250
},
{
"epoch": 3.865979381443299,
"eval_loss": 0.04320518299937248,
"eval_pearson_cosine": 0.7953389681810563,
"eval_pearson_dot": 0.6355398535557981,
"eval_pearson_euclidean": 0.7518228241740452,
"eval_pearson_manhattan": 0.7507054666151873,
"eval_runtime": 3.6594,
"eval_samples_per_second": 409.902,
"eval_spearman_cosine": 0.799221389686279,
"eval_spearman_dot": 0.6411224962967852,
"eval_spearman_euclidean": 0.7674570737836293,
"eval_spearman_manhattan": 0.7673149114040088,
"eval_steps_per_second": 25.687,
"step": 8250
},
{
"epoch": 3.8706654170571695,
"grad_norm": 1.8069274425506592,
"learning_rate": 1.7580834114339272e-05,
"loss": 0.1293,
"step": 8260
},
{
"epoch": 3.8753514526710404,
"grad_norm": 1.9214448928833008,
"learning_rate": 1.7577905342080602e-05,
"loss": 0.1235,
"step": 8270
},
{
"epoch": 3.880037488284911,
"grad_norm": 2.0332281589508057,
"learning_rate": 1.757497656982193e-05,
"loss": 0.1303,
"step": 8280
},
{
"epoch": 3.8847235238987814,
"grad_norm": 1.8587048053741455,
"learning_rate": 1.7572047797563264e-05,
"loss": 0.1272,
"step": 8290
},
{
"epoch": 3.8894095595126523,
"grad_norm": 1.7040314674377441,
"learning_rate": 1.7569119025304594e-05,
"loss": 0.1304,
"step": 8300
},
{
"epoch": 3.894095595126523,
"grad_norm": 1.7979313135147095,
"learning_rate": 1.7566190253045923e-05,
"loss": 0.1226,
"step": 8310
},
{
"epoch": 3.8987816307403937,
"grad_norm": 1.6295264959335327,
"learning_rate": 1.7563261480787256e-05,
"loss": 0.1198,
"step": 8320
},
{
"epoch": 3.9034676663542642,
"grad_norm": 1.764858603477478,
"learning_rate": 1.7560332708528586e-05,
"loss": 0.1294,
"step": 8330
},
{
"epoch": 3.9081537019681347,
"grad_norm": 2.3997533321380615,
"learning_rate": 1.7557403936269915e-05,
"loss": 0.1288,
"step": 8340
},
{
"epoch": 3.9128397375820057,
"grad_norm": 2.302992820739746,
"learning_rate": 1.7554475164011248e-05,
"loss": 0.1482,
"step": 8350
},
{
"epoch": 3.917525773195876,
"grad_norm": 1.8705153465270996,
"learning_rate": 1.7551546391752578e-05,
"loss": 0.1203,
"step": 8360
},
{
"epoch": 3.922211808809747,
"grad_norm": 2.1490349769592285,
"learning_rate": 1.754861761949391e-05,
"loss": 0.1074,
"step": 8370
},
{
"epoch": 3.9268978444236176,
"grad_norm": 1.5012431144714355,
"learning_rate": 1.754568884723524e-05,
"loss": 0.1202,
"step": 8380
},
{
"epoch": 3.931583880037488,
"grad_norm": 1.2775022983551025,
"learning_rate": 1.754276007497657e-05,
"loss": 0.1241,
"step": 8390
},
{
"epoch": 3.936269915651359,
"grad_norm": 2.361064910888672,
"learning_rate": 1.7539831302717902e-05,
"loss": 0.1393,
"step": 8400
},
{
"epoch": 3.9409559512652295,
"grad_norm": 1.8726896047592163,
"learning_rate": 1.7536902530459232e-05,
"loss": 0.1191,
"step": 8410
},
{
"epoch": 3.9456419868791004,
"grad_norm": 1.8277250528335571,
"learning_rate": 1.7533973758200565e-05,
"loss": 0.1265,
"step": 8420
},
{
"epoch": 3.950328022492971,
"grad_norm": 1.534006118774414,
"learning_rate": 1.7531044985941894e-05,
"loss": 0.1153,
"step": 8430
},
{
"epoch": 3.9550140581068414,
"grad_norm": 1.9085185527801514,
"learning_rate": 1.7528116213683227e-05,
"loss": 0.1216,
"step": 8440
},
{
"epoch": 3.9597000937207123,
"grad_norm": 2.059544563293457,
"learning_rate": 1.7525187441424557e-05,
"loss": 0.1185,
"step": 8450
},
{
"epoch": 3.964386129334583,
"grad_norm": 2.6308438777923584,
"learning_rate": 1.752225866916589e-05,
"loss": 0.1137,
"step": 8460
},
{
"epoch": 3.9690721649484537,
"grad_norm": 1.2617682218551636,
"learning_rate": 1.751932989690722e-05,
"loss": 0.143,
"step": 8470
},
{
"epoch": 3.973758200562324,
"grad_norm": 2.1921794414520264,
"learning_rate": 1.751640112464855e-05,
"loss": 0.1086,
"step": 8480
},
{
"epoch": 3.9784442361761947,
"grad_norm": 1.7036564350128174,
"learning_rate": 1.751347235238988e-05,
"loss": 0.127,
"step": 8490
},
{
"epoch": 3.9831302717900656,
"grad_norm": 1.6190659999847412,
"learning_rate": 1.751054358013121e-05,
"loss": 0.1232,
"step": 8500
},
{
"epoch": 3.9831302717900656,
"eval_loss": 0.04382430762052536,
"eval_pearson_cosine": 0.7946660833476784,
"eval_pearson_dot": 0.6408322008008298,
"eval_pearson_euclidean": 0.7506581872281402,
"eval_pearson_manhattan": 0.7492952268314212,
"eval_runtime": 3.2595,
"eval_samples_per_second": 460.2,
"eval_spearman_cosine": 0.7991836067006112,
"eval_spearman_dot": 0.6471776019534929,
"eval_spearman_euclidean": 0.7665561093361127,
"eval_spearman_manhattan": 0.7655283844207366,
"eval_steps_per_second": 28.839,
"step": 8500
},
{
"epoch": 3.987816307403936,
"grad_norm": 1.5553314685821533,
"learning_rate": 1.750761480787254e-05,
"loss": 0.1199,
"step": 8510
},
{
"epoch": 3.992502343017807,
"grad_norm": 1.8198939561843872,
"learning_rate": 1.750468603561387e-05,
"loss": 0.1387,
"step": 8520
},
{
"epoch": 3.9971883786316775,
"grad_norm": 1.5737895965576172,
"learning_rate": 1.7501757263355203e-05,
"loss": 0.1291,
"step": 8530
}
],
"logging_steps": 10,
"max_steps": 8536,
"num_input_tokens_seen": 0,
"num_train_epochs": 4,
"save_steps": 1000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 0.0,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}