{ "best_global_step": 2148, "best_metric": 0.06230008, "best_model_checkpoint": "/home/mithil/PycharmProjects/HebrewSearch/output/Qwen3-Embedding-/v10-20250927-032234/checkpoint-2148", "epoch": 2.0, "eval_steps": 500, "global_step": 2148, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0009315323707498836, "grad_norm": 5.125, "learning_rate": 5.999996791349807e-05, "loss": 0.22629377245903015, "step": 1 }, { "epoch": 0.004657661853749418, "grad_norm": 2.734375, "learning_rate": 5.999919784088342e-05, "loss": 0.11207550764083862, "step": 5 }, { "epoch": 0.009315323707498836, "grad_norm": 2.328125, "learning_rate": 5.999679140643096e-05, "loss": 0.1241973876953125, "step": 10 }, { "epoch": 0.013972985561248253, "grad_norm": 1.0, "learning_rate": 5.999278082533218e-05, "loss": 0.08484302759170533, "step": 15 }, { "epoch": 0.018630647414997672, "grad_norm": 1.1484375, "learning_rate": 5.998716631206202e-05, "loss": 0.11840656995773316, "step": 20 }, { "epoch": 0.02328830926874709, "grad_norm": 1.3125, "learning_rate": 5.997994816686935e-05, "loss": 0.10533121824264527, "step": 25 }, { "epoch": 0.027945971122496506, "grad_norm": 1.96875, "learning_rate": 5.997112677576091e-05, "loss": 0.07531467080116272, "step": 30 }, { "epoch": 0.032603632976245925, "grad_norm": 1.4765625, "learning_rate": 5.9960702610480635e-05, "loss": 0.08655669689178466, "step": 35 }, { "epoch": 0.037261294829995344, "grad_norm": 1.140625, "learning_rate": 5.994867622848448e-05, "loss": 0.08251949548721313, "step": 40 }, { "epoch": 0.04191895668374476, "grad_norm": 1.3515625, "learning_rate": 5.993504827291059e-05, "loss": 0.07267643213272094, "step": 45 }, { "epoch": 0.04657661853749418, "grad_norm": 1.0078125, "learning_rate": 5.991981947254487e-05, "loss": 0.08904569149017334, "step": 50 }, { "epoch": 0.05123428039124359, "grad_norm": 1.4140625, "learning_rate": 5.990299064178205e-05, "loss": 0.06884298324584961, "step": 55 }, { "epoch": 0.05589194224499301, "grad_norm": 1.7734375, "learning_rate": 5.9884562680582146e-05, "loss": 0.08851982355117798, "step": 60 }, { "epoch": 0.06054960409874243, "grad_norm": 4.4375, "learning_rate": 5.98645365744223e-05, "loss": 0.08651072978973388, "step": 65 }, { "epoch": 0.06520726595249185, "grad_norm": 1.0859375, "learning_rate": 5.9842913394244063e-05, "loss": 0.07889164686203003, "step": 70 }, { "epoch": 0.06986492780624126, "grad_norm": 1.34375, "learning_rate": 5.9819694296396204e-05, "loss": 0.0799069881439209, "step": 75 }, { "epoch": 0.07452258965999069, "grad_norm": 1.890625, "learning_rate": 5.979488052257277e-05, "loss": 0.10144302845001221, "step": 80 }, { "epoch": 0.0791802515137401, "grad_norm": 1.1953125, "learning_rate": 5.9768473399746766e-05, "loss": 0.07646682262420654, "step": 85 }, { "epoch": 0.08383791336748952, "grad_norm": 1.6640625, "learning_rate": 5.974047434009914e-05, "loss": 0.08959404826164245, "step": 90 }, { "epoch": 0.08849557522123894, "grad_norm": 1.1796875, "learning_rate": 5.9710884840943296e-05, "loss": 0.07164722084999084, "step": 95 }, { "epoch": 0.09315323707498836, "grad_norm": 1.359375, "learning_rate": 5.967970648464499e-05, "loss": 0.07039438486099243, "step": 100 }, { "epoch": 0.09781089892873777, "grad_norm": 1.1953125, "learning_rate": 5.964694093853774e-05, "loss": 0.06741951704025269, "step": 105 }, { "epoch": 0.10246856078248719, "grad_norm": 1.0625, "learning_rate": 5.9612589954833654e-05, "loss": 0.059288203716278076, "step": 110 }, { "epoch": 0.10712622263623661, "grad_norm": 2.28125, "learning_rate": 5.957665537052972e-05, "loss": 0.08943802118301392, "step": 115 }, { "epoch": 0.11178388448998602, "grad_norm": 1.75, "learning_rate": 5.953913910730955e-05, "loss": 0.0767773985862732, "step": 120 }, { "epoch": 0.11644154634373545, "grad_norm": 1.625, "learning_rate": 5.9500043171440655e-05, "loss": 0.061524766683578494, "step": 125 }, { "epoch": 0.12109920819748486, "grad_norm": 1.859375, "learning_rate": 5.9459369653667134e-05, "loss": 0.10620454549789429, "step": 130 }, { "epoch": 0.1257568700512343, "grad_norm": 1.4140625, "learning_rate": 5.9417120729097845e-05, "loss": 0.04981007874011993, "step": 135 }, { "epoch": 0.1304145319049837, "grad_norm": 2.1875, "learning_rate": 5.937329865709012e-05, "loss": 0.06490777730941773, "step": 140 }, { "epoch": 0.1350721937587331, "grad_norm": 1.375, "learning_rate": 5.932790578112896e-05, "loss": 0.06334985494613647, "step": 145 }, { "epoch": 0.13972985561248252, "grad_norm": 1.078125, "learning_rate": 5.928094452870161e-05, "loss": 0.07352197170257568, "step": 150 }, { "epoch": 0.14438751746623196, "grad_norm": 1.75, "learning_rate": 5.923241741116789e-05, "loss": 0.07104206681251526, "step": 155 }, { "epoch": 0.14904517931998137, "grad_norm": 1.0078125, "learning_rate": 5.918232702362575e-05, "loss": 0.05300582647323608, "step": 160 }, { "epoch": 0.1537028411737308, "grad_norm": 2.078125, "learning_rate": 5.913067604477262e-05, "loss": 0.13233035802841187, "step": 165 }, { "epoch": 0.1583605030274802, "grad_norm": 1.3125, "learning_rate": 5.907746723676205e-05, "loss": 0.07034454345703126, "step": 170 }, { "epoch": 0.1630181648812296, "grad_norm": 1.0, "learning_rate": 5.902270344505608e-05, "loss": 0.06918042302131652, "step": 175 }, { "epoch": 0.16767582673497905, "grad_norm": 0.96484375, "learning_rate": 5.896638759827303e-05, "loss": 0.061054134368896486, "step": 180 }, { "epoch": 0.17233348858872846, "grad_norm": 1.6015625, "learning_rate": 5.8908522708030884e-05, "loss": 0.10253015756607056, "step": 185 }, { "epoch": 0.17699115044247787, "grad_norm": 1.3125, "learning_rate": 5.884911186878626e-05, "loss": 0.09637802839279175, "step": 190 }, { "epoch": 0.18164881229622729, "grad_norm": 1.3125, "learning_rate": 5.878815825766892e-05, "loss": 0.09289785027503968, "step": 195 }, { "epoch": 0.18630647414997673, "grad_norm": 3.484375, "learning_rate": 5.872566513431184e-05, "loss": 0.08359139561653137, "step": 200 }, { "epoch": 0.19096413600372614, "grad_norm": 7.6875, "learning_rate": 5.8661635840676944e-05, "loss": 0.07835670113563538, "step": 205 }, { "epoch": 0.19562179785747555, "grad_norm": 1.1171875, "learning_rate": 5.859607380087634e-05, "loss": 0.06977214813232421, "step": 210 }, { "epoch": 0.20027945971122496, "grad_norm": 1.1953125, "learning_rate": 5.852898252098921e-05, "loss": 0.0683005690574646, "step": 215 }, { "epoch": 0.20493712156497437, "grad_norm": 1.0859375, "learning_rate": 5.846036558887435e-05, "loss": 0.06400226354598999, "step": 220 }, { "epoch": 0.2095947834187238, "grad_norm": 1.0078125, "learning_rate": 5.839022667397827e-05, "loss": 0.06587435007095337, "step": 225 }, { "epoch": 0.21425244527247322, "grad_norm": 1.3125, "learning_rate": 5.831856952713896e-05, "loss": 0.08793356418609619, "step": 230 }, { "epoch": 0.21891010712622264, "grad_norm": 1.53125, "learning_rate": 5.8245397980385344e-05, "loss": 0.06810299754142761, "step": 235 }, { "epoch": 0.22356776897997205, "grad_norm": 2.921875, "learning_rate": 5.817071594673229e-05, "loss": 0.06869109869003295, "step": 240 }, { "epoch": 0.22822543083372146, "grad_norm": 1.5234375, "learning_rate": 5.809452741997143e-05, "loss": 0.08056961297988892, "step": 245 }, { "epoch": 0.2328830926874709, "grad_norm": 1.6171875, "learning_rate": 5.801683647445749e-05, "loss": 0.08139086961746216, "step": 250 }, { "epoch": 0.2375407545412203, "grad_norm": 0.9375, "learning_rate": 5.7937647264890514e-05, "loss": 0.06604200601577759, "step": 255 }, { "epoch": 0.24219841639496972, "grad_norm": 3.53125, "learning_rate": 5.7856964026093575e-05, "loss": 0.08705702424049377, "step": 260 }, { "epoch": 0.24685607824871914, "grad_norm": 0.9765625, "learning_rate": 5.777479107278639e-05, "loss": 0.08005684614181519, "step": 265 }, { "epoch": 0.2515137401024686, "grad_norm": 1.734375, "learning_rate": 5.769113279935452e-05, "loss": 0.06209390759468079, "step": 270 }, { "epoch": 0.25617140195621796, "grad_norm": 1.9609375, "learning_rate": 5.760599367961442e-05, "loss": 0.10012856721878052, "step": 275 }, { "epoch": 0.2608290638099674, "grad_norm": 1.0234375, "learning_rate": 5.751937826657417e-05, "loss": 0.09347758889198303, "step": 280 }, { "epoch": 0.26548672566371684, "grad_norm": 1.3203125, "learning_rate": 5.743129119218996e-05, "loss": 0.08649051785469056, "step": 285 }, { "epoch": 0.2701443875174662, "grad_norm": 1.640625, "learning_rate": 5.734173716711847e-05, "loss": 0.06381095051765442, "step": 290 }, { "epoch": 0.27480204937121566, "grad_norm": 1.59375, "learning_rate": 5.725072098046486e-05, "loss": 0.06688931584358215, "step": 295 }, { "epoch": 0.27945971122496505, "grad_norm": 1.921875, "learning_rate": 5.715824749952672e-05, "loss": 0.07226124405860901, "step": 300 }, { "epoch": 0.2841173730787145, "grad_norm": 2.125, "learning_rate": 5.706432166953378e-05, "loss": 0.07504152059555054, "step": 305 }, { "epoch": 0.2887750349324639, "grad_norm": 0.95703125, "learning_rate": 5.696894851338341e-05, "loss": 0.049904930591583255, "step": 310 }, { "epoch": 0.2934326967862133, "grad_norm": 1.5390625, "learning_rate": 5.687213313137208e-05, "loss": 0.07915764451026916, "step": 315 }, { "epoch": 0.29809035863996275, "grad_norm": 1.9296875, "learning_rate": 5.677388070092252e-05, "loss": 0.06789354085922242, "step": 320 }, { "epoch": 0.30274802049371213, "grad_norm": 1.6015625, "learning_rate": 5.667419647630693e-05, "loss": 0.07174413204193116, "step": 325 }, { "epoch": 0.3074056823474616, "grad_norm": 1.3671875, "learning_rate": 5.657308578836594e-05, "loss": 0.0827239990234375, "step": 330 }, { "epoch": 0.312063344201211, "grad_norm": 1.640625, "learning_rate": 5.647055404422357e-05, "loss": 0.06073644757270813, "step": 335 }, { "epoch": 0.3167210060549604, "grad_norm": 1.328125, "learning_rate": 5.636660672699803e-05, "loss": 0.050183039903640744, "step": 340 }, { "epoch": 0.32137866790870984, "grad_norm": 2.875, "learning_rate": 5.626124939550853e-05, "loss": 0.08814806938171386, "step": 345 }, { "epoch": 0.3260363297624592, "grad_norm": 1.921875, "learning_rate": 5.6154487683978e-05, "loss": 0.07444313168525696, "step": 350 }, { "epoch": 0.33069399161620866, "grad_norm": 1.7734375, "learning_rate": 5.60463273017318e-05, "loss": 0.05980492830276489, "step": 355 }, { "epoch": 0.3353516534699581, "grad_norm": 0.80078125, "learning_rate": 5.593677403289235e-05, "loss": 0.05909588932991028, "step": 360 }, { "epoch": 0.3400093153237075, "grad_norm": 1.296875, "learning_rate": 5.5825833736069904e-05, "loss": 0.08176165819168091, "step": 365 }, { "epoch": 0.3446669771774569, "grad_norm": 1.640625, "learning_rate": 5.571351234404913e-05, "loss": 0.07512840628623962, "step": 370 }, { "epoch": 0.3493246390312063, "grad_norm": 2.0, "learning_rate": 5.559981586347196e-05, "loss": 0.07859928011894227, "step": 375 }, { "epoch": 0.35398230088495575, "grad_norm": 1.9296875, "learning_rate": 5.548475037451627e-05, "loss": 0.07391362190246582, "step": 380 }, { "epoch": 0.3586399627387052, "grad_norm": 2.5625, "learning_rate": 5.536832203057081e-05, "loss": 0.08393895626068115, "step": 385 }, { "epoch": 0.36329762459245457, "grad_norm": 2.9375, "learning_rate": 5.525053705790606e-05, "loss": 0.06296343803405761, "step": 390 }, { "epoch": 0.367955286446204, "grad_norm": 1.2734375, "learning_rate": 5.5131401755341354e-05, "loss": 0.08504929542541503, "step": 395 }, { "epoch": 0.37261294829995345, "grad_norm": 1.1953125, "learning_rate": 5.501092249390793e-05, "loss": 0.08101065158843994, "step": 400 }, { "epoch": 0.37727061015370283, "grad_norm": 7.5, "learning_rate": 5.488910571650835e-05, "loss": 0.07101285457611084, "step": 405 }, { "epoch": 0.3819282720074523, "grad_norm": 1.90625, "learning_rate": 5.4765957937571824e-05, "loss": 0.08065653443336487, "step": 410 }, { "epoch": 0.38658593386120166, "grad_norm": 1.4921875, "learning_rate": 5.464148574270594e-05, "loss": 0.07403583526611328, "step": 415 }, { "epoch": 0.3912435957149511, "grad_norm": 1.46875, "learning_rate": 5.451569578834441e-05, "loss": 0.06334596872329712, "step": 420 }, { "epoch": 0.39590125756870054, "grad_norm": 1.6875, "learning_rate": 5.4388594801391146e-05, "loss": 0.06494083404541015, "step": 425 }, { "epoch": 0.4005589194224499, "grad_norm": 3.25, "learning_rate": 5.426018957886052e-05, "loss": 0.06925193071365357, "step": 430 }, { "epoch": 0.40521658127619936, "grad_norm": 2.84375, "learning_rate": 5.413048698751386e-05, "loss": 0.08767553567886352, "step": 435 }, { "epoch": 0.40987424312994875, "grad_norm": 1.6796875, "learning_rate": 5.399949396349221e-05, "loss": 0.07317562699317932, "step": 440 }, { "epoch": 0.4145319049836982, "grad_norm": 0.984375, "learning_rate": 5.386721751194549e-05, "loss": 0.07249165177345276, "step": 445 }, { "epoch": 0.4191895668374476, "grad_norm": 1.3359375, "learning_rate": 5.37336647066578e-05, "loss": 0.08126922845840454, "step": 450 }, { "epoch": 0.423847228691197, "grad_norm": 1.3359375, "learning_rate": 5.3598842689669155e-05, "loss": 0.06756317615509033, "step": 455 }, { "epoch": 0.42850489054494645, "grad_norm": 1.671875, "learning_rate": 5.346275867089355e-05, "loss": 0.07897292971611022, "step": 460 }, { "epoch": 0.43316255239869583, "grad_norm": 0.8046875, "learning_rate": 5.332541992773342e-05, "loss": 0.049942368268966676, "step": 465 }, { "epoch": 0.43782021425244527, "grad_norm": 1.0859375, "learning_rate": 5.3186833804690415e-05, "loss": 0.05567145347595215, "step": 470 }, { "epoch": 0.4424778761061947, "grad_norm": 1.1796875, "learning_rate": 5.304700771297267e-05, "loss": 0.06544985771179199, "step": 475 }, { "epoch": 0.4471355379599441, "grad_norm": 1.671875, "learning_rate": 5.2905949130098465e-05, "loss": 0.07286288738250732, "step": 480 }, { "epoch": 0.45179319981369354, "grad_norm": 1.6328125, "learning_rate": 5.276366559949635e-05, "loss": 0.08477526903152466, "step": 485 }, { "epoch": 0.4564508616674429, "grad_norm": 1.3125, "learning_rate": 5.2620164730101746e-05, "loss": 0.07429101467132568, "step": 490 }, { "epoch": 0.46110852352119236, "grad_norm": 1.078125, "learning_rate": 5.2475454195950024e-05, "loss": 0.0752301573753357, "step": 495 }, { "epoch": 0.4657661853749418, "grad_norm": 1.171875, "learning_rate": 5.2329541735766116e-05, "loss": 0.06806610822677613, "step": 500 }, { "epoch": 0.4704238472286912, "grad_norm": 3.5, "learning_rate": 5.218243515255072e-05, "loss": 0.07269450426101684, "step": 505 }, { "epoch": 0.4750815090824406, "grad_norm": 1.6640625, "learning_rate": 5.203414231316295e-05, "loss": 0.0846606969833374, "step": 510 }, { "epoch": 0.47973917093619, "grad_norm": 1.125, "learning_rate": 5.1884671147899686e-05, "loss": 0.05865778923034668, "step": 515 }, { "epoch": 0.48439683278993945, "grad_norm": 1.5234375, "learning_rate": 5.1734029650071437e-05, "loss": 0.06631482243537903, "step": 520 }, { "epoch": 0.4890544946436889, "grad_norm": 1.8359375, "learning_rate": 5.158222587557493e-05, "loss": 0.09194896817207336, "step": 525 }, { "epoch": 0.49371215649743827, "grad_norm": 1.3671875, "learning_rate": 5.142926794246228e-05, "loss": 0.06618914604187012, "step": 530 }, { "epoch": 0.4983698183511877, "grad_norm": 1.7734375, "learning_rate": 5.1275164030506855e-05, "loss": 0.05849605202674866, "step": 535 }, { "epoch": 0.5030274802049371, "grad_norm": 1.71875, "learning_rate": 5.111992238076583e-05, "loss": 0.06805847287178039, "step": 540 }, { "epoch": 0.5076851420586865, "grad_norm": 1.328125, "learning_rate": 5.096355129513954e-05, "loss": 0.0659984827041626, "step": 545 }, { "epoch": 0.5123428039124359, "grad_norm": 2.421875, "learning_rate": 5.0806059135927425e-05, "loss": 0.08051547408103943, "step": 550 }, { "epoch": 0.5170004657661854, "grad_norm": 1.0234375, "learning_rate": 5.06474543253809e-05, "loss": 0.05917655229568482, "step": 555 }, { "epoch": 0.5216581276199348, "grad_norm": 1.4765625, "learning_rate": 5.048774534525296e-05, "loss": 0.07989615201950073, "step": 560 }, { "epoch": 0.5263157894736842, "grad_norm": 1.53125, "learning_rate": 5.032694073634456e-05, "loss": 0.0757764756679535, "step": 565 }, { "epoch": 0.5309734513274337, "grad_norm": 1.296875, "learning_rate": 5.0165049098047905e-05, "loss": 0.07939087748527526, "step": 570 }, { "epoch": 0.5356311131811831, "grad_norm": 1.0703125, "learning_rate": 5.000207908788657e-05, "loss": 0.06915702223777771, "step": 575 }, { "epoch": 0.5402887750349324, "grad_norm": 1.2890625, "learning_rate": 4.9838039421052495e-05, "loss": 0.05958174467086792, "step": 580 }, { "epoch": 0.5449464368886818, "grad_norm": 0.765625, "learning_rate": 4.9672938869939985e-05, "loss": 0.053989958763122556, "step": 585 }, { "epoch": 0.5496040987424313, "grad_norm": 0.96484375, "learning_rate": 4.95067862636765e-05, "loss": 0.07627325057983399, "step": 590 }, { "epoch": 0.5542617605961807, "grad_norm": 0.86328125, "learning_rate": 4.933959048765059e-05, "loss": 0.05423608422279358, "step": 595 }, { "epoch": 0.5589194224499301, "grad_norm": 1.3515625, "learning_rate": 4.9171360483036624e-05, "loss": 0.06786616444587708, "step": 600 }, { "epoch": 0.5635770843036796, "grad_norm": 0.80078125, "learning_rate": 4.900210524631676e-05, "loss": 0.06529110670089722, "step": 605 }, { "epoch": 0.568234746157429, "grad_norm": 1.3359375, "learning_rate": 4.883183382879971e-05, "loss": 0.06635728478431702, "step": 610 }, { "epoch": 0.5728924080111784, "grad_norm": 1.2578125, "learning_rate": 4.8660555336136816e-05, "loss": 0.07374628186225891, "step": 615 }, { "epoch": 0.5775500698649279, "grad_norm": 1.109375, "learning_rate": 4.848827892783503e-05, "loss": 0.058982133865356445, "step": 620 }, { "epoch": 0.5822077317186772, "grad_norm": 1.328125, "learning_rate": 4.831501381676712e-05, "loss": 0.08937416672706604, "step": 625 }, { "epoch": 0.5868653935724266, "grad_norm": 1.0703125, "learning_rate": 4.814076926867899e-05, "loss": 0.06815096139907836, "step": 630 }, { "epoch": 0.5915230554261761, "grad_norm": 0.77734375, "learning_rate": 4.796555460169415e-05, "loss": 0.06007786989212036, "step": 635 }, { "epoch": 0.5961807172799255, "grad_norm": 1.1796875, "learning_rate": 4.778937918581542e-05, "loss": 0.06676498651504517, "step": 640 }, { "epoch": 0.6008383791336749, "grad_norm": 0.8828125, "learning_rate": 4.761225244242389e-05, "loss": 0.06312382817268372, "step": 645 }, { "epoch": 0.6054960409874243, "grad_norm": 1.875, "learning_rate": 4.7434183843775e-05, "loss": 0.07485262155532837, "step": 650 }, { "epoch": 0.6101537028411738, "grad_norm": 1.2890625, "learning_rate": 4.7255182912492086e-05, "loss": 0.05705668330192566, "step": 655 }, { "epoch": 0.6148113646949231, "grad_norm": 1.3359375, "learning_rate": 4.707525922105708e-05, "loss": 0.08206185102462768, "step": 660 }, { "epoch": 0.6194690265486725, "grad_norm": 1.1875, "learning_rate": 4.689442239129858e-05, "loss": 0.06421056389808655, "step": 665 }, { "epoch": 0.624126688402422, "grad_norm": 0.71484375, "learning_rate": 4.6712682093877395e-05, "loss": 0.0625627338886261, "step": 670 }, { "epoch": 0.6287843502561714, "grad_norm": 1.2265625, "learning_rate": 4.653004804776925e-05, "loss": 0.08293673396110535, "step": 675 }, { "epoch": 0.6334420121099208, "grad_norm": 1.6015625, "learning_rate": 4.634653001974519e-05, "loss": 0.07818044424057007, "step": 680 }, { "epoch": 0.6380996739636703, "grad_norm": 1.1953125, "learning_rate": 4.6162137823849125e-05, "loss": 0.06400628089904785, "step": 685 }, { "epoch": 0.6427573358174197, "grad_norm": 1.296875, "learning_rate": 4.597688132087314e-05, "loss": 0.05182795524597168, "step": 690 }, { "epoch": 0.6474149976711691, "grad_norm": 1.4609375, "learning_rate": 4.5790770417830084e-05, "loss": 0.08495147228240967, "step": 695 }, { "epoch": 0.6520726595249184, "grad_norm": 1.3515625, "learning_rate": 4.56038150674238e-05, "loss": 0.061500102281570435, "step": 700 }, { "epoch": 0.6567303213786679, "grad_norm": 3.25, "learning_rate": 4.5416025267516866e-05, "loss": 0.069045090675354, "step": 705 }, { "epoch": 0.6613879832324173, "grad_norm": 1.328125, "learning_rate": 4.522741106059595e-05, "loss": 0.059973156452178954, "step": 710 }, { "epoch": 0.6660456450861667, "grad_norm": 1.171875, "learning_rate": 4.5037982533234734e-05, "loss": 0.07730792760848999, "step": 715 }, { "epoch": 0.6707033069399162, "grad_norm": 2.25, "learning_rate": 4.48477498155546e-05, "loss": 0.06633560657501221, "step": 720 }, { "epoch": 0.6753609687936656, "grad_norm": 1.5546875, "learning_rate": 4.465672308068278e-05, "loss": 0.05775619745254516, "step": 725 }, { "epoch": 0.680018630647415, "grad_norm": 1.5703125, "learning_rate": 4.446491254420841e-05, "loss": 0.07142780423164367, "step": 730 }, { "epoch": 0.6846762925011645, "grad_norm": 2.15625, "learning_rate": 4.427232846363617e-05, "loss": 0.06803997159004212, "step": 735 }, { "epoch": 0.6893339543549138, "grad_norm": 1.1015625, "learning_rate": 4.4078981137837804e-05, "loss": 0.06891953945159912, "step": 740 }, { "epoch": 0.6939916162086632, "grad_norm": 1.1328125, "learning_rate": 4.388488090650132e-05, "loss": 0.07929319143295288, "step": 745 }, { "epoch": 0.6986492780624126, "grad_norm": 1.3125, "learning_rate": 4.3690038149578044e-05, "loss": 0.06713913083076477, "step": 750 }, { "epoch": 0.7033069399161621, "grad_norm": 2.171875, "learning_rate": 4.349446328672756e-05, "loss": 0.0739738404750824, "step": 755 }, { "epoch": 0.7079646017699115, "grad_norm": 1.1328125, "learning_rate": 4.329816677676049e-05, "loss": 0.058427393436431885, "step": 760 }, { "epoch": 0.7126222636236609, "grad_norm": 1.3046875, "learning_rate": 4.310115911707918e-05, "loss": 0.06763650178909301, "step": 765 }, { "epoch": 0.7172799254774104, "grad_norm": 1.984375, "learning_rate": 4.2903450843116284e-05, "loss": 0.07517719864845276, "step": 770 }, { "epoch": 0.7219375873311598, "grad_norm": 1.4765625, "learning_rate": 4.2705052527771444e-05, "loss": 0.08248428702354431, "step": 775 }, { "epoch": 0.7265952491849091, "grad_norm": 0.93359375, "learning_rate": 4.250597478084583e-05, "loss": 0.05707501769065857, "step": 780 }, { "epoch": 0.7312529110386586, "grad_norm": 1.8671875, "learning_rate": 4.230622824847474e-05, "loss": 0.07538543343544006, "step": 785 }, { "epoch": 0.735910572892408, "grad_norm": 1.4453125, "learning_rate": 4.2105823612558285e-05, "loss": 0.054673463106155396, "step": 790 }, { "epoch": 0.7405682347461574, "grad_norm": 1.8984375, "learning_rate": 4.19047715901902e-05, "loss": 0.07854291796684265, "step": 795 }, { "epoch": 0.7452258965999069, "grad_norm": 1.453125, "learning_rate": 4.1703082933084666e-05, "loss": 0.05171064734458923, "step": 800 }, { "epoch": 0.7498835584536563, "grad_norm": 1.5546875, "learning_rate": 4.1500768427001337e-05, "loss": 0.05756812691688538, "step": 805 }, { "epoch": 0.7545412203074057, "grad_norm": 1.4453125, "learning_rate": 4.1297838891168575e-05, "loss": 0.07100930213928222, "step": 810 }, { "epoch": 0.759198882161155, "grad_norm": 1.203125, "learning_rate": 4.109430517770487e-05, "loss": 0.06920987367630005, "step": 815 }, { "epoch": 0.7638565440149045, "grad_norm": 0.92578125, "learning_rate": 4.089017817103847e-05, "loss": 0.05224462747573853, "step": 820 }, { "epoch": 0.7685142058686539, "grad_norm": 2.0625, "learning_rate": 4.0685468787325325e-05, "loss": 0.06810380220413208, "step": 825 }, { "epoch": 0.7731718677224033, "grad_norm": 2.484375, "learning_rate": 4.048018797386533e-05, "loss": 0.07529953122138977, "step": 830 }, { "epoch": 0.7778295295761528, "grad_norm": 1.3203125, "learning_rate": 4.02743467085169e-05, "loss": 0.07324286699295043, "step": 835 }, { "epoch": 0.7824871914299022, "grad_norm": 1.796875, "learning_rate": 4.006795599910986e-05, "loss": 0.06480486392974853, "step": 840 }, { "epoch": 0.7871448532836516, "grad_norm": 2.140625, "learning_rate": 3.986102688285681e-05, "loss": 0.07136549353599549, "step": 845 }, { "epoch": 0.7918025151374011, "grad_norm": 1.4453125, "learning_rate": 3.9653570425762894e-05, "loss": 0.0683221161365509, "step": 850 }, { "epoch": 0.7964601769911505, "grad_norm": 1.1953125, "learning_rate": 3.9445597722034024e-05, "loss": 0.07045792937278747, "step": 855 }, { "epoch": 0.8011178388448998, "grad_norm": 1.9765625, "learning_rate": 3.923711989348352e-05, "loss": 0.06577153205871582, "step": 860 }, { "epoch": 0.8057755006986492, "grad_norm": 1.9921875, "learning_rate": 3.9028148088937454e-05, "loss": 0.07191576361656189, "step": 865 }, { "epoch": 0.8104331625523987, "grad_norm": 1.4453125, "learning_rate": 3.881869348363836e-05, "loss": 0.05743700861930847, "step": 870 }, { "epoch": 0.8150908244061481, "grad_norm": 1.078125, "learning_rate": 3.8608767278647655e-05, "loss": 0.06495893597602845, "step": 875 }, { "epoch": 0.8197484862598975, "grad_norm": 1.796875, "learning_rate": 3.83983807002466e-05, "loss": 0.06839098334312439, "step": 880 }, { "epoch": 0.824406148113647, "grad_norm": 1.0546875, "learning_rate": 3.818754499933601e-05, "loss": 0.06613754630088806, "step": 885 }, { "epoch": 0.8290638099673964, "grad_norm": 1.2421875, "learning_rate": 3.7976271450834504e-05, "loss": 0.06127074360847473, "step": 890 }, { "epoch": 0.8337214718211458, "grad_norm": 1.3046875, "learning_rate": 3.776457135307562e-05, "loss": 0.05446965098381042, "step": 895 }, { "epoch": 0.8383791336748952, "grad_norm": 1.03125, "learning_rate": 3.75524560272036e-05, "loss": 0.057818031311035155, "step": 900 }, { "epoch": 0.8430367955286446, "grad_norm": 1.1640625, "learning_rate": 3.733993681656792e-05, "loss": 0.06393643021583557, "step": 905 }, { "epoch": 0.847694457382394, "grad_norm": 1.4921875, "learning_rate": 3.7127025086116736e-05, "loss": 0.06902187466621398, "step": 910 }, { "epoch": 0.8523521192361434, "grad_norm": 1.71875, "learning_rate": 3.691373222178909e-05, "loss": 0.06524651050567627, "step": 915 }, { "epoch": 0.8570097810898929, "grad_norm": 1.5234375, "learning_rate": 3.670006962990604e-05, "loss": 0.06683992147445679, "step": 920 }, { "epoch": 0.8616674429436423, "grad_norm": 0.97265625, "learning_rate": 3.6486048736560614e-05, "loss": 0.07013018727302552, "step": 925 }, { "epoch": 0.8663251047973917, "grad_norm": 1.3203125, "learning_rate": 3.627168098700688e-05, "loss": 0.06274893283843994, "step": 930 }, { "epoch": 0.8709827666511412, "grad_norm": 1.4140625, "learning_rate": 3.605697784504782e-05, "loss": 0.07435029745101929, "step": 935 }, { "epoch": 0.8756404285048905, "grad_norm": 2.640625, "learning_rate": 3.584195079242227e-05, "loss": 0.04883918762207031, "step": 940 }, { "epoch": 0.8802980903586399, "grad_norm": 1.328125, "learning_rate": 3.562661132819093e-05, "loss": 0.0625451922416687, "step": 945 }, { "epoch": 0.8849557522123894, "grad_norm": 1.875, "learning_rate": 3.541097096812144e-05, "loss": 0.0689104974269867, "step": 950 }, { "epoch": 0.8896134140661388, "grad_norm": 2.21875, "learning_rate": 3.5195041244072506e-05, "loss": 0.05478867292404175, "step": 955 }, { "epoch": 0.8942710759198882, "grad_norm": 1.6640625, "learning_rate": 3.497883370337724e-05, "loss": 0.07447841763496399, "step": 960 }, { "epoch": 0.8989287377736377, "grad_norm": 1.2734375, "learning_rate": 3.476235990822563e-05, "loss": 0.07308604717254638, "step": 965 }, { "epoch": 0.9035863996273871, "grad_norm": 1.3125, "learning_rate": 3.4545631435046233e-05, "loss": 0.06406825184822082, "step": 970 }, { "epoch": 0.9082440614811365, "grad_norm": 1.2421875, "learning_rate": 3.43286598738871e-05, "loss": 0.06342694759368897, "step": 975 }, { "epoch": 0.9129017233348858, "grad_norm": 1.625, "learning_rate": 3.411145682779593e-05, "loss": 0.06270037293434143, "step": 980 }, { "epoch": 0.9175593851886353, "grad_norm": 4.25, "learning_rate": 3.389403391219963e-05, "loss": 0.07241315245628357, "step": 985 }, { "epoch": 0.9222170470423847, "grad_norm": 0.8515625, "learning_rate": 3.3676402754283145e-05, "loss": 0.0640365719795227, "step": 990 }, { "epoch": 0.9268747088961341, "grad_norm": 1.2265625, "learning_rate": 3.345857499236761e-05, "loss": 0.07342595458030701, "step": 995 }, { "epoch": 0.9315323707498836, "grad_norm": 1.8203125, "learning_rate": 3.324056227528805e-05, "loss": 0.059070253372192384, "step": 1000 }, { "epoch": 0.936190032603633, "grad_norm": 1.046875, "learning_rate": 3.302237626177036e-05, "loss": 0.06199865341186524, "step": 1005 }, { "epoch": 0.9408476944573824, "grad_norm": 1.2421875, "learning_rate": 3.2804028619807865e-05, "loss": 0.04795067310333252, "step": 1010 }, { "epoch": 0.9455053563111319, "grad_norm": 1.1015625, "learning_rate": 3.2585531026037325e-05, "loss": 0.05880788564682007, "step": 1015 }, { "epoch": 0.9501630181648812, "grad_norm": 1.4140625, "learning_rate": 3.2366895165114534e-05, "loss": 0.06886560916900634, "step": 1020 }, { "epoch": 0.9548206800186306, "grad_norm": 0.890625, "learning_rate": 3.2148132729089436e-05, "loss": 0.05063498020172119, "step": 1025 }, { "epoch": 0.95947834187238, "grad_norm": 1.0390625, "learning_rate": 3.192925541678085e-05, "loss": 0.05080598592758179, "step": 1030 }, { "epoch": 0.9641360037261295, "grad_norm": 1.1015625, "learning_rate": 3.171027493315089e-05, "loss": 0.07384888529777527, "step": 1035 }, { "epoch": 0.9687936655798789, "grad_norm": 1.546875, "learning_rate": 3.149120298867895e-05, "loss": 0.06471606492996215, "step": 1040 }, { "epoch": 0.9734513274336283, "grad_norm": 6.40625, "learning_rate": 3.127205129873555e-05, "loss": 0.08100690245628357, "step": 1045 }, { "epoch": 0.9781089892873778, "grad_norm": 1.2109375, "learning_rate": 3.105283158295575e-05, "loss": 0.07905409932136535, "step": 1050 }, { "epoch": 0.9827666511411272, "grad_norm": 1.2734375, "learning_rate": 3.083355556461244e-05, "loss": 0.05518416166305542, "step": 1055 }, { "epoch": 0.9874243129948765, "grad_norm": 0.90625, "learning_rate": 3.0614234969989445e-05, "loss": 0.07515279650688171, "step": 1060 }, { "epoch": 0.992081974848626, "grad_norm": 1.3515625, "learning_rate": 3.039488152775439e-05, "loss": 0.059355854988098145, "step": 1065 }, { "epoch": 0.9967396367023754, "grad_norm": 1.109375, "learning_rate": 3.017550696833151e-05, "loss": 0.07503612041473388, "step": 1070 }, { "epoch": 1.0, "eval_loss": 0.06487495452165604, "eval_pearson_cosine": 0.5299683213233948, "eval_pearson_dot_product": 0.5299683213233948, "eval_pearson_euclidean": 0.536834180355072, "eval_pearson_manhattan": 0.534827470779419, "eval_runtime": 17.3569, "eval_samples_per_second": 117.186, "eval_spearman_cosine": 0.5408944697862549, "eval_spearman_dot_product": 0.5408955665177739, "eval_spearman_euclidean": 0.5408944701719207, "eval_spearman_manhattan": 0.5392575148709198, "eval_steps_per_second": 19.531, "step": 1074 }, { "epoch": 1.00093153237075, "grad_norm": 1.0234375, "learning_rate": 2.995612302327431e-05, "loss": 0.04589760899543762, "step": 1075 }, { "epoch": 1.0055891942244992, "grad_norm": 0.9609375, "learning_rate": 2.9736741424638238e-05, "loss": 0.049758344888687134, "step": 1080 }, { "epoch": 1.0102468560782487, "grad_norm": 1.34375, "learning_rate": 2.951737390435325e-05, "loss": 0.0830036759376526, "step": 1085 }, { "epoch": 1.0149045179319982, "grad_norm": 1.1484375, "learning_rate": 2.9298032193596438e-05, "loss": 0.07352387309074401, "step": 1090 }, { "epoch": 1.0195621797857475, "grad_norm": 1.71875, "learning_rate": 2.9078728022164646e-05, "loss": 0.050916272401809695, "step": 1095 }, { "epoch": 1.024219841639497, "grad_norm": 1.15625, "learning_rate": 2.8859473117847253e-05, "loss": 0.05987505912780762, "step": 1100 }, { "epoch": 1.0288775034932465, "grad_norm": 1.7109375, "learning_rate": 2.864027920579893e-05, "loss": 0.05799069404602051, "step": 1105 }, { "epoch": 1.0335351653469957, "grad_norm": 1.484375, "learning_rate": 2.8421158007912687e-05, "loss": 0.06869199872016907, "step": 1110 }, { "epoch": 1.0381928272007452, "grad_norm": 1.09375, "learning_rate": 2.8202121242192954e-05, "loss": 0.05415593981742859, "step": 1115 }, { "epoch": 1.0428504890544947, "grad_norm": 1.078125, "learning_rate": 2.798318062212895e-05, "loss": 0.06130121946334839, "step": 1120 }, { "epoch": 1.047508150908244, "grad_norm": 0.80859375, "learning_rate": 2.7764347856068322e-05, "loss": 0.05010392665863037, "step": 1125 }, { "epoch": 1.0521658127619935, "grad_norm": 1.3671875, "learning_rate": 2.7545634646590937e-05, "loss": 0.06199369430541992, "step": 1130 }, { "epoch": 1.056823474615743, "grad_norm": 1.78125, "learning_rate": 2.7327052689883128e-05, "loss": 0.05268788933753967, "step": 1135 }, { "epoch": 1.0614811364694923, "grad_norm": 1.328125, "learning_rate": 2.7108613675112173e-05, "loss": 0.059926819801330564, "step": 1140 }, { "epoch": 1.0661387983232418, "grad_norm": 1.109375, "learning_rate": 2.6890329283801237e-05, "loss": 0.06648289561271667, "step": 1145 }, { "epoch": 1.0707964601769913, "grad_norm": 1.328125, "learning_rate": 2.6672211189204604e-05, "loss": 0.0454281896352768, "step": 1150 }, { "epoch": 1.0754541220307405, "grad_norm": 1.1796875, "learning_rate": 2.6454271055683478e-05, "loss": 0.06076472997665405, "step": 1155 }, { "epoch": 1.08011178388449, "grad_norm": 0.921875, "learning_rate": 2.6236520538082193e-05, "loss": 0.0678758978843689, "step": 1160 }, { "epoch": 1.0847694457382393, "grad_norm": 1.53125, "learning_rate": 2.601897128110494e-05, "loss": 0.05920932292938232, "step": 1165 }, { "epoch": 1.0894271075919888, "grad_norm": 1.0703125, "learning_rate": 2.5801634918693032e-05, "loss": 0.07065654397010804, "step": 1170 }, { "epoch": 1.0940847694457383, "grad_norm": 0.94921875, "learning_rate": 2.558452307340276e-05, "loss": 0.05684482455253601, "step": 1175 }, { "epoch": 1.0987424312994876, "grad_norm": 1.0390625, "learning_rate": 2.5367647355783883e-05, "loss": 0.04538664221763611, "step": 1180 }, { "epoch": 1.103400093153237, "grad_norm": 1.3828125, "learning_rate": 2.5151019363758657e-05, "loss": 0.04917815029621124, "step": 1185 }, { "epoch": 1.1080577550069866, "grad_norm": 1.390625, "learning_rate": 2.493465068200167e-05, "loss": 0.0480605810880661, "step": 1190 }, { "epoch": 1.1127154168607358, "grad_norm": 7.3125, "learning_rate": 2.4718552881320275e-05, "loss": 0.07286378145217895, "step": 1195 }, { "epoch": 1.1173730787144853, "grad_norm": 1.0, "learning_rate": 2.450273751803589e-05, "loss": 0.05467050075531006, "step": 1200 }, { "epoch": 1.1220307405682348, "grad_norm": 0.96875, "learning_rate": 2.4287216133365914e-05, "loss": 0.046616628766059875, "step": 1205 }, { "epoch": 1.126688402421984, "grad_norm": 1.078125, "learning_rate": 2.407200025280657e-05, "loss": 0.04863499999046326, "step": 1210 }, { "epoch": 1.1313460642757336, "grad_norm": 1.5, "learning_rate": 2.3857101385516585e-05, "loss": 0.07006229758262635, "step": 1215 }, { "epoch": 1.136003726129483, "grad_norm": 1.1953125, "learning_rate": 2.3642531023701663e-05, "loss": 0.054609501361846925, "step": 1220 }, { "epoch": 1.1406613879832324, "grad_norm": 1.5390625, "learning_rate": 2.342830064199992e-05, "loss": 0.05989115834236145, "step": 1225 }, { "epoch": 1.1453190498369819, "grad_norm": 1.0625, "learning_rate": 2.3214421696868273e-05, "loss": 0.059736084938049314, "step": 1230 }, { "epoch": 1.1499767116907313, "grad_norm": 1.71875, "learning_rate": 2.3000905625969768e-05, "loss": 0.06554105281829833, "step": 1235 }, { "epoch": 1.1546343735444806, "grad_norm": 1.0546875, "learning_rate": 2.278776384756193e-05, "loss": 0.057540059089660645, "step": 1240 }, { "epoch": 1.1592920353982301, "grad_norm": 1.390625, "learning_rate": 2.257500775988612e-05, "loss": 0.050140655040740965, "step": 1245 }, { "epoch": 1.1639496972519794, "grad_norm": 1.1796875, "learning_rate": 2.2362648740558042e-05, "loss": 0.056704151630401614, "step": 1250 }, { "epoch": 1.1686073591057289, "grad_norm": 1.0546875, "learning_rate": 2.215069814595926e-05, "loss": 0.06525537371635437, "step": 1255 }, { "epoch": 1.1732650209594784, "grad_norm": 1.375, "learning_rate": 2.193916731062988e-05, "loss": 0.05826699137687683, "step": 1260 }, { "epoch": 1.1779226828132279, "grad_norm": 1.171875, "learning_rate": 2.1728067546662422e-05, "loss": 0.046400585770606996, "step": 1265 }, { "epoch": 1.1825803446669771, "grad_norm": 1.0546875, "learning_rate": 2.1517410143096922e-05, "loss": 0.0535780668258667, "step": 1270 }, { "epoch": 1.1872380065207266, "grad_norm": 1.09375, "learning_rate": 2.1307206365317133e-05, "loss": 0.046080824732780454, "step": 1275 }, { "epoch": 1.191895668374476, "grad_norm": 1.53125, "learning_rate": 2.10974674544482e-05, "loss": 0.07175366878509522, "step": 1280 }, { "epoch": 1.1965533302282254, "grad_norm": 1.0234375, "learning_rate": 2.0888204626755408e-05, "loss": 0.06331425905227661, "step": 1285 }, { "epoch": 1.201210992081975, "grad_norm": 1.203125, "learning_rate": 2.0679429073044408e-05, "loss": 0.0636473536491394, "step": 1290 }, { "epoch": 1.2058686539357242, "grad_norm": 1.015625, "learning_rate": 2.0471151958062806e-05, "loss": 0.03870426416397095, "step": 1295 }, { "epoch": 1.2105263157894737, "grad_norm": 0.99609375, "learning_rate": 2.0263384419903027e-05, "loss": 0.03741786777973175, "step": 1300 }, { "epoch": 1.2151839776432232, "grad_norm": 1.3359375, "learning_rate": 2.0056137569406722e-05, "loss": 0.05014714598655701, "step": 1305 }, { "epoch": 1.2198416394969724, "grad_norm": 1.0546875, "learning_rate": 1.9849422489570593e-05, "loss": 0.042108118534088135, "step": 1310 }, { "epoch": 1.224499301350722, "grad_norm": 1.5703125, "learning_rate": 1.9643250234953714e-05, "loss": 0.04169891178607941, "step": 1315 }, { "epoch": 1.2291569632044714, "grad_norm": 1.1484375, "learning_rate": 1.943763183108631e-05, "loss": 0.06241189241409302, "step": 1320 }, { "epoch": 1.2338146250582207, "grad_norm": 0.99609375, "learning_rate": 1.9232578273880195e-05, "loss": 0.04898407459259033, "step": 1325 }, { "epoch": 1.2384722869119702, "grad_norm": 1.296875, "learning_rate": 1.9028100529040728e-05, "loss": 0.05555582642555237, "step": 1330 }, { "epoch": 1.2431299487657197, "grad_norm": 1.25, "learning_rate": 1.8824209531480384e-05, "loss": 0.05224438309669495, "step": 1335 }, { "epoch": 1.247787610619469, "grad_norm": 1.40625, "learning_rate": 1.8620916184733993e-05, "loss": 0.04730784296989441, "step": 1340 }, { "epoch": 1.2524452724732185, "grad_norm": 1.2578125, "learning_rate": 1.8418231360375647e-05, "loss": 0.045959821343421935, "step": 1345 }, { "epoch": 1.257102934326968, "grad_norm": 0.921875, "learning_rate": 1.8216165897437335e-05, "loss": 0.046253901720046994, "step": 1350 }, { "epoch": 1.2617605961807172, "grad_norm": 1.1484375, "learning_rate": 1.8014730601829263e-05, "loss": 0.05002530813217163, "step": 1355 }, { "epoch": 1.2664182580344667, "grad_norm": 0.9921875, "learning_rate": 1.781393624576203e-05, "loss": 0.058111125230789186, "step": 1360 }, { "epoch": 1.271075919888216, "grad_norm": 1.2734375, "learning_rate": 1.7613793567170495e-05, "loss": 0.057514190673828125, "step": 1365 }, { "epoch": 1.2757335817419655, "grad_norm": 1.046875, "learning_rate": 1.741431326913963e-05, "loss": 0.05811060070991516, "step": 1370 }, { "epoch": 1.280391243595715, "grad_norm": 1.0078125, "learning_rate": 1.7215506019332065e-05, "loss": 0.04376150071620941, "step": 1375 }, { "epoch": 1.2850489054494645, "grad_norm": 1.2734375, "learning_rate": 1.7017382449417665e-05, "loss": 0.05309507846832275, "step": 1380 }, { "epoch": 1.2897065673032138, "grad_norm": 0.984375, "learning_rate": 1.6819953154504952e-05, "loss": 0.04199279248714447, "step": 1385 }, { "epoch": 1.2943642291569633, "grad_norm": 0.98046875, "learning_rate": 1.6623228692574524e-05, "loss": 0.044170570373535153, "step": 1390 }, { "epoch": 1.2990218910107125, "grad_norm": 0.81640625, "learning_rate": 1.6427219583914396e-05, "loss": 0.040523186326026917, "step": 1395 }, { "epoch": 1.303679552864462, "grad_norm": 1.3671875, "learning_rate": 1.6231936310557472e-05, "loss": 0.06709733009338378, "step": 1400 }, { "epoch": 1.3083372147182115, "grad_norm": 1.4609375, "learning_rate": 1.6037389315720967e-05, "loss": 0.06956819295883179, "step": 1405 }, { "epoch": 1.312994876571961, "grad_norm": 1.5078125, "learning_rate": 1.5843589003247902e-05, "loss": 0.0630099892616272, "step": 1410 }, { "epoch": 1.3176525384257103, "grad_norm": 0.98046875, "learning_rate": 1.565054573705076e-05, "loss": 0.038538819551467894, "step": 1415 }, { "epoch": 1.3223102002794598, "grad_norm": 1.2890625, "learning_rate": 1.545826984055731e-05, "loss": 0.038901329040527344, "step": 1420 }, { "epoch": 1.326967862133209, "grad_norm": 1.140625, "learning_rate": 1.52667715961584e-05, "loss": 0.07003207802772522, "step": 1425 }, { "epoch": 1.3316255239869585, "grad_norm": 1.953125, "learning_rate": 1.5076061244658204e-05, "loss": 0.049628344178199765, "step": 1430 }, { "epoch": 1.336283185840708, "grad_norm": 1.96875, "learning_rate": 1.4886148984726551e-05, "loss": 0.05139703750610351, "step": 1435 }, { "epoch": 1.3409408476944573, "grad_norm": 1.015625, "learning_rate": 1.4697044972353454e-05, "loss": 0.05404648780822754, "step": 1440 }, { "epoch": 1.3455985095482068, "grad_norm": 1.015625, "learning_rate": 1.4508759320306097e-05, "loss": 0.042486587166786195, "step": 1445 }, { "epoch": 1.350256171401956, "grad_norm": 2.515625, "learning_rate": 1.4321302097587964e-05, "loss": 0.05852026343345642, "step": 1450 }, { "epoch": 1.3549138332557056, "grad_norm": 1.53125, "learning_rate": 1.4134683328900397e-05, "loss": 0.062414920330047606, "step": 1455 }, { "epoch": 1.359571495109455, "grad_norm": 0.83984375, "learning_rate": 1.394891299410652e-05, "loss": 0.0492287278175354, "step": 1460 }, { "epoch": 1.3642291569632046, "grad_norm": 0.9765625, "learning_rate": 1.3764001027697486e-05, "loss": 0.057495498657226564, "step": 1465 }, { "epoch": 1.3688868188169538, "grad_norm": 1.734375, "learning_rate": 1.357995731826128e-05, "loss": 0.04979708790779114, "step": 1470 }, { "epoch": 1.3735444806707033, "grad_norm": 1.109375, "learning_rate": 1.3396791707953873e-05, "loss": 0.05228898525238037, "step": 1475 }, { "epoch": 1.3782021425244526, "grad_norm": 1.265625, "learning_rate": 1.321451399197284e-05, "loss": 0.06390035152435303, "step": 1480 }, { "epoch": 1.382859804378202, "grad_norm": 0.91015625, "learning_rate": 1.3033133918033669e-05, "loss": 0.053880298137664796, "step": 1485 }, { "epoch": 1.3875174662319516, "grad_norm": 1.765625, "learning_rate": 1.2852661185848319e-05, "loss": 0.06999391913414002, "step": 1490 }, { "epoch": 1.392175128085701, "grad_norm": 1.53125, "learning_rate": 1.2673105446606623e-05, "loss": 0.05712270736694336, "step": 1495 }, { "epoch": 1.3968327899394504, "grad_norm": 0.85546875, "learning_rate": 1.2494476302460142e-05, "loss": 0.048511195182800296, "step": 1500 }, { "epoch": 1.4014904517931999, "grad_norm": 1.578125, "learning_rate": 1.2316783306008618e-05, "loss": 0.05195020437240601, "step": 1505 }, { "epoch": 1.4061481136469491, "grad_norm": 2.703125, "learning_rate": 1.2140035959789195e-05, "loss": 0.06756129860877991, "step": 1510 }, { "epoch": 1.4108057755006986, "grad_norm": 2.859375, "learning_rate": 1.1964243715768218e-05, "loss": 0.06034356355667114, "step": 1515 }, { "epoch": 1.4154634373544481, "grad_norm": 1.2578125, "learning_rate": 1.1789415974835764e-05, "loss": 0.0402948796749115, "step": 1520 }, { "epoch": 1.4201210992081974, "grad_norm": 1.3125, "learning_rate": 1.1615562086302912e-05, "loss": 0.06786430478096009, "step": 1525 }, { "epoch": 1.424778761061947, "grad_norm": 0.859375, "learning_rate": 1.1442691347401783e-05, "loss": 0.04251951277256012, "step": 1530 }, { "epoch": 1.4294364229156964, "grad_norm": 1.53125, "learning_rate": 1.1270813002788303e-05, "loss": 0.06105688214302063, "step": 1535 }, { "epoch": 1.4340940847694457, "grad_norm": 1.25, "learning_rate": 1.1099936244047885e-05, "loss": 0.055726665258407596, "step": 1540 }, { "epoch": 1.4387517466231952, "grad_norm": 1.4375, "learning_rate": 1.0930070209203867e-05, "loss": 0.05416175723075867, "step": 1545 }, { "epoch": 1.4434094084769447, "grad_norm": 1.1484375, "learning_rate": 1.0761223982228775e-05, "loss": 0.05864529609680176, "step": 1550 }, { "epoch": 1.448067070330694, "grad_norm": 1.984375, "learning_rate": 1.0593406592558674e-05, "loss": 0.06089619398117065, "step": 1555 }, { "epoch": 1.4527247321844434, "grad_norm": 0.9375, "learning_rate": 1.042662701461013e-05, "loss": 0.047282624244689944, "step": 1560 }, { "epoch": 1.4573823940381927, "grad_norm": 0.8828125, "learning_rate": 1.0260894167300419e-05, "loss": 0.0594075083732605, "step": 1565 }, { "epoch": 1.4620400558919422, "grad_norm": 0.828125, "learning_rate": 1.0096216913570506e-05, "loss": 0.057616645097732545, "step": 1570 }, { "epoch": 1.4666977177456917, "grad_norm": 1.125, "learning_rate": 9.932604059911067e-06, "loss": 0.03644887804985046, "step": 1575 }, { "epoch": 1.4713553795994412, "grad_norm": 0.90234375, "learning_rate": 9.770064355891585e-06, "loss": 0.0486981213092804, "step": 1580 }, { "epoch": 1.4760130414531905, "grad_norm": 0.859375, "learning_rate": 9.608606493692426e-06, "loss": 0.04411686062812805, "step": 1585 }, { "epoch": 1.48067070330694, "grad_norm": 1.8671875, "learning_rate": 9.448239107639987e-06, "loss": 0.04628075659275055, "step": 1590 }, { "epoch": 1.4853283651606892, "grad_norm": 1.3046875, "learning_rate": 9.288970773744997e-06, "loss": 0.05574510097503662, "step": 1595 }, { "epoch": 1.4899860270144387, "grad_norm": 1.796875, "learning_rate": 9.13081000924384e-06, "loss": 0.05554535984992981, "step": 1600 }, { "epoch": 1.4946436888681882, "grad_norm": 1.2265625, "learning_rate": 8.973765272143129e-06, "loss": 0.04478040039539337, "step": 1605 }, { "epoch": 1.4993013507219377, "grad_norm": 0.87890625, "learning_rate": 8.817844960767386e-06, "loss": 0.04596472382545471, "step": 1610 }, { "epoch": 1.503959012575687, "grad_norm": 6.875, "learning_rate": 8.663057413309858e-06, "loss": 0.05729023814201355, "step": 1615 }, { "epoch": 1.5086166744294365, "grad_norm": 1.0, "learning_rate": 8.509410907386743e-06, "loss": 0.04612794816493988, "step": 1620 }, { "epoch": 1.5132743362831858, "grad_norm": 1.1171875, "learning_rate": 8.356913659594373e-06, "loss": 0.04824492633342743, "step": 1625 }, { "epoch": 1.5179319981369352, "grad_norm": 4.0, "learning_rate": 8.20557382506993e-06, "loss": 0.05382657051086426, "step": 1630 }, { "epoch": 1.5225896599906847, "grad_norm": 1.078125, "learning_rate": 8.05539949705529e-06, "loss": 0.042561736702919004, "step": 1635 }, { "epoch": 1.5272473218444342, "grad_norm": 1.1171875, "learning_rate": 7.90639870646418e-06, "loss": 0.04216844141483307, "step": 1640 }, { "epoch": 1.5319049836981835, "grad_norm": 0.7890625, "learning_rate": 7.758579421452786e-06, "loss": 0.06024673581123352, "step": 1645 }, { "epoch": 1.5365626455519328, "grad_norm": 0.76953125, "learning_rate": 7.611949546993579e-06, "loss": 0.03899524211883545, "step": 1650 }, { "epoch": 1.5412203074056823, "grad_norm": 0.92578125, "learning_rate": 7.466516924452593e-06, "loss": 0.05587719082832336, "step": 1655 }, { "epoch": 1.5458779692594318, "grad_norm": 1.0078125, "learning_rate": 7.322289331170099e-06, "loss": 0.06056018471717835, "step": 1660 }, { "epoch": 1.5505356311131813, "grad_norm": 1.3046875, "learning_rate": 7.179274480044695e-06, "loss": 0.06227887272834778, "step": 1665 }, { "epoch": 1.5551932929669308, "grad_norm": 1.578125, "learning_rate": 7.037480019120803e-06, "loss": 0.05833619236946106, "step": 1670 }, { "epoch": 1.55985095482068, "grad_norm": 1.3515625, "learning_rate": 6.896913531179733e-06, "loss": 0.06514707803726197, "step": 1675 }, { "epoch": 1.5645086166744293, "grad_norm": 1.3515625, "learning_rate": 6.757582533334142e-06, "loss": 0.0441954642534256, "step": 1680 }, { "epoch": 1.5691662785281788, "grad_norm": 1.296875, "learning_rate": 6.619494476626022e-06, "loss": 0.04353019595146179, "step": 1685 }, { "epoch": 1.5738239403819283, "grad_norm": 1.4453125, "learning_rate": 6.4826567456283095e-06, "loss": 0.06212781667709351, "step": 1690 }, { "epoch": 1.5784816022356778, "grad_norm": 1.25, "learning_rate": 6.347076658049872e-06, "loss": 0.04992106556892395, "step": 1695 }, { "epoch": 1.583139264089427, "grad_norm": 1.4375, "learning_rate": 6.212761464344267e-06, "loss": 0.05654107928276062, "step": 1700 }, { "epoch": 1.5877969259431766, "grad_norm": 1.171875, "learning_rate": 6.079718347321983e-06, "loss": 0.04213224053382873, "step": 1705 }, { "epoch": 1.5924545877969258, "grad_norm": 1.3359375, "learning_rate": 5.947954421766284e-06, "loss": 0.05712945461273193, "step": 1710 }, { "epoch": 1.5971122496506753, "grad_norm": 1.5, "learning_rate": 5.81747673405278e-06, "loss": 0.06145240068435669, "step": 1715 }, { "epoch": 1.6017699115044248, "grad_norm": 2.859375, "learning_rate": 5.688292261772593e-06, "loss": 0.05003199577331543, "step": 1720 }, { "epoch": 1.6064275733581743, "grad_norm": 1.0078125, "learning_rate": 5.560407913359198e-06, "loss": 0.05077915191650391, "step": 1725 }, { "epoch": 1.6110852352119236, "grad_norm": 1.0234375, "learning_rate": 5.4338305277189995e-06, "loss": 0.05965242385864258, "step": 1730 }, { "epoch": 1.6157428970656729, "grad_norm": 0.9921875, "learning_rate": 5.308566873865567e-06, "loss": 0.05326489806175232, "step": 1735 }, { "epoch": 1.6204005589194224, "grad_norm": 0.79296875, "learning_rate": 5.184623650557708e-06, "loss": 0.04797676503658295, "step": 1740 }, { "epoch": 1.6250582207731719, "grad_norm": 1.453125, "learning_rate": 5.062007485941196e-06, "loss": 0.05377109050750732, "step": 1745 }, { "epoch": 1.6297158826269214, "grad_norm": 1.3671875, "learning_rate": 4.9407249371943e-06, "loss": 0.06788500547409057, "step": 1750 }, { "epoch": 1.6343735444806708, "grad_norm": 1.3828125, "learning_rate": 4.820782490177166e-06, "loss": 0.0571264922618866, "step": 1755 }, { "epoch": 1.6390312063344201, "grad_norm": 1.4140625, "learning_rate": 4.702186559084972e-06, "loss": 0.05916694402694702, "step": 1760 }, { "epoch": 1.6436888681881694, "grad_norm": 1.0703125, "learning_rate": 4.584943486104851e-06, "loss": 0.051477736234664916, "step": 1765 }, { "epoch": 1.648346530041919, "grad_norm": 1.9296875, "learning_rate": 4.469059541076806e-06, "loss": 0.05232837200164795, "step": 1770 }, { "epoch": 1.6530041918956684, "grad_norm": 1.0, "learning_rate": 4.3545409211583554e-06, "loss": 0.047325053811073305, "step": 1775 }, { "epoch": 1.6576618537494179, "grad_norm": 1.4140625, "learning_rate": 4.24139375049317e-06, "loss": 0.052236026525497435, "step": 1780 }, { "epoch": 1.6623195156031674, "grad_norm": 1.03125, "learning_rate": 4.129624079883549e-06, "loss": 0.055283254384994505, "step": 1785 }, { "epoch": 1.6669771774569166, "grad_norm": 1.140625, "learning_rate": 4.019237886466839e-06, "loss": 0.044303908944129944, "step": 1790 }, { "epoch": 1.671634839310666, "grad_norm": 1.34375, "learning_rate": 3.910241073395805e-06, "loss": 0.06972193121910095, "step": 1795 }, { "epoch": 1.6762925011644154, "grad_norm": 1.6640625, "learning_rate": 3.8026394695229316e-06, "loss": 0.05040590763092041, "step": 1800 }, { "epoch": 1.680950163018165, "grad_norm": 1.1328125, "learning_rate": 3.696438829088713e-06, "loss": 0.060897302627563474, "step": 1805 }, { "epoch": 1.6856078248719144, "grad_norm": 1.0859375, "learning_rate": 3.591644831413946e-06, "loss": 0.06547480821609497, "step": 1810 }, { "epoch": 1.6902654867256637, "grad_norm": 0.875, "learning_rate": 3.48826308059601e-06, "loss": 0.040547126531600954, "step": 1815 }, { "epoch": 1.6949231485794132, "grad_norm": 1.3125, "learning_rate": 3.3862991052091508e-06, "loss": 0.04342193007469177, "step": 1820 }, { "epoch": 1.6995808104331624, "grad_norm": 1.03125, "learning_rate": 3.2857583580088814e-06, "loss": 0.040316405892372134, "step": 1825 }, { "epoch": 1.704238472286912, "grad_norm": 1.0234375, "learning_rate": 3.1866462156403177e-06, "loss": 0.049350231885910034, "step": 1830 }, { "epoch": 1.7088961341406614, "grad_norm": 1.2578125, "learning_rate": 3.088967978350704e-06, "loss": 0.06080756783485412, "step": 1835 }, { "epoch": 1.713553795994411, "grad_norm": 1.140625, "learning_rate": 2.9927288697059475e-06, "loss": 0.0580963134765625, "step": 1840 }, { "epoch": 1.7182114578481602, "grad_norm": 1.1640625, "learning_rate": 2.8979340363112592e-06, "loss": 0.05519581437110901, "step": 1845 }, { "epoch": 1.7228691197019095, "grad_norm": 1.4453125, "learning_rate": 2.804588547535968e-06, "loss": 0.04703545570373535, "step": 1850 }, { "epoch": 1.727526781555659, "grad_norm": 1.234375, "learning_rate": 2.7126973952423927e-06, "loss": 0.06626067757606506, "step": 1855 }, { "epoch": 1.7321844434094085, "grad_norm": 1.171875, "learning_rate": 2.622265493518907e-06, "loss": 0.06088282465934754, "step": 1860 }, { "epoch": 1.736842105263158, "grad_norm": 0.953125, "learning_rate": 2.533297678417137e-06, "loss": 0.04956701695919037, "step": 1865 }, { "epoch": 1.7414997671169075, "grad_norm": 1.375, "learning_rate": 2.445798707693344e-06, "loss": 0.06391797065734864, "step": 1870 }, { "epoch": 1.7461574289706567, "grad_norm": 1.421875, "learning_rate": 2.3597732605539945e-06, "loss": 0.054262596368789676, "step": 1875 }, { "epoch": 1.750815090824406, "grad_norm": 1.03125, "learning_rate": 2.2752259374055473e-06, "loss": 0.04359312951564789, "step": 1880 }, { "epoch": 1.7554727526781555, "grad_norm": 0.84375, "learning_rate": 2.1921612596083917e-06, "loss": 0.053888702392578126, "step": 1885 }, { "epoch": 1.760130414531905, "grad_norm": 2.578125, "learning_rate": 2.110583669235098e-06, "loss": 0.05314147472381592, "step": 1890 }, { "epoch": 1.7647880763856545, "grad_norm": 1.171875, "learning_rate": 2.030497528832872e-06, "loss": 0.05475490689277649, "step": 1895 }, { "epoch": 1.7694457382394038, "grad_norm": 1.6875, "learning_rate": 1.951907121190202e-06, "loss": 0.05292813777923584, "step": 1900 }, { "epoch": 1.7741034000931533, "grad_norm": 1.15625, "learning_rate": 1.8748166491078978e-06, "loss": 0.04046391844749451, "step": 1905 }, { "epoch": 1.7787610619469025, "grad_norm": 1.1640625, "learning_rate": 1.799230235174294e-06, "loss": 0.05409615635871887, "step": 1910 }, { "epoch": 1.783418723800652, "grad_norm": 1.3203125, "learning_rate": 1.7251519215447809e-06, "loss": 0.05164237022399902, "step": 1915 }, { "epoch": 1.7880763856544015, "grad_norm": 1.359375, "learning_rate": 1.6525856697256748e-06, "loss": 0.04969433844089508, "step": 1920 }, { "epoch": 1.792734047508151, "grad_norm": 1.1484375, "learning_rate": 1.581535360362334e-06, "loss": 0.05390878915786743, "step": 1925 }, { "epoch": 1.7973917093619003, "grad_norm": 1.421875, "learning_rate": 1.512004793031655e-06, "loss": 0.04839572012424469, "step": 1930 }, { "epoch": 1.8020493712156498, "grad_norm": 1.4375, "learning_rate": 1.4439976860388715e-06, "loss": 0.052365976572036746, "step": 1935 }, { "epoch": 1.806707033069399, "grad_norm": 0.8671875, "learning_rate": 1.3775176762187036e-06, "loss": 0.05246726870536804, "step": 1940 }, { "epoch": 1.8113646949231486, "grad_norm": 1.6015625, "learning_rate": 1.3125683187408798e-06, "loss": 0.06147383451461792, "step": 1945 }, { "epoch": 1.816022356776898, "grad_norm": 2.609375, "learning_rate": 1.2491530869200196e-06, "loss": 0.0531234622001648, "step": 1950 }, { "epoch": 1.8206800186306475, "grad_norm": 1.0859375, "learning_rate": 1.187275372029868e-06, "loss": 0.04805976152420044, "step": 1955 }, { "epoch": 1.8253376804843968, "grad_norm": 1.0703125, "learning_rate": 1.1269384831219843e-06, "loss": 0.056858569383621216, "step": 1960 }, { "epoch": 1.829995342338146, "grad_norm": 1.1171875, "learning_rate": 1.0681456468487172e-06, "loss": 0.058970791101455686, "step": 1965 }, { "epoch": 1.8346530041918956, "grad_norm": 1.453125, "learning_rate": 1.010900007290716e-06, "loss": 0.0543209969997406, "step": 1970 }, { "epoch": 1.839310666045645, "grad_norm": 2.59375, "learning_rate": 9.552046257887614e-07, "loss": 0.047796967625617984, "step": 1975 }, { "epoch": 1.8439683278993946, "grad_norm": 1.0546875, "learning_rate": 9.010624807800471e-07, "loss": 0.05391690135002136, "step": 1980 }, { "epoch": 1.848625989753144, "grad_norm": 1.359375, "learning_rate": 8.484764676389201e-07, "loss": 0.07577978372573853, "step": 1985 }, { "epoch": 1.8532836516068933, "grad_norm": 1.140625, "learning_rate": 7.97449398522041e-07, "loss": 0.06300070285797119, "step": 1990 }, { "epoch": 1.8579413134606426, "grad_norm": 1.5703125, "learning_rate": 7.479840022179884e-07, "loss": 0.07114969491958618, "step": 1995 }, { "epoch": 1.8625989753143921, "grad_norm": 1.453125, "learning_rate": 7.000829240013384e-07, "loss": 0.07656757831573487, "step": 2000 }, { "epoch": 1.8672566371681416, "grad_norm": 0.9921875, "learning_rate": 6.537487254911911e-07, "loss": 0.05285425186157226, "step": 2005 }, { "epoch": 1.871914299021891, "grad_norm": 1.6171875, "learning_rate": 6.089838845141971e-07, "loss": 0.05697638392448425, "step": 2010 }, { "epoch": 1.8765719608756404, "grad_norm": 1.3203125, "learning_rate": 5.657907949720464e-07, "loss": 0.05292525291442871, "step": 2015 }, { "epoch": 1.8812296227293899, "grad_norm": 1.359375, "learning_rate": 5.241717667134382e-07, "loss": 0.040792858600616454, "step": 2020 }, { "epoch": 1.8858872845831391, "grad_norm": 9.125, "learning_rate": 4.84129025410569e-07, "loss": 0.07155822515487671, "step": 2025 }, { "epoch": 1.8905449464368886, "grad_norm": 1.3984375, "learning_rate": 4.456647124401081e-07, "loss": 0.059972846508026124, "step": 2030 }, { "epoch": 1.8952026082906381, "grad_norm": 1.2890625, "learning_rate": 4.0878088476867246e-07, "loss": 0.05164983868598938, "step": 2035 }, { "epoch": 1.8998602701443876, "grad_norm": 1.171875, "learning_rate": 3.7347951484283825e-07, "loss": 0.061475759744644164, "step": 2040 }, { "epoch": 1.904517931998137, "grad_norm": 3.75, "learning_rate": 3.397624904836549e-07, "loss": 0.06030047535896301, "step": 2045 }, { "epoch": 1.9091755938518864, "grad_norm": 0.984375, "learning_rate": 3.076316147856828e-07, "loss": 0.06555002927780151, "step": 2050 }, { "epoch": 1.9138332557056357, "grad_norm": 1.71875, "learning_rate": 2.770886060205835e-07, "loss": 0.03764947056770325, "step": 2055 }, { "epoch": 1.9184909175593852, "grad_norm": 1.296875, "learning_rate": 2.481350975452168e-07, "loss": 0.04990624189376831, "step": 2060 }, { "epoch": 1.9231485794131347, "grad_norm": 1.390625, "learning_rate": 2.2077263771430378e-07, "loss": 0.06039892435073853, "step": 2065 }, { "epoch": 1.9278062412668842, "grad_norm": 1.234375, "learning_rate": 1.9500268979761982e-07, "loss": 0.06764861345291137, "step": 2070 }, { "epoch": 1.9324639031206334, "grad_norm": 1.5234375, "learning_rate": 1.7082663190173376e-07, "loss": 0.05397939085960388, "step": 2075 }, { "epoch": 1.9371215649743827, "grad_norm": 1.4453125, "learning_rate": 1.482457568963369e-07, "loss": 0.058798480033874514, "step": 2080 }, { "epoch": 1.9417792268281322, "grad_norm": 1.4140625, "learning_rate": 1.27261272345075e-07, "loss": 0.036405691504478456, "step": 2085 }, { "epoch": 1.9464368886818817, "grad_norm": 1.1640625, "learning_rate": 1.0787430044098656e-07, "loss": 0.05242518186569214, "step": 2090 }, { "epoch": 1.9510945505356312, "grad_norm": 1.265625, "learning_rate": 9.008587794649081e-08, "loss": 0.06891694664955139, "step": 2095 }, { "epoch": 1.9557522123893807, "grad_norm": 1.6796875, "learning_rate": 7.389695613793879e-08, "loss": 0.060939884185791014, "step": 2100 }, { "epoch": 1.96040987424313, "grad_norm": 1.90625, "learning_rate": 5.9308400754747304e-08, "loss": 0.056053650379180905, "step": 2105 }, { "epoch": 1.9650675360968792, "grad_norm": 1.1484375, "learning_rate": 4.63209919530927e-08, "loss": 0.04752773642539978, "step": 2110 }, { "epoch": 1.9697251979506287, "grad_norm": 2.0625, "learning_rate": 3.493542426419749e-08, "loss": 0.05915160179138183, "step": 2115 }, { "epoch": 1.9743828598043782, "grad_norm": 1.28125, "learning_rate": 2.5152306557196802e-08, "loss": 0.052164262533187865, "step": 2120 }, { "epoch": 1.9790405216581277, "grad_norm": 1.15625, "learning_rate": 1.697216200655438e-08, "loss": 0.05432420372962952, "step": 2125 }, { "epoch": 1.983698183511877, "grad_norm": 1.4453125, "learning_rate": 1.0395428064108315e-08, "loss": 0.05430280566215515, "step": 2130 }, { "epoch": 1.9883558453656265, "grad_norm": 1.203125, "learning_rate": 5.422456435666412e-09, "loss": 0.04979383051395416, "step": 2135 }, { "epoch": 1.9930135072193758, "grad_norm": 1.0390625, "learning_rate": 2.05351306219459e-09, "loss": 0.05388739109039307, "step": 2140 }, { "epoch": 1.9976711690731253, "grad_norm": 1.0625, "learning_rate": 2.887781056015726e-10, "loss": 0.06473076939582825, "step": 2145 }, { "epoch": 2.0, "eval_loss": 0.06230008229613304, "eval_pearson_cosine": 0.5569300055503845, "eval_pearson_dot_product": 0.5569300651550293, "eval_pearson_euclidean": 0.5638938546180725, "eval_pearson_manhattan": 0.5624173879623413, "eval_runtime": 17.219, "eval_samples_per_second": 118.126, "eval_spearman_cosine": 0.5637662854873432, "eval_spearman_dot_product": 0.5637662852863563, "eval_spearman_euclidean": 0.5637662854873432, "eval_spearman_manhattan": 0.5633536421411317, "eval_steps_per_second": 19.688, "step": 2148 }, { "epoch": 2.0, "eval_loss": 0.06230008229613304, "eval_pearson_cosine": 0.5569300055503845, "eval_pearson_dot_product": 0.5569300651550293, "eval_pearson_euclidean": 0.5638938546180725, "eval_pearson_manhattan": 0.5624173879623413, "eval_runtime": 17.4857, "eval_samples_per_second": 116.324, "eval_spearman_cosine": 0.5637662854873432, "eval_spearman_dot_product": 0.5637662852863563, "eval_spearman_euclidean": 0.5637662854873432, "eval_spearman_manhattan": 0.5633536421411317, "eval_steps_per_second": 19.387, "step": 2148 } ], "logging_steps": 5, "max_steps": 2148, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.5428717251657728e+17, "train_batch_size": 6, "trial_name": null, "trial_params": null }