FIRM-Gen-8B / trainer_state.json
zpy777's picture
Upload folder using huggingface_hub
7e157e9 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 500,
"global_step": 3624,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0027597626604112047,
"grad_norm": 9.687899712924036,
"learning_rate": 2.4793388429752067e-07,
"loss": 1.2489,
"step": 10
},
{
"epoch": 0.005519525320822409,
"grad_norm": 7.993583387794843,
"learning_rate": 5.234159779614326e-07,
"loss": 1.23,
"step": 20
},
{
"epoch": 0.008279287981233613,
"grad_norm": 4.419911118410825,
"learning_rate": 7.988980716253444e-07,
"loss": 1.134,
"step": 30
},
{
"epoch": 0.011039050641644819,
"grad_norm": 3.0732247545437965,
"learning_rate": 1.0743801652892562e-06,
"loss": 0.993,
"step": 40
},
{
"epoch": 0.013798813302056023,
"grad_norm": 2.005789581836949,
"learning_rate": 1.3498622589531682e-06,
"loss": 0.8782,
"step": 50
},
{
"epoch": 0.016558575962467226,
"grad_norm": 1.3028620164242184,
"learning_rate": 1.62534435261708e-06,
"loss": 0.8004,
"step": 60
},
{
"epoch": 0.019318338622878434,
"grad_norm": 1.2871645472343145,
"learning_rate": 1.900826446280992e-06,
"loss": 0.7618,
"step": 70
},
{
"epoch": 0.022078101283289638,
"grad_norm": 2.0977280137118854,
"learning_rate": 2.1763085399449038e-06,
"loss": 0.7363,
"step": 80
},
{
"epoch": 0.02483786394370084,
"grad_norm": 1.235679470482907,
"learning_rate": 2.4517906336088157e-06,
"loss": 0.7218,
"step": 90
},
{
"epoch": 0.027597626604112045,
"grad_norm": 1.2008847267657137,
"learning_rate": 2.7272727272727272e-06,
"loss": 0.7039,
"step": 100
},
{
"epoch": 0.030357389264523253,
"grad_norm": 1.2865692124934534,
"learning_rate": 3.002754820936639e-06,
"loss": 0.7016,
"step": 110
},
{
"epoch": 0.03311715192493445,
"grad_norm": 1.225519692241417,
"learning_rate": 3.278236914600551e-06,
"loss": 0.6911,
"step": 120
},
{
"epoch": 0.035876914585345664,
"grad_norm": 1.1074744380119828,
"learning_rate": 3.553719008264463e-06,
"loss": 0.6837,
"step": 130
},
{
"epoch": 0.03863667724575687,
"grad_norm": 1.2255853901709166,
"learning_rate": 3.8292011019283746e-06,
"loss": 0.6733,
"step": 140
},
{
"epoch": 0.04139643990616807,
"grad_norm": 1.2160790653325682,
"learning_rate": 4.104683195592287e-06,
"loss": 0.6663,
"step": 150
},
{
"epoch": 0.044156202566579275,
"grad_norm": 1.2160126107257938,
"learning_rate": 4.3801652892561984e-06,
"loss": 0.6672,
"step": 160
},
{
"epoch": 0.04691596522699048,
"grad_norm": 1.3030345619512333,
"learning_rate": 4.655647382920111e-06,
"loss": 0.6604,
"step": 170
},
{
"epoch": 0.04967572788740168,
"grad_norm": 1.1675762652939756,
"learning_rate": 4.931129476584022e-06,
"loss": 0.654,
"step": 180
},
{
"epoch": 0.05243549054781289,
"grad_norm": 1.3421701645033057,
"learning_rate": 5.206611570247935e-06,
"loss": 0.653,
"step": 190
},
{
"epoch": 0.05519525320822409,
"grad_norm": 1.1987056035659454,
"learning_rate": 5.482093663911846e-06,
"loss": 0.6508,
"step": 200
},
{
"epoch": 0.057955015868635294,
"grad_norm": 1.3869845490974586,
"learning_rate": 5.7575757575757586e-06,
"loss": 0.6498,
"step": 210
},
{
"epoch": 0.060714778529046505,
"grad_norm": 1.116588718762935,
"learning_rate": 6.03305785123967e-06,
"loss": 0.6453,
"step": 220
},
{
"epoch": 0.06347454118945771,
"grad_norm": 1.0960389471025322,
"learning_rate": 6.3085399449035824e-06,
"loss": 0.6414,
"step": 230
},
{
"epoch": 0.0662343038498689,
"grad_norm": 1.2133969264114732,
"learning_rate": 6.584022038567494e-06,
"loss": 0.6345,
"step": 240
},
{
"epoch": 0.06899406651028012,
"grad_norm": 1.2700545101041685,
"learning_rate": 6.859504132231406e-06,
"loss": 0.6386,
"step": 250
},
{
"epoch": 0.07175382917069133,
"grad_norm": 1.2174406212987252,
"learning_rate": 7.134986225895317e-06,
"loss": 0.6403,
"step": 260
},
{
"epoch": 0.07451359183110252,
"grad_norm": 1.1500811137175568,
"learning_rate": 7.410468319559229e-06,
"loss": 0.6361,
"step": 270
},
{
"epoch": 0.07727335449151373,
"grad_norm": 1.2075741428461082,
"learning_rate": 7.685950413223142e-06,
"loss": 0.6348,
"step": 280
},
{
"epoch": 0.08003311715192493,
"grad_norm": 1.6661536273928046,
"learning_rate": 7.961432506887054e-06,
"loss": 0.6263,
"step": 290
},
{
"epoch": 0.08279287981233614,
"grad_norm": 1.2692327811914956,
"learning_rate": 8.236914600550965e-06,
"loss": 0.6294,
"step": 300
},
{
"epoch": 0.08555264247274734,
"grad_norm": 1.3688551805953726,
"learning_rate": 8.512396694214877e-06,
"loss": 0.6268,
"step": 310
},
{
"epoch": 0.08831240513315855,
"grad_norm": 1.3470949176246652,
"learning_rate": 8.787878787878788e-06,
"loss": 0.6247,
"step": 320
},
{
"epoch": 0.09107216779356975,
"grad_norm": 1.136787185698281,
"learning_rate": 9.063360881542702e-06,
"loss": 0.619,
"step": 330
},
{
"epoch": 0.09383193045398096,
"grad_norm": 1.1593643608420736,
"learning_rate": 9.338842975206613e-06,
"loss": 0.6262,
"step": 340
},
{
"epoch": 0.09659169311439217,
"grad_norm": 1.1708021309308343,
"learning_rate": 9.614325068870525e-06,
"loss": 0.6241,
"step": 350
},
{
"epoch": 0.09935145577480337,
"grad_norm": 1.1394208261112413,
"learning_rate": 9.889807162534436e-06,
"loss": 0.6247,
"step": 360
},
{
"epoch": 0.10211121843521458,
"grad_norm": 1.243982861233763,
"learning_rate": 9.999916470583429e-06,
"loss": 0.6223,
"step": 370
},
{
"epoch": 0.10487098109562577,
"grad_norm": 1.0950836157642558,
"learning_rate": 9.999406023144514e-06,
"loss": 0.6172,
"step": 380
},
{
"epoch": 0.10763074375603698,
"grad_norm": 1.1223718246628647,
"learning_rate": 9.998431580815314e-06,
"loss": 0.6178,
"step": 390
},
{
"epoch": 0.11039050641644818,
"grad_norm": 1.2801215739716154,
"learning_rate": 9.996993234033826e-06,
"loss": 0.6187,
"step": 400
},
{
"epoch": 0.11315026907685939,
"grad_norm": 1.1553035959732303,
"learning_rate": 9.995091116293022e-06,
"loss": 0.6195,
"step": 410
},
{
"epoch": 0.11591003173727059,
"grad_norm": 1.163353982074495,
"learning_rate": 9.992725404128452e-06,
"loss": 0.6162,
"step": 420
},
{
"epoch": 0.1186697943976818,
"grad_norm": 1.0553152985762932,
"learning_rate": 9.989896317101873e-06,
"loss": 0.6075,
"step": 430
},
{
"epoch": 0.12142955705809301,
"grad_norm": 1.061674081293769,
"learning_rate": 9.986604117780861e-06,
"loss": 0.6157,
"step": 440
},
{
"epoch": 0.1241893197185042,
"grad_norm": 1.0865591957408132,
"learning_rate": 9.982849111714445e-06,
"loss": 0.6147,
"step": 450
},
{
"epoch": 0.12694908237891542,
"grad_norm": 1.039091554305776,
"learning_rate": 9.978631647404755e-06,
"loss": 0.6089,
"step": 460
},
{
"epoch": 0.12970884503932661,
"grad_norm": 1.1565672987905995,
"learning_rate": 9.973952116274664e-06,
"loss": 0.6132,
"step": 470
},
{
"epoch": 0.1324686076997378,
"grad_norm": 1.141107303982074,
"learning_rate": 9.968810952631473e-06,
"loss": 0.6077,
"step": 480
},
{
"epoch": 0.13522837036014904,
"grad_norm": 1.1858850784459474,
"learning_rate": 9.9632086336266e-06,
"loss": 0.6116,
"step": 490
},
{
"epoch": 0.13798813302056023,
"grad_norm": 1.0965450940865213,
"learning_rate": 9.957145679211288e-06,
"loss": 0.6131,
"step": 500
},
{
"epoch": 0.13798813302056023,
"eval_loss": 0.6089209318161011,
"eval_runtime": 49.9455,
"eval_samples_per_second": 58.624,
"eval_steps_per_second": 3.664,
"step": 500
},
{
"epoch": 0.14074789568097143,
"grad_norm": 1.076032821296556,
"learning_rate": 9.95062265208836e-06,
"loss": 0.606,
"step": 510
},
{
"epoch": 0.14350765834138265,
"grad_norm": 1.0371790631769926,
"learning_rate": 9.943640157659984e-06,
"loss": 0.6071,
"step": 520
},
{
"epoch": 0.14626742100179385,
"grad_norm": 1.0166966431904405,
"learning_rate": 9.936198843971493e-06,
"loss": 0.6021,
"step": 530
},
{
"epoch": 0.14902718366220505,
"grad_norm": 1.3219793417334544,
"learning_rate": 9.928299401651236e-06,
"loss": 0.5981,
"step": 540
},
{
"epoch": 0.15178694632261625,
"grad_norm": 1.0389071117117548,
"learning_rate": 9.919942563846482e-06,
"loss": 0.6005,
"step": 550
},
{
"epoch": 0.15454670898302747,
"grad_norm": 1.0246893171620508,
"learning_rate": 9.911129106155375e-06,
"loss": 0.6046,
"step": 560
},
{
"epoch": 0.15730647164343867,
"grad_norm": 1.0445099757660354,
"learning_rate": 9.901859846554955e-06,
"loss": 0.6053,
"step": 570
},
{
"epoch": 0.16006623430384986,
"grad_norm": 1.064106651611097,
"learning_rate": 9.892135645325238e-06,
"loss": 0.5984,
"step": 580
},
{
"epoch": 0.16282599696426106,
"grad_norm": 1.0462870772071302,
"learning_rate": 9.881957404969373e-06,
"loss": 0.5978,
"step": 590
},
{
"epoch": 0.16558575962467229,
"grad_norm": 1.0000683469984344,
"learning_rate": 9.871326070129885e-06,
"loss": 0.6094,
"step": 600
},
{
"epoch": 0.16834552228508348,
"grad_norm": 0.9977951236432633,
"learning_rate": 9.860242627500994e-06,
"loss": 0.5974,
"step": 610
},
{
"epoch": 0.17110528494549468,
"grad_norm": 4.8609383522246405,
"learning_rate": 9.848708105737049e-06,
"loss": 0.598,
"step": 620
},
{
"epoch": 0.1738650476059059,
"grad_norm": 1.0005981539720457,
"learning_rate": 9.836723575357056e-06,
"loss": 0.5988,
"step": 630
},
{
"epoch": 0.1766248102663171,
"grad_norm": 1.0638070414980612,
"learning_rate": 9.824290148645322e-06,
"loss": 0.5957,
"step": 640
},
{
"epoch": 0.1793845729267283,
"grad_norm": 1.0406376026434327,
"learning_rate": 9.811408979548219e-06,
"loss": 0.591,
"step": 650
},
{
"epoch": 0.1821443355871395,
"grad_norm": 1.07624118905585,
"learning_rate": 9.7980812635671e-06,
"loss": 0.5948,
"step": 660
},
{
"epoch": 0.18490409824755072,
"grad_norm": 1.049133255644815,
"learning_rate": 9.784308237647329e-06,
"loss": 0.5954,
"step": 670
},
{
"epoch": 0.18766386090796192,
"grad_norm": 1.0143310556909468,
"learning_rate": 9.770091180063489e-06,
"loss": 0.596,
"step": 680
},
{
"epoch": 0.1904236235683731,
"grad_norm": 1.0508407539599804,
"learning_rate": 9.755431410300743e-06,
"loss": 0.5871,
"step": 690
},
{
"epoch": 0.19318338622878434,
"grad_norm": 1.2937338319267142,
"learning_rate": 9.740330288932379e-06,
"loss": 0.5948,
"step": 700
},
{
"epoch": 0.19594314888919553,
"grad_norm": 1.0318704519944248,
"learning_rate": 9.724789217493514e-06,
"loss": 0.5898,
"step": 710
},
{
"epoch": 0.19870291154960673,
"grad_norm": 0.9605096481233693,
"learning_rate": 9.708809638351048e-06,
"loss": 0.5925,
"step": 720
},
{
"epoch": 0.20146267421001793,
"grad_norm": 1.0371369229103382,
"learning_rate": 9.692393034569776e-06,
"loss": 0.5917,
"step": 730
},
{
"epoch": 0.20422243687042915,
"grad_norm": 0.97367889824368,
"learning_rate": 9.675540929774751e-06,
"loss": 0.5911,
"step": 740
},
{
"epoch": 0.20698219953084035,
"grad_norm": 0.9914313419740074,
"learning_rate": 9.658254888009877e-06,
"loss": 0.5862,
"step": 750
},
{
"epoch": 0.20974196219125155,
"grad_norm": 0.9822682480321101,
"learning_rate": 9.64053651359275e-06,
"loss": 0.5871,
"step": 760
},
{
"epoch": 0.21250172485166274,
"grad_norm": 1.001214024547401,
"learning_rate": 9.622387450965758e-06,
"loss": 0.5851,
"step": 770
},
{
"epoch": 0.21526148751207397,
"grad_norm": 0.9608822251160972,
"learning_rate": 9.603809384543472e-06,
"loss": 0.5937,
"step": 780
},
{
"epoch": 0.21802125017248516,
"grad_norm": 1.003270353697929,
"learning_rate": 9.584804038556297e-06,
"loss": 0.5834,
"step": 790
},
{
"epoch": 0.22078101283289636,
"grad_norm": 1.0623920143347612,
"learning_rate": 9.56537317689046e-06,
"loss": 0.5903,
"step": 800
},
{
"epoch": 0.2235407754933076,
"grad_norm": 0.9490080408477698,
"learning_rate": 9.5455186029243e-06,
"loss": 0.5872,
"step": 810
},
{
"epoch": 0.22630053815371878,
"grad_norm": 0.944874262643946,
"learning_rate": 9.525242159360897e-06,
"loss": 0.5888,
"step": 820
},
{
"epoch": 0.22906030081412998,
"grad_norm": 0.9755675580087483,
"learning_rate": 9.504545728057046e-06,
"loss": 0.5822,
"step": 830
},
{
"epoch": 0.23182006347454118,
"grad_norm": 0.9892178439036814,
"learning_rate": 9.483431229848607e-06,
"loss": 0.5847,
"step": 840
},
{
"epoch": 0.2345798261349524,
"grad_norm": 1.3473013721931224,
"learning_rate": 9.461900624372233e-06,
"loss": 0.5763,
"step": 850
},
{
"epoch": 0.2373395887953636,
"grad_norm": 1.1368854612150283,
"learning_rate": 9.439955909883493e-06,
"loss": 0.5815,
"step": 860
},
{
"epoch": 0.2400993514557748,
"grad_norm": 1.0103375787124518,
"learning_rate": 9.417599123071417e-06,
"loss": 0.5838,
"step": 870
},
{
"epoch": 0.24285911411618602,
"grad_norm": 1.0432433126635938,
"learning_rate": 9.39483233886946e-06,
"loss": 0.5785,
"step": 880
},
{
"epoch": 0.24561887677659722,
"grad_norm": 1.1847765522018567,
"learning_rate": 9.371657670262947e-06,
"loss": 0.5742,
"step": 890
},
{
"epoch": 0.2483786394370084,
"grad_norm": 0.9243729444752771,
"learning_rate": 9.348077268092951e-06,
"loss": 0.5763,
"step": 900
},
{
"epoch": 0.2511384020974196,
"grad_norm": 1.3733927460707915,
"learning_rate": 9.324093320856679e-06,
"loss": 0.5777,
"step": 910
},
{
"epoch": 0.25389816475783084,
"grad_norm": 1.098267274159629,
"learning_rate": 9.299708054504355e-06,
"loss": 0.5781,
"step": 920
},
{
"epoch": 0.256657927418242,
"grad_norm": 0.940463162328034,
"learning_rate": 9.274923732232635e-06,
"loss": 0.5813,
"step": 930
},
{
"epoch": 0.25941769007865323,
"grad_norm": 0.936901862695869,
"learning_rate": 9.249742654274554e-06,
"loss": 0.576,
"step": 940
},
{
"epoch": 0.26217745273906445,
"grad_norm": 0.9160606268594537,
"learning_rate": 9.224167157686044e-06,
"loss": 0.5756,
"step": 950
},
{
"epoch": 0.2649372153994756,
"grad_norm": 0.902157881662283,
"learning_rate": 9.198199616129033e-06,
"loss": 0.5718,
"step": 960
},
{
"epoch": 0.26769697805988685,
"grad_norm": 0.9251021746256725,
"learning_rate": 9.171842439651143e-06,
"loss": 0.5722,
"step": 970
},
{
"epoch": 0.27045674072029807,
"grad_norm": 0.9049498340659763,
"learning_rate": 9.145098074462012e-06,
"loss": 0.577,
"step": 980
},
{
"epoch": 0.27321650338070924,
"grad_norm": 1.0139196145776603,
"learning_rate": 9.117969002706267e-06,
"loss": 0.5738,
"step": 990
},
{
"epoch": 0.27597626604112047,
"grad_norm": 0.9648188962102778,
"learning_rate": 9.090457742233152e-06,
"loss": 0.5714,
"step": 1000
},
{
"epoch": 0.27597626604112047,
"eval_loss": 0.5767696499824524,
"eval_runtime": 48.8421,
"eval_samples_per_second": 59.948,
"eval_steps_per_second": 3.747,
"step": 1000
},
{
"epoch": 0.2787360287015317,
"grad_norm": 0.994421928125615,
"learning_rate": 9.062566846362843e-06,
"loss": 0.5746,
"step": 1010
},
{
"epoch": 0.28149579136194286,
"grad_norm": 0.9627083740767366,
"learning_rate": 9.034298903649485e-06,
"loss": 0.5731,
"step": 1020
},
{
"epoch": 0.2842555540223541,
"grad_norm": 0.9143645503640928,
"learning_rate": 9.005656537640942e-06,
"loss": 0.574,
"step": 1030
},
{
"epoch": 0.2870153166827653,
"grad_norm": 1.0384970550188914,
"learning_rate": 8.976642406635295e-06,
"loss": 0.5725,
"step": 1040
},
{
"epoch": 0.2897750793431765,
"grad_norm": 0.9669853254759909,
"learning_rate": 8.947259203434147e-06,
"loss": 0.5737,
"step": 1050
},
{
"epoch": 0.2925348420035877,
"grad_norm": 0.965075042317439,
"learning_rate": 8.917509655092691e-06,
"loss": 0.5723,
"step": 1060
},
{
"epoch": 0.29529460466399887,
"grad_norm": 0.9720003134521186,
"learning_rate": 8.887396522666608e-06,
"loss": 0.566,
"step": 1070
},
{
"epoch": 0.2980543673244101,
"grad_norm": 1.0399992404707865,
"learning_rate": 8.85692260095582e-06,
"loss": 0.5683,
"step": 1080
},
{
"epoch": 0.3008141299848213,
"grad_norm": 1.0679186267160017,
"learning_rate": 8.826090718245112e-06,
"loss": 0.5661,
"step": 1090
},
{
"epoch": 0.3035738926452325,
"grad_norm": 0.9643258915292706,
"learning_rate": 8.794903736041622e-06,
"loss": 0.5717,
"step": 1100
},
{
"epoch": 0.3063336553056437,
"grad_norm": 0.9136180676106652,
"learning_rate": 8.763364548809279e-06,
"loss": 0.5658,
"step": 1110
},
{
"epoch": 0.30909341796605494,
"grad_norm": 0.9512590668833296,
"learning_rate": 8.731476083700154e-06,
"loss": 0.5679,
"step": 1120
},
{
"epoch": 0.3118531806264661,
"grad_norm": 0.8874320591828697,
"learning_rate": 8.699241300282806e-06,
"loss": 0.5684,
"step": 1130
},
{
"epoch": 0.31461294328687733,
"grad_norm": 0.8922502034099749,
"learning_rate": 8.666663190267596e-06,
"loss": 0.5621,
"step": 1140
},
{
"epoch": 0.31737270594728856,
"grad_norm": 1.6201987068583557,
"learning_rate": 8.633744777229029e-06,
"loss": 0.569,
"step": 1150
},
{
"epoch": 0.3201324686076997,
"grad_norm": 0.9202299543891491,
"learning_rate": 8.600489116325128e-06,
"loss": 0.5678,
"step": 1160
},
{
"epoch": 0.32289223126811095,
"grad_norm": 1.007642513849446,
"learning_rate": 8.566899294013901e-06,
"loss": 0.5696,
"step": 1170
},
{
"epoch": 0.3256519939285221,
"grad_norm": 0.9837333100849195,
"learning_rate": 8.53297842776687e-06,
"loss": 0.5655,
"step": 1180
},
{
"epoch": 0.32841175658893335,
"grad_norm": 0.9874781114751451,
"learning_rate": 8.498729665779751e-06,
"loss": 0.5653,
"step": 1190
},
{
"epoch": 0.33117151924934457,
"grad_norm": 0.9230978368811615,
"learning_rate": 8.464156186680262e-06,
"loss": 0.5665,
"step": 1200
},
{
"epoch": 0.33393128190975574,
"grad_norm": 0.8844537320977038,
"learning_rate": 8.429261199233114e-06,
"loss": 0.5633,
"step": 1210
},
{
"epoch": 0.33669104457016696,
"grad_norm": 0.8743072862956215,
"learning_rate": 8.394047942042215e-06,
"loss": 0.5648,
"step": 1220
},
{
"epoch": 0.3394508072305782,
"grad_norm": 0.9571103987457272,
"learning_rate": 8.358519683250087e-06,
"loss": 0.5628,
"step": 1230
},
{
"epoch": 0.34221056989098936,
"grad_norm": 0.9109271244972179,
"learning_rate": 8.322679720234553e-06,
"loss": 0.5665,
"step": 1240
},
{
"epoch": 0.3449703325514006,
"grad_norm": 0.9812446794213074,
"learning_rate": 8.286531379302703e-06,
"loss": 0.5615,
"step": 1250
},
{
"epoch": 0.3477300952118118,
"grad_norm": 0.9841621501852277,
"learning_rate": 8.25007801538218e-06,
"loss": 0.5626,
"step": 1260
},
{
"epoch": 0.350489857872223,
"grad_norm": 0.9203538923510197,
"learning_rate": 8.21332301170982e-06,
"loss": 0.5645,
"step": 1270
},
{
"epoch": 0.3532496205326342,
"grad_norm": 1.0479420531686792,
"learning_rate": 8.17626977951764e-06,
"loss": 0.5622,
"step": 1280
},
{
"epoch": 0.3560093831930454,
"grad_norm": 0.9787960431652523,
"learning_rate": 8.138921757716245e-06,
"loss": 0.5581,
"step": 1290
},
{
"epoch": 0.3587691458534566,
"grad_norm": 0.9530332373900086,
"learning_rate": 8.101282412575673e-06,
"loss": 0.5618,
"step": 1300
},
{
"epoch": 0.3615289085138678,
"grad_norm": 1.0048234119883286,
"learning_rate": 8.063355237403672e-06,
"loss": 0.5555,
"step": 1310
},
{
"epoch": 0.364288671174279,
"grad_norm": 0.9557255183834055,
"learning_rate": 8.0251437522215e-06,
"loss": 0.5588,
"step": 1320
},
{
"epoch": 0.3670484338346902,
"grad_norm": 0.8586667700859966,
"learning_rate": 7.986651503437233e-06,
"loss": 0.5638,
"step": 1330
},
{
"epoch": 0.36980819649510144,
"grad_norm": 0.910151087145653,
"learning_rate": 7.947882063516612e-06,
"loss": 0.561,
"step": 1340
},
{
"epoch": 0.3725679591555126,
"grad_norm": 1.055588127662,
"learning_rate": 7.908839030651488e-06,
"loss": 0.558,
"step": 1350
},
{
"epoch": 0.37532772181592383,
"grad_norm": 0.9158993525267682,
"learning_rate": 7.869526028425878e-06,
"loss": 0.5587,
"step": 1360
},
{
"epoch": 0.37808748447633506,
"grad_norm": 0.9374100988045402,
"learning_rate": 7.829946705479654e-06,
"loss": 0.5584,
"step": 1370
},
{
"epoch": 0.3808472471367462,
"grad_norm": 0.9220438433292831,
"learning_rate": 7.790104735169915e-06,
"loss": 0.5576,
"step": 1380
},
{
"epoch": 0.38360700979715745,
"grad_norm": 0.987945504497949,
"learning_rate": 7.750003815230062e-06,
"loss": 0.5558,
"step": 1390
},
{
"epoch": 0.3863667724575687,
"grad_norm": 0.9218901797189658,
"learning_rate": 7.70964766742662e-06,
"loss": 0.5611,
"step": 1400
},
{
"epoch": 0.38912653511797984,
"grad_norm": 0.964522451572826,
"learning_rate": 7.669040037213795e-06,
"loss": 0.5559,
"step": 1410
},
{
"epoch": 0.39188629777839107,
"grad_norm": 0.8962914985254243,
"learning_rate": 7.628184693385896e-06,
"loss": 0.5601,
"step": 1420
},
{
"epoch": 0.39464606043880224,
"grad_norm": 1.0900746038682583,
"learning_rate": 7.587085427727523e-06,
"loss": 0.5555,
"step": 1430
},
{
"epoch": 0.39740582309921346,
"grad_norm": 0.8919597256872009,
"learning_rate": 7.54574605466166e-06,
"loss": 0.554,
"step": 1440
},
{
"epoch": 0.4001655857596247,
"grad_norm": 0.937111912961035,
"learning_rate": 7.504170410895668e-06,
"loss": 0.5576,
"step": 1450
},
{
"epoch": 0.40292534842003586,
"grad_norm": 0.9458249661943667,
"learning_rate": 7.462362355065189e-06,
"loss": 0.5531,
"step": 1460
},
{
"epoch": 0.4056851110804471,
"grad_norm": 0.9326011700869207,
"learning_rate": 7.420325767376026e-06,
"loss": 0.5536,
"step": 1470
},
{
"epoch": 0.4084448737408583,
"grad_norm": 1.0186083692765828,
"learning_rate": 7.378064549244031e-06,
"loss": 0.5512,
"step": 1480
},
{
"epoch": 0.4112046364012695,
"grad_norm": 1.2256145317450289,
"learning_rate": 7.335582622933e-06,
"loss": 0.5601,
"step": 1490
},
{
"epoch": 0.4139643990616807,
"grad_norm": 0.8742188645170639,
"learning_rate": 7.292883931190667e-06,
"loss": 0.5524,
"step": 1500
},
{
"epoch": 0.4139643990616807,
"eval_loss": 0.5561984777450562,
"eval_runtime": 49.4336,
"eval_samples_per_second": 59.231,
"eval_steps_per_second": 3.702,
"step": 1500
},
{
"epoch": 0.4167241617220919,
"grad_norm": 1.2121655526375013,
"learning_rate": 7.249972436882756e-06,
"loss": 0.5542,
"step": 1510
},
{
"epoch": 0.4194839243825031,
"grad_norm": 0.9357492214275386,
"learning_rate": 7.206852122625203e-06,
"loss": 0.5524,
"step": 1520
},
{
"epoch": 0.4222436870429143,
"grad_norm": 0.9433912153799177,
"learning_rate": 7.163526990414522e-06,
"loss": 0.5542,
"step": 1530
},
{
"epoch": 0.4250034497033255,
"grad_norm": 1.1073143594605637,
"learning_rate": 7.120001061256387e-06,
"loss": 0.5525,
"step": 1540
},
{
"epoch": 0.4277632123637367,
"grad_norm": 0.9302028752309601,
"learning_rate": 7.076278374792429e-06,
"loss": 0.5541,
"step": 1550
},
{
"epoch": 0.43052297502414794,
"grad_norm": 0.9606456277899846,
"learning_rate": 7.032362988925332e-06,
"loss": 0.5553,
"step": 1560
},
{
"epoch": 0.4332827376845591,
"grad_norm": 0.9251439989697735,
"learning_rate": 6.9882589794422105e-06,
"loss": 0.5489,
"step": 1570
},
{
"epoch": 0.43604250034497033,
"grad_norm": 0.9323884014701264,
"learning_rate": 6.943970439636336e-06,
"loss": 0.5494,
"step": 1580
},
{
"epoch": 0.43880226300538155,
"grad_norm": 1.3158449326861952,
"learning_rate": 6.899501479927242e-06,
"loss": 0.5484,
"step": 1590
},
{
"epoch": 0.4415620256657927,
"grad_norm": 0.8775790961507058,
"learning_rate": 6.8548562274792325e-06,
"loss": 0.5453,
"step": 1600
},
{
"epoch": 0.44432178832620395,
"grad_norm": 0.9478549256736298,
"learning_rate": 6.81003882581834e-06,
"loss": 0.5482,
"step": 1610
},
{
"epoch": 0.4470815509866152,
"grad_norm": 0.9723342894189143,
"learning_rate": 6.765053434447769e-06,
"loss": 0.5487,
"step": 1620
},
{
"epoch": 0.44984131364702634,
"grad_norm": 0.974749959194784,
"learning_rate": 6.7199042284618484e-06,
"loss": 0.5505,
"step": 1630
},
{
"epoch": 0.45260107630743757,
"grad_norm": 0.9255671399322642,
"learning_rate": 6.674595398158541e-06,
"loss": 0.5493,
"step": 1640
},
{
"epoch": 0.4553608389678488,
"grad_norm": 0.9465763497145431,
"learning_rate": 6.629131148650543e-06,
"loss": 0.5444,
"step": 1650
},
{
"epoch": 0.45812060162825996,
"grad_norm": 0.9292686081943923,
"learning_rate": 6.583515699475009e-06,
"loss": 0.5496,
"step": 1660
},
{
"epoch": 0.4608803642886712,
"grad_norm": 1.108323416936679,
"learning_rate": 6.537753284201935e-06,
"loss": 0.5487,
"step": 1670
},
{
"epoch": 0.46364012694908235,
"grad_norm": 0.9042384031621171,
"learning_rate": 6.491848150041242e-06,
"loss": 0.5477,
"step": 1680
},
{
"epoch": 0.4663998896094936,
"grad_norm": 0.850329034972769,
"learning_rate": 6.4458045574485875e-06,
"loss": 0.5519,
"step": 1690
},
{
"epoch": 0.4691596522699048,
"grad_norm": 0.9490531745735549,
"learning_rate": 6.399626779729959e-06,
"loss": 0.5439,
"step": 1700
},
{
"epoch": 0.471919414930316,
"grad_norm": 0.9352622534171564,
"learning_rate": 6.353319102645069e-06,
"loss": 0.5465,
"step": 1710
},
{
"epoch": 0.4746791775907272,
"grad_norm": 0.9191680266246547,
"learning_rate": 6.306885824009585e-06,
"loss": 0.547,
"step": 1720
},
{
"epoch": 0.4774389402511384,
"grad_norm": 0.8597116149610817,
"learning_rate": 6.260331253296259e-06,
"loss": 0.5495,
"step": 1730
},
{
"epoch": 0.4801987029115496,
"grad_norm": 0.929502998942292,
"learning_rate": 6.213659711234958e-06,
"loss": 0.5481,
"step": 1740
},
{
"epoch": 0.4829584655719608,
"grad_norm": 0.9675046561306946,
"learning_rate": 6.1668755294116655e-06,
"loss": 0.5468,
"step": 1750
},
{
"epoch": 0.48571822823237204,
"grad_norm": 0.9073333178919157,
"learning_rate": 6.119983049866456e-06,
"loss": 0.5438,
"step": 1760
},
{
"epoch": 0.4884779908927832,
"grad_norm": 0.9109565773453232,
"learning_rate": 6.072986624690516e-06,
"loss": 0.5478,
"step": 1770
},
{
"epoch": 0.49123775355319443,
"grad_norm": 0.8955027642684514,
"learning_rate": 6.025890615622233e-06,
"loss": 0.5415,
"step": 1780
},
{
"epoch": 0.4939975162136056,
"grad_norm": 0.8648546291027864,
"learning_rate": 5.97869939364237e-06,
"loss": 0.5411,
"step": 1790
},
{
"epoch": 0.4967572788740168,
"grad_norm": 0.9523614008291538,
"learning_rate": 5.9314173385683986e-06,
"loss": 0.5453,
"step": 1800
},
{
"epoch": 0.49951704153442805,
"grad_norm": 0.9383064266388607,
"learning_rate": 5.884048838648017e-06,
"loss": 0.5384,
"step": 1810
},
{
"epoch": 0.5022768041948392,
"grad_norm": 0.9021624786717238,
"learning_rate": 5.836598290151866e-06,
"loss": 0.5456,
"step": 1820
},
{
"epoch": 0.5050365668552504,
"grad_norm": 0.9088357380698893,
"learning_rate": 5.789070096965514e-06,
"loss": 0.5426,
"step": 1830
},
{
"epoch": 0.5077963295156617,
"grad_norm": 0.922862032580903,
"learning_rate": 5.741468670180737e-06,
"loss": 0.5434,
"step": 1840
},
{
"epoch": 0.5105560921760729,
"grad_norm": 1.4926573419318607,
"learning_rate": 5.6937984276861195e-06,
"loss": 0.5441,
"step": 1850
},
{
"epoch": 0.513315854836484,
"grad_norm": 0.8580464247033338,
"learning_rate": 5.646063793757028e-06,
"loss": 0.5413,
"step": 1860
},
{
"epoch": 0.5160756174968952,
"grad_norm": 0.8836024449571962,
"learning_rate": 5.598269198645008e-06,
"loss": 0.5457,
"step": 1870
},
{
"epoch": 0.5188353801573065,
"grad_norm": 0.8951578721801148,
"learning_rate": 5.550419078166594e-06,
"loss": 0.5452,
"step": 1880
},
{
"epoch": 0.5215951428177177,
"grad_norm": 0.9955192132312739,
"learning_rate": 5.502517873291632e-06,
"loss": 0.5467,
"step": 1890
},
{
"epoch": 0.5243549054781289,
"grad_norm": 1.1988654703686727,
"learning_rate": 5.454570029731115e-06,
"loss": 0.5413,
"step": 1900
},
{
"epoch": 0.5271146681385401,
"grad_norm": 0.8839237136665734,
"learning_rate": 5.406579997524567e-06,
"loss": 0.5411,
"step": 1910
},
{
"epoch": 0.5298744307989512,
"grad_norm": 0.9047836025418462,
"learning_rate": 5.358552230627044e-06,
"loss": 0.539,
"step": 1920
},
{
"epoch": 0.5326341934593625,
"grad_norm": 0.875356281556686,
"learning_rate": 5.310491186495757e-06,
"loss": 0.5429,
"step": 1930
},
{
"epoch": 0.5353939561197737,
"grad_norm": 0.9167892739496736,
"learning_rate": 5.262401325676378e-06,
"loss": 0.5384,
"step": 1940
},
{
"epoch": 0.5381537187801849,
"grad_norm": 0.9525976300842104,
"learning_rate": 5.214287111389057e-06,
"loss": 0.5362,
"step": 1950
},
{
"epoch": 0.5409134814405961,
"grad_norm": 0.885862636010332,
"learning_rate": 5.166153009114188e-06,
"loss": 0.5378,
"step": 1960
},
{
"epoch": 0.5436732441010073,
"grad_norm": 0.9013246458040829,
"learning_rate": 5.1180034861779685e-06,
"loss": 0.5399,
"step": 1970
},
{
"epoch": 0.5464330067614185,
"grad_norm": 0.9017690594117211,
"learning_rate": 5.069843011337789e-06,
"loss": 0.5412,
"step": 1980
},
{
"epoch": 0.5491927694218297,
"grad_norm": 0.9063217665862804,
"learning_rate": 5.0216760543674855e-06,
"loss": 0.5415,
"step": 1990
},
{
"epoch": 0.5519525320822409,
"grad_norm": 0.8411472472974292,
"learning_rate": 4.973507085642502e-06,
"loss": 0.537,
"step": 2000
},
{
"epoch": 0.5519525320822409,
"eval_loss": 0.540687620639801,
"eval_runtime": 48.7225,
"eval_samples_per_second": 60.095,
"eval_steps_per_second": 3.756,
"step": 2000
},
{
"epoch": 0.5547122947426522,
"grad_norm": 0.8884146283634436,
"learning_rate": 4.92534057572499e-06,
"loss": 0.5406,
"step": 2010
},
{
"epoch": 0.5574720574030634,
"grad_norm": 0.8914897264034541,
"learning_rate": 4.8771809949489056e-06,
"loss": 0.5385,
"step": 2020
},
{
"epoch": 0.5602318200634745,
"grad_norm": 0.936304320970765,
"learning_rate": 4.829032813005103e-06,
"loss": 0.542,
"step": 2030
},
{
"epoch": 0.5629915827238857,
"grad_norm": 0.9310122808691009,
"learning_rate": 4.780900498526515e-06,
"loss": 0.5365,
"step": 2040
},
{
"epoch": 0.5657513453842969,
"grad_norm": 0.9518666584186597,
"learning_rate": 4.732788518673418e-06,
"loss": 0.5353,
"step": 2050
},
{
"epoch": 0.5685111080447082,
"grad_norm": 0.9510071761987918,
"learning_rate": 4.684701338718825e-06,
"loss": 0.5317,
"step": 2060
},
{
"epoch": 0.5712708707051194,
"grad_norm": 0.9432991523879628,
"learning_rate": 4.636643421634075e-06,
"loss": 0.5371,
"step": 2070
},
{
"epoch": 0.5740306333655306,
"grad_norm": 0.8853811386420297,
"learning_rate": 4.588619227674619e-06,
"loss": 0.5365,
"step": 2080
},
{
"epoch": 0.5767903960259417,
"grad_norm": 0.9290294341366813,
"learning_rate": 4.540633213966064e-06,
"loss": 0.5334,
"step": 2090
},
{
"epoch": 0.579550158686353,
"grad_norm": 0.9327668087903835,
"learning_rate": 4.492689834090508e-06,
"loss": 0.5341,
"step": 2100
},
{
"epoch": 0.5823099213467642,
"grad_norm": 1.3176571259191412,
"learning_rate": 4.444793537673204e-06,
"loss": 0.5306,
"step": 2110
},
{
"epoch": 0.5850696840071754,
"grad_norm": 0.9485574547153368,
"learning_rate": 4.396948769969587e-06,
"loss": 0.5379,
"step": 2120
},
{
"epoch": 0.5878294466675866,
"grad_norm": 0.951959101524703,
"learning_rate": 4.3491599714527115e-06,
"loss": 0.5348,
"step": 2130
},
{
"epoch": 0.5905892093279977,
"grad_norm": 0.8736078038744113,
"learning_rate": 4.301431577401136e-06,
"loss": 0.5323,
"step": 2140
},
{
"epoch": 0.593348971988409,
"grad_norm": 0.8979881907613325,
"learning_rate": 4.253768017487275e-06,
"loss": 0.5389,
"step": 2150
},
{
"epoch": 0.5961087346488202,
"grad_norm": 0.9371779488004714,
"learning_rate": 4.206173715366289e-06,
"loss": 0.5343,
"step": 2160
},
{
"epoch": 0.5988684973092314,
"grad_norm": 0.9563875289878281,
"learning_rate": 4.1586530882655226e-06,
"loss": 0.5352,
"step": 2170
},
{
"epoch": 0.6016282599696426,
"grad_norm": 0.9747450112581771,
"learning_rate": 4.111210546574545e-06,
"loss": 0.534,
"step": 2180
},
{
"epoch": 0.6043880226300539,
"grad_norm": 0.9351891869474309,
"learning_rate": 4.063850493435808e-06,
"loss": 0.5298,
"step": 2190
},
{
"epoch": 0.607147785290465,
"grad_norm": 0.8980425451702817,
"learning_rate": 4.0165773243360105e-06,
"loss": 0.5338,
"step": 2200
},
{
"epoch": 0.6099075479508762,
"grad_norm": 0.9224682367333663,
"learning_rate": 3.96939542669814e-06,
"loss": 0.5306,
"step": 2210
},
{
"epoch": 0.6126673106112874,
"grad_norm": 0.8491438680952316,
"learning_rate": 3.922309179474279e-06,
"loss": 0.5306,
"step": 2220
},
{
"epoch": 0.6154270732716987,
"grad_norm": 0.9550543136397366,
"learning_rate": 3.875322952739196e-06,
"loss": 0.5348,
"step": 2230
},
{
"epoch": 0.6181868359321099,
"grad_norm": 0.8390020003541692,
"learning_rate": 3.828441107284755e-06,
"loss": 0.5343,
"step": 2240
},
{
"epoch": 0.620946598592521,
"grad_norm": 0.892945059961568,
"learning_rate": 3.7816679942151945e-06,
"loss": 0.523,
"step": 2250
},
{
"epoch": 0.6237063612529322,
"grad_norm": 0.9210249115479134,
"learning_rate": 3.7350079545433014e-06,
"loss": 0.5279,
"step": 2260
},
{
"epoch": 0.6264661239133434,
"grad_norm": 0.9570746572961814,
"learning_rate": 3.6884653187875193e-06,
"loss": 0.5293,
"step": 2270
},
{
"epoch": 0.6292258865737547,
"grad_norm": 0.8534492122472642,
"learning_rate": 3.642044406570031e-06,
"loss": 0.5256,
"step": 2280
},
{
"epoch": 0.6319856492341659,
"grad_norm": 0.8732168431838571,
"learning_rate": 3.595749526215862e-06,
"loss": 0.5328,
"step": 2290
},
{
"epoch": 0.6347454118945771,
"grad_norm": 0.8962727112777485,
"learning_rate": 3.549584974353018e-06,
"loss": 0.5336,
"step": 2300
},
{
"epoch": 0.6375051745549882,
"grad_norm": 0.8730073595406601,
"learning_rate": 3.5035550355137156e-06,
"loss": 0.5318,
"step": 2310
},
{
"epoch": 0.6402649372153995,
"grad_norm": 0.9162256052176202,
"learning_rate": 3.457663981736739e-06,
"loss": 0.5346,
"step": 2320
},
{
"epoch": 0.6430246998758107,
"grad_norm": 0.8555722066919174,
"learning_rate": 3.411916072170946e-06,
"loss": 0.5318,
"step": 2330
},
{
"epoch": 0.6457844625362219,
"grad_norm": 0.8407652045731654,
"learning_rate": 3.3663155526799827e-06,
"loss": 0.5306,
"step": 2340
},
{
"epoch": 0.6485442251966331,
"grad_norm": 0.876217339145573,
"learning_rate": 3.3208666554482216e-06,
"loss": 0.5291,
"step": 2350
},
{
"epoch": 0.6513039878570442,
"grad_norm": 0.897083649993733,
"learning_rate": 3.275573598587969e-06,
"loss": 0.5229,
"step": 2360
},
{
"epoch": 0.6540637505174555,
"grad_norm": 0.7995897085360415,
"learning_rate": 3.230440585747991e-06,
"loss": 0.527,
"step": 2370
},
{
"epoch": 0.6568235131778667,
"grad_norm": 0.9014903537041772,
"learning_rate": 3.185471805723365e-06,
"loss": 0.5261,
"step": 2380
},
{
"epoch": 0.6595832758382779,
"grad_norm": 0.856492997294296,
"learning_rate": 3.140671432066719e-06,
"loss": 0.5308,
"step": 2390
},
{
"epoch": 0.6623430384986891,
"grad_norm": 0.888225217505466,
"learning_rate": 3.096043622700888e-06,
"loss": 0.5294,
"step": 2400
},
{
"epoch": 0.6651028011591004,
"grad_norm": 0.9531940854665482,
"learning_rate": 3.0515925195330148e-06,
"loss": 0.5283,
"step": 2410
},
{
"epoch": 0.6678625638195115,
"grad_norm": 1.2539256252009454,
"learning_rate": 3.0073222480701354e-06,
"loss": 0.5275,
"step": 2420
},
{
"epoch": 0.6706223264799227,
"grad_norm": 0.9137548818626098,
"learning_rate": 2.9632369170362977e-06,
"loss": 0.5235,
"step": 2430
},
{
"epoch": 0.6733820891403339,
"grad_norm": 1.6494495739553416,
"learning_rate": 2.9193406179912297e-06,
"loss": 0.5271,
"step": 2440
},
{
"epoch": 0.6761418518007452,
"grad_norm": 0.8590622849369576,
"learning_rate": 2.875637424950595e-06,
"loss": 0.522,
"step": 2450
},
{
"epoch": 0.6789016144611564,
"grad_norm": 0.9186749142898574,
"learning_rate": 2.832131394007891e-06,
"loss": 0.5221,
"step": 2460
},
{
"epoch": 0.6816613771215675,
"grad_norm": 0.9090102002494171,
"learning_rate": 2.788826562958e-06,
"loss": 0.5289,
"step": 2470
},
{
"epoch": 0.6844211397819787,
"grad_norm": 1.0599147672457048,
"learning_rate": 2.745726950922444e-06,
"loss": 0.5241,
"step": 2480
},
{
"epoch": 0.6871809024423899,
"grad_norm": 0.8881000835172156,
"learning_rate": 2.7028365579763606e-06,
"loss": 0.5298,
"step": 2490
},
{
"epoch": 0.6899406651028012,
"grad_norm": 0.8992940208666876,
"learning_rate": 2.6601593647772696e-06,
"loss": 0.5282,
"step": 2500
},
{
"epoch": 0.6899406651028012,
"eval_loss": 0.5283246636390686,
"eval_runtime": 51.449,
"eval_samples_per_second": 56.911,
"eval_steps_per_second": 3.557,
"step": 2500
},
{
"epoch": 0.6927004277632124,
"grad_norm": 0.9769386150489252,
"learning_rate": 2.6176993321956185e-06,
"loss": 0.5253,
"step": 2510
},
{
"epoch": 0.6954601904236236,
"grad_norm": 0.9191890198831723,
"learning_rate": 2.5754604009471786e-06,
"loss": 0.5229,
"step": 2520
},
{
"epoch": 0.6982199530840347,
"grad_norm": 0.8879005162140707,
"learning_rate": 2.533446491227305e-06,
"loss": 0.5247,
"step": 2530
},
{
"epoch": 0.700979715744446,
"grad_norm": 0.9109715281747162,
"learning_rate": 2.491661502347106e-06,
"loss": 0.5218,
"step": 2540
},
{
"epoch": 0.7037394784048572,
"grad_norm": 0.8969876737597425,
"learning_rate": 2.4501093123715395e-06,
"loss": 0.5238,
"step": 2550
},
{
"epoch": 0.7064992410652684,
"grad_norm": 0.9123939456680997,
"learning_rate": 2.408793777759504e-06,
"loss": 0.5234,
"step": 2560
},
{
"epoch": 0.7092590037256796,
"grad_norm": 0.8767338790553574,
"learning_rate": 2.3677187330059084e-06,
"loss": 0.5247,
"step": 2570
},
{
"epoch": 0.7120187663860909,
"grad_norm": 0.9605486388857951,
"learning_rate": 2.3268879902857978e-06,
"loss": 0.5194,
"step": 2580
},
{
"epoch": 0.714778529046502,
"grad_norm": 0.9895359728969572,
"learning_rate": 2.2863053391005462e-06,
"loss": 0.5236,
"step": 2590
},
{
"epoch": 0.7175382917069132,
"grad_norm": 0.957733912839131,
"learning_rate": 2.245974545926152e-06,
"loss": 0.5179,
"step": 2600
},
{
"epoch": 0.7202980543673244,
"grad_norm": 0.9601154451397089,
"learning_rate": 2.205899353863665e-06,
"loss": 0.5224,
"step": 2610
},
{
"epoch": 0.7230578170277356,
"grad_norm": 0.9084441231449814,
"learning_rate": 2.166083482291801e-06,
"loss": 0.5254,
"step": 2620
},
{
"epoch": 0.7258175796881469,
"grad_norm": 1.0841389588139543,
"learning_rate": 2.1265306265217382e-06,
"loss": 0.5214,
"step": 2630
},
{
"epoch": 0.728577342348558,
"grad_norm": 1.451146075094232,
"learning_rate": 2.0872444574541574e-06,
"loss": 0.5298,
"step": 2640
},
{
"epoch": 0.7313371050089692,
"grad_norm": 0.9009140743152368,
"learning_rate": 2.048228621238547e-06,
"loss": 0.5209,
"step": 2650
},
{
"epoch": 0.7340968676693804,
"grad_norm": 0.8967623610111647,
"learning_rate": 2.0094867389347982e-06,
"loss": 0.5248,
"step": 2660
},
{
"epoch": 0.7368566303297917,
"grad_norm": 0.9029822617189499,
"learning_rate": 1.971022406177142e-06,
"loss": 0.5227,
"step": 2670
},
{
"epoch": 0.7396163929902029,
"grad_norm": 0.9009675358712146,
"learning_rate": 1.932839192840436e-06,
"loss": 0.522,
"step": 2680
},
{
"epoch": 0.7423761556506141,
"grad_norm": 0.8816707412986489,
"learning_rate": 1.8949406427088407e-06,
"loss": 0.5223,
"step": 2690
},
{
"epoch": 0.7451359183110252,
"grad_norm": 0.8605202814409114,
"learning_rate": 1.8573302731469255e-06,
"loss": 0.5241,
"step": 2700
},
{
"epoch": 0.7478956809714364,
"grad_norm": 0.9612849756128029,
"learning_rate": 1.820011574773221e-06,
"loss": 0.5264,
"step": 2710
},
{
"epoch": 0.7506554436318477,
"grad_norm": 0.8793842539758803,
"learning_rate": 1.7829880111362486e-06,
"loss": 0.5181,
"step": 2720
},
{
"epoch": 0.7534152062922589,
"grad_norm": 1.2081068952668312,
"learning_rate": 1.746263018393079e-06,
"loss": 0.5292,
"step": 2730
},
{
"epoch": 0.7561749689526701,
"grad_norm": 0.8860653822887559,
"learning_rate": 1.7098400049904163e-06,
"loss": 0.5219,
"step": 2740
},
{
"epoch": 0.7589347316130812,
"grad_norm": 0.8901286343549357,
"learning_rate": 1.6737223513482591e-06,
"loss": 0.5259,
"step": 2750
},
{
"epoch": 0.7616944942734925,
"grad_norm": 0.9081512876330592,
"learning_rate": 1.6379134095461673e-06,
"loss": 0.5171,
"step": 2760
},
{
"epoch": 0.7644542569339037,
"grad_norm": 0.854771397996212,
"learning_rate": 1.6024165030121542e-06,
"loss": 0.521,
"step": 2770
},
{
"epoch": 0.7672140195943149,
"grad_norm": 0.9522518174715512,
"learning_rate": 1.567234926214236e-06,
"loss": 0.5149,
"step": 2780
},
{
"epoch": 0.7699737822547261,
"grad_norm": 0.855171965732468,
"learning_rate": 1.5323719443546785e-06,
"loss": 0.5217,
"step": 2790
},
{
"epoch": 0.7727335449151373,
"grad_norm": 0.846551026714045,
"learning_rate": 1.4978307930669483e-06,
"loss": 0.5174,
"step": 2800
},
{
"epoch": 0.7754933075755485,
"grad_norm": 0.9109328397942944,
"learning_rate": 1.4636146781154164e-06,
"loss": 0.5262,
"step": 2810
},
{
"epoch": 0.7782530702359597,
"grad_norm": 0.8916543609749644,
"learning_rate": 1.4297267750978277e-06,
"loss": 0.5258,
"step": 2820
},
{
"epoch": 0.7810128328963709,
"grad_norm": 0.839376705554234,
"learning_rate": 1.3961702291505791e-06,
"loss": 0.5233,
"step": 2830
},
{
"epoch": 0.7837725955567821,
"grad_norm": 0.9530767315659909,
"learning_rate": 1.3629481546568163e-06,
"loss": 0.5153,
"step": 2840
},
{
"epoch": 0.7865323582171934,
"grad_norm": 0.9433520691735623,
"learning_rate": 1.3300636349573882e-06,
"loss": 0.5169,
"step": 2850
},
{
"epoch": 0.7892921208776045,
"grad_norm": 0.8633774616839734,
"learning_rate": 1.2975197220646807e-06,
"loss": 0.5211,
"step": 2860
},
{
"epoch": 0.7920518835380157,
"grad_norm": 0.8830161483028809,
"learning_rate": 1.2653194363793642e-06,
"loss": 0.5169,
"step": 2870
},
{
"epoch": 0.7948116461984269,
"grad_norm": 0.905868276953366,
"learning_rate": 1.2334657664100614e-06,
"loss": 0.5167,
"step": 2880
},
{
"epoch": 0.7975714088588381,
"grad_norm": 0.8402578283356454,
"learning_rate": 1.2019616684959934e-06,
"loss": 0.5207,
"step": 2890
},
{
"epoch": 0.8003311715192494,
"grad_norm": 1.0070145612320756,
"learning_rate": 1.1708100665325967e-06,
"loss": 0.5176,
"step": 2900
},
{
"epoch": 0.8030909341796606,
"grad_norm": 0.8631246623697408,
"learning_rate": 1.1400138517001564e-06,
"loss": 0.5147,
"step": 2910
},
{
"epoch": 0.8058506968400717,
"grad_norm": 0.8746596064396687,
"learning_rate": 1.1095758821954788e-06,
"loss": 0.5212,
"step": 2920
},
{
"epoch": 0.8086104595004829,
"grad_norm": 0.8440888494194767,
"learning_rate": 1.0794989829666197e-06,
"loss": 0.5154,
"step": 2930
},
{
"epoch": 0.8113702221608942,
"grad_norm": 0.8369885923181033,
"learning_rate": 1.049785945450697e-06,
"loss": 0.516,
"step": 2940
},
{
"epoch": 0.8141299848213054,
"grad_norm": 1.5997002594596106,
"learning_rate": 1.0204395273148277e-06,
"loss": 0.5197,
"step": 2950
},
{
"epoch": 0.8168897474817166,
"grad_norm": 0.8873807689235719,
"learning_rate": 9.914624522001792e-07,
"loss": 0.5232,
"step": 2960
},
{
"epoch": 0.8196495101421277,
"grad_norm": 0.8697811520680214,
"learning_rate": 9.628574094691945e-07,
"loss": 0.5139,
"step": 2970
},
{
"epoch": 0.822409272802539,
"grad_norm": 0.8378502425029508,
"learning_rate": 9.346270539559882e-07,
"loss": 0.52,
"step": 2980
},
{
"epoch": 0.8251690354629502,
"grad_norm": 0.8849196731646373,
"learning_rate": 9.067740057199514e-07,
"loss": 0.518,
"step": 2990
},
{
"epoch": 0.8279287981233614,
"grad_norm": 0.8664792504731555,
"learning_rate": 8.793008498025879e-07,
"loss": 0.5155,
"step": 3000
},
{
"epoch": 0.8279287981233614,
"eval_loss": 0.5207065939903259,
"eval_runtime": 49.1882,
"eval_samples_per_second": 59.526,
"eval_steps_per_second": 3.72,
"step": 3000
},
{
"epoch": 0.8306885607837726,
"grad_norm": 0.8700644465815897,
"learning_rate": 8.522101359875934e-07,
"loss": 0.5131,
"step": 3010
},
{
"epoch": 0.8334483234441838,
"grad_norm": 0.9097989981675693,
"learning_rate": 8.255043785642108e-07,
"loss": 0.5125,
"step": 3020
},
{
"epoch": 0.836208086104595,
"grad_norm": 0.8631047164247316,
"learning_rate": 7.991860560938786e-07,
"loss": 0.5234,
"step": 3030
},
{
"epoch": 0.8389678487650062,
"grad_norm": 1.2358190890284166,
"learning_rate": 7.732576111801982e-07,
"loss": 0.5176,
"step": 3040
},
{
"epoch": 0.8417276114254174,
"grad_norm": 0.8514248482095018,
"learning_rate": 7.477214502422281e-07,
"loss": 0.5223,
"step": 3050
},
{
"epoch": 0.8444873740858286,
"grad_norm": 0.8563129413552018,
"learning_rate": 7.225799432911557e-07,
"loss": 0.5219,
"step": 3060
},
{
"epoch": 0.8472471367462399,
"grad_norm": 1.6151093899813862,
"learning_rate": 6.978354237103264e-07,
"loss": 0.516,
"step": 3070
},
{
"epoch": 0.850006899406651,
"grad_norm": 2.00845392683435,
"learning_rate": 6.734901880386896e-07,
"loss": 0.5185,
"step": 3080
},
{
"epoch": 0.8527666620670622,
"grad_norm": 1.1862873972091095,
"learning_rate": 6.495464957576508e-07,
"loss": 0.5122,
"step": 3090
},
{
"epoch": 0.8555264247274734,
"grad_norm": 0.8470639388291085,
"learning_rate": 6.260065690813754e-07,
"loss": 0.5158,
"step": 3100
},
{
"epoch": 0.8582861873878846,
"grad_norm": 0.9219107715988072,
"learning_rate": 6.028725927505369e-07,
"loss": 0.5201,
"step": 3110
},
{
"epoch": 0.8610459500482959,
"grad_norm": 0.8827729120692012,
"learning_rate": 5.801467138295597e-07,
"loss": 0.5189,
"step": 3120
},
{
"epoch": 0.8638057127087071,
"grad_norm": 0.8610445224310359,
"learning_rate": 5.578310415073451e-07,
"loss": 0.515,
"step": 3130
},
{
"epoch": 0.8665654753691182,
"grad_norm": 0.870963273635567,
"learning_rate": 5.359276469015179e-07,
"loss": 0.5151,
"step": 3140
},
{
"epoch": 0.8693252380295294,
"grad_norm": 0.8396590023282043,
"learning_rate": 5.14438562866208e-07,
"loss": 0.5223,
"step": 3150
},
{
"epoch": 0.8720850006899407,
"grad_norm": 0.8992961235238134,
"learning_rate": 4.933657838033795e-07,
"loss": 0.5144,
"step": 3160
},
{
"epoch": 0.8748447633503519,
"grad_norm": 0.9038948906592709,
"learning_rate": 4.7271126547772773e-07,
"loss": 0.5165,
"step": 3170
},
{
"epoch": 0.8776045260107631,
"grad_norm": 0.8686028788473747,
"learning_rate": 4.524769248351718e-07,
"loss": 0.5157,
"step": 3180
},
{
"epoch": 0.8803642886711743,
"grad_norm": 0.8744713706157983,
"learning_rate": 4.3266463982493566e-07,
"loss": 0.5167,
"step": 3190
},
{
"epoch": 0.8831240513315854,
"grad_norm": 0.8953682081464681,
"learning_rate": 4.132762492252601e-07,
"loss": 0.5229,
"step": 3200
},
{
"epoch": 0.8858838139919967,
"grad_norm": 1.0500886926956579,
"learning_rate": 3.943135524727448e-07,
"loss": 0.5168,
"step": 3210
},
{
"epoch": 0.8886435766524079,
"grad_norm": 0.8891911066911182,
"learning_rate": 3.757783094953382e-07,
"loss": 0.52,
"step": 3220
},
{
"epoch": 0.8914033393128191,
"grad_norm": 0.8469428073562058,
"learning_rate": 3.5767224054900687e-07,
"loss": 0.514,
"step": 3230
},
{
"epoch": 0.8941631019732303,
"grad_norm": 0.8985979256834127,
"learning_rate": 3.3999702605807203e-07,
"loss": 0.5153,
"step": 3240
},
{
"epoch": 0.8969228646336415,
"grad_norm": 0.8740663379412131,
"learning_rate": 3.227543064592514e-07,
"loss": 0.5179,
"step": 3250
},
{
"epoch": 0.8996826272940527,
"grad_norm": 0.9253549886032714,
"learning_rate": 3.059456820494111e-07,
"loss": 0.5187,
"step": 3260
},
{
"epoch": 0.9024423899544639,
"grad_norm": 0.8657401705396123,
"learning_rate": 2.8957271283704067e-07,
"loss": 0.5188,
"step": 3270
},
{
"epoch": 0.9052021526148751,
"grad_norm": 0.8317058461633765,
"learning_rate": 2.736369183974685e-07,
"loss": 0.5161,
"step": 3280
},
{
"epoch": 0.9079619152752864,
"grad_norm": 0.8738336902844653,
"learning_rate": 2.5813977773183175e-07,
"loss": 0.515,
"step": 3290
},
{
"epoch": 0.9107216779356976,
"grad_norm": 0.9137477549092036,
"learning_rate": 2.430827291298099e-07,
"loss": 0.5198,
"step": 3300
},
{
"epoch": 0.9134814405961087,
"grad_norm": 0.8521777662643962,
"learning_rate": 2.2846717003613462e-07,
"loss": 0.5221,
"step": 3310
},
{
"epoch": 0.9162412032565199,
"grad_norm": 1.12701228509693,
"learning_rate": 2.1429445692089712e-07,
"loss": 0.5154,
"step": 3320
},
{
"epoch": 0.9190009659169311,
"grad_norm": 0.9027916863842183,
"learning_rate": 2.0056590515365016e-07,
"loss": 0.5159,
"step": 3330
},
{
"epoch": 0.9217607285773424,
"grad_norm": 0.839347519350046,
"learning_rate": 1.8728278888132944e-07,
"loss": 0.518,
"step": 3340
},
{
"epoch": 0.9245204912377536,
"grad_norm": 0.8627735870099915,
"learning_rate": 1.744463409100039e-07,
"loss": 0.5189,
"step": 3350
},
{
"epoch": 0.9272802538981647,
"grad_norm": 0.8677279393067534,
"learning_rate": 1.620577525904532e-07,
"loss": 0.5099,
"step": 3360
},
{
"epoch": 0.9300400165585759,
"grad_norm": 0.8705917700670992,
"learning_rate": 1.501181737076035e-07,
"loss": 0.5154,
"step": 3370
},
{
"epoch": 0.9327997792189872,
"grad_norm": 0.9046790697526798,
"learning_rate": 1.3862871237381004e-07,
"loss": 0.5108,
"step": 3380
},
{
"epoch": 0.9355595418793984,
"grad_norm": 0.8203655507439958,
"learning_rate": 1.2759043492601986e-07,
"loss": 0.5136,
"step": 3390
},
{
"epoch": 0.9383193045398096,
"grad_norm": 0.8949093577414237,
"learning_rate": 1.1700436582680108e-07,
"loss": 0.5192,
"step": 3400
},
{
"epoch": 0.9410790672002208,
"grad_norm": 0.8425742363423367,
"learning_rate": 1.068714875692628e-07,
"loss": 0.5116,
"step": 3410
},
{
"epoch": 0.943838829860632,
"grad_norm": 0.9030687400394452,
"learning_rate": 9.719274058587247e-08,
"loss": 0.5125,
"step": 3420
},
{
"epoch": 0.9465985925210432,
"grad_norm": 0.8753228488205499,
"learning_rate": 8.796902316117018e-08,
"loss": 0.5156,
"step": 3430
},
{
"epoch": 0.9493583551814544,
"grad_norm": 0.8396184463940205,
"learning_rate": 7.920119134840199e-08,
"loss": 0.5141,
"step": 3440
},
{
"epoch": 0.9521181178418656,
"grad_norm": 0.882889766020658,
"learning_rate": 7.08900588900685e-08,
"loss": 0.5131,
"step": 3450
},
{
"epoch": 0.9548778805022768,
"grad_norm": 0.8504174898031396,
"learning_rate": 6.303639714240196e-08,
"loss": 0.5193,
"step": 3460
},
{
"epoch": 0.957637643162688,
"grad_norm": 0.8595498992279648,
"learning_rate": 5.564093500377732e-08,
"loss": 0.515,
"step": 3470
},
{
"epoch": 0.9603974058230992,
"grad_norm": 0.8244758548316357,
"learning_rate": 4.870435884705982e-08,
"loss": 0.5135,
"step": 3480
},
{
"epoch": 0.9631571684835104,
"grad_norm": 0.9192549923625648,
"learning_rate": 4.22273124559075e-08,
"loss": 0.5165,
"step": 3490
},
{
"epoch": 0.9659169311439216,
"grad_norm": 1.8467401343991516,
"learning_rate": 3.621039696501794e-08,
"loss": 0.5106,
"step": 3500
},
{
"epoch": 0.9659169311439216,
"eval_loss": 0.5180693864822388,
"eval_runtime": 48.9508,
"eval_samples_per_second": 59.815,
"eval_steps_per_second": 3.738,
"step": 3500
},
{
"epoch": 0.9686766938043329,
"grad_norm": 0.9145993584663505,
"learning_rate": 3.065417080433841e-08,
"loss": 0.517,
"step": 3510
},
{
"epoch": 0.9714364564647441,
"grad_norm": 0.8570121000540161,
"learning_rate": 2.555914964723849e-08,
"loss": 0.5142,
"step": 3520
},
{
"epoch": 0.9741962191251552,
"grad_norm": 0.8678615083324009,
"learning_rate": 2.0925806362648847e-08,
"loss": 0.5198,
"step": 3530
},
{
"epoch": 0.9769559817855664,
"grad_norm": 0.8381539016578547,
"learning_rate": 1.6754570971176944e-08,
"loss": 0.5125,
"step": 3540
},
{
"epoch": 0.9797157444459776,
"grad_norm": 0.9056379885360033,
"learning_rate": 1.3045830605192266e-08,
"loss": 0.5192,
"step": 3550
},
{
"epoch": 0.9824755071063889,
"grad_norm": 0.8696346972561173,
"learning_rate": 9.799929472902315e-09,
"loss": 0.5205,
"step": 3560
},
{
"epoch": 0.9852352697668001,
"grad_norm": 3.314581793197866,
"learning_rate": 7.017168826401466e-09,
"loss": 0.5169,
"step": 3570
},
{
"epoch": 0.9879950324272112,
"grad_norm": 0.8566695475770818,
"learning_rate": 4.697806933715021e-09,
"loss": 0.5095,
"step": 3580
},
{
"epoch": 0.9907547950876224,
"grad_norm": 0.8718028954806468,
"learning_rate": 2.8420590548294825e-09,
"loss": 0.5175,
"step": 3590
},
{
"epoch": 0.9935145577480337,
"grad_norm": 0.8678996243220692,
"learning_rate": 1.450097421710206e-09,
"loss": 0.519,
"step": 3600
},
{
"epoch": 0.9962743204084449,
"grad_norm": 0.8913548884246177,
"learning_rate": 5.220512223219621e-10,
"loss": 0.5135,
"step": 3610
},
{
"epoch": 0.9990340830688561,
"grad_norm": 0.8708615108897616,
"learning_rate": 5.8006588636305704e-11,
"loss": 0.5151,
"step": 3620
},
{
"epoch": 1.0,
"step": 3624,
"total_flos": 1884305306419200.0,
"train_loss": 0.5639046454929622,
"train_runtime": 12600.7671,
"train_samples_per_second": 23.004,
"train_steps_per_second": 0.288
}
],
"logging_steps": 10,
"max_steps": 3624,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 200,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1884305306419200.0,
"train_batch_size": 5,
"trial_name": null,
"trial_params": null
}