embedding-4b-sql / checkpoint-2400 /trainer_state.json
thanhdath's picture
Upload folder using huggingface_hub
fb7ede6 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.10710221567708682,
"eval_steps": 200.0,
"global_step": 2400,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 4.4625923198786177e-05,
"grad_norm": 0.38751721382141113,
"learning_rate": 2.677376171352075e-09,
"loss": 5.556500434875488,
"step": 1
},
{
"epoch": 0.00022312961599393088,
"grad_norm": 0.37884023785591125,
"learning_rate": 1.3386880856760375e-08,
"loss": 5.5404791831970215,
"step": 5
},
{
"epoch": 0.00044625923198786177,
"grad_norm": 0.38282614946365356,
"learning_rate": 2.677376171352075e-08,
"loss": 5.552686309814453,
"step": 10
},
{
"epoch": 0.0006693888479817927,
"grad_norm": 0.3849342167377472,
"learning_rate": 4.016064257028113e-08,
"loss": 5.491200637817383,
"step": 15
},
{
"epoch": 0.0008925184639757235,
"grad_norm": 0.3901270925998688,
"learning_rate": 5.35475234270415e-08,
"loss": 5.454954147338867,
"step": 20
},
{
"epoch": 0.0011156480799696543,
"grad_norm": 0.38169801235198975,
"learning_rate": 6.693440428380188e-08,
"loss": 5.510653686523438,
"step": 25
},
{
"epoch": 0.0013387776959635854,
"grad_norm": 0.37212520837783813,
"learning_rate": 8.032128514056226e-08,
"loss": 5.510681915283203,
"step": 30
},
{
"epoch": 0.0015619073119575162,
"grad_norm": 0.37682151794433594,
"learning_rate": 9.370816599732262e-08,
"loss": 5.651955413818359,
"step": 35
},
{
"epoch": 0.001785036927951447,
"grad_norm": 0.4001241624355316,
"learning_rate": 1.07095046854083e-07,
"loss": 5.452754211425781,
"step": 40
},
{
"epoch": 0.002008166543945378,
"grad_norm": 0.405489057302475,
"learning_rate": 1.2048192771084337e-07,
"loss": 5.520745849609375,
"step": 45
},
{
"epoch": 0.0022312961599393086,
"grad_norm": 0.39249682426452637,
"learning_rate": 1.3386880856760375e-07,
"loss": 5.529154205322266,
"step": 50
},
{
"epoch": 0.0024544257759332396,
"grad_norm": 0.3897538483142853,
"learning_rate": 1.4725568942436413e-07,
"loss": 5.533833312988281,
"step": 55
},
{
"epoch": 0.0026775553919271707,
"grad_norm": 0.3601232171058655,
"learning_rate": 1.606425702811245e-07,
"loss": 5.433098220825196,
"step": 60
},
{
"epoch": 0.0029006850079211013,
"grad_norm": 0.3851630985736847,
"learning_rate": 1.7402945113788486e-07,
"loss": 5.608365631103515,
"step": 65
},
{
"epoch": 0.0031238146239150324,
"grad_norm": 0.37249451875686646,
"learning_rate": 1.8741633199464525e-07,
"loss": 5.5082344055175785,
"step": 70
},
{
"epoch": 0.003346944239908963,
"grad_norm": 0.37606358528137207,
"learning_rate": 2.0080321285140563e-07,
"loss": 5.465315246582032,
"step": 75
},
{
"epoch": 0.003570073855902894,
"grad_norm": 0.3777235448360443,
"learning_rate": 2.14190093708166e-07,
"loss": 5.494733428955078,
"step": 80
},
{
"epoch": 0.0037932034718968248,
"grad_norm": 0.3441930115222931,
"learning_rate": 2.2757697456492636e-07,
"loss": 5.509701538085937,
"step": 85
},
{
"epoch": 0.004016333087890756,
"grad_norm": 0.41749539971351624,
"learning_rate": 2.4096385542168674e-07,
"loss": 5.484718704223633,
"step": 90
},
{
"epoch": 0.004239462703884687,
"grad_norm": 0.426409512758255,
"learning_rate": 2.543507362784471e-07,
"loss": 5.485423278808594,
"step": 95
},
{
"epoch": 0.004462592319878617,
"grad_norm": 0.3890352249145508,
"learning_rate": 2.677376171352075e-07,
"loss": 5.415613174438477,
"step": 100
},
{
"epoch": 0.004685721935872548,
"grad_norm": 0.3934517502784729,
"learning_rate": 2.811244979919679e-07,
"loss": 5.502951049804688,
"step": 105
},
{
"epoch": 0.004908851551866479,
"grad_norm": 0.418043315410614,
"learning_rate": 2.9451137884872826e-07,
"loss": 5.533256149291992,
"step": 110
},
{
"epoch": 0.00513198116786041,
"grad_norm": 0.38715749979019165,
"learning_rate": 3.078982597054886e-07,
"loss": 5.538555908203125,
"step": 115
},
{
"epoch": 0.005355110783854341,
"grad_norm": 0.38706645369529724,
"learning_rate": 3.21285140562249e-07,
"loss": 5.495933151245117,
"step": 120
},
{
"epoch": 0.005578240399848272,
"grad_norm": 0.37968552112579346,
"learning_rate": 3.346720214190094e-07,
"loss": 5.507209014892578,
"step": 125
},
{
"epoch": 0.005801370015842203,
"grad_norm": 0.37994489073753357,
"learning_rate": 3.4805890227576973e-07,
"loss": 5.505420684814453,
"step": 130
},
{
"epoch": 0.006024499631836134,
"grad_norm": 0.37164536118507385,
"learning_rate": 3.614457831325301e-07,
"loss": 5.491921615600586,
"step": 135
},
{
"epoch": 0.006247629247830065,
"grad_norm": 0.3722690939903259,
"learning_rate": 3.748326639892905e-07,
"loss": 5.4893798828125,
"step": 140
},
{
"epoch": 0.006470758863823995,
"grad_norm": 0.37924614548683167,
"learning_rate": 3.882195448460509e-07,
"loss": 5.498228073120117,
"step": 145
},
{
"epoch": 0.006693888479817926,
"grad_norm": 0.4063311517238617,
"learning_rate": 4.0160642570281125e-07,
"loss": 5.492959213256836,
"step": 150
},
{
"epoch": 0.006917018095811857,
"grad_norm": 0.40316569805145264,
"learning_rate": 4.149933065595716e-07,
"loss": 5.330196762084961,
"step": 155
},
{
"epoch": 0.007140147711805788,
"grad_norm": 0.38558679819107056,
"learning_rate": 4.28380187416332e-07,
"loss": 5.483318328857422,
"step": 160
},
{
"epoch": 0.0073632773277997185,
"grad_norm": 0.3990952968597412,
"learning_rate": 4.417670682730924e-07,
"loss": 5.4153087615966795,
"step": 165
},
{
"epoch": 0.0075864069437936495,
"grad_norm": 0.41763484477996826,
"learning_rate": 4.551539491298527e-07,
"loss": 5.46466178894043,
"step": 170
},
{
"epoch": 0.007809536559787581,
"grad_norm": 0.41940048336982727,
"learning_rate": 4.6854082998661315e-07,
"loss": 5.480670166015625,
"step": 175
},
{
"epoch": 0.008032666175781512,
"grad_norm": 0.4029385447502136,
"learning_rate": 4.819277108433735e-07,
"loss": 5.44397201538086,
"step": 180
},
{
"epoch": 0.008255795791775442,
"grad_norm": 0.4002819061279297,
"learning_rate": 4.953145917001339e-07,
"loss": 5.355976867675781,
"step": 185
},
{
"epoch": 0.008478925407769374,
"grad_norm": 0.41026008129119873,
"learning_rate": 5.087014725568942e-07,
"loss": 5.4827728271484375,
"step": 190
},
{
"epoch": 0.008702055023763304,
"grad_norm": 0.4057953655719757,
"learning_rate": 5.220883534136546e-07,
"loss": 5.4723457336425785,
"step": 195
},
{
"epoch": 0.008925184639757234,
"grad_norm": 0.39414113759994507,
"learning_rate": 5.35475234270415e-07,
"loss": 5.458871841430664,
"step": 200
},
{
"epoch": 0.009148314255751166,
"grad_norm": 0.40346527099609375,
"learning_rate": 5.488621151271754e-07,
"loss": 5.332700729370117,
"step": 205
},
{
"epoch": 0.009371443871745096,
"grad_norm": 0.4204488694667816,
"learning_rate": 5.622489959839358e-07,
"loss": 5.45002326965332,
"step": 210
},
{
"epoch": 0.009594573487739028,
"grad_norm": 0.4042298495769501,
"learning_rate": 5.756358768406961e-07,
"loss": 5.469032287597656,
"step": 215
},
{
"epoch": 0.009817703103732959,
"grad_norm": 0.42059728503227234,
"learning_rate": 5.890227576974565e-07,
"loss": 5.416038513183594,
"step": 220
},
{
"epoch": 0.010040832719726889,
"grad_norm": 0.4257930517196655,
"learning_rate": 6.024096385542169e-07,
"loss": 5.365918731689453,
"step": 225
},
{
"epoch": 0.01026396233572082,
"grad_norm": 0.4297005534172058,
"learning_rate": 6.157965194109772e-07,
"loss": 5.386648941040039,
"step": 230
},
{
"epoch": 0.010487091951714751,
"grad_norm": 0.44034355878829956,
"learning_rate": 6.291834002677377e-07,
"loss": 5.4175865173339846,
"step": 235
},
{
"epoch": 0.010710221567708683,
"grad_norm": 0.4148600995540619,
"learning_rate": 6.42570281124498e-07,
"loss": 5.456137084960938,
"step": 240
},
{
"epoch": 0.010933351183702613,
"grad_norm": 0.3927938938140869,
"learning_rate": 6.559571619812583e-07,
"loss": 5.421617889404297,
"step": 245
},
{
"epoch": 0.011156480799696543,
"grad_norm": 0.46364349126815796,
"learning_rate": 6.693440428380188e-07,
"loss": 5.263519287109375,
"step": 250
},
{
"epoch": 0.011379610415690475,
"grad_norm": 0.4380209147930145,
"learning_rate": 6.827309236947791e-07,
"loss": 5.4163257598876955,
"step": 255
},
{
"epoch": 0.011602740031684405,
"grad_norm": 0.44802144169807434,
"learning_rate": 6.961178045515395e-07,
"loss": 5.307229232788086,
"step": 260
},
{
"epoch": 0.011825869647678336,
"grad_norm": 0.48721882700920105,
"learning_rate": 7.095046854082999e-07,
"loss": 5.361153030395508,
"step": 265
},
{
"epoch": 0.012048999263672268,
"grad_norm": 0.4634721875190735,
"learning_rate": 7.228915662650602e-07,
"loss": 5.335968017578125,
"step": 270
},
{
"epoch": 0.012272128879666198,
"grad_norm": 0.44632282853126526,
"learning_rate": 7.362784471218206e-07,
"loss": 5.377736282348633,
"step": 275
},
{
"epoch": 0.01249525849566013,
"grad_norm": 0.46110817790031433,
"learning_rate": 7.49665327978581e-07,
"loss": 5.379433059692383,
"step": 280
},
{
"epoch": 0.01271838811165406,
"grad_norm": 0.4558963477611542,
"learning_rate": 7.630522088353414e-07,
"loss": 5.252085876464844,
"step": 285
},
{
"epoch": 0.01294151772764799,
"grad_norm": 0.4590746760368347,
"learning_rate": 7.764390896921018e-07,
"loss": 5.320807647705078,
"step": 290
},
{
"epoch": 0.013164647343641922,
"grad_norm": 0.458065927028656,
"learning_rate": 7.898259705488621e-07,
"loss": 5.342620086669922,
"step": 295
},
{
"epoch": 0.013387776959635852,
"grad_norm": 0.47713929414749146,
"learning_rate": 8.032128514056225e-07,
"loss": 5.309723281860352,
"step": 300
},
{
"epoch": 0.013610906575629784,
"grad_norm": 0.4549182653427124,
"learning_rate": 8.16599732262383e-07,
"loss": 5.263495635986328,
"step": 305
},
{
"epoch": 0.013834036191623714,
"grad_norm": 0.45380067825317383,
"learning_rate": 8.299866131191432e-07,
"loss": 5.270233535766602,
"step": 310
},
{
"epoch": 0.014057165807617645,
"grad_norm": 0.4339616894721985,
"learning_rate": 8.433734939759036e-07,
"loss": 5.273446655273437,
"step": 315
},
{
"epoch": 0.014280295423611577,
"grad_norm": 0.44467630982398987,
"learning_rate": 8.56760374832664e-07,
"loss": 5.210543823242188,
"step": 320
},
{
"epoch": 0.014503425039605507,
"grad_norm": 0.4964953064918518,
"learning_rate": 8.701472556894243e-07,
"loss": 5.216452026367188,
"step": 325
},
{
"epoch": 0.014726554655599437,
"grad_norm": 0.4745877683162689,
"learning_rate": 8.835341365461848e-07,
"loss": 5.237066650390625,
"step": 330
},
{
"epoch": 0.014949684271593369,
"grad_norm": 0.4923485815525055,
"learning_rate": 8.969210174029452e-07,
"loss": 5.16167106628418,
"step": 335
},
{
"epoch": 0.015172813887587299,
"grad_norm": 0.45548155903816223,
"learning_rate": 9.103078982597054e-07,
"loss": 5.184793090820312,
"step": 340
},
{
"epoch": 0.015395943503581231,
"grad_norm": 0.5323435068130493,
"learning_rate": 9.236947791164659e-07,
"loss": 5.1162841796875,
"step": 345
},
{
"epoch": 0.015619073119575161,
"grad_norm": 0.5220831036567688,
"learning_rate": 9.370816599732263e-07,
"loss": 5.135643005371094,
"step": 350
},
{
"epoch": 0.01584220273556909,
"grad_norm": 0.5357551574707031,
"learning_rate": 9.504685408299866e-07,
"loss": 5.001235580444336,
"step": 355
},
{
"epoch": 0.016065332351563023,
"grad_norm": 0.49958014488220215,
"learning_rate": 9.63855421686747e-07,
"loss": 5.096822357177734,
"step": 360
},
{
"epoch": 0.016288461967556955,
"grad_norm": 0.53453129529953,
"learning_rate": 9.772423025435074e-07,
"loss": 5.057343673706055,
"step": 365
},
{
"epoch": 0.016511591583550884,
"grad_norm": 0.49969372153282166,
"learning_rate": 9.906291834002677e-07,
"loss": 5.097013092041015,
"step": 370
},
{
"epoch": 0.016734721199544816,
"grad_norm": 0.477762907743454,
"learning_rate": 1.0040160642570282e-06,
"loss": 5.05908317565918,
"step": 375
},
{
"epoch": 0.016957850815538748,
"grad_norm": 0.5632966756820679,
"learning_rate": 1.0174029451137885e-06,
"loss": 4.880051040649414,
"step": 380
},
{
"epoch": 0.017180980431532676,
"grad_norm": 0.4806019961833954,
"learning_rate": 1.0307898259705488e-06,
"loss": 4.938255310058594,
"step": 385
},
{
"epoch": 0.017404110047526608,
"grad_norm": 0.5882935523986816,
"learning_rate": 1.0441767068273092e-06,
"loss": 4.972189331054688,
"step": 390
},
{
"epoch": 0.01762723966352054,
"grad_norm": 0.583991289138794,
"learning_rate": 1.0575635876840697e-06,
"loss": 4.8074909210205075,
"step": 395
},
{
"epoch": 0.01785036927951447,
"grad_norm": 0.6130750179290771,
"learning_rate": 1.07095046854083e-06,
"loss": 4.851003265380859,
"step": 400
},
{
"epoch": 0.0180734988955084,
"grad_norm": 0.630670428276062,
"learning_rate": 1.0843373493975905e-06,
"loss": 4.742978286743164,
"step": 405
},
{
"epoch": 0.018296628511502332,
"grad_norm": 0.6021527647972107,
"learning_rate": 1.0977242302543508e-06,
"loss": 4.795417022705078,
"step": 410
},
{
"epoch": 0.018519758127496264,
"grad_norm": 0.6027496457099915,
"learning_rate": 1.111111111111111e-06,
"loss": 4.799094009399414,
"step": 415
},
{
"epoch": 0.018742887743490193,
"grad_norm": 0.62542325258255,
"learning_rate": 1.1244979919678715e-06,
"loss": 4.734822845458984,
"step": 420
},
{
"epoch": 0.018966017359484125,
"grad_norm": 0.6744217872619629,
"learning_rate": 1.137884872824632e-06,
"loss": 4.470468902587891,
"step": 425
},
{
"epoch": 0.019189146975478057,
"grad_norm": 0.6994063854217529,
"learning_rate": 1.1512717536813923e-06,
"loss": 4.7334716796875,
"step": 430
},
{
"epoch": 0.019412276591471985,
"grad_norm": 0.6043664813041687,
"learning_rate": 1.1646586345381526e-06,
"loss": 4.6335186004638675,
"step": 435
},
{
"epoch": 0.019635406207465917,
"grad_norm": 0.600983738899231,
"learning_rate": 1.178045515394913e-06,
"loss": 4.521899795532226,
"step": 440
},
{
"epoch": 0.01985853582345985,
"grad_norm": 0.632734477519989,
"learning_rate": 1.1914323962516733e-06,
"loss": 4.455641555786133,
"step": 445
},
{
"epoch": 0.020081665439453777,
"grad_norm": 0.6841812133789062,
"learning_rate": 1.2048192771084338e-06,
"loss": 4.552269363403321,
"step": 450
},
{
"epoch": 0.02030479505544771,
"grad_norm": 0.7014154195785522,
"learning_rate": 1.2182061579651943e-06,
"loss": 4.4612682342529295,
"step": 455
},
{
"epoch": 0.02052792467144164,
"grad_norm": 0.7957334518432617,
"learning_rate": 1.2315930388219544e-06,
"loss": 4.16539421081543,
"step": 460
},
{
"epoch": 0.02075105428743557,
"grad_norm": 0.6679075360298157,
"learning_rate": 1.2449799196787148e-06,
"loss": 4.380799102783203,
"step": 465
},
{
"epoch": 0.020974183903429502,
"grad_norm": 0.6911934614181519,
"learning_rate": 1.2583668005354753e-06,
"loss": 4.188001251220703,
"step": 470
},
{
"epoch": 0.021197313519423434,
"grad_norm": 0.6941277384757996,
"learning_rate": 1.2717536813922356e-06,
"loss": 4.30323600769043,
"step": 475
},
{
"epoch": 0.021420443135417366,
"grad_norm": 0.6411834359169006,
"learning_rate": 1.285140562248996e-06,
"loss": 4.210566711425781,
"step": 480
},
{
"epoch": 0.021643572751411294,
"grad_norm": 0.6447572112083435,
"learning_rate": 1.2985274431057564e-06,
"loss": 4.145897674560547,
"step": 485
},
{
"epoch": 0.021866702367405226,
"grad_norm": 0.6956306099891663,
"learning_rate": 1.3119143239625166e-06,
"loss": 4.1259113311767575,
"step": 490
},
{
"epoch": 0.022089831983399158,
"grad_norm": 0.6369218826293945,
"learning_rate": 1.3253012048192771e-06,
"loss": 4.050044250488281,
"step": 495
},
{
"epoch": 0.022312961599393086,
"grad_norm": 0.6199172735214233,
"learning_rate": 1.3386880856760376e-06,
"loss": 3.9885379791259767,
"step": 500
},
{
"epoch": 0.02253609121538702,
"grad_norm": 0.6004146933555603,
"learning_rate": 1.3520749665327979e-06,
"loss": 3.9876686096191407,
"step": 505
},
{
"epoch": 0.02275922083138095,
"grad_norm": 0.6573876142501831,
"learning_rate": 1.3654618473895582e-06,
"loss": 3.860630416870117,
"step": 510
},
{
"epoch": 0.02298235044737488,
"grad_norm": 0.6570749878883362,
"learning_rate": 1.3788487282463186e-06,
"loss": 3.987852096557617,
"step": 515
},
{
"epoch": 0.02320548006336881,
"grad_norm": 0.6052077412605286,
"learning_rate": 1.392235609103079e-06,
"loss": 3.773631286621094,
"step": 520
},
{
"epoch": 0.023428609679362743,
"grad_norm": 0.6024802923202515,
"learning_rate": 1.4056224899598394e-06,
"loss": 3.8334468841552733,
"step": 525
},
{
"epoch": 0.02365173929535667,
"grad_norm": 0.6448367238044739,
"learning_rate": 1.4190093708165999e-06,
"loss": 3.709330749511719,
"step": 530
},
{
"epoch": 0.023874868911350603,
"grad_norm": 0.5610556602478027,
"learning_rate": 1.4323962516733602e-06,
"loss": 3.608018493652344,
"step": 535
},
{
"epoch": 0.024097998527344535,
"grad_norm": 0.5798413157463074,
"learning_rate": 1.4457831325301204e-06,
"loss": 3.513960266113281,
"step": 540
},
{
"epoch": 0.024321128143338467,
"grad_norm": 0.5836862921714783,
"learning_rate": 1.459170013386881e-06,
"loss": 3.5928466796875,
"step": 545
},
{
"epoch": 0.024544257759332395,
"grad_norm": 0.5705676078796387,
"learning_rate": 1.4725568942436412e-06,
"loss": 3.5890869140625,
"step": 550
},
{
"epoch": 0.024767387375326327,
"grad_norm": 0.6377549767494202,
"learning_rate": 1.4859437751004017e-06,
"loss": 3.6042083740234374,
"step": 555
},
{
"epoch": 0.02499051699132026,
"grad_norm": 0.5612762570381165,
"learning_rate": 1.499330655957162e-06,
"loss": 3.511037826538086,
"step": 560
},
{
"epoch": 0.025213646607314188,
"grad_norm": 0.5943495035171509,
"learning_rate": 1.5127175368139222e-06,
"loss": 3.5007259368896486,
"step": 565
},
{
"epoch": 0.02543677622330812,
"grad_norm": 0.5872607827186584,
"learning_rate": 1.5261044176706827e-06,
"loss": 3.3742324829101564,
"step": 570
},
{
"epoch": 0.02565990583930205,
"grad_norm": 0.5494210124015808,
"learning_rate": 1.5394912985274432e-06,
"loss": 3.3373321533203124,
"step": 575
},
{
"epoch": 0.02588303545529598,
"grad_norm": 0.5138423442840576,
"learning_rate": 1.5528781793842037e-06,
"loss": 3.346783447265625,
"step": 580
},
{
"epoch": 0.026106165071289912,
"grad_norm": 0.5539407730102539,
"learning_rate": 1.566265060240964e-06,
"loss": 3.183103561401367,
"step": 585
},
{
"epoch": 0.026329294687283844,
"grad_norm": 0.5888996720314026,
"learning_rate": 1.5796519410977242e-06,
"loss": 3.357350540161133,
"step": 590
},
{
"epoch": 0.026552424303277773,
"grad_norm": 0.5445035696029663,
"learning_rate": 1.5930388219544845e-06,
"loss": 3.374500274658203,
"step": 595
},
{
"epoch": 0.026775553919271704,
"grad_norm": 0.5013962388038635,
"learning_rate": 1.606425702811245e-06,
"loss": 3.3687782287597656,
"step": 600
},
{
"epoch": 0.026998683535265636,
"grad_norm": 0.5150508284568787,
"learning_rate": 1.6198125836680055e-06,
"loss": 3.2590232849121095,
"step": 605
},
{
"epoch": 0.02722181315125957,
"grad_norm": 0.603961169719696,
"learning_rate": 1.633199464524766e-06,
"loss": 3.1635900497436524,
"step": 610
},
{
"epoch": 0.027444942767253497,
"grad_norm": 0.5365360975265503,
"learning_rate": 1.6465863453815263e-06,
"loss": 3.3600418090820314,
"step": 615
},
{
"epoch": 0.02766807238324743,
"grad_norm": 0.5401721000671387,
"learning_rate": 1.6599732262382863e-06,
"loss": 3.183880424499512,
"step": 620
},
{
"epoch": 0.02789120199924136,
"grad_norm": 0.5595034956932068,
"learning_rate": 1.6733601070950468e-06,
"loss": 3.0762868881225587,
"step": 625
},
{
"epoch": 0.02811433161523529,
"grad_norm": 0.4972860813140869,
"learning_rate": 1.6867469879518073e-06,
"loss": 3.0724456787109373,
"step": 630
},
{
"epoch": 0.02833746123122922,
"grad_norm": 0.44132477045059204,
"learning_rate": 1.7001338688085678e-06,
"loss": 3.1175127029418945,
"step": 635
},
{
"epoch": 0.028560590847223153,
"grad_norm": 0.5078781843185425,
"learning_rate": 1.713520749665328e-06,
"loss": 3.0849445343017576,
"step": 640
},
{
"epoch": 0.02878372046321708,
"grad_norm": 0.4757557809352875,
"learning_rate": 1.7269076305220885e-06,
"loss": 3.2029571533203125,
"step": 645
},
{
"epoch": 0.029006850079211013,
"grad_norm": 0.4932602345943451,
"learning_rate": 1.7402945113788486e-06,
"loss": 3.0668067932128906,
"step": 650
},
{
"epoch": 0.029229979695204945,
"grad_norm": 0.5687287449836731,
"learning_rate": 1.753681392235609e-06,
"loss": 3.113470268249512,
"step": 655
},
{
"epoch": 0.029453109311198874,
"grad_norm": 0.4471919536590576,
"learning_rate": 1.7670682730923696e-06,
"loss": 2.9105451583862303,
"step": 660
},
{
"epoch": 0.029676238927192806,
"grad_norm": 0.5096343159675598,
"learning_rate": 1.7804551539491298e-06,
"loss": 2.94341926574707,
"step": 665
},
{
"epoch": 0.029899368543186738,
"grad_norm": 0.5599822402000427,
"learning_rate": 1.7938420348058903e-06,
"loss": 2.892704391479492,
"step": 670
},
{
"epoch": 0.03012249815918067,
"grad_norm": 0.40055856108665466,
"learning_rate": 1.8072289156626508e-06,
"loss": 2.896807861328125,
"step": 675
},
{
"epoch": 0.030345627775174598,
"grad_norm": 0.47082021832466125,
"learning_rate": 1.8206157965194109e-06,
"loss": 2.7966148376464846,
"step": 680
},
{
"epoch": 0.03056875739116853,
"grad_norm": 0.42042815685272217,
"learning_rate": 1.8340026773761714e-06,
"loss": 2.839730644226074,
"step": 685
},
{
"epoch": 0.030791887007162462,
"grad_norm": 0.49198001623153687,
"learning_rate": 1.8473895582329318e-06,
"loss": 2.9913705825805663,
"step": 690
},
{
"epoch": 0.03101501662315639,
"grad_norm": 0.4656424820423126,
"learning_rate": 1.8607764390896921e-06,
"loss": 3.0195621490478515,
"step": 695
},
{
"epoch": 0.031238146239150322,
"grad_norm": 0.46259376406669617,
"learning_rate": 1.8741633199464526e-06,
"loss": 2.948585319519043,
"step": 700
},
{
"epoch": 0.03146127585514425,
"grad_norm": 0.4671732783317566,
"learning_rate": 1.887550200803213e-06,
"loss": 2.7285741806030273,
"step": 705
},
{
"epoch": 0.03168440547113818,
"grad_norm": 0.6493679285049438,
"learning_rate": 1.9009370816599732e-06,
"loss": 2.9263893127441407,
"step": 710
},
{
"epoch": 0.031907535087132115,
"grad_norm": 0.46389710903167725,
"learning_rate": 1.9143239625167336e-06,
"loss": 2.634868049621582,
"step": 715
},
{
"epoch": 0.03213066470312605,
"grad_norm": 0.44173645973205566,
"learning_rate": 1.927710843373494e-06,
"loss": 2.7959117889404297,
"step": 720
},
{
"epoch": 0.03235379431911998,
"grad_norm": 0.45440003275871277,
"learning_rate": 1.9410977242302546e-06,
"loss": 2.9056529998779297,
"step": 725
},
{
"epoch": 0.03257692393511391,
"grad_norm": 0.44541046023368835,
"learning_rate": 1.954484605087015e-06,
"loss": 2.7214975357055664,
"step": 730
},
{
"epoch": 0.032800053551107836,
"grad_norm": 0.4217804968357086,
"learning_rate": 1.967871485943775e-06,
"loss": 2.7198978424072267,
"step": 735
},
{
"epoch": 0.03302318316710177,
"grad_norm": 0.46086645126342773,
"learning_rate": 1.9812583668005354e-06,
"loss": 2.7366127014160155,
"step": 740
},
{
"epoch": 0.0332463127830957,
"grad_norm": 0.41817471385002136,
"learning_rate": 1.9946452476572957e-06,
"loss": 2.857570457458496,
"step": 745
},
{
"epoch": 0.03346944239908963,
"grad_norm": 0.44163694977760315,
"learning_rate": 2.0080321285140564e-06,
"loss": 2.8251710891723634,
"step": 750
},
{
"epoch": 0.03369257201508356,
"grad_norm": 0.46645739674568176,
"learning_rate": 2.0214190093708167e-06,
"loss": 2.548258399963379,
"step": 755
},
{
"epoch": 0.033915701631077495,
"grad_norm": 0.3971084654331207,
"learning_rate": 2.034805890227577e-06,
"loss": 2.682722473144531,
"step": 760
},
{
"epoch": 0.03413883124707143,
"grad_norm": 0.5373950004577637,
"learning_rate": 2.0481927710843377e-06,
"loss": 2.8539506912231447,
"step": 765
},
{
"epoch": 0.03436196086306535,
"grad_norm": 0.6133913397789001,
"learning_rate": 2.0615796519410975e-06,
"loss": 2.7016387939453126,
"step": 770
},
{
"epoch": 0.034585090479059284,
"grad_norm": 0.5247603058815002,
"learning_rate": 2.074966532797858e-06,
"loss": 2.645602226257324,
"step": 775
},
{
"epoch": 0.034808220095053216,
"grad_norm": 0.4487764537334442,
"learning_rate": 2.0883534136546185e-06,
"loss": 2.699404716491699,
"step": 780
},
{
"epoch": 0.03503134971104715,
"grad_norm": 0.5962596535682678,
"learning_rate": 2.1017402945113788e-06,
"loss": 2.7891347885131834,
"step": 785
},
{
"epoch": 0.03525447932704108,
"grad_norm": 0.5410242080688477,
"learning_rate": 2.1151271753681395e-06,
"loss": 2.675504684448242,
"step": 790
},
{
"epoch": 0.03547760894303501,
"grad_norm": 0.4576985239982605,
"learning_rate": 2.1285140562248997e-06,
"loss": 2.642281341552734,
"step": 795
},
{
"epoch": 0.03570073855902894,
"grad_norm": 0.45449239015579224,
"learning_rate": 2.14190093708166e-06,
"loss": 2.7088804244995117,
"step": 800
},
{
"epoch": 0.03592386817502287,
"grad_norm": 0.4782140552997589,
"learning_rate": 2.1552878179384203e-06,
"loss": 2.698718452453613,
"step": 805
},
{
"epoch": 0.0361469977910168,
"grad_norm": 0.43745940923690796,
"learning_rate": 2.168674698795181e-06,
"loss": 2.7042917251586913,
"step": 810
},
{
"epoch": 0.03637012740701073,
"grad_norm": 0.6170194149017334,
"learning_rate": 2.1820615796519413e-06,
"loss": 2.6548912048339846,
"step": 815
},
{
"epoch": 0.036593257023004665,
"grad_norm": 0.5126772522926331,
"learning_rate": 2.1954484605087015e-06,
"loss": 2.663498306274414,
"step": 820
},
{
"epoch": 0.0368163866389986,
"grad_norm": 0.42503371834754944,
"learning_rate": 2.2088353413654622e-06,
"loss": 2.766114616394043,
"step": 825
},
{
"epoch": 0.03703951625499253,
"grad_norm": 0.5707629323005676,
"learning_rate": 2.222222222222222e-06,
"loss": 2.6172439575195314,
"step": 830
},
{
"epoch": 0.037262645870986454,
"grad_norm": 0.45169416069984436,
"learning_rate": 2.2356091030789828e-06,
"loss": 2.5655603408813477,
"step": 835
},
{
"epoch": 0.037485775486980386,
"grad_norm": 0.44726496934890747,
"learning_rate": 2.248995983935743e-06,
"loss": 2.5273290634155274,
"step": 840
},
{
"epoch": 0.03770890510297432,
"grad_norm": 0.5287439823150635,
"learning_rate": 2.2623828647925033e-06,
"loss": 2.6541069030761717,
"step": 845
},
{
"epoch": 0.03793203471896825,
"grad_norm": 0.46616867184638977,
"learning_rate": 2.275769745649264e-06,
"loss": 2.56517333984375,
"step": 850
},
{
"epoch": 0.03815516433496218,
"grad_norm": 0.4369581937789917,
"learning_rate": 2.2891566265060243e-06,
"loss": 2.4222312927246095,
"step": 855
},
{
"epoch": 0.03837829395095611,
"grad_norm": 0.4172717332839966,
"learning_rate": 2.3025435073627846e-06,
"loss": 2.549334716796875,
"step": 860
},
{
"epoch": 0.03860142356695004,
"grad_norm": 0.46124762296676636,
"learning_rate": 2.315930388219545e-06,
"loss": 2.5675996780395507,
"step": 865
},
{
"epoch": 0.03882455318294397,
"grad_norm": 0.4226182699203491,
"learning_rate": 2.329317269076305e-06,
"loss": 2.4144569396972657,
"step": 870
},
{
"epoch": 0.0390476827989379,
"grad_norm": 0.4673754870891571,
"learning_rate": 2.342704149933066e-06,
"loss": 2.3829999923706056,
"step": 875
},
{
"epoch": 0.039270812414931834,
"grad_norm": 0.5152049660682678,
"learning_rate": 2.356091030789826e-06,
"loss": 2.5030202865600586,
"step": 880
},
{
"epoch": 0.039493942030925766,
"grad_norm": 0.5377767086029053,
"learning_rate": 2.3694779116465864e-06,
"loss": 2.7714466094970702,
"step": 885
},
{
"epoch": 0.0397170716469197,
"grad_norm": 0.45861950516700745,
"learning_rate": 2.3828647925033466e-06,
"loss": 2.461492729187012,
"step": 890
},
{
"epoch": 0.03994020126291363,
"grad_norm": 0.4814665615558624,
"learning_rate": 2.396251673360107e-06,
"loss": 2.4756641387939453,
"step": 895
},
{
"epoch": 0.040163330878907555,
"grad_norm": 0.5024599432945251,
"learning_rate": 2.4096385542168676e-06,
"loss": 2.5811389923095702,
"step": 900
},
{
"epoch": 0.04038646049490149,
"grad_norm": 0.45141535997390747,
"learning_rate": 2.423025435073628e-06,
"loss": 2.686309814453125,
"step": 905
},
{
"epoch": 0.04060959011089542,
"grad_norm": 0.4806381165981293,
"learning_rate": 2.4364123159303886e-06,
"loss": 2.530971717834473,
"step": 910
},
{
"epoch": 0.04083271972688935,
"grad_norm": 0.7574505805969238,
"learning_rate": 2.449799196787149e-06,
"loss": 2.5287572860717775,
"step": 915
},
{
"epoch": 0.04105584934288328,
"grad_norm": 0.39056211709976196,
"learning_rate": 2.4631860776439087e-06,
"loss": 2.4373926162719726,
"step": 920
},
{
"epoch": 0.041278978958877215,
"grad_norm": 0.4549945890903473,
"learning_rate": 2.4765729585006694e-06,
"loss": 2.508747100830078,
"step": 925
},
{
"epoch": 0.04150210857487114,
"grad_norm": 0.49253398180007935,
"learning_rate": 2.4899598393574297e-06,
"loss": 2.5109813690185545,
"step": 930
},
{
"epoch": 0.04172523819086507,
"grad_norm": 0.4185622036457062,
"learning_rate": 2.5033467202141904e-06,
"loss": 2.5602550506591797,
"step": 935
},
{
"epoch": 0.041948367806859004,
"grad_norm": 0.5255184769630432,
"learning_rate": 2.5167336010709507e-06,
"loss": 2.5018032073974608,
"step": 940
},
{
"epoch": 0.042171497422852935,
"grad_norm": 0.40105360746383667,
"learning_rate": 2.530120481927711e-06,
"loss": 2.3703737258911133,
"step": 945
},
{
"epoch": 0.04239462703884687,
"grad_norm": 0.5396221876144409,
"learning_rate": 2.543507362784471e-06,
"loss": 2.46767520904541,
"step": 950
},
{
"epoch": 0.0426177566548408,
"grad_norm": 0.4686439633369446,
"learning_rate": 2.5568942436412315e-06,
"loss": 2.535739517211914,
"step": 955
},
{
"epoch": 0.04284088627083473,
"grad_norm": 0.5110803246498108,
"learning_rate": 2.570281124497992e-06,
"loss": 2.5248756408691406,
"step": 960
},
{
"epoch": 0.043064015886828656,
"grad_norm": 0.5164591670036316,
"learning_rate": 2.5836680053547524e-06,
"loss": 2.4275962829589846,
"step": 965
},
{
"epoch": 0.04328714550282259,
"grad_norm": 0.42096105217933655,
"learning_rate": 2.5970548862115127e-06,
"loss": 2.4086238861083986,
"step": 970
},
{
"epoch": 0.04351027511881652,
"grad_norm": 0.5388748645782471,
"learning_rate": 2.6104417670682734e-06,
"loss": 2.340771293640137,
"step": 975
},
{
"epoch": 0.04373340473481045,
"grad_norm": 0.4667651653289795,
"learning_rate": 2.6238286479250333e-06,
"loss": 2.506967544555664,
"step": 980
},
{
"epoch": 0.043956534350804384,
"grad_norm": 0.5090363621711731,
"learning_rate": 2.637215528781794e-06,
"loss": 2.251785469055176,
"step": 985
},
{
"epoch": 0.044179663966798316,
"grad_norm": 0.4600473940372467,
"learning_rate": 2.6506024096385542e-06,
"loss": 2.3269046783447265,
"step": 990
},
{
"epoch": 0.04440279358279224,
"grad_norm": 0.4453408718109131,
"learning_rate": 2.6639892904953145e-06,
"loss": 2.4940771102905273,
"step": 995
},
{
"epoch": 0.04462592319878617,
"grad_norm": 0.4637933671474457,
"learning_rate": 2.6773761713520752e-06,
"loss": 2.578083801269531,
"step": 1000
},
{
"epoch": 0.044849052814780105,
"grad_norm": 0.48008954524993896,
"learning_rate": 2.6907630522088355e-06,
"loss": 2.509678077697754,
"step": 1005
},
{
"epoch": 0.04507218243077404,
"grad_norm": 0.5186890363693237,
"learning_rate": 2.7041499330655958e-06,
"loss": 2.420572280883789,
"step": 1010
},
{
"epoch": 0.04529531204676797,
"grad_norm": 0.5090295076370239,
"learning_rate": 2.717536813922356e-06,
"loss": 2.42071533203125,
"step": 1015
},
{
"epoch": 0.0455184416627619,
"grad_norm": 0.5519959926605225,
"learning_rate": 2.7309236947791163e-06,
"loss": 2.458280563354492,
"step": 1020
},
{
"epoch": 0.04574157127875583,
"grad_norm": 0.5525882840156555,
"learning_rate": 2.744310575635877e-06,
"loss": 2.3400331497192384,
"step": 1025
},
{
"epoch": 0.04596470089474976,
"grad_norm": 0.48508089780807495,
"learning_rate": 2.7576974564926373e-06,
"loss": 2.4403892517089845,
"step": 1030
},
{
"epoch": 0.04618783051074369,
"grad_norm": 0.5230780839920044,
"learning_rate": 2.771084337349398e-06,
"loss": 2.3652227401733397,
"step": 1035
},
{
"epoch": 0.04641096012673762,
"grad_norm": 0.5132386088371277,
"learning_rate": 2.784471218206158e-06,
"loss": 2.313191604614258,
"step": 1040
},
{
"epoch": 0.04663408974273155,
"grad_norm": 0.490017294883728,
"learning_rate": 2.7978580990629185e-06,
"loss": 2.24790096282959,
"step": 1045
},
{
"epoch": 0.046857219358725485,
"grad_norm": 0.5250972509384155,
"learning_rate": 2.811244979919679e-06,
"loss": 2.1867441177368163,
"step": 1050
},
{
"epoch": 0.04708034897471942,
"grad_norm": 0.45831558108329773,
"learning_rate": 2.824631860776439e-06,
"loss": 2.379189300537109,
"step": 1055
},
{
"epoch": 0.04730347859071334,
"grad_norm": 0.5892948508262634,
"learning_rate": 2.8380187416331998e-06,
"loss": 2.2735191345214845,
"step": 1060
},
{
"epoch": 0.047526608206707274,
"grad_norm": 0.5409689545631409,
"learning_rate": 2.85140562248996e-06,
"loss": 2.3979333877563476,
"step": 1065
},
{
"epoch": 0.047749737822701206,
"grad_norm": 0.5346503257751465,
"learning_rate": 2.8647925033467203e-06,
"loss": 2.4819768905639648,
"step": 1070
},
{
"epoch": 0.04797286743869514,
"grad_norm": 0.6150287985801697,
"learning_rate": 2.8781793842034806e-06,
"loss": 2.5218339920043946,
"step": 1075
},
{
"epoch": 0.04819599705468907,
"grad_norm": 0.4347957968711853,
"learning_rate": 2.891566265060241e-06,
"loss": 2.1929386138916014,
"step": 1080
},
{
"epoch": 0.048419126670683,
"grad_norm": 0.5238802433013916,
"learning_rate": 2.9049531459170016e-06,
"loss": 2.215384292602539,
"step": 1085
},
{
"epoch": 0.048642256286676934,
"grad_norm": 0.5371809601783752,
"learning_rate": 2.918340026773762e-06,
"loss": 2.4019641876220703,
"step": 1090
},
{
"epoch": 0.04886538590267086,
"grad_norm": 0.5125853419303894,
"learning_rate": 2.931726907630522e-06,
"loss": 2.3410377502441406,
"step": 1095
},
{
"epoch": 0.04908851551866479,
"grad_norm": 0.5087509751319885,
"learning_rate": 2.9451137884872824e-06,
"loss": 2.3634616851806642,
"step": 1100
},
{
"epoch": 0.04931164513465872,
"grad_norm": 0.5605607628822327,
"learning_rate": 2.9585006693440427e-06,
"loss": 2.239984130859375,
"step": 1105
},
{
"epoch": 0.049534774750652655,
"grad_norm": 0.6371492743492126,
"learning_rate": 2.9718875502008034e-06,
"loss": 2.499210739135742,
"step": 1110
},
{
"epoch": 0.04975790436664659,
"grad_norm": 0.5526962280273438,
"learning_rate": 2.9852744310575636e-06,
"loss": 2.433857536315918,
"step": 1115
},
{
"epoch": 0.04998103398264052,
"grad_norm": 0.5811582803726196,
"learning_rate": 2.998661311914324e-06,
"loss": 2.372517967224121,
"step": 1120
},
{
"epoch": 0.050204163598634444,
"grad_norm": 0.47710540890693665,
"learning_rate": 3.0120481927710846e-06,
"loss": 2.441071891784668,
"step": 1125
},
{
"epoch": 0.050427293214628376,
"grad_norm": 0.49463921785354614,
"learning_rate": 3.0254350736278445e-06,
"loss": 2.218737030029297,
"step": 1130
},
{
"epoch": 0.05065042283062231,
"grad_norm": 0.7574100494384766,
"learning_rate": 3.0388219544846056e-06,
"loss": 2.1490055084228517,
"step": 1135
},
{
"epoch": 0.05087355244661624,
"grad_norm": 0.5574637651443481,
"learning_rate": 3.0522088353413654e-06,
"loss": 2.3484785079956056,
"step": 1140
},
{
"epoch": 0.05109668206261017,
"grad_norm": 0.5128051042556763,
"learning_rate": 3.0655957161981257e-06,
"loss": 2.2358184814453126,
"step": 1145
},
{
"epoch": 0.0513198116786041,
"grad_norm": 0.4747551679611206,
"learning_rate": 3.0789825970548864e-06,
"loss": 2.2787382125854494,
"step": 1150
},
{
"epoch": 0.051542941294598035,
"grad_norm": 0.4365915358066559,
"learning_rate": 3.0923694779116467e-06,
"loss": 2.327534484863281,
"step": 1155
},
{
"epoch": 0.05176607091059196,
"grad_norm": 0.6433975696563721,
"learning_rate": 3.1057563587684074e-06,
"loss": 2.288431167602539,
"step": 1160
},
{
"epoch": 0.05198920052658589,
"grad_norm": 0.6102743744850159,
"learning_rate": 3.1191432396251672e-06,
"loss": 2.24114933013916,
"step": 1165
},
{
"epoch": 0.052212330142579824,
"grad_norm": 0.5493948459625244,
"learning_rate": 3.132530120481928e-06,
"loss": 2.317499542236328,
"step": 1170
},
{
"epoch": 0.052435459758573756,
"grad_norm": 0.5411773324012756,
"learning_rate": 3.145917001338688e-06,
"loss": 2.1778676986694334,
"step": 1175
},
{
"epoch": 0.05265858937456769,
"grad_norm": 0.5491811633110046,
"learning_rate": 3.1593038821954485e-06,
"loss": 2.260644721984863,
"step": 1180
},
{
"epoch": 0.05288171899056162,
"grad_norm": 0.5138102769851685,
"learning_rate": 3.172690763052209e-06,
"loss": 2.2743204116821287,
"step": 1185
},
{
"epoch": 0.053104848606555545,
"grad_norm": 0.5310954451560974,
"learning_rate": 3.186077643908969e-06,
"loss": 2.257237434387207,
"step": 1190
},
{
"epoch": 0.05332797822254948,
"grad_norm": 0.5704927444458008,
"learning_rate": 3.1994645247657297e-06,
"loss": 2.0418323516845702,
"step": 1195
},
{
"epoch": 0.05355110783854341,
"grad_norm": 0.5295515656471252,
"learning_rate": 3.21285140562249e-06,
"loss": 2.282415771484375,
"step": 1200
},
{
"epoch": 0.05377423745453734,
"grad_norm": 0.6476492881774902,
"learning_rate": 3.2262382864792503e-06,
"loss": 2.306402015686035,
"step": 1205
},
{
"epoch": 0.05399736707053127,
"grad_norm": 0.6115656495094299,
"learning_rate": 3.239625167336011e-06,
"loss": 2.1535377502441406,
"step": 1210
},
{
"epoch": 0.054220496686525205,
"grad_norm": 0.5802296996116638,
"learning_rate": 3.253012048192771e-06,
"loss": 2.2679920196533203,
"step": 1215
},
{
"epoch": 0.05444362630251914,
"grad_norm": 0.5182514190673828,
"learning_rate": 3.266398929049532e-06,
"loss": 2.3105010986328125,
"step": 1220
},
{
"epoch": 0.05466675591851306,
"grad_norm": 0.6288827657699585,
"learning_rate": 3.279785809906292e-06,
"loss": 2.0794536590576174,
"step": 1225
},
{
"epoch": 0.054889885534506994,
"grad_norm": 0.7542652487754822,
"learning_rate": 3.2931726907630525e-06,
"loss": 2.3161176681518554,
"step": 1230
},
{
"epoch": 0.055113015150500926,
"grad_norm": 0.7416955232620239,
"learning_rate": 3.3065595716198128e-06,
"loss": 2.279362106323242,
"step": 1235
},
{
"epoch": 0.05533614476649486,
"grad_norm": 0.5682216882705688,
"learning_rate": 3.3199464524765726e-06,
"loss": 2.2426963806152345,
"step": 1240
},
{
"epoch": 0.05555927438248879,
"grad_norm": 0.6715278625488281,
"learning_rate": 3.3333333333333337e-06,
"loss": 2.1546775817871096,
"step": 1245
},
{
"epoch": 0.05578240399848272,
"grad_norm": 0.5482577085494995,
"learning_rate": 3.3467202141900936e-06,
"loss": 2.1484096527099608,
"step": 1250
},
{
"epoch": 0.056005533614476646,
"grad_norm": 0.6009266972541809,
"learning_rate": 3.3601070950468543e-06,
"loss": 2.221162796020508,
"step": 1255
},
{
"epoch": 0.05622866323047058,
"grad_norm": 0.6640043258666992,
"learning_rate": 3.3734939759036146e-06,
"loss": 2.338100242614746,
"step": 1260
},
{
"epoch": 0.05645179284646451,
"grad_norm": 0.6514598727226257,
"learning_rate": 3.3868808567603744e-06,
"loss": 2.2825199127197267,
"step": 1265
},
{
"epoch": 0.05667492246245844,
"grad_norm": 0.6053763031959534,
"learning_rate": 3.4002677376171355e-06,
"loss": 2.280612754821777,
"step": 1270
},
{
"epoch": 0.056898052078452374,
"grad_norm": 0.8854132294654846,
"learning_rate": 3.4136546184738954e-06,
"loss": 2.3481237411499025,
"step": 1275
},
{
"epoch": 0.057121181694446306,
"grad_norm": 0.6433872580528259,
"learning_rate": 3.427041499330656e-06,
"loss": 2.2561552047729494,
"step": 1280
},
{
"epoch": 0.05734431131044024,
"grad_norm": 0.709402859210968,
"learning_rate": 3.4404283801874164e-06,
"loss": 2.264519691467285,
"step": 1285
},
{
"epoch": 0.05756744092643416,
"grad_norm": 0.5431029796600342,
"learning_rate": 3.453815261044177e-06,
"loss": 2.163587760925293,
"step": 1290
},
{
"epoch": 0.057790570542428095,
"grad_norm": 0.7146894931793213,
"learning_rate": 3.4672021419009373e-06,
"loss": 2.2624624252319334,
"step": 1295
},
{
"epoch": 0.05801370015842203,
"grad_norm": 0.6029372215270996,
"learning_rate": 3.480589022757697e-06,
"loss": 2.0209115982055663,
"step": 1300
},
{
"epoch": 0.05823682977441596,
"grad_norm": 0.6099830865859985,
"learning_rate": 3.493975903614458e-06,
"loss": 2.177568054199219,
"step": 1305
},
{
"epoch": 0.05845995939040989,
"grad_norm": 0.8421804308891296,
"learning_rate": 3.507362784471218e-06,
"loss": 2.1980712890625,
"step": 1310
},
{
"epoch": 0.05868308900640382,
"grad_norm": 0.7849622368812561,
"learning_rate": 3.520749665327979e-06,
"loss": 2.2047344207763673,
"step": 1315
},
{
"epoch": 0.05890621862239775,
"grad_norm": 0.5906243324279785,
"learning_rate": 3.534136546184739e-06,
"loss": 2.3746465682983398,
"step": 1320
},
{
"epoch": 0.05912934823839168,
"grad_norm": 0.9020318388938904,
"learning_rate": 3.547523427041499e-06,
"loss": 2.2612667083740234,
"step": 1325
},
{
"epoch": 0.05935247785438561,
"grad_norm": 0.6542187929153442,
"learning_rate": 3.5609103078982597e-06,
"loss": 2.2060800552368165,
"step": 1330
},
{
"epoch": 0.059575607470379544,
"grad_norm": 0.7428557872772217,
"learning_rate": 3.57429718875502e-06,
"loss": 2.0877670288085937,
"step": 1335
},
{
"epoch": 0.059798737086373475,
"grad_norm": 0.7468170523643494,
"learning_rate": 3.5876840696117807e-06,
"loss": 2.3160980224609373,
"step": 1340
},
{
"epoch": 0.06002186670236741,
"grad_norm": 0.8021371960639954,
"learning_rate": 3.601070950468541e-06,
"loss": 2.1923311233520506,
"step": 1345
},
{
"epoch": 0.06024499631836134,
"grad_norm": 0.6904175281524658,
"learning_rate": 3.6144578313253016e-06,
"loss": 2.1911346435546877,
"step": 1350
},
{
"epoch": 0.060468125934355264,
"grad_norm": 0.7336163520812988,
"learning_rate": 3.627844712182062e-06,
"loss": 2.2488683700561523,
"step": 1355
},
{
"epoch": 0.060691255550349196,
"grad_norm": 0.7925560474395752,
"learning_rate": 3.6412315930388218e-06,
"loss": 2.174961280822754,
"step": 1360
},
{
"epoch": 0.06091438516634313,
"grad_norm": 0.8825748562812805,
"learning_rate": 3.6546184738955825e-06,
"loss": 2.012922668457031,
"step": 1365
},
{
"epoch": 0.06113751478233706,
"grad_norm": 0.6754929423332214,
"learning_rate": 3.6680053547523427e-06,
"loss": 2.216470146179199,
"step": 1370
},
{
"epoch": 0.06136064439833099,
"grad_norm": 0.6616571545600891,
"learning_rate": 3.6813922356091034e-06,
"loss": 2.067160415649414,
"step": 1375
},
{
"epoch": 0.061583774014324924,
"grad_norm": 0.7033371925354004,
"learning_rate": 3.6947791164658637e-06,
"loss": 2.2197628021240234,
"step": 1380
},
{
"epoch": 0.06180690363031885,
"grad_norm": 0.7256376147270203,
"learning_rate": 3.7081659973226235e-06,
"loss": 2.2435529708862303,
"step": 1385
},
{
"epoch": 0.06203003324631278,
"grad_norm": 0.6585844159126282,
"learning_rate": 3.7215528781793842e-06,
"loss": 2.113719177246094,
"step": 1390
},
{
"epoch": 0.06225316286230671,
"grad_norm": 0.7192341685295105,
"learning_rate": 3.7349397590361445e-06,
"loss": 2.2271860122680662,
"step": 1395
},
{
"epoch": 0.062476292478300645,
"grad_norm": 0.7313777804374695,
"learning_rate": 3.7483266398929052e-06,
"loss": 2.312770462036133,
"step": 1400
},
{
"epoch": 0.06269942209429458,
"grad_norm": 0.7106947302818298,
"learning_rate": 3.7617135207496655e-06,
"loss": 2.041255760192871,
"step": 1405
},
{
"epoch": 0.0629225517102885,
"grad_norm": 0.7265296578407288,
"learning_rate": 3.775100401606426e-06,
"loss": 2.2530691146850588,
"step": 1410
},
{
"epoch": 0.06314568132628244,
"grad_norm": 0.7085316777229309,
"learning_rate": 3.788487282463186e-06,
"loss": 2.03436222076416,
"step": 1415
},
{
"epoch": 0.06336881094227637,
"grad_norm": 0.7872986793518066,
"learning_rate": 3.8018741633199463e-06,
"loss": 2.0966476440429687,
"step": 1420
},
{
"epoch": 0.0635919405582703,
"grad_norm": 0.8961313366889954,
"learning_rate": 3.8152610441767074e-06,
"loss": 2.1432809829711914,
"step": 1425
},
{
"epoch": 0.06381507017426423,
"grad_norm": 0.7127321362495422,
"learning_rate": 3.828647925033467e-06,
"loss": 2.2022947311401366,
"step": 1430
},
{
"epoch": 0.06403819979025815,
"grad_norm": 0.8228124976158142,
"learning_rate": 3.842034805890228e-06,
"loss": 2.39965705871582,
"step": 1435
},
{
"epoch": 0.0642613294062521,
"grad_norm": 0.775909960269928,
"learning_rate": 3.855421686746988e-06,
"loss": 2.1313138961791993,
"step": 1440
},
{
"epoch": 0.06448445902224602,
"grad_norm": 0.7945120334625244,
"learning_rate": 3.8688085676037485e-06,
"loss": 2.074580955505371,
"step": 1445
},
{
"epoch": 0.06470758863823996,
"grad_norm": 1.014404535293579,
"learning_rate": 3.882195448460509e-06,
"loss": 2.26824893951416,
"step": 1450
},
{
"epoch": 0.06493071825423388,
"grad_norm": 0.986544132232666,
"learning_rate": 3.895582329317269e-06,
"loss": 2.207651138305664,
"step": 1455
},
{
"epoch": 0.06515384787022782,
"grad_norm": 0.7337960004806519,
"learning_rate": 3.90896921017403e-06,
"loss": 2.2294179916381838,
"step": 1460
},
{
"epoch": 0.06537697748622175,
"grad_norm": 1.2697299718856812,
"learning_rate": 3.92235609103079e-06,
"loss": 2.114596366882324,
"step": 1465
},
{
"epoch": 0.06560010710221567,
"grad_norm": 0.8556989431381226,
"learning_rate": 3.93574297188755e-06,
"loss": 2.21254940032959,
"step": 1470
},
{
"epoch": 0.06582323671820961,
"grad_norm": 0.9477710723876953,
"learning_rate": 3.949129852744311e-06,
"loss": 1.977022361755371,
"step": 1475
},
{
"epoch": 0.06604636633420354,
"grad_norm": 1.7189115285873413,
"learning_rate": 3.962516733601071e-06,
"loss": 2.012314224243164,
"step": 1480
},
{
"epoch": 0.06626949595019747,
"grad_norm": 0.826131284236908,
"learning_rate": 3.975903614457832e-06,
"loss": 2.193516731262207,
"step": 1485
},
{
"epoch": 0.0664926255661914,
"grad_norm": 1.0023162364959717,
"learning_rate": 3.9892904953145914e-06,
"loss": 2.2499979019165037,
"step": 1490
},
{
"epoch": 0.06671575518218534,
"grad_norm": 0.7376646399497986,
"learning_rate": 4.002677376171352e-06,
"loss": 2.1386489868164062,
"step": 1495
},
{
"epoch": 0.06693888479817926,
"grad_norm": 0.7226489782333374,
"learning_rate": 4.016064257028113e-06,
"loss": 2.11199836730957,
"step": 1500
},
{
"epoch": 0.06716201441417319,
"grad_norm": 0.8427999019622803,
"learning_rate": 4.029451137884873e-06,
"loss": 2.034942626953125,
"step": 1505
},
{
"epoch": 0.06738514403016713,
"grad_norm": 0.79557204246521,
"learning_rate": 4.042838018741633e-06,
"loss": 2.145208549499512,
"step": 1510
},
{
"epoch": 0.06760827364616105,
"grad_norm": 0.7423805594444275,
"learning_rate": 4.056224899598393e-06,
"loss": 2.2282032012939452,
"step": 1515
},
{
"epoch": 0.06783140326215499,
"grad_norm": 0.8701015114784241,
"learning_rate": 4.069611780455154e-06,
"loss": 2.0600002288818358,
"step": 1520
},
{
"epoch": 0.06805453287814892,
"grad_norm": 0.9967679977416992,
"learning_rate": 4.082998661311915e-06,
"loss": 2.20194034576416,
"step": 1525
},
{
"epoch": 0.06827766249414285,
"grad_norm": 1.0514639616012573,
"learning_rate": 4.096385542168675e-06,
"loss": 2.250165557861328,
"step": 1530
},
{
"epoch": 0.06850079211013678,
"grad_norm": 0.7402302026748657,
"learning_rate": 4.109772423025435e-06,
"loss": 2.0741525650024415,
"step": 1535
},
{
"epoch": 0.0687239217261307,
"grad_norm": 1.0764678716659546,
"learning_rate": 4.123159303882195e-06,
"loss": 2.068874549865723,
"step": 1540
},
{
"epoch": 0.06894705134212464,
"grad_norm": 0.9718510508537292,
"learning_rate": 4.136546184738956e-06,
"loss": 1.9750284194946288,
"step": 1545
},
{
"epoch": 0.06917018095811857,
"grad_norm": 0.7534152269363403,
"learning_rate": 4.149933065595716e-06,
"loss": 2.1882530212402345,
"step": 1550
},
{
"epoch": 0.06939331057411251,
"grad_norm": 0.8103408813476562,
"learning_rate": 4.163319946452477e-06,
"loss": 2.076101875305176,
"step": 1555
},
{
"epoch": 0.06961644019010643,
"grad_norm": 0.7058753967285156,
"learning_rate": 4.176706827309237e-06,
"loss": 2.1349681854248046,
"step": 1560
},
{
"epoch": 0.06983956980610036,
"grad_norm": 0.664568305015564,
"learning_rate": 4.190093708165997e-06,
"loss": 2.1840320587158204,
"step": 1565
},
{
"epoch": 0.0700626994220943,
"grad_norm": 0.9274365901947021,
"learning_rate": 4.2034805890227575e-06,
"loss": 2.0162349700927735,
"step": 1570
},
{
"epoch": 0.07028582903808822,
"grad_norm": 0.806480348110199,
"learning_rate": 4.216867469879518e-06,
"loss": 2.1628509521484376,
"step": 1575
},
{
"epoch": 0.07050895865408216,
"grad_norm": 0.9540587067604065,
"learning_rate": 4.230254350736279e-06,
"loss": 2.239594841003418,
"step": 1580
},
{
"epoch": 0.07073208827007609,
"grad_norm": 0.8032299876213074,
"learning_rate": 4.243641231593039e-06,
"loss": 2.170275115966797,
"step": 1585
},
{
"epoch": 0.07095521788607002,
"grad_norm": 0.9516310095787048,
"learning_rate": 4.2570281124497995e-06,
"loss": 2.035562515258789,
"step": 1590
},
{
"epoch": 0.07117834750206395,
"grad_norm": 0.8803707361221313,
"learning_rate": 4.270414993306559e-06,
"loss": 2.1310283660888674,
"step": 1595
},
{
"epoch": 0.07140147711805787,
"grad_norm": 1.022939682006836,
"learning_rate": 4.28380187416332e-06,
"loss": 2.1093021392822267,
"step": 1600
},
{
"epoch": 0.07162460673405181,
"grad_norm": 0.8254362344741821,
"learning_rate": 4.297188755020081e-06,
"loss": 1.8178255081176757,
"step": 1605
},
{
"epoch": 0.07184773635004574,
"grad_norm": 0.779062807559967,
"learning_rate": 4.3105756358768406e-06,
"loss": 2.261147880554199,
"step": 1610
},
{
"epoch": 0.07207086596603968,
"grad_norm": 0.8894750475883484,
"learning_rate": 4.323962516733601e-06,
"loss": 2.0283786773681642,
"step": 1615
},
{
"epoch": 0.0722939955820336,
"grad_norm": 0.9776813387870789,
"learning_rate": 4.337349397590362e-06,
"loss": 2.177045440673828,
"step": 1620
},
{
"epoch": 0.07251712519802754,
"grad_norm": 1.2759140729904175,
"learning_rate": 4.350736278447122e-06,
"loss": 2.0527717590332033,
"step": 1625
},
{
"epoch": 0.07274025481402147,
"grad_norm": 0.9953216314315796,
"learning_rate": 4.3641231593038825e-06,
"loss": 2.05969352722168,
"step": 1630
},
{
"epoch": 0.07296338443001539,
"grad_norm": 0.8319527506828308,
"learning_rate": 4.377510040160642e-06,
"loss": 2.003727340698242,
"step": 1635
},
{
"epoch": 0.07318651404600933,
"grad_norm": 0.8335089087486267,
"learning_rate": 4.390896921017403e-06,
"loss": 2.0173160552978517,
"step": 1640
},
{
"epoch": 0.07340964366200325,
"grad_norm": 1.1071332693099976,
"learning_rate": 4.404283801874164e-06,
"loss": 2.1184307098388673,
"step": 1645
},
{
"epoch": 0.0736327732779972,
"grad_norm": 1.2880672216415405,
"learning_rate": 4.4176706827309244e-06,
"loss": 2.0427175521850587,
"step": 1650
},
{
"epoch": 0.07385590289399112,
"grad_norm": 1.1434367895126343,
"learning_rate": 4.431057563587684e-06,
"loss": 2.0279747009277345,
"step": 1655
},
{
"epoch": 0.07407903250998506,
"grad_norm": 0.7727090716362,
"learning_rate": 4.444444444444444e-06,
"loss": 2.1546749114990233,
"step": 1660
},
{
"epoch": 0.07430216212597898,
"grad_norm": 1.0890729427337646,
"learning_rate": 4.457831325301205e-06,
"loss": 2.1247718811035154,
"step": 1665
},
{
"epoch": 0.07452529174197291,
"grad_norm": 1.0179014205932617,
"learning_rate": 4.4712182061579655e-06,
"loss": 2.146767234802246,
"step": 1670
},
{
"epoch": 0.07474842135796685,
"grad_norm": 0.745704174041748,
"learning_rate": 4.484605087014726e-06,
"loss": 2.224173736572266,
"step": 1675
},
{
"epoch": 0.07497155097396077,
"grad_norm": 0.771865963935852,
"learning_rate": 4.497991967871486e-06,
"loss": 2.2265625,
"step": 1680
},
{
"epoch": 0.07519468058995471,
"grad_norm": 1.352388620376587,
"learning_rate": 4.511378848728246e-06,
"loss": 2.261865425109863,
"step": 1685
},
{
"epoch": 0.07541781020594863,
"grad_norm": 0.8781198859214783,
"learning_rate": 4.524765729585007e-06,
"loss": 2.0096199035644533,
"step": 1690
},
{
"epoch": 0.07564093982194256,
"grad_norm": 1.0159834623336792,
"learning_rate": 4.538152610441767e-06,
"loss": 2.0243366241455076,
"step": 1695
},
{
"epoch": 0.0758640694379365,
"grad_norm": 0.8101127743721008,
"learning_rate": 4.551539491298528e-06,
"loss": 2.119925117492676,
"step": 1700
},
{
"epoch": 0.07608719905393042,
"grad_norm": 1.0311274528503418,
"learning_rate": 4.564926372155288e-06,
"loss": 1.9274934768676757,
"step": 1705
},
{
"epoch": 0.07631032866992436,
"grad_norm": 1.315838098526001,
"learning_rate": 4.578313253012049e-06,
"loss": 2.058757209777832,
"step": 1710
},
{
"epoch": 0.07653345828591829,
"grad_norm": 1.016562819480896,
"learning_rate": 4.5917001338688084e-06,
"loss": 2.0910503387451174,
"step": 1715
},
{
"epoch": 0.07675658790191223,
"grad_norm": 1.1127480268478394,
"learning_rate": 4.605087014725569e-06,
"loss": 2.0274993896484377,
"step": 1720
},
{
"epoch": 0.07697971751790615,
"grad_norm": 0.7386570572853088,
"learning_rate": 4.61847389558233e-06,
"loss": 2.03863468170166,
"step": 1725
},
{
"epoch": 0.07720284713390008,
"grad_norm": 0.9372514486312866,
"learning_rate": 4.63186077643909e-06,
"loss": 1.9647506713867187,
"step": 1730
},
{
"epoch": 0.07742597674989402,
"grad_norm": 0.866057276725769,
"learning_rate": 4.64524765729585e-06,
"loss": 2.0213361740112306,
"step": 1735
},
{
"epoch": 0.07764910636588794,
"grad_norm": 0.8000425100326538,
"learning_rate": 4.65863453815261e-06,
"loss": 2.009931755065918,
"step": 1740
},
{
"epoch": 0.07787223598188188,
"grad_norm": 1.4270923137664795,
"learning_rate": 4.672021419009371e-06,
"loss": 2.0473087310791014,
"step": 1745
},
{
"epoch": 0.0780953655978758,
"grad_norm": 0.7464487552642822,
"learning_rate": 4.685408299866132e-06,
"loss": 2.087724304199219,
"step": 1750
},
{
"epoch": 0.07831849521386974,
"grad_norm": 1.0895863771438599,
"learning_rate": 4.6987951807228915e-06,
"loss": 2.0931013107299803,
"step": 1755
},
{
"epoch": 0.07854162482986367,
"grad_norm": 1.2510039806365967,
"learning_rate": 4.712182061579652e-06,
"loss": 2.0668258666992188,
"step": 1760
},
{
"epoch": 0.0787647544458576,
"grad_norm": 0.9064348340034485,
"learning_rate": 4.725568942436412e-06,
"loss": 1.9942918777465821,
"step": 1765
},
{
"epoch": 0.07898788406185153,
"grad_norm": 1.0130219459533691,
"learning_rate": 4.738955823293173e-06,
"loss": 2.0953414916992186,
"step": 1770
},
{
"epoch": 0.07921101367784546,
"grad_norm": 1.0138216018676758,
"learning_rate": 4.7523427041499334e-06,
"loss": 2.097544479370117,
"step": 1775
},
{
"epoch": 0.0794341432938394,
"grad_norm": 0.9159922003746033,
"learning_rate": 4.765729585006693e-06,
"loss": 2.024347686767578,
"step": 1780
},
{
"epoch": 0.07965727290983332,
"grad_norm": 0.6978763937950134,
"learning_rate": 4.779116465863454e-06,
"loss": 2.032564163208008,
"step": 1785
},
{
"epoch": 0.07988040252582726,
"grad_norm": 0.9015496969223022,
"learning_rate": 4.792503346720214e-06,
"loss": 2.091436767578125,
"step": 1790
},
{
"epoch": 0.08010353214182118,
"grad_norm": 0.8910471796989441,
"learning_rate": 4.8058902275769745e-06,
"loss": 2.095396041870117,
"step": 1795
},
{
"epoch": 0.08032666175781511,
"grad_norm": 0.8550083637237549,
"learning_rate": 4.819277108433735e-06,
"loss": 2.061022186279297,
"step": 1800
},
{
"epoch": 0.08054979137380905,
"grad_norm": 0.888156533241272,
"learning_rate": 4.832663989290495e-06,
"loss": 2.0070331573486326,
"step": 1805
},
{
"epoch": 0.08077292098980297,
"grad_norm": 1.0453161001205444,
"learning_rate": 4.846050870147256e-06,
"loss": 2.0485980987548826,
"step": 1810
},
{
"epoch": 0.08099605060579691,
"grad_norm": 0.835533857345581,
"learning_rate": 4.859437751004016e-06,
"loss": 2.0737443923950196,
"step": 1815
},
{
"epoch": 0.08121918022179084,
"grad_norm": 1.3074698448181152,
"learning_rate": 4.872824631860777e-06,
"loss": 2.204619216918945,
"step": 1820
},
{
"epoch": 0.08144230983778476,
"grad_norm": 0.952622652053833,
"learning_rate": 4.886211512717537e-06,
"loss": 1.8864990234375,
"step": 1825
},
{
"epoch": 0.0816654394537787,
"grad_norm": 1.069469690322876,
"learning_rate": 4.899598393574298e-06,
"loss": 2.0791854858398438,
"step": 1830
},
{
"epoch": 0.08188856906977263,
"grad_norm": 1.1299378871917725,
"learning_rate": 4.9129852744310576e-06,
"loss": 2.0667266845703125,
"step": 1835
},
{
"epoch": 0.08211169868576657,
"grad_norm": 0.7093334794044495,
"learning_rate": 4.926372155287817e-06,
"loss": 1.9636056900024415,
"step": 1840
},
{
"epoch": 0.08233482830176049,
"grad_norm": 1.1765010356903076,
"learning_rate": 4.939759036144579e-06,
"loss": 2.0537750244140627,
"step": 1845
},
{
"epoch": 0.08255795791775443,
"grad_norm": 0.8516542315483093,
"learning_rate": 4.953145917001339e-06,
"loss": 2.1657791137695312,
"step": 1850
},
{
"epoch": 0.08278108753374835,
"grad_norm": 1.0495437383651733,
"learning_rate": 4.9665327978580995e-06,
"loss": 2.1656772613525392,
"step": 1855
},
{
"epoch": 0.08300421714974228,
"grad_norm": 0.8000126481056213,
"learning_rate": 4.979919678714859e-06,
"loss": 2.020354461669922,
"step": 1860
},
{
"epoch": 0.08322734676573622,
"grad_norm": 1.2887593507766724,
"learning_rate": 4.993306559571619e-06,
"loss": 1.9938453674316405,
"step": 1865
},
{
"epoch": 0.08345047638173014,
"grad_norm": 0.8419036865234375,
"learning_rate": 5.006693440428381e-06,
"loss": 2.0833532333374025,
"step": 1870
},
{
"epoch": 0.08367360599772408,
"grad_norm": 0.9748121500015259,
"learning_rate": 5.020080321285141e-06,
"loss": 2.0065855026245116,
"step": 1875
},
{
"epoch": 0.08389673561371801,
"grad_norm": 0.8467640280723572,
"learning_rate": 5.033467202141901e-06,
"loss": 2.2140008926391603,
"step": 1880
},
{
"epoch": 0.08411986522971195,
"grad_norm": 0.8260283470153809,
"learning_rate": 5.046854082998661e-06,
"loss": 1.973113441467285,
"step": 1885
},
{
"epoch": 0.08434299484570587,
"grad_norm": 1.2561461925506592,
"learning_rate": 5.060240963855422e-06,
"loss": 1.9699825286865233,
"step": 1890
},
{
"epoch": 0.0845661244616998,
"grad_norm": 0.7021331191062927,
"learning_rate": 5.0736278447121826e-06,
"loss": 1.9100543975830078,
"step": 1895
},
{
"epoch": 0.08478925407769373,
"grad_norm": 0.9285181760787964,
"learning_rate": 5.087014725568942e-06,
"loss": 2.100686454772949,
"step": 1900
},
{
"epoch": 0.08501238369368766,
"grad_norm": 1.0512628555297852,
"learning_rate": 5.100401606425703e-06,
"loss": 2.092105484008789,
"step": 1905
},
{
"epoch": 0.0852355133096816,
"grad_norm": 0.9993311166763306,
"learning_rate": 5.113788487282463e-06,
"loss": 2.178700256347656,
"step": 1910
},
{
"epoch": 0.08545864292567552,
"grad_norm": 0.8007418513298035,
"learning_rate": 5.127175368139224e-06,
"loss": 2.061174774169922,
"step": 1915
},
{
"epoch": 0.08568177254166946,
"grad_norm": 0.9044376611709595,
"learning_rate": 5.140562248995984e-06,
"loss": 1.905235481262207,
"step": 1920
},
{
"epoch": 0.08590490215766339,
"grad_norm": 1.1532313823699951,
"learning_rate": 5.153949129852744e-06,
"loss": 2.0893394470214846,
"step": 1925
},
{
"epoch": 0.08612803177365731,
"grad_norm": 1.0485821962356567,
"learning_rate": 5.167336010709505e-06,
"loss": 1.9315324783325196,
"step": 1930
},
{
"epoch": 0.08635116138965125,
"grad_norm": 1.1166177988052368,
"learning_rate": 5.180722891566265e-06,
"loss": 1.9648147583007813,
"step": 1935
},
{
"epoch": 0.08657429100564518,
"grad_norm": 1.4419307708740234,
"learning_rate": 5.1941097724230254e-06,
"loss": 2.0460626602172853,
"step": 1940
},
{
"epoch": 0.08679742062163912,
"grad_norm": 1.0400481224060059,
"learning_rate": 5.207496653279786e-06,
"loss": 2.034720230102539,
"step": 1945
},
{
"epoch": 0.08702055023763304,
"grad_norm": 1.041756510734558,
"learning_rate": 5.220883534136547e-06,
"loss": 1.9372877120971679,
"step": 1950
},
{
"epoch": 0.08724367985362697,
"grad_norm": 0.9398750066757202,
"learning_rate": 5.234270414993307e-06,
"loss": 1.9075811386108399,
"step": 1955
},
{
"epoch": 0.0874668094696209,
"grad_norm": 0.9996817708015442,
"learning_rate": 5.2476572958500665e-06,
"loss": 2.04040412902832,
"step": 1960
},
{
"epoch": 0.08768993908561483,
"grad_norm": 1.1434073448181152,
"learning_rate": 5.261044176706827e-06,
"loss": 2.05865535736084,
"step": 1965
},
{
"epoch": 0.08791306870160877,
"grad_norm": 1.1526947021484375,
"learning_rate": 5.274431057563588e-06,
"loss": 2.0584930419921874,
"step": 1970
},
{
"epoch": 0.0881361983176027,
"grad_norm": 1.1225361824035645,
"learning_rate": 5.287817938420349e-06,
"loss": 1.869145965576172,
"step": 1975
},
{
"epoch": 0.08835932793359663,
"grad_norm": 1.0828170776367188,
"learning_rate": 5.3012048192771085e-06,
"loss": 2.0549264907836915,
"step": 1980
},
{
"epoch": 0.08858245754959056,
"grad_norm": 0.9480776786804199,
"learning_rate": 5.314591700133868e-06,
"loss": 1.987673568725586,
"step": 1985
},
{
"epoch": 0.08880558716558448,
"grad_norm": 0.9799356460571289,
"learning_rate": 5.327978580990629e-06,
"loss": 1.9657482147216796,
"step": 1990
},
{
"epoch": 0.08902871678157842,
"grad_norm": 1.1955480575561523,
"learning_rate": 5.34136546184739e-06,
"loss": 1.9616188049316405,
"step": 1995
},
{
"epoch": 0.08925184639757235,
"grad_norm": 0.9528502225875854,
"learning_rate": 5.3547523427041504e-06,
"loss": 2.1342418670654295,
"step": 2000
},
{
"epoch": 0.08947497601356628,
"grad_norm": 1.1190009117126465,
"learning_rate": 5.36813922356091e-06,
"loss": 2.1064796447753906,
"step": 2005
},
{
"epoch": 0.08969810562956021,
"grad_norm": 1.0113600492477417,
"learning_rate": 5.381526104417671e-06,
"loss": 1.9123802185058594,
"step": 2010
},
{
"epoch": 0.08992123524555415,
"grad_norm": 0.9729359745979309,
"learning_rate": 5.394912985274431e-06,
"loss": 1.9950223922729493,
"step": 2015
},
{
"epoch": 0.09014436486154807,
"grad_norm": 1.035287618637085,
"learning_rate": 5.4082998661311915e-06,
"loss": 1.968986701965332,
"step": 2020
},
{
"epoch": 0.090367494477542,
"grad_norm": 0.7954384684562683,
"learning_rate": 5.421686746987952e-06,
"loss": 2.075417327880859,
"step": 2025
},
{
"epoch": 0.09059062409353594,
"grad_norm": 1.2406619787216187,
"learning_rate": 5.435073627844712e-06,
"loss": 2.097009468078613,
"step": 2030
},
{
"epoch": 0.09081375370952986,
"grad_norm": 0.8746753931045532,
"learning_rate": 5.448460508701473e-06,
"loss": 2.04738826751709,
"step": 2035
},
{
"epoch": 0.0910368833255238,
"grad_norm": 1.137431025505066,
"learning_rate": 5.461847389558233e-06,
"loss": 1.9636581420898438,
"step": 2040
},
{
"epoch": 0.09126001294151773,
"grad_norm": 1.4577034711837769,
"learning_rate": 5.475234270414993e-06,
"loss": 1.8939685821533203,
"step": 2045
},
{
"epoch": 0.09148314255751167,
"grad_norm": 1.3480346202850342,
"learning_rate": 5.488621151271754e-06,
"loss": 1.9858078002929687,
"step": 2050
},
{
"epoch": 0.09170627217350559,
"grad_norm": 0.8342795372009277,
"learning_rate": 5.502008032128514e-06,
"loss": 1.7992578506469727,
"step": 2055
},
{
"epoch": 0.09192940178949952,
"grad_norm": 1.1130149364471436,
"learning_rate": 5.5153949129852746e-06,
"loss": 2.0617441177368163,
"step": 2060
},
{
"epoch": 0.09215253140549345,
"grad_norm": 1.0582903623580933,
"learning_rate": 5.528781793842035e-06,
"loss": 2.081201934814453,
"step": 2065
},
{
"epoch": 0.09237566102148738,
"grad_norm": 0.8622370958328247,
"learning_rate": 5.542168674698796e-06,
"loss": 2.184007263183594,
"step": 2070
},
{
"epoch": 0.09259879063748132,
"grad_norm": 1.247605800628662,
"learning_rate": 5.555555555555556e-06,
"loss": 2.2668291091918946,
"step": 2075
},
{
"epoch": 0.09282192025347524,
"grad_norm": 0.9048439860343933,
"learning_rate": 5.568942436412316e-06,
"loss": 2.0036815643310546,
"step": 2080
},
{
"epoch": 0.09304504986946917,
"grad_norm": 1.5477705001831055,
"learning_rate": 5.582329317269076e-06,
"loss": 2.1254167556762695,
"step": 2085
},
{
"epoch": 0.0932681794854631,
"grad_norm": 0.898073673248291,
"learning_rate": 5.595716198125837e-06,
"loss": 2.0748970031738283,
"step": 2090
},
{
"epoch": 0.09349130910145703,
"grad_norm": 0.976369321346283,
"learning_rate": 5.609103078982598e-06,
"loss": 1.9387639999389648,
"step": 2095
},
{
"epoch": 0.09371443871745097,
"grad_norm": 1.229954719543457,
"learning_rate": 5.622489959839358e-06,
"loss": 1.9117536544799805,
"step": 2100
},
{
"epoch": 0.0939375683334449,
"grad_norm": 1.0723400115966797,
"learning_rate": 5.6358768406961175e-06,
"loss": 1.935175323486328,
"step": 2105
},
{
"epoch": 0.09416069794943883,
"grad_norm": 1.1824326515197754,
"learning_rate": 5.649263721552878e-06,
"loss": 2.126773452758789,
"step": 2110
},
{
"epoch": 0.09438382756543276,
"grad_norm": 0.8257484436035156,
"learning_rate": 5.662650602409639e-06,
"loss": 2.01248836517334,
"step": 2115
},
{
"epoch": 0.09460695718142668,
"grad_norm": 1.11427903175354,
"learning_rate": 5.6760374832663996e-06,
"loss": 2.1134754180908204,
"step": 2120
},
{
"epoch": 0.09483008679742062,
"grad_norm": 1.5741039514541626,
"learning_rate": 5.689424364123159e-06,
"loss": 2.0824302673339843,
"step": 2125
},
{
"epoch": 0.09505321641341455,
"grad_norm": 0.9258787631988525,
"learning_rate": 5.70281124497992e-06,
"loss": 1.9319812774658203,
"step": 2130
},
{
"epoch": 0.09527634602940849,
"grad_norm": 1.09774911403656,
"learning_rate": 5.71619812583668e-06,
"loss": 1.9680803298950196,
"step": 2135
},
{
"epoch": 0.09549947564540241,
"grad_norm": 0.9626898765563965,
"learning_rate": 5.729585006693441e-06,
"loss": 1.925799560546875,
"step": 2140
},
{
"epoch": 0.09572260526139635,
"grad_norm": 0.9046334624290466,
"learning_rate": 5.742971887550201e-06,
"loss": 1.9070989608764648,
"step": 2145
},
{
"epoch": 0.09594573487739028,
"grad_norm": 1.0075236558914185,
"learning_rate": 5.756358768406961e-06,
"loss": 1.9298274993896485,
"step": 2150
},
{
"epoch": 0.0961688644933842,
"grad_norm": 0.8876293301582336,
"learning_rate": 5.769745649263722e-06,
"loss": 1.995884323120117,
"step": 2155
},
{
"epoch": 0.09639199410937814,
"grad_norm": 0.9397478699684143,
"learning_rate": 5.783132530120482e-06,
"loss": 2.1033966064453127,
"step": 2160
},
{
"epoch": 0.09661512372537207,
"grad_norm": 1.1887969970703125,
"learning_rate": 5.7965194109772425e-06,
"loss": 2.2321205139160156,
"step": 2165
},
{
"epoch": 0.096838253341366,
"grad_norm": 0.9165719151496887,
"learning_rate": 5.809906291834003e-06,
"loss": 1.9332220077514648,
"step": 2170
},
{
"epoch": 0.09706138295735993,
"grad_norm": 1.0620633363723755,
"learning_rate": 5.823293172690763e-06,
"loss": 1.942098617553711,
"step": 2175
},
{
"epoch": 0.09728451257335387,
"grad_norm": 1.1249985694885254,
"learning_rate": 5.836680053547524e-06,
"loss": 2.004232406616211,
"step": 2180
},
{
"epoch": 0.09750764218934779,
"grad_norm": 1.009824514389038,
"learning_rate": 5.8500669344042836e-06,
"loss": 2.060811424255371,
"step": 2185
},
{
"epoch": 0.09773077180534172,
"grad_norm": 1.1177529096603394,
"learning_rate": 5.863453815261044e-06,
"loss": 1.9775056838989258,
"step": 2190
},
{
"epoch": 0.09795390142133566,
"grad_norm": 0.8883247375488281,
"learning_rate": 5.876840696117805e-06,
"loss": 2.0491350173950194,
"step": 2195
},
{
"epoch": 0.09817703103732958,
"grad_norm": 1.0781581401824951,
"learning_rate": 5.890227576974565e-06,
"loss": 2.130450439453125,
"step": 2200
},
{
"epoch": 0.09840016065332352,
"grad_norm": 1.022125244140625,
"learning_rate": 5.9036144578313255e-06,
"loss": 1.9054424285888671,
"step": 2205
},
{
"epoch": 0.09862329026931745,
"grad_norm": 1.5722389221191406,
"learning_rate": 5.917001338688085e-06,
"loss": 1.9462669372558594,
"step": 2210
},
{
"epoch": 0.09884641988531137,
"grad_norm": 0.8767709732055664,
"learning_rate": 5.930388219544846e-06,
"loss": 1.9716548919677734,
"step": 2215
},
{
"epoch": 0.09906954950130531,
"grad_norm": 0.8978337049484253,
"learning_rate": 5.943775100401607e-06,
"loss": 1.932182502746582,
"step": 2220
},
{
"epoch": 0.09929267911729923,
"grad_norm": 1.1649725437164307,
"learning_rate": 5.957161981258367e-06,
"loss": 1.949960708618164,
"step": 2225
},
{
"epoch": 0.09951580873329317,
"grad_norm": 0.9707964062690735,
"learning_rate": 5.970548862115127e-06,
"loss": 1.9415960311889648,
"step": 2230
},
{
"epoch": 0.0997389383492871,
"grad_norm": 0.8075286149978638,
"learning_rate": 5.983935742971887e-06,
"loss": 1.9884584426879883,
"step": 2235
},
{
"epoch": 0.09996206796528104,
"grad_norm": 0.919001579284668,
"learning_rate": 5.997322623828648e-06,
"loss": 1.9047340393066405,
"step": 2240
},
{
"epoch": 0.10018519758127496,
"grad_norm": 1.1452890634536743,
"learning_rate": 5.999999417648353e-06,
"loss": 1.8911170959472656,
"step": 2245
},
{
"epoch": 0.10040832719726889,
"grad_norm": 0.9075614213943481,
"learning_rate": 5.999997051845177e-06,
"loss": 1.9537294387817383,
"step": 2250
},
{
"epoch": 0.10063145681326283,
"grad_norm": 0.9411978721618652,
"learning_rate": 5.999992866194927e-06,
"loss": 2.0128080368041994,
"step": 2255
},
{
"epoch": 0.10085458642925675,
"grad_norm": 1.7404388189315796,
"learning_rate": 5.999986860700142e-06,
"loss": 1.9837495803833007,
"step": 2260
},
{
"epoch": 0.10107771604525069,
"grad_norm": 0.990151047706604,
"learning_rate": 5.999979035364464e-06,
"loss": 1.9975740432739257,
"step": 2265
},
{
"epoch": 0.10130084566124462,
"grad_norm": 1.042856216430664,
"learning_rate": 5.9999693901926435e-06,
"loss": 2.016056251525879,
"step": 2270
},
{
"epoch": 0.10152397527723855,
"grad_norm": 0.8757387399673462,
"learning_rate": 5.999957925190527e-06,
"loss": 1.8990396499633788,
"step": 2275
},
{
"epoch": 0.10174710489323248,
"grad_norm": 1.4858282804489136,
"learning_rate": 5.999944640365074e-06,
"loss": 2.0470705032348633,
"step": 2280
},
{
"epoch": 0.1019702345092264,
"grad_norm": 1.0206114053726196,
"learning_rate": 5.999929535724339e-06,
"loss": 2.063922882080078,
"step": 2285
},
{
"epoch": 0.10219336412522034,
"grad_norm": 1.2412887811660767,
"learning_rate": 5.999912611277487e-06,
"loss": 1.9137586593627929,
"step": 2290
},
{
"epoch": 0.10241649374121427,
"grad_norm": 1.0442379713058472,
"learning_rate": 5.999893867034784e-06,
"loss": 1.9131647109985352,
"step": 2295
},
{
"epoch": 0.1026396233572082,
"grad_norm": 1.0354846715927124,
"learning_rate": 5.9998733030076015e-06,
"loss": 1.943178367614746,
"step": 2300
},
{
"epoch": 0.10286275297320213,
"grad_norm": 0.9171436429023743,
"learning_rate": 5.999850919208413e-06,
"loss": 1.9243003845214843,
"step": 2305
},
{
"epoch": 0.10308588258919607,
"grad_norm": 1.3587474822998047,
"learning_rate": 5.999826715650797e-06,
"loss": 1.7801437377929688,
"step": 2310
},
{
"epoch": 0.10330901220519,
"grad_norm": 1.6726102828979492,
"learning_rate": 5.999800692349435e-06,
"loss": 2.0044769287109374,
"step": 2315
},
{
"epoch": 0.10353214182118392,
"grad_norm": 0.909557580947876,
"learning_rate": 5.999772849320115e-06,
"loss": 1.8790843963623047,
"step": 2320
},
{
"epoch": 0.10375527143717786,
"grad_norm": 1.1021777391433716,
"learning_rate": 5.9997431865797255e-06,
"loss": 1.9645397186279296,
"step": 2325
},
{
"epoch": 0.10397840105317178,
"grad_norm": 1.0416226387023926,
"learning_rate": 5.999711704146261e-06,
"loss": 1.9316511154174805,
"step": 2330
},
{
"epoch": 0.10420153066916572,
"grad_norm": 1.014394998550415,
"learning_rate": 5.99967840203882e-06,
"loss": 1.9602357864379882,
"step": 2335
},
{
"epoch": 0.10442466028515965,
"grad_norm": 1.0431203842163086,
"learning_rate": 5.999643280277603e-06,
"loss": 1.9662353515625,
"step": 2340
},
{
"epoch": 0.10464778990115357,
"grad_norm": 0.9976010322570801,
"learning_rate": 5.999606338883915e-06,
"loss": 1.8156768798828125,
"step": 2345
},
{
"epoch": 0.10487091951714751,
"grad_norm": 1.3542364835739136,
"learning_rate": 5.999567577880167e-06,
"loss": 1.9991931915283203,
"step": 2350
},
{
"epoch": 0.10509404913314144,
"grad_norm": 0.9118264317512512,
"learning_rate": 5.99952699728987e-06,
"loss": 1.6772186279296875,
"step": 2355
},
{
"epoch": 0.10531717874913538,
"grad_norm": 1.431563138961792,
"learning_rate": 5.9994845971376435e-06,
"loss": 2.0468074798583986,
"step": 2360
},
{
"epoch": 0.1055403083651293,
"grad_norm": 0.8818404078483582,
"learning_rate": 5.999440377449205e-06,
"loss": 1.9281387329101562,
"step": 2365
},
{
"epoch": 0.10576343798112324,
"grad_norm": 0.905672013759613,
"learning_rate": 5.999394338251381e-06,
"loss": 2.0013042449951173,
"step": 2370
},
{
"epoch": 0.10598656759711717,
"grad_norm": 1.0039812326431274,
"learning_rate": 5.999346479572099e-06,
"loss": 2.0309635162353517,
"step": 2375
},
{
"epoch": 0.10620969721311109,
"grad_norm": 1.1300805807113647,
"learning_rate": 5.999296801440391e-06,
"loss": 2.013144111633301,
"step": 2380
},
{
"epoch": 0.10643282682910503,
"grad_norm": 1.052232027053833,
"learning_rate": 5.999245303886392e-06,
"loss": 2.089672660827637,
"step": 2385
},
{
"epoch": 0.10665595644509895,
"grad_norm": 0.8320952653884888,
"learning_rate": 5.999191986941342e-06,
"loss": 2.0204065322875975,
"step": 2390
},
{
"epoch": 0.10687908606109289,
"grad_norm": 0.8853346705436707,
"learning_rate": 5.999136850637584e-06,
"loss": 1.9383895874023438,
"step": 2395
},
{
"epoch": 0.10710221567708682,
"grad_norm": 1.3795983791351318,
"learning_rate": 5.999079895008564e-06,
"loss": 1.808138084411621,
"step": 2400
}
],
"logging_steps": 5,
"max_steps": 22409,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 200,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 3.2148094686056153e+19,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}