nemo_nano_science / trainer_state.json
sedrickkeh's picture
End of training
757bbf4 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 4.9891808346213296,
"eval_steps": 500,
"global_step": 805,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0061823802163833074,
"grad_norm": 10.693047278702467,
"learning_rate": 9.876543209876544e-07,
"loss": 1.6775,
"step": 1
},
{
"epoch": 0.012364760432766615,
"grad_norm": 10.815416954217403,
"learning_rate": 1.9753086419753087e-06,
"loss": 1.6824,
"step": 2
},
{
"epoch": 0.01854714064914992,
"grad_norm": 10.616338898894972,
"learning_rate": 2.962962962962963e-06,
"loss": 1.6745,
"step": 3
},
{
"epoch": 0.02472952086553323,
"grad_norm": 9.861459123224664,
"learning_rate": 3.9506172839506175e-06,
"loss": 1.6421,
"step": 4
},
{
"epoch": 0.030911901081916538,
"grad_norm": 7.614357546616699,
"learning_rate": 4.938271604938272e-06,
"loss": 1.5691,
"step": 5
},
{
"epoch": 0.03709428129829984,
"grad_norm": 4.203625583627038,
"learning_rate": 5.925925925925926e-06,
"loss": 1.5004,
"step": 6
},
{
"epoch": 0.04327666151468315,
"grad_norm": 3.5662282715378195,
"learning_rate": 6.913580246913581e-06,
"loss": 1.4777,
"step": 7
},
{
"epoch": 0.04945904173106646,
"grad_norm": 5.739743938641111,
"learning_rate": 7.901234567901235e-06,
"loss": 1.4747,
"step": 8
},
{
"epoch": 0.05564142194744977,
"grad_norm": 6.180000116352135,
"learning_rate": 8.888888888888888e-06,
"loss": 1.4418,
"step": 9
},
{
"epoch": 0.061823802163833076,
"grad_norm": 6.739405843412405,
"learning_rate": 9.876543209876543e-06,
"loss": 1.3952,
"step": 10
},
{
"epoch": 0.06800618238021638,
"grad_norm": 7.52997521103117,
"learning_rate": 1.0864197530864198e-05,
"loss": 1.3959,
"step": 11
},
{
"epoch": 0.07418856259659969,
"grad_norm": 6.385085000826414,
"learning_rate": 1.1851851851851852e-05,
"loss": 1.363,
"step": 12
},
{
"epoch": 0.080370942812983,
"grad_norm": 5.007200781174352,
"learning_rate": 1.2839506172839507e-05,
"loss": 1.3392,
"step": 13
},
{
"epoch": 0.0865533230293663,
"grad_norm": 4.058377749470976,
"learning_rate": 1.3827160493827162e-05,
"loss": 1.3045,
"step": 14
},
{
"epoch": 0.09273570324574962,
"grad_norm": 3.5316422870540793,
"learning_rate": 1.4814814814814815e-05,
"loss": 1.2877,
"step": 15
},
{
"epoch": 0.09891808346213292,
"grad_norm": 3.842638477890943,
"learning_rate": 1.580246913580247e-05,
"loss": 1.2683,
"step": 16
},
{
"epoch": 0.10510046367851623,
"grad_norm": 3.1304998105929815,
"learning_rate": 1.6790123456790123e-05,
"loss": 1.2722,
"step": 17
},
{
"epoch": 0.11128284389489954,
"grad_norm": 1.831350110427916,
"learning_rate": 1.7777777777777777e-05,
"loss": 1.2427,
"step": 18
},
{
"epoch": 0.11746522411128284,
"grad_norm": 2.1448411883286593,
"learning_rate": 1.8765432098765433e-05,
"loss": 1.238,
"step": 19
},
{
"epoch": 0.12364760432766615,
"grad_norm": 2.0898956083682263,
"learning_rate": 1.9753086419753087e-05,
"loss": 1.2214,
"step": 20
},
{
"epoch": 0.12982998454404945,
"grad_norm": 1.5783782113338631,
"learning_rate": 2.074074074074074e-05,
"loss": 1.21,
"step": 21
},
{
"epoch": 0.13601236476043277,
"grad_norm": 1.7643935663724537,
"learning_rate": 2.1728395061728397e-05,
"loss": 1.1865,
"step": 22
},
{
"epoch": 0.14219474497681608,
"grad_norm": 2.0831771497504783,
"learning_rate": 2.271604938271605e-05,
"loss": 1.195,
"step": 23
},
{
"epoch": 0.14837712519319937,
"grad_norm": 1.7488536795684462,
"learning_rate": 2.3703703703703703e-05,
"loss": 1.1696,
"step": 24
},
{
"epoch": 0.1545595054095827,
"grad_norm": 1.5931909595273852,
"learning_rate": 2.469135802469136e-05,
"loss": 1.1892,
"step": 25
},
{
"epoch": 0.160741885625966,
"grad_norm": 1.396445568089242,
"learning_rate": 2.5679012345679013e-05,
"loss": 1.1793,
"step": 26
},
{
"epoch": 0.16692426584234932,
"grad_norm": 1.0145848862012221,
"learning_rate": 2.6666666666666667e-05,
"loss": 1.1664,
"step": 27
},
{
"epoch": 0.1731066460587326,
"grad_norm": 2.4324574107146226,
"learning_rate": 2.7654320987654323e-05,
"loss": 1.1573,
"step": 28
},
{
"epoch": 0.17928902627511592,
"grad_norm": 1.4236022981017393,
"learning_rate": 2.8641975308641977e-05,
"loss": 1.1501,
"step": 29
},
{
"epoch": 0.18547140649149924,
"grad_norm": 2.1562503786156153,
"learning_rate": 2.962962962962963e-05,
"loss": 1.1608,
"step": 30
},
{
"epoch": 0.19165378670788252,
"grad_norm": 1.9018824527792042,
"learning_rate": 3.061728395061729e-05,
"loss": 1.1451,
"step": 31
},
{
"epoch": 0.19783616692426584,
"grad_norm": 1.5703835911829096,
"learning_rate": 3.160493827160494e-05,
"loss": 1.1408,
"step": 32
},
{
"epoch": 0.20401854714064915,
"grad_norm": 8.652718407803869,
"learning_rate": 3.259259259259259e-05,
"loss": 1.1346,
"step": 33
},
{
"epoch": 0.21020092735703247,
"grad_norm": 6.435439968019908,
"learning_rate": 3.3580246913580247e-05,
"loss": 1.1582,
"step": 34
},
{
"epoch": 0.21638330757341576,
"grad_norm": 105.3495618548623,
"learning_rate": 3.45679012345679e-05,
"loss": 1.1558,
"step": 35
},
{
"epoch": 0.22256568778979907,
"grad_norm": 2.5364224403663056,
"learning_rate": 3.555555555555555e-05,
"loss": 1.1431,
"step": 36
},
{
"epoch": 0.2287480680061824,
"grad_norm": 1.9024835350187577,
"learning_rate": 3.654320987654321e-05,
"loss": 1.1463,
"step": 37
},
{
"epoch": 0.23493044822256567,
"grad_norm": 2.0426678600887254,
"learning_rate": 3.7530864197530867e-05,
"loss": 1.1307,
"step": 38
},
{
"epoch": 0.241112828438949,
"grad_norm": 1.8582429580354745,
"learning_rate": 3.851851851851852e-05,
"loss": 1.1272,
"step": 39
},
{
"epoch": 0.2472952086553323,
"grad_norm": 1.860203917219744,
"learning_rate": 3.950617283950617e-05,
"loss": 1.1137,
"step": 40
},
{
"epoch": 0.2534775888717156,
"grad_norm": 10.27861275542122,
"learning_rate": 4.049382716049383e-05,
"loss": 1.1247,
"step": 41
},
{
"epoch": 0.2596599690880989,
"grad_norm": 5.066950295263267,
"learning_rate": 4.148148148148148e-05,
"loss": 1.1334,
"step": 42
},
{
"epoch": 0.26584234930448225,
"grad_norm": 3.502140270278529,
"learning_rate": 4.246913580246914e-05,
"loss": 1.1114,
"step": 43
},
{
"epoch": 0.27202472952086554,
"grad_norm": 3.203814615327836,
"learning_rate": 4.345679012345679e-05,
"loss": 1.1182,
"step": 44
},
{
"epoch": 0.2782071097372488,
"grad_norm": 2.4989709311447954,
"learning_rate": 4.444444444444445e-05,
"loss": 1.1157,
"step": 45
},
{
"epoch": 0.28438948995363217,
"grad_norm": 3.2014585800746813,
"learning_rate": 4.54320987654321e-05,
"loss": 1.1158,
"step": 46
},
{
"epoch": 0.29057187017001546,
"grad_norm": 2.3255133597202073,
"learning_rate": 4.641975308641976e-05,
"loss": 1.1214,
"step": 47
},
{
"epoch": 0.29675425038639874,
"grad_norm": 3.340427993360326,
"learning_rate": 4.7407407407407407e-05,
"loss": 1.1072,
"step": 48
},
{
"epoch": 0.3029366306027821,
"grad_norm": 2.6053156370494137,
"learning_rate": 4.8395061728395067e-05,
"loss": 1.1125,
"step": 49
},
{
"epoch": 0.3091190108191654,
"grad_norm": 3.416491408495564,
"learning_rate": 4.938271604938272e-05,
"loss": 1.1076,
"step": 50
},
{
"epoch": 0.31530139103554866,
"grad_norm": 3.0240291240770016,
"learning_rate": 5.037037037037037e-05,
"loss": 1.1039,
"step": 51
},
{
"epoch": 0.321483771251932,
"grad_norm": 2.6974546966855173,
"learning_rate": 5.1358024691358027e-05,
"loss": 1.1083,
"step": 52
},
{
"epoch": 0.3276661514683153,
"grad_norm": 2.1712600314640174,
"learning_rate": 5.234567901234569e-05,
"loss": 1.0969,
"step": 53
},
{
"epoch": 0.33384853168469864,
"grad_norm": 2.43061865655575,
"learning_rate": 5.333333333333333e-05,
"loss": 1.1009,
"step": 54
},
{
"epoch": 0.3400309119010819,
"grad_norm": 1.4299217962960098,
"learning_rate": 5.432098765432099e-05,
"loss": 1.1071,
"step": 55
},
{
"epoch": 0.3462132921174652,
"grad_norm": 3.4472743528033942,
"learning_rate": 5.5308641975308647e-05,
"loss": 1.1034,
"step": 56
},
{
"epoch": 0.35239567233384855,
"grad_norm": 2.2533692308911486,
"learning_rate": 5.62962962962963e-05,
"loss": 1.0926,
"step": 57
},
{
"epoch": 0.35857805255023184,
"grad_norm": 3.124786132702191,
"learning_rate": 5.728395061728395e-05,
"loss": 1.0998,
"step": 58
},
{
"epoch": 0.36476043276661513,
"grad_norm": 2.9945765774059847,
"learning_rate": 5.827160493827161e-05,
"loss": 1.0936,
"step": 59
},
{
"epoch": 0.37094281298299847,
"grad_norm": 1.9029030448259223,
"learning_rate": 5.925925925925926e-05,
"loss": 1.0865,
"step": 60
},
{
"epoch": 0.37712519319938176,
"grad_norm": 2.5191915533993434,
"learning_rate": 6.024691358024692e-05,
"loss": 1.094,
"step": 61
},
{
"epoch": 0.38330757341576505,
"grad_norm": 2.377743288984034,
"learning_rate": 6.123456790123457e-05,
"loss": 1.0724,
"step": 62
},
{
"epoch": 0.3894899536321484,
"grad_norm": 3.418329883653554,
"learning_rate": 6.222222222222223e-05,
"loss": 1.1097,
"step": 63
},
{
"epoch": 0.3956723338485317,
"grad_norm": 1.8243621455646457,
"learning_rate": 6.320987654320988e-05,
"loss": 1.0972,
"step": 64
},
{
"epoch": 0.401854714064915,
"grad_norm": 3.3653254310211804,
"learning_rate": 6.419753086419754e-05,
"loss": 1.0954,
"step": 65
},
{
"epoch": 0.4080370942812983,
"grad_norm": 2.859591151453805,
"learning_rate": 6.518518518518519e-05,
"loss": 1.0887,
"step": 66
},
{
"epoch": 0.4142194744976816,
"grad_norm": 2.132914874530937,
"learning_rate": 6.617283950617285e-05,
"loss": 1.0847,
"step": 67
},
{
"epoch": 0.42040185471406494,
"grad_norm": 3.0306427822925386,
"learning_rate": 6.716049382716049e-05,
"loss": 1.1009,
"step": 68
},
{
"epoch": 0.4265842349304482,
"grad_norm": 2.0609891283339477,
"learning_rate": 6.814814814814815e-05,
"loss": 1.092,
"step": 69
},
{
"epoch": 0.4327666151468315,
"grad_norm": 2.9169289764667603,
"learning_rate": 6.91358024691358e-05,
"loss": 1.0721,
"step": 70
},
{
"epoch": 0.43894899536321486,
"grad_norm": 2.170819034455064,
"learning_rate": 7.012345679012346e-05,
"loss": 1.0752,
"step": 71
},
{
"epoch": 0.44513137557959814,
"grad_norm": 2.6561925596391256,
"learning_rate": 7.11111111111111e-05,
"loss": 1.0763,
"step": 72
},
{
"epoch": 0.45131375579598143,
"grad_norm": 2.0412946622837453,
"learning_rate": 7.209876543209877e-05,
"loss": 1.0855,
"step": 73
},
{
"epoch": 0.4574961360123648,
"grad_norm": 1.7224050848012167,
"learning_rate": 7.308641975308643e-05,
"loss": 1.0781,
"step": 74
},
{
"epoch": 0.46367851622874806,
"grad_norm": 2.0509259917647564,
"learning_rate": 7.407407407407409e-05,
"loss": 1.0864,
"step": 75
},
{
"epoch": 0.46986089644513135,
"grad_norm": 3.9723221221808225,
"learning_rate": 7.506172839506173e-05,
"loss": 1.077,
"step": 76
},
{
"epoch": 0.4760432766615147,
"grad_norm": 2.5313197271865464,
"learning_rate": 7.60493827160494e-05,
"loss": 1.0873,
"step": 77
},
{
"epoch": 0.482225656877898,
"grad_norm": 2.6058497426158342,
"learning_rate": 7.703703703703704e-05,
"loss": 1.0912,
"step": 78
},
{
"epoch": 0.4884080370942813,
"grad_norm": 4.045410961666709,
"learning_rate": 7.80246913580247e-05,
"loss": 1.0905,
"step": 79
},
{
"epoch": 0.4945904173106646,
"grad_norm": 2.085038720402762,
"learning_rate": 7.901234567901235e-05,
"loss": 1.0796,
"step": 80
},
{
"epoch": 0.500772797527048,
"grad_norm": 4.492244112618374,
"learning_rate": 8e-05,
"loss": 1.094,
"step": 81
},
{
"epoch": 0.5069551777434312,
"grad_norm": 3.6362050618093122,
"learning_rate": 7.999962342461635e-05,
"loss": 1.1031,
"step": 82
},
{
"epoch": 0.5131375579598145,
"grad_norm": 3.1038756568216317,
"learning_rate": 7.999849370555585e-05,
"loss": 1.0726,
"step": 83
},
{
"epoch": 0.5193199381761978,
"grad_norm": 2.526971760311998,
"learning_rate": 7.999661086408972e-05,
"loss": 1.0824,
"step": 84
},
{
"epoch": 0.5255023183925811,
"grad_norm": 1.8821681779156758,
"learning_rate": 7.999397493566954e-05,
"loss": 1.0752,
"step": 85
},
{
"epoch": 0.5316846986089645,
"grad_norm": 2.3448656117874913,
"learning_rate": 7.99905859699266e-05,
"loss": 1.0908,
"step": 86
},
{
"epoch": 0.5378670788253478,
"grad_norm": 2.0611523416179542,
"learning_rate": 7.998644403067097e-05,
"loss": 1.0721,
"step": 87
},
{
"epoch": 0.5440494590417311,
"grad_norm": 2.9983162968300983,
"learning_rate": 7.998154919589024e-05,
"loss": 1.0776,
"step": 88
},
{
"epoch": 0.5502318392581144,
"grad_norm": 2.1806569564777987,
"learning_rate": 7.997590155774815e-05,
"loss": 1.0733,
"step": 89
},
{
"epoch": 0.5564142194744977,
"grad_norm": 2.1571411866423533,
"learning_rate": 7.996950122258277e-05,
"loss": 1.0577,
"step": 90
},
{
"epoch": 0.5625965996908809,
"grad_norm": 3.999074777477232,
"learning_rate": 7.996234831090451e-05,
"loss": 1.0793,
"step": 91
},
{
"epoch": 0.5687789799072643,
"grad_norm": 2.2312489091174172,
"learning_rate": 7.995444295739394e-05,
"loss": 1.0775,
"step": 92
},
{
"epoch": 0.5749613601236476,
"grad_norm": 6.134653577370118,
"learning_rate": 7.994578531089908e-05,
"loss": 1.0927,
"step": 93
},
{
"epoch": 0.5811437403400309,
"grad_norm": 6.274513062714145,
"learning_rate": 7.99363755344328e-05,
"loss": 1.1068,
"step": 94
},
{
"epoch": 0.5873261205564142,
"grad_norm": 2.3500100638359287,
"learning_rate": 7.99262138051696e-05,
"loss": 1.0885,
"step": 95
},
{
"epoch": 0.5935085007727975,
"grad_norm": 4.48097545076724,
"learning_rate": 7.991530031444231e-05,
"loss": 1.0838,
"step": 96
},
{
"epoch": 0.5996908809891809,
"grad_norm": 3.749050135435733,
"learning_rate": 7.990363526773856e-05,
"loss": 1.0947,
"step": 97
},
{
"epoch": 0.6058732612055642,
"grad_norm": 2.7226697465582674,
"learning_rate": 7.98912188846968e-05,
"loss": 1.0835,
"step": 98
},
{
"epoch": 0.6120556414219475,
"grad_norm": 2.7748343898198073,
"learning_rate": 7.987805139910226e-05,
"loss": 1.0601,
"step": 99
},
{
"epoch": 0.6182380216383307,
"grad_norm": 2.0786601854908087,
"learning_rate": 7.986413305888247e-05,
"loss": 1.0628,
"step": 100
},
{
"epoch": 0.624420401854714,
"grad_norm": 3.0450749679460842,
"learning_rate": 7.984946412610265e-05,
"loss": 1.0746,
"step": 101
},
{
"epoch": 0.6306027820710973,
"grad_norm": 2.0666897869515273,
"learning_rate": 7.983404487696076e-05,
"loss": 1.0712,
"step": 102
},
{
"epoch": 0.6367851622874807,
"grad_norm": 3.3981180649215212,
"learning_rate": 7.981787560178227e-05,
"loss": 1.0555,
"step": 103
},
{
"epoch": 0.642967542503864,
"grad_norm": 2.793840699546338,
"learning_rate": 7.980095660501473e-05,
"loss": 1.0705,
"step": 104
},
{
"epoch": 0.6491499227202473,
"grad_norm": 3.0934299956993754,
"learning_rate": 7.978328820522204e-05,
"loss": 1.0593,
"step": 105
},
{
"epoch": 0.6553323029366306,
"grad_norm": 2.2997166417283186,
"learning_rate": 7.976487073507839e-05,
"loss": 1.0485,
"step": 106
},
{
"epoch": 0.6615146831530139,
"grad_norm": 3.3687080136243903,
"learning_rate": 7.974570454136212e-05,
"loss": 1.0623,
"step": 107
},
{
"epoch": 0.6676970633693973,
"grad_norm": 2.668689204618569,
"learning_rate": 7.972578998494903e-05,
"loss": 1.06,
"step": 108
},
{
"epoch": 0.6738794435857806,
"grad_norm": 3.259938106278256,
"learning_rate": 7.970512744080571e-05,
"loss": 1.0646,
"step": 109
},
{
"epoch": 0.6800618238021638,
"grad_norm": 2.539712415638497,
"learning_rate": 7.968371729798246e-05,
"loss": 1.0578,
"step": 110
},
{
"epoch": 0.6862442040185471,
"grad_norm": 2.875087869638967,
"learning_rate": 7.96615599596059e-05,
"loss": 1.0495,
"step": 111
},
{
"epoch": 0.6924265842349304,
"grad_norm": 2.171964418083895,
"learning_rate": 7.963865584287142e-05,
"loss": 1.0482,
"step": 112
},
{
"epoch": 0.6986089644513137,
"grad_norm": 2.4445515677858776,
"learning_rate": 7.96150053790354e-05,
"loss": 1.0478,
"step": 113
},
{
"epoch": 0.7047913446676971,
"grad_norm": 1.4736364847794041,
"learning_rate": 7.959060901340691e-05,
"loss": 1.0644,
"step": 114
},
{
"epoch": 0.7109737248840804,
"grad_norm": 3.2793219071142574,
"learning_rate": 7.956546720533952e-05,
"loss": 1.0581,
"step": 115
},
{
"epoch": 0.7171561051004637,
"grad_norm": 2.1935733687249157,
"learning_rate": 7.953958042822252e-05,
"loss": 1.0577,
"step": 116
},
{
"epoch": 0.723338485316847,
"grad_norm": 1.9736309619709513,
"learning_rate": 7.951294916947206e-05,
"loss": 1.0562,
"step": 117
},
{
"epoch": 0.7295208655332303,
"grad_norm": 2.1690441232660724,
"learning_rate": 7.948557393052195e-05,
"loss": 1.052,
"step": 118
},
{
"epoch": 0.7357032457496137,
"grad_norm": 2.703798771440302,
"learning_rate": 7.945745522681429e-05,
"loss": 1.0645,
"step": 119
},
{
"epoch": 0.7418856259659969,
"grad_norm": 2.6860226290594427,
"learning_rate": 7.942859358778961e-05,
"loss": 1.0469,
"step": 120
},
{
"epoch": 0.7480680061823802,
"grad_norm": 1.2516089611393217,
"learning_rate": 7.939898955687709e-05,
"loss": 1.0442,
"step": 121
},
{
"epoch": 0.7542503863987635,
"grad_norm": 3.4267893638027878,
"learning_rate": 7.936864369148418e-05,
"loss": 1.0608,
"step": 122
},
{
"epoch": 0.7604327666151468,
"grad_norm": 2.225834600604942,
"learning_rate": 7.933755656298616e-05,
"loss": 1.0668,
"step": 123
},
{
"epoch": 0.7666151468315301,
"grad_norm": 1.6048920414465118,
"learning_rate": 7.930572875671542e-05,
"loss": 1.046,
"step": 124
},
{
"epoch": 0.7727975270479135,
"grad_norm": 3.4705093098033957,
"learning_rate": 7.927316087195035e-05,
"loss": 1.0703,
"step": 125
},
{
"epoch": 0.7789799072642968,
"grad_norm": 2.280254489129258,
"learning_rate": 7.923985352190416e-05,
"loss": 1.0632,
"step": 126
},
{
"epoch": 0.7851622874806801,
"grad_norm": 2.8077251065232884,
"learning_rate": 7.920580733371325e-05,
"loss": 1.0637,
"step": 127
},
{
"epoch": 0.7913446676970634,
"grad_norm": 2.79897007061372,
"learning_rate": 7.917102294842545e-05,
"loss": 1.0511,
"step": 128
},
{
"epoch": 0.7975270479134466,
"grad_norm": 2.356217446775361,
"learning_rate": 7.913550102098789e-05,
"loss": 1.0621,
"step": 129
},
{
"epoch": 0.80370942812983,
"grad_norm": 2.837553637508471,
"learning_rate": 7.909924222023478e-05,
"loss": 1.0498,
"step": 130
},
{
"epoch": 0.8098918083462133,
"grad_norm": 2.074863492575953,
"learning_rate": 7.906224722887468e-05,
"loss": 1.0555,
"step": 131
},
{
"epoch": 0.8160741885625966,
"grad_norm": 3.906632921125056,
"learning_rate": 7.902451674347778e-05,
"loss": 1.0547,
"step": 132
},
{
"epoch": 0.8222565687789799,
"grad_norm": 3.098700348501345,
"learning_rate": 7.898605147446262e-05,
"loss": 1.0459,
"step": 133
},
{
"epoch": 0.8284389489953632,
"grad_norm": 2.847790553655261,
"learning_rate": 7.894685214608293e-05,
"loss": 1.0442,
"step": 134
},
{
"epoch": 0.8346213292117465,
"grad_norm": 2.908201865083222,
"learning_rate": 7.89069194964138e-05,
"loss": 1.0553,
"step": 135
},
{
"epoch": 0.8408037094281299,
"grad_norm": 2.1751989062888146,
"learning_rate": 7.886625427733786e-05,
"loss": 1.0479,
"step": 136
},
{
"epoch": 0.8469860896445132,
"grad_norm": 1.729338903416391,
"learning_rate": 7.882485725453114e-05,
"loss": 1.0315,
"step": 137
},
{
"epoch": 0.8531684698608965,
"grad_norm": 2.2703882009096827,
"learning_rate": 7.878272920744862e-05,
"loss": 1.0496,
"step": 138
},
{
"epoch": 0.8593508500772797,
"grad_norm": 1.3426280628545364,
"learning_rate": 7.87398709293096e-05,
"loss": 1.0407,
"step": 139
},
{
"epoch": 0.865533230293663,
"grad_norm": 2.68728781786679,
"learning_rate": 7.869628322708269e-05,
"loss": 1.0551,
"step": 140
},
{
"epoch": 0.8717156105100463,
"grad_norm": 1.7262218207414903,
"learning_rate": 7.865196692147066e-05,
"loss": 1.0483,
"step": 141
},
{
"epoch": 0.8778979907264297,
"grad_norm": 2.5755795419575156,
"learning_rate": 7.860692284689502e-05,
"loss": 1.0657,
"step": 142
},
{
"epoch": 0.884080370942813,
"grad_norm": 2.1034039247877856,
"learning_rate": 7.856115185148025e-05,
"loss": 1.056,
"step": 143
},
{
"epoch": 0.8902627511591963,
"grad_norm": 1.9719606878351836,
"learning_rate": 7.851465479703785e-05,
"loss": 1.0549,
"step": 144
},
{
"epoch": 0.8964451313755796,
"grad_norm": 2.066417220460305,
"learning_rate": 7.846743255905014e-05,
"loss": 1.0483,
"step": 145
},
{
"epoch": 0.9026275115919629,
"grad_norm": 1.7106893433354307,
"learning_rate": 7.841948602665373e-05,
"loss": 1.0423,
"step": 146
},
{
"epoch": 0.9088098918083463,
"grad_norm": 2.4989235217550676,
"learning_rate": 7.83708161026228e-05,
"loss": 1.0486,
"step": 147
},
{
"epoch": 0.9149922720247295,
"grad_norm": 2.354839754481779,
"learning_rate": 7.832142370335215e-05,
"loss": 1.0445,
"step": 148
},
{
"epoch": 0.9211746522411128,
"grad_norm": 1.702601699332555,
"learning_rate": 7.827130975883982e-05,
"loss": 1.0206,
"step": 149
},
{
"epoch": 0.9273570324574961,
"grad_norm": 3.284158052097331,
"learning_rate": 7.822047521266973e-05,
"loss": 1.028,
"step": 150
},
{
"epoch": 0.9335394126738794,
"grad_norm": 2.5496411598803648,
"learning_rate": 7.816892102199382e-05,
"loss": 1.0367,
"step": 151
},
{
"epoch": 0.9397217928902627,
"grad_norm": 1.842181874139785,
"learning_rate": 7.811664815751404e-05,
"loss": 1.0415,
"step": 152
},
{
"epoch": 0.9459041731066461,
"grad_norm": 3.2108389564617887,
"learning_rate": 7.806365760346409e-05,
"loss": 1.0379,
"step": 153
},
{
"epoch": 0.9520865533230294,
"grad_norm": 1.8227204697842778,
"learning_rate": 7.800995035759087e-05,
"loss": 1.0498,
"step": 154
},
{
"epoch": 0.9582689335394127,
"grad_norm": 3.4947702832902476,
"learning_rate": 7.795552743113573e-05,
"loss": 1.0479,
"step": 155
},
{
"epoch": 0.964451313755796,
"grad_norm": 1.8114431864984537,
"learning_rate": 7.79003898488154e-05,
"loss": 1.0313,
"step": 156
},
{
"epoch": 0.9706336939721792,
"grad_norm": 3.535447837519456,
"learning_rate": 7.784453864880267e-05,
"loss": 1.0474,
"step": 157
},
{
"epoch": 0.9768160741885626,
"grad_norm": 1.7365198353975386,
"learning_rate": 7.77879748827069e-05,
"loss": 1.0511,
"step": 158
},
{
"epoch": 0.9829984544049459,
"grad_norm": 3.594055596899371,
"learning_rate": 7.77306996155542e-05,
"loss": 1.0603,
"step": 159
},
{
"epoch": 0.9891808346213292,
"grad_norm": 2.5313017940876086,
"learning_rate": 7.767271392576732e-05,
"loss": 1.0626,
"step": 160
},
{
"epoch": 0.9953632148377125,
"grad_norm": 3.40567680255142,
"learning_rate": 7.761401890514547e-05,
"loss": 1.0638,
"step": 161
},
{
"epoch": 1.0046367851622875,
"grad_norm": 4.5174674178938945,
"learning_rate": 7.755461565884362e-05,
"loss": 1.8415,
"step": 162
},
{
"epoch": 1.010819165378671,
"grad_norm": 3.0038968679186957,
"learning_rate": 7.749450530535179e-05,
"loss": 1.0416,
"step": 163
},
{
"epoch": 1.017001545595054,
"grad_norm": 2.217689466421883,
"learning_rate": 7.743368897647395e-05,
"loss": 1.0335,
"step": 164
},
{
"epoch": 1.0231839258114375,
"grad_norm": 3.387468691972088,
"learning_rate": 7.737216781730673e-05,
"loss": 1.0285,
"step": 165
},
{
"epoch": 1.0293663060278206,
"grad_norm": 2.9834708542560433,
"learning_rate": 7.730994298621781e-05,
"loss": 1.0462,
"step": 166
},
{
"epoch": 1.035548686244204,
"grad_norm": 2.33967613172349,
"learning_rate": 7.72470156548242e-05,
"loss": 1.0315,
"step": 167
},
{
"epoch": 1.0417310664605872,
"grad_norm": 1.98032140135135,
"learning_rate": 7.718338700797007e-05,
"loss": 1.021,
"step": 168
},
{
"epoch": 1.0479134466769706,
"grad_norm": 1.8602974186471763,
"learning_rate": 7.711905824370457e-05,
"loss": 1.0253,
"step": 169
},
{
"epoch": 1.054095826893354,
"grad_norm": 1.571944862888458,
"learning_rate": 7.70540305732591e-05,
"loss": 1.0096,
"step": 170
},
{
"epoch": 1.0602782071097372,
"grad_norm": 1.2148625487373597,
"learning_rate": 7.69883052210247e-05,
"loss": 1.0156,
"step": 171
},
{
"epoch": 1.0664605873261206,
"grad_norm": 2.1277306224703865,
"learning_rate": 7.692188342452885e-05,
"loss": 1.0201,
"step": 172
},
{
"epoch": 1.0726429675425038,
"grad_norm": 1.92583372296243,
"learning_rate": 7.685476643441219e-05,
"loss": 1.0207,
"step": 173
},
{
"epoch": 1.0788253477588872,
"grad_norm": 1.8133100096935075,
"learning_rate": 7.678695551440506e-05,
"loss": 1.0181,
"step": 174
},
{
"epoch": 1.0850077279752706,
"grad_norm": 1.4080085200242758,
"learning_rate": 7.671845194130363e-05,
"loss": 1.0219,
"step": 175
},
{
"epoch": 1.0911901081916537,
"grad_norm": 2.2832079127892326,
"learning_rate": 7.664925700494585e-05,
"loss": 1.0199,
"step": 176
},
{
"epoch": 1.0973724884080371,
"grad_norm": 2.150797505157169,
"learning_rate": 7.657937200818722e-05,
"loss": 1.0173,
"step": 177
},
{
"epoch": 1.1035548686244203,
"grad_norm": 0.8849916762716471,
"learning_rate": 7.65087982668762e-05,
"loss": 1.005,
"step": 178
},
{
"epoch": 1.1097372488408037,
"grad_norm": 3.229134861969822,
"learning_rate": 7.643753710982949e-05,
"loss": 1.0133,
"step": 179
},
{
"epoch": 1.1159196290571871,
"grad_norm": 1.7293989048925642,
"learning_rate": 7.636558987880694e-05,
"loss": 1.0353,
"step": 180
},
{
"epoch": 1.1221020092735703,
"grad_norm": 2.7325244810866747,
"learning_rate": 7.629295792848639e-05,
"loss": 1.0188,
"step": 181
},
{
"epoch": 1.1282843894899537,
"grad_norm": 1.9813373605063669,
"learning_rate": 7.621964262643805e-05,
"loss": 1.0331,
"step": 182
},
{
"epoch": 1.1344667697063369,
"grad_norm": 1.9157662111283729,
"learning_rate": 7.614564535309882e-05,
"loss": 0.9959,
"step": 183
},
{
"epoch": 1.1406491499227203,
"grad_norm": 2.0892926569694685,
"learning_rate": 7.607096750174629e-05,
"loss": 1.0097,
"step": 184
},
{
"epoch": 1.1468315301391034,
"grad_norm": 1.821781934019597,
"learning_rate": 7.599561047847247e-05,
"loss": 1.0213,
"step": 185
},
{
"epoch": 1.1530139103554868,
"grad_norm": 1.8556200827944644,
"learning_rate": 7.591957570215735e-05,
"loss": 1.0084,
"step": 186
},
{
"epoch": 1.1591962905718702,
"grad_norm": 1.3594082674263888,
"learning_rate": 7.584286460444222e-05,
"loss": 1.0044,
"step": 187
},
{
"epoch": 1.1653786707882534,
"grad_norm": 2.2744478173451594,
"learning_rate": 7.576547862970261e-05,
"loss": 0.993,
"step": 188
},
{
"epoch": 1.1715610510046368,
"grad_norm": 1.3757650290844639,
"learning_rate": 7.568741923502118e-05,
"loss": 1.0055,
"step": 189
},
{
"epoch": 1.1777434312210202,
"grad_norm": 1.854325739268096,
"learning_rate": 7.560868789016024e-05,
"loss": 1.0125,
"step": 190
},
{
"epoch": 1.1839258114374034,
"grad_norm": 1.5111560222547549,
"learning_rate": 7.552928607753414e-05,
"loss": 1.0005,
"step": 191
},
{
"epoch": 1.1901081916537868,
"grad_norm": 1.811472101825589,
"learning_rate": 7.544921529218125e-05,
"loss": 1.004,
"step": 192
},
{
"epoch": 1.19629057187017,
"grad_norm": 1.4223817808791495,
"learning_rate": 7.536847704173593e-05,
"loss": 1.0014,
"step": 193
},
{
"epoch": 1.2024729520865534,
"grad_norm": 1.9843008102793314,
"learning_rate": 7.528707284640004e-05,
"loss": 1.0142,
"step": 194
},
{
"epoch": 1.2086553323029365,
"grad_norm": 1.5650426744555666,
"learning_rate": 7.520500423891442e-05,
"loss": 1.0053,
"step": 195
},
{
"epoch": 1.21483771251932,
"grad_norm": 1.5008249395490572,
"learning_rate": 7.512227276452989e-05,
"loss": 1.0001,
"step": 196
},
{
"epoch": 1.2210200927357033,
"grad_norm": 1.8158622037471164,
"learning_rate": 7.503887998097833e-05,
"loss": 1.0046,
"step": 197
},
{
"epoch": 1.2272024729520865,
"grad_norm": 1.594408802837267,
"learning_rate": 7.495482745844317e-05,
"loss": 0.9926,
"step": 198
},
{
"epoch": 1.23338485316847,
"grad_norm": 1.6151725290749885,
"learning_rate": 7.487011677953e-05,
"loss": 1.0135,
"step": 199
},
{
"epoch": 1.239567233384853,
"grad_norm": 1.6326603710438912,
"learning_rate": 7.478474953923662e-05,
"loss": 1.0011,
"step": 200
},
{
"epoch": 1.2457496136012365,
"grad_norm": 1.8016191114443714,
"learning_rate": 7.469872734492308e-05,
"loss": 0.9952,
"step": 201
},
{
"epoch": 1.2519319938176197,
"grad_norm": 1.1900057306211553,
"learning_rate": 7.461205181628143e-05,
"loss": 0.9903,
"step": 202
},
{
"epoch": 1.258114374034003,
"grad_norm": 2.4753152288862763,
"learning_rate": 7.452472458530522e-05,
"loss": 1.0031,
"step": 203
},
{
"epoch": 1.2642967542503865,
"grad_norm": 1.156802595501098,
"learning_rate": 7.443674729625869e-05,
"loss": 1.0059,
"step": 204
},
{
"epoch": 1.2704791344667696,
"grad_norm": 1.41497847335521,
"learning_rate": 7.434812160564592e-05,
"loss": 1.0144,
"step": 205
},
{
"epoch": 1.276661514683153,
"grad_norm": 2.3844147534681155,
"learning_rate": 7.425884918217958e-05,
"loss": 1.0058,
"step": 206
},
{
"epoch": 1.2828438948995364,
"grad_norm": 1.3271435145317263,
"learning_rate": 7.416893170674954e-05,
"loss": 1.0165,
"step": 207
},
{
"epoch": 1.2890262751159196,
"grad_norm": 1.6975589849591926,
"learning_rate": 7.407837087239117e-05,
"loss": 0.99,
"step": 208
},
{
"epoch": 1.295208655332303,
"grad_norm": 1.8519388740645097,
"learning_rate": 7.398716838425352e-05,
"loss": 1.0089,
"step": 209
},
{
"epoch": 1.3013910355486862,
"grad_norm": 1.6947179785568054,
"learning_rate": 7.38953259595672e-05,
"loss": 0.9879,
"step": 210
},
{
"epoch": 1.3075734157650696,
"grad_norm": 1.3528548666192206,
"learning_rate": 7.380284532761202e-05,
"loss": 1.0024,
"step": 211
},
{
"epoch": 1.3137557959814528,
"grad_norm": 1.6098695630590318,
"learning_rate": 7.370972822968446e-05,
"loss": 0.9804,
"step": 212
},
{
"epoch": 1.3199381761978362,
"grad_norm": 1.5753154362328263,
"learning_rate": 7.361597641906486e-05,
"loss": 0.9905,
"step": 213
},
{
"epoch": 1.3261205564142196,
"grad_norm": 1.4448197507546439,
"learning_rate": 7.352159166098441e-05,
"loss": 0.9847,
"step": 214
},
{
"epoch": 1.3323029366306027,
"grad_norm": 2.188314280055232,
"learning_rate": 7.342657573259194e-05,
"loss": 0.9992,
"step": 215
},
{
"epoch": 1.3384853168469861,
"grad_norm": 1.2657997745237497,
"learning_rate": 7.333093042292044e-05,
"loss": 0.9904,
"step": 216
},
{
"epoch": 1.3446676970633695,
"grad_norm": 1.2714200260276027,
"learning_rate": 7.323465753285337e-05,
"loss": 1.0057,
"step": 217
},
{
"epoch": 1.3508500772797527,
"grad_norm": 1.3287688093801202,
"learning_rate": 7.313775887509075e-05,
"loss": 1.0102,
"step": 218
},
{
"epoch": 1.3570324574961359,
"grad_norm": 3.081273929769149,
"learning_rate": 7.304023627411505e-05,
"loss": 1.0116,
"step": 219
},
{
"epoch": 1.3632148377125193,
"grad_norm": 1.767890752318478,
"learning_rate": 7.29420915661568e-05,
"loss": 0.9999,
"step": 220
},
{
"epoch": 1.3693972179289027,
"grad_norm": 3.3832241258249143,
"learning_rate": 7.284332659916006e-05,
"loss": 1.0146,
"step": 221
},
{
"epoch": 1.3755795981452859,
"grad_norm": 2.792347707620742,
"learning_rate": 7.274394323274761e-05,
"loss": 1.038,
"step": 222
},
{
"epoch": 1.3817619783616693,
"grad_norm": 2.717608779519406,
"learning_rate": 7.26439433381859e-05,
"loss": 0.9968,
"step": 223
},
{
"epoch": 1.3879443585780527,
"grad_norm": 2.6449000513172023,
"learning_rate": 7.254332879834986e-05,
"loss": 1.0202,
"step": 224
},
{
"epoch": 1.3941267387944358,
"grad_norm": 2.5816179353247386,
"learning_rate": 7.244210150768745e-05,
"loss": 1.0126,
"step": 225
},
{
"epoch": 1.4003091190108192,
"grad_norm": 1.8819265089556003,
"learning_rate": 7.234026337218395e-05,
"loss": 1.0044,
"step": 226
},
{
"epoch": 1.4064914992272024,
"grad_norm": 2.778434728345896,
"learning_rate": 7.223781630932612e-05,
"loss": 1.0112,
"step": 227
},
{
"epoch": 1.4126738794435858,
"grad_norm": 2.3953953560776164,
"learning_rate": 7.213476224806604e-05,
"loss": 1.007,
"step": 228
},
{
"epoch": 1.418856259659969,
"grad_norm": 2.1884861738889554,
"learning_rate": 7.203110312878487e-05,
"loss": 0.9997,
"step": 229
},
{
"epoch": 1.4250386398763524,
"grad_norm": 1.9108092994214398,
"learning_rate": 7.192684090325621e-05,
"loss": 0.9961,
"step": 230
},
{
"epoch": 1.4312210200927358,
"grad_norm": 2.070150415301729,
"learning_rate": 7.182197753460947e-05,
"loss": 0.9928,
"step": 231
},
{
"epoch": 1.437403400309119,
"grad_norm": 1.2896742775847363,
"learning_rate": 7.171651499729278e-05,
"loss": 0.9843,
"step": 232
},
{
"epoch": 1.4435857805255023,
"grad_norm": 2.4892589479787914,
"learning_rate": 7.161045527703593e-05,
"loss": 0.997,
"step": 233
},
{
"epoch": 1.4497681607418857,
"grad_norm": 1.9172236222391295,
"learning_rate": 7.150380037081293e-05,
"loss": 0.9986,
"step": 234
},
{
"epoch": 1.455950540958269,
"grad_norm": 2.2862715410270282,
"learning_rate": 7.139655228680438e-05,
"loss": 1.0075,
"step": 235
},
{
"epoch": 1.4621329211746523,
"grad_norm": 1.8213511562680735,
"learning_rate": 7.128871304435968e-05,
"loss": 1.0109,
"step": 236
},
{
"epoch": 1.4683153013910355,
"grad_norm": 2.1357271598290186,
"learning_rate": 7.118028467395904e-05,
"loss": 0.9899,
"step": 237
},
{
"epoch": 1.474497681607419,
"grad_norm": 1.51723941617764,
"learning_rate": 7.107126921717526e-05,
"loss": 0.999,
"step": 238
},
{
"epoch": 1.480680061823802,
"grad_norm": 2.216906523864941,
"learning_rate": 7.096166872663516e-05,
"loss": 0.9908,
"step": 239
},
{
"epoch": 1.4868624420401855,
"grad_norm": 1.6475567811924245,
"learning_rate": 7.085148526598109e-05,
"loss": 0.9846,
"step": 240
},
{
"epoch": 1.4930448222565689,
"grad_norm": 1.8561898420792513,
"learning_rate": 7.074072090983201e-05,
"loss": 0.9885,
"step": 241
},
{
"epoch": 1.499227202472952,
"grad_norm": 1.4076817648604656,
"learning_rate": 7.062937774374444e-05,
"loss": 0.9827,
"step": 242
},
{
"epoch": 1.5054095826893354,
"grad_norm": 2.043899369455665,
"learning_rate": 7.051745786417309e-05,
"loss": 1.0022,
"step": 243
},
{
"epoch": 1.5115919629057188,
"grad_norm": 1.7025584839030323,
"learning_rate": 7.04049633784316e-05,
"loss": 0.9935,
"step": 244
},
{
"epoch": 1.517774343122102,
"grad_norm": 1.8180249630171987,
"learning_rate": 7.029189640465266e-05,
"loss": 0.9948,
"step": 245
},
{
"epoch": 1.5239567233384852,
"grad_norm": 1.617153539919805,
"learning_rate": 7.017825907174819e-05,
"loss": 0.9946,
"step": 246
},
{
"epoch": 1.5301391035548686,
"grad_norm": 1.9584695599096857,
"learning_rate": 7.006405351936935e-05,
"loss": 0.981,
"step": 247
},
{
"epoch": 1.536321483771252,
"grad_norm": 1.5120141733299006,
"learning_rate": 6.99492818978661e-05,
"loss": 0.9957,
"step": 248
},
{
"epoch": 1.5425038639876352,
"grad_norm": 1.7224427233243673,
"learning_rate": 6.983394636824681e-05,
"loss": 0.985,
"step": 249
},
{
"epoch": 1.5486862442040186,
"grad_norm": 1.352522335955625,
"learning_rate": 6.971804910213753e-05,
"loss": 0.9947,
"step": 250
},
{
"epoch": 1.554868624420402,
"grad_norm": 1.9649504513247107,
"learning_rate": 6.960159228174118e-05,
"loss": 0.9802,
"step": 251
},
{
"epoch": 1.5610510046367851,
"grad_norm": 1.400091397431256,
"learning_rate": 6.94845780997963e-05,
"loss": 0.9869,
"step": 252
},
{
"epoch": 1.5672333848531683,
"grad_norm": 1.809258707764459,
"learning_rate": 6.936700875953593e-05,
"loss": 0.9951,
"step": 253
},
{
"epoch": 1.573415765069552,
"grad_norm": 1.544914399926865,
"learning_rate": 6.924888647464606e-05,
"loss": 0.9866,
"step": 254
},
{
"epoch": 1.5795981452859351,
"grad_norm": 1.6678917783004827,
"learning_rate": 6.91302134692239e-05,
"loss": 0.9839,
"step": 255
},
{
"epoch": 1.5857805255023183,
"grad_norm": 1.4327533793298342,
"learning_rate": 6.90109919777361e-05,
"loss": 0.9842,
"step": 256
},
{
"epoch": 1.5919629057187017,
"grad_norm": 1.4415951253324433,
"learning_rate": 6.889122424497659e-05,
"loss": 0.9888,
"step": 257
},
{
"epoch": 1.598145285935085,
"grad_norm": 1.305272535568148,
"learning_rate": 6.877091252602437e-05,
"loss": 0.9925,
"step": 258
},
{
"epoch": 1.6043276661514683,
"grad_norm": 2.4748168628602016,
"learning_rate": 6.865005908620102e-05,
"loss": 0.9862,
"step": 259
},
{
"epoch": 1.6105100463678517,
"grad_norm": 1.5958180723817172,
"learning_rate": 6.852866620102808e-05,
"loss": 1.0002,
"step": 260
},
{
"epoch": 1.616692426584235,
"grad_norm": 1.6536900940347805,
"learning_rate": 6.840673615618416e-05,
"loss": 0.9708,
"step": 261
},
{
"epoch": 1.6228748068006182,
"grad_norm": 2.6056238946624335,
"learning_rate": 6.828427124746191e-05,
"loss": 0.9896,
"step": 262
},
{
"epoch": 1.6290571870170014,
"grad_norm": 1.090301990739562,
"learning_rate": 6.816127378072484e-05,
"loss": 0.9909,
"step": 263
},
{
"epoch": 1.6352395672333848,
"grad_norm": 2.9921825238289768,
"learning_rate": 6.803774607186389e-05,
"loss": 1.014,
"step": 264
},
{
"epoch": 1.6414219474497682,
"grad_norm": 2.2996006999305822,
"learning_rate": 6.791369044675372e-05,
"loss": 1.0089,
"step": 265
},
{
"epoch": 1.6476043276661514,
"grad_norm": 2.472497102419325,
"learning_rate": 6.778910924120913e-05,
"loss": 0.9898,
"step": 266
},
{
"epoch": 1.6537867078825348,
"grad_norm": 1.4712820809315457,
"learning_rate": 6.766400480094084e-05,
"loss": 0.9982,
"step": 267
},
{
"epoch": 1.6599690880989182,
"grad_norm": 1.7900944541125188,
"learning_rate": 6.753837948151148e-05,
"loss": 1.0049,
"step": 268
},
{
"epoch": 1.6661514683153014,
"grad_norm": 1.5422419041247095,
"learning_rate": 6.74122356482912e-05,
"loss": 0.984,
"step": 269
},
{
"epoch": 1.6723338485316845,
"grad_norm": 1.7666863054997888,
"learning_rate": 6.728557567641313e-05,
"loss": 0.9982,
"step": 270
},
{
"epoch": 1.6785162287480682,
"grad_norm": 1.0188448221447397,
"learning_rate": 6.715840195072862e-05,
"loss": 0.9927,
"step": 271
},
{
"epoch": 1.6846986089644513,
"grad_norm": 1.4323614508510398,
"learning_rate": 6.703071686576243e-05,
"loss": 0.9933,
"step": 272
},
{
"epoch": 1.6908809891808345,
"grad_norm": 1.3733086668334178,
"learning_rate": 6.690252282566753e-05,
"loss": 0.98,
"step": 273
},
{
"epoch": 1.697063369397218,
"grad_norm": 1.7979930738646992,
"learning_rate": 6.677382224417991e-05,
"loss": 0.9915,
"step": 274
},
{
"epoch": 1.7032457496136013,
"grad_norm": 1.5018213587724738,
"learning_rate": 6.664461754457313e-05,
"loss": 0.9789,
"step": 275
},
{
"epoch": 1.7094281298299845,
"grad_norm": 0.8313817964266842,
"learning_rate": 6.651491115961264e-05,
"loss": 0.9898,
"step": 276
},
{
"epoch": 1.7156105100463679,
"grad_norm": 1.563766264104193,
"learning_rate": 6.638470553151003e-05,
"loss": 0.9839,
"step": 277
},
{
"epoch": 1.7217928902627513,
"grad_norm": 1.3394469302914662,
"learning_rate": 6.625400311187701e-05,
"loss": 0.9862,
"step": 278
},
{
"epoch": 1.7279752704791345,
"grad_norm": 0.9069310750596258,
"learning_rate": 6.61228063616793e-05,
"loss": 0.9772,
"step": 279
},
{
"epoch": 1.7341576506955176,
"grad_norm": 1.1937761300056076,
"learning_rate": 6.599111775119019e-05,
"loss": 0.9834,
"step": 280
},
{
"epoch": 1.740340030911901,
"grad_norm": 2.0959839523048722,
"learning_rate": 6.585893975994416e-05,
"loss": 0.9908,
"step": 281
},
{
"epoch": 1.7465224111282844,
"grad_norm": 1.4068229617185941,
"learning_rate": 6.572627487669005e-05,
"loss": 0.9857,
"step": 282
},
{
"epoch": 1.7527047913446676,
"grad_norm": 1.497838436021524,
"learning_rate": 6.55931255993444e-05,
"loss": 0.9982,
"step": 283
},
{
"epoch": 1.758887171561051,
"grad_norm": 1.2049062127509398,
"learning_rate": 6.545949443494416e-05,
"loss": 0.9816,
"step": 284
},
{
"epoch": 1.7650695517774344,
"grad_norm": 2.293459748858112,
"learning_rate": 6.53253838995997e-05,
"loss": 0.9939,
"step": 285
},
{
"epoch": 1.7712519319938176,
"grad_norm": 1.196771206420521,
"learning_rate": 6.519079651844733e-05,
"loss": 0.9917,
"step": 286
},
{
"epoch": 1.7774343122102008,
"grad_norm": 2.6719101899588464,
"learning_rate": 6.505573482560181e-05,
"loss": 1.0024,
"step": 287
},
{
"epoch": 1.7836166924265844,
"grad_norm": 2.151184032156713,
"learning_rate": 6.492020136410854e-05,
"loss": 1.0023,
"step": 288
},
{
"epoch": 1.7897990726429676,
"grad_norm": 2.049945153207459,
"learning_rate": 6.478419868589582e-05,
"loss": 0.9942,
"step": 289
},
{
"epoch": 1.7959814528593507,
"grad_norm": 1.7030760467943136,
"learning_rate": 6.464772935172666e-05,
"loss": 0.9972,
"step": 290
},
{
"epoch": 1.8021638330757341,
"grad_norm": 1.6124042099877869,
"learning_rate": 6.451079593115065e-05,
"loss": 0.9859,
"step": 291
},
{
"epoch": 1.8083462132921175,
"grad_norm": 1.7086055086974385,
"learning_rate": 6.437340100245558e-05,
"loss": 0.9875,
"step": 292
},
{
"epoch": 1.8145285935085007,
"grad_norm": 1.1804574569777302,
"learning_rate": 6.423554715261885e-05,
"loss": 0.9863,
"step": 293
},
{
"epoch": 1.820710973724884,
"grad_norm": 2.072000983521134,
"learning_rate": 6.409723697725876e-05,
"loss": 0.984,
"step": 294
},
{
"epoch": 1.8268933539412675,
"grad_norm": 1.5605771368628958,
"learning_rate": 6.395847308058569e-05,
"loss": 1.0016,
"step": 295
},
{
"epoch": 1.8330757341576507,
"grad_norm": 2.0470560137227944,
"learning_rate": 6.381925807535302e-05,
"loss": 0.9797,
"step": 296
},
{
"epoch": 1.8392581143740339,
"grad_norm": 1.5269600972745945,
"learning_rate": 6.367959458280795e-05,
"loss": 0.9952,
"step": 297
},
{
"epoch": 1.8454404945904173,
"grad_norm": 1.6911790008170018,
"learning_rate": 6.353948523264216e-05,
"loss": 0.9948,
"step": 298
},
{
"epoch": 1.8516228748068007,
"grad_norm": 1.5053272401335374,
"learning_rate": 6.339893266294223e-05,
"loss": 0.9851,
"step": 299
},
{
"epoch": 1.8578052550231838,
"grad_norm": 1.3443219469962477,
"learning_rate": 6.325793952014007e-05,
"loss": 0.9819,
"step": 300
},
{
"epoch": 1.8639876352395672,
"grad_norm": 1.3516905581796814,
"learning_rate": 6.311650845896303e-05,
"loss": 0.9957,
"step": 301
},
{
"epoch": 1.8701700154559506,
"grad_norm": 1.2720069275023291,
"learning_rate": 6.297464214238391e-05,
"loss": 0.9773,
"step": 302
},
{
"epoch": 1.8763523956723338,
"grad_norm": 1.281381485288164,
"learning_rate": 6.283234324157084e-05,
"loss": 0.9751,
"step": 303
},
{
"epoch": 1.8825347758887172,
"grad_norm": 1.109592460056133,
"learning_rate": 6.268961443583696e-05,
"loss": 0.9908,
"step": 304
},
{
"epoch": 1.8887171561051006,
"grad_norm": 2.4545190255394225,
"learning_rate": 6.254645841259005e-05,
"loss": 0.9889,
"step": 305
},
{
"epoch": 1.8948995363214838,
"grad_norm": 1.4595972385019333,
"learning_rate": 6.24028778672818e-05,
"loss": 0.987,
"step": 306
},
{
"epoch": 1.901081916537867,
"grad_norm": 3.0482536802039535,
"learning_rate": 6.225887550335715e-05,
"loss": 0.9898,
"step": 307
},
{
"epoch": 1.9072642967542504,
"grad_norm": 2.1361756083703503,
"learning_rate": 6.21144540322034e-05,
"loss": 0.9899,
"step": 308
},
{
"epoch": 1.9134466769706338,
"grad_norm": 2.820913172877751,
"learning_rate": 6.196961617309908e-05,
"loss": 0.9906,
"step": 309
},
{
"epoch": 1.919629057187017,
"grad_norm": 2.1277865542014536,
"learning_rate": 6.182436465316279e-05,
"loss": 0.9755,
"step": 310
},
{
"epoch": 1.9258114374034003,
"grad_norm": 2.3636934602350372,
"learning_rate": 6.16787022073019e-05,
"loss": 0.981,
"step": 311
},
{
"epoch": 1.9319938176197837,
"grad_norm": 1.5489010186147583,
"learning_rate": 6.153263157816098e-05,
"loss": 0.9714,
"step": 312
},
{
"epoch": 1.938176197836167,
"grad_norm": 2.4665556380164366,
"learning_rate": 6.138615551607017e-05,
"loss": 0.9852,
"step": 313
},
{
"epoch": 1.94435857805255,
"grad_norm": 1.769365451752886,
"learning_rate": 6.123927677899344e-05,
"loss": 0.9817,
"step": 314
},
{
"epoch": 1.9505409582689337,
"grad_norm": 2.276315268584059,
"learning_rate": 6.109199813247666e-05,
"loss": 0.98,
"step": 315
},
{
"epoch": 1.9567233384853169,
"grad_norm": 1.7457681230603967,
"learning_rate": 6.0944322349595426e-05,
"loss": 0.9943,
"step": 316
},
{
"epoch": 1.9629057187017,
"grad_norm": 1.897451826245397,
"learning_rate": 6.0796252210902986e-05,
"loss": 0.9707,
"step": 317
},
{
"epoch": 1.9690880989180835,
"grad_norm": 1.7758674602809486,
"learning_rate": 6.06477905043778e-05,
"loss": 0.986,
"step": 318
},
{
"epoch": 1.9752704791344669,
"grad_norm": 1.4801297268108475,
"learning_rate": 6.049894002537109e-05,
"loss": 0.9624,
"step": 319
},
{
"epoch": 1.98145285935085,
"grad_norm": 1.1171987537741563,
"learning_rate": 6.034970357655415e-05,
"loss": 0.9712,
"step": 320
},
{
"epoch": 1.9876352395672334,
"grad_norm": 1.5190947129294494,
"learning_rate": 6.020008396786562e-05,
"loss": 0.9823,
"step": 321
},
{
"epoch": 1.9938176197836168,
"grad_norm": 1.006947536865789,
"learning_rate": 6.00500840164586e-05,
"loss": 0.9771,
"step": 322
},
{
"epoch": 2.003091190108192,
"grad_norm": 4.006886737490512,
"learning_rate": 5.989970654664756e-05,
"loss": 1.6791,
"step": 323
},
{
"epoch": 2.009273570324575,
"grad_norm": 1.8558698221909793,
"learning_rate": 5.974895438985514e-05,
"loss": 0.952,
"step": 324
},
{
"epoch": 2.015455950540958,
"grad_norm": 1.455258316310302,
"learning_rate": 5.959783038455891e-05,
"loss": 0.9402,
"step": 325
},
{
"epoch": 2.021638330757342,
"grad_norm": 1.7859640511983568,
"learning_rate": 5.944633737623791e-05,
"loss": 0.9346,
"step": 326
},
{
"epoch": 2.027820710973725,
"grad_norm": 0.9891580259760341,
"learning_rate": 5.929447821731899e-05,
"loss": 0.9467,
"step": 327
},
{
"epoch": 2.034003091190108,
"grad_norm": 1.2780787887464713,
"learning_rate": 5.914225576712323e-05,
"loss": 0.9364,
"step": 328
},
{
"epoch": 2.0401854714064913,
"grad_norm": 1.747509422985778,
"learning_rate": 5.8989672891811987e-05,
"loss": 0.9337,
"step": 329
},
{
"epoch": 2.046367851622875,
"grad_norm": 1.2249047782739366,
"learning_rate": 5.883673246433302e-05,
"loss": 0.9498,
"step": 330
},
{
"epoch": 2.052550231839258,
"grad_norm": 2.003614583053326,
"learning_rate": 5.8683437364366316e-05,
"loss": 0.9535,
"step": 331
},
{
"epoch": 2.0587326120556413,
"grad_norm": 1.226972934974554,
"learning_rate": 5.852979047826996e-05,
"loss": 0.9329,
"step": 332
},
{
"epoch": 2.064914992272025,
"grad_norm": 2.172547322337205,
"learning_rate": 5.837579469902567e-05,
"loss": 0.9479,
"step": 333
},
{
"epoch": 2.071097372488408,
"grad_norm": 1.6402327078514485,
"learning_rate": 5.822145292618444e-05,
"loss": 0.9462,
"step": 334
},
{
"epoch": 2.0772797527047913,
"grad_norm": 1.6699665609409802,
"learning_rate": 5.806676806581192e-05,
"loss": 0.9431,
"step": 335
},
{
"epoch": 2.0834621329211744,
"grad_norm": 1.3197835509782132,
"learning_rate": 5.7911743030433596e-05,
"loss": 0.9557,
"step": 336
},
{
"epoch": 2.089644513137558,
"grad_norm": 1.1660288371337755,
"learning_rate": 5.77563807389801e-05,
"loss": 0.9356,
"step": 337
},
{
"epoch": 2.0958268933539412,
"grad_norm": 1.1228481962160324,
"learning_rate": 5.760068411673215e-05,
"loss": 0.9417,
"step": 338
},
{
"epoch": 2.1020092735703244,
"grad_norm": 1.372197582933058,
"learning_rate": 5.7444656095265506e-05,
"loss": 0.9269,
"step": 339
},
{
"epoch": 2.108191653786708,
"grad_norm": 1.0621673720020575,
"learning_rate": 5.7288299612395784e-05,
"loss": 0.9376,
"step": 340
},
{
"epoch": 2.114374034003091,
"grad_norm": 1.290738782393939,
"learning_rate": 5.713161761212309e-05,
"loss": 0.9484,
"step": 341
},
{
"epoch": 2.1205564142194744,
"grad_norm": 1.5545485725757495,
"learning_rate": 5.697461304457667e-05,
"loss": 0.9451,
"step": 342
},
{
"epoch": 2.126738794435858,
"grad_norm": 0.9984947527180189,
"learning_rate": 5.681728886595926e-05,
"loss": 0.9264,
"step": 343
},
{
"epoch": 2.132921174652241,
"grad_norm": 1.435440933520763,
"learning_rate": 5.665964803849152e-05,
"loss": 0.9576,
"step": 344
},
{
"epoch": 2.1391035548686244,
"grad_norm": 1.0914760893879398,
"learning_rate": 5.65016935303562e-05,
"loss": 0.9341,
"step": 345
},
{
"epoch": 2.1452859350850075,
"grad_norm": 0.9611538837750422,
"learning_rate": 5.634342831564228e-05,
"loss": 0.9408,
"step": 346
},
{
"epoch": 2.151468315301391,
"grad_norm": 1.1965180402340152,
"learning_rate": 5.6184855374288946e-05,
"loss": 0.9415,
"step": 347
},
{
"epoch": 2.1576506955177743,
"grad_norm": 1.0359352454914963,
"learning_rate": 5.602597769202953e-05,
"loss": 0.9506,
"step": 348
},
{
"epoch": 2.1638330757341575,
"grad_norm": 1.8769371801728494,
"learning_rate": 5.586679826033521e-05,
"loss": 0.9369,
"step": 349
},
{
"epoch": 2.170015455950541,
"grad_norm": 0.9836827672447646,
"learning_rate": 5.5707320076358776e-05,
"loss": 0.9452,
"step": 350
},
{
"epoch": 2.1761978361669243,
"grad_norm": 1.32097187927076,
"learning_rate": 5.5547546142878164e-05,
"loss": 0.9464,
"step": 351
},
{
"epoch": 2.1823802163833075,
"grad_norm": 1.3411663056222534,
"learning_rate": 5.538747946823987e-05,
"loss": 0.9321,
"step": 352
},
{
"epoch": 2.1885625965996907,
"grad_norm": 1.3570047609173206,
"learning_rate": 5.522712306630236e-05,
"loss": 0.9373,
"step": 353
},
{
"epoch": 2.1947449768160743,
"grad_norm": 0.9761232794419572,
"learning_rate": 5.506647995637932e-05,
"loss": 0.9462,
"step": 354
},
{
"epoch": 2.2009273570324575,
"grad_norm": 1.430074471656494,
"learning_rate": 5.490555316318279e-05,
"loss": 0.9485,
"step": 355
},
{
"epoch": 2.2071097372488406,
"grad_norm": 1.059127482296428,
"learning_rate": 5.474434571676622e-05,
"loss": 0.9422,
"step": 356
},
{
"epoch": 2.2132921174652243,
"grad_norm": 1.8097837664747738,
"learning_rate": 5.4582860652467385e-05,
"loss": 0.938,
"step": 357
},
{
"epoch": 2.2194744976816074,
"grad_norm": 1.125902108018908,
"learning_rate": 5.4421101010851315e-05,
"loss": 0.9334,
"step": 358
},
{
"epoch": 2.2256568778979906,
"grad_norm": 1.8628860336453976,
"learning_rate": 5.425906983765295e-05,
"loss": 0.9462,
"step": 359
},
{
"epoch": 2.2318392581143742,
"grad_norm": 1.5709465886819862,
"learning_rate": 5.409677018371986e-05,
"loss": 0.9563,
"step": 360
},
{
"epoch": 2.2380216383307574,
"grad_norm": 1.5452591189158589,
"learning_rate": 5.393420510495476e-05,
"loss": 0.9542,
"step": 361
},
{
"epoch": 2.2442040185471406,
"grad_norm": 1.215273856313518,
"learning_rate": 5.377137766225799e-05,
"loss": 0.9401,
"step": 362
},
{
"epoch": 2.250386398763524,
"grad_norm": 1.2233276485983613,
"learning_rate": 5.36082909214699e-05,
"loss": 0.9448,
"step": 363
},
{
"epoch": 2.2565687789799074,
"grad_norm": 1.3147561966681955,
"learning_rate": 5.344494795331308e-05,
"loss": 0.9465,
"step": 364
},
{
"epoch": 2.2627511591962906,
"grad_norm": 1.4155564384468768,
"learning_rate": 5.328135183333456e-05,
"loss": 0.951,
"step": 365
},
{
"epoch": 2.2689335394126737,
"grad_norm": 0.6114114300419576,
"learning_rate": 5.311750564184795e-05,
"loss": 0.9396,
"step": 366
},
{
"epoch": 2.2751159196290573,
"grad_norm": 1.5798139980221138,
"learning_rate": 5.2953412463875364e-05,
"loss": 0.9506,
"step": 367
},
{
"epoch": 2.2812982998454405,
"grad_norm": 0.9988799351566259,
"learning_rate": 5.278907538908936e-05,
"loss": 0.9526,
"step": 368
},
{
"epoch": 2.2874806800618237,
"grad_norm": 1.05377706000738,
"learning_rate": 5.26244975117548e-05,
"loss": 0.942,
"step": 369
},
{
"epoch": 2.293663060278207,
"grad_norm": 1.5699703997946644,
"learning_rate": 5.2459681930670536e-05,
"loss": 0.9511,
"step": 370
},
{
"epoch": 2.2998454404945905,
"grad_norm": 0.9947848025404337,
"learning_rate": 5.2294631749111125e-05,
"loss": 0.9151,
"step": 371
},
{
"epoch": 2.3060278207109737,
"grad_norm": 1.1765237704089029,
"learning_rate": 5.2129350074768316e-05,
"loss": 0.9467,
"step": 372
},
{
"epoch": 2.312210200927357,
"grad_norm": 1.5204036418963962,
"learning_rate": 5.1963840019692616e-05,
"loss": 0.953,
"step": 373
},
{
"epoch": 2.3183925811437405,
"grad_norm": 0.9961157368195384,
"learning_rate": 5.1798104700234655e-05,
"loss": 0.9353,
"step": 374
},
{
"epoch": 2.3245749613601236,
"grad_norm": 1.2662517756577527,
"learning_rate": 5.16321472369865e-05,
"loss": 0.9313,
"step": 375
},
{
"epoch": 2.330757341576507,
"grad_norm": 1.0077333440409275,
"learning_rate": 5.146597075472293e-05,
"loss": 0.949,
"step": 376
},
{
"epoch": 2.3369397217928904,
"grad_norm": 1.1569892416058336,
"learning_rate": 5.129957838234255e-05,
"loss": 0.9421,
"step": 377
},
{
"epoch": 2.3431221020092736,
"grad_norm": 1.0327382337486777,
"learning_rate": 5.113297325280896e-05,
"loss": 0.9359,
"step": 378
},
{
"epoch": 2.349304482225657,
"grad_norm": 1.5413059311272268,
"learning_rate": 5.0966158503091673e-05,
"loss": 0.9308,
"step": 379
},
{
"epoch": 2.3554868624420404,
"grad_norm": 1.0041025469926124,
"learning_rate": 5.079913727410712e-05,
"loss": 0.9472,
"step": 380
},
{
"epoch": 2.3616692426584236,
"grad_norm": 1.0449355058196719,
"learning_rate": 5.063191271065945e-05,
"loss": 0.9507,
"step": 381
},
{
"epoch": 2.3678516228748068,
"grad_norm": 1.4231290368942768,
"learning_rate": 5.046448796138138e-05,
"loss": 0.9617,
"step": 382
},
{
"epoch": 2.37403400309119,
"grad_norm": 1.4686945371549769,
"learning_rate": 5.029686617867488e-05,
"loss": 0.9487,
"step": 383
},
{
"epoch": 2.3802163833075736,
"grad_norm": 0.7321343928729324,
"learning_rate": 5.012905051865179e-05,
"loss": 0.9134,
"step": 384
},
{
"epoch": 2.3863987635239567,
"grad_norm": 1.7871707923975175,
"learning_rate": 4.9961044141074454e-05,
"loss": 0.9454,
"step": 385
},
{
"epoch": 2.39258114374034,
"grad_norm": 0.8554156480675892,
"learning_rate": 4.979285020929617e-05,
"loss": 0.958,
"step": 386
},
{
"epoch": 2.398763523956723,
"grad_norm": 2.0222183736349177,
"learning_rate": 4.9624471890201656e-05,
"loss": 0.9398,
"step": 387
},
{
"epoch": 2.4049459041731067,
"grad_norm": 1.1083392384698756,
"learning_rate": 4.945591235414741e-05,
"loss": 0.9353,
"step": 388
},
{
"epoch": 2.41112828438949,
"grad_norm": 2.249237349775883,
"learning_rate": 4.9287174774902034e-05,
"loss": 0.9299,
"step": 389
},
{
"epoch": 2.417310664605873,
"grad_norm": 1.7547197652519158,
"learning_rate": 4.911826232958647e-05,
"loss": 0.9465,
"step": 390
},
{
"epoch": 2.4234930448222567,
"grad_norm": 1.816970919376788,
"learning_rate": 4.894917819861416e-05,
"loss": 0.9425,
"step": 391
},
{
"epoch": 2.42967542503864,
"grad_norm": 1.5766998688721199,
"learning_rate": 4.8779925565631166e-05,
"loss": 0.9572,
"step": 392
},
{
"epoch": 2.435857805255023,
"grad_norm": 1.6988141319269177,
"learning_rate": 4.8610507617456265e-05,
"loss": 0.9401,
"step": 393
},
{
"epoch": 2.4420401854714067,
"grad_norm": 1.4658219077550854,
"learning_rate": 4.8440927544020896e-05,
"loss": 0.9429,
"step": 394
},
{
"epoch": 2.44822256568779,
"grad_norm": 1.4942830281619395,
"learning_rate": 4.82711885383091e-05,
"loss": 0.9305,
"step": 395
},
{
"epoch": 2.454404945904173,
"grad_norm": 1.4911047290042807,
"learning_rate": 4.810129379629745e-05,
"loss": 0.9509,
"step": 396
},
{
"epoch": 2.4605873261205566,
"grad_norm": 1.0945238549084377,
"learning_rate": 4.793124651689484e-05,
"loss": 0.9319,
"step": 397
},
{
"epoch": 2.46676970633694,
"grad_norm": 1.6671980576283978,
"learning_rate": 4.7761049901882227e-05,
"loss": 0.939,
"step": 398
},
{
"epoch": 2.472952086553323,
"grad_norm": 0.92852786359375,
"learning_rate": 4.7590707155852406e-05,
"loss": 0.954,
"step": 399
},
{
"epoch": 2.479134466769706,
"grad_norm": 1.7875978184704922,
"learning_rate": 4.74202214861496e-05,
"loss": 0.9381,
"step": 400
},
{
"epoch": 2.48531684698609,
"grad_norm": 1.212284622798296,
"learning_rate": 4.7249596102809164e-05,
"loss": 0.9467,
"step": 401
},
{
"epoch": 2.491499227202473,
"grad_norm": 1.7168328470834615,
"learning_rate": 4.707883421849703e-05,
"loss": 0.9496,
"step": 402
},
{
"epoch": 2.497681607418856,
"grad_norm": 1.3407496449046934,
"learning_rate": 4.6907939048449314e-05,
"loss": 0.9486,
"step": 403
},
{
"epoch": 2.5038639876352393,
"grad_norm": 1.4996984373050763,
"learning_rate": 4.673691381041173e-05,
"loss": 0.9298,
"step": 404
},
{
"epoch": 2.510046367851623,
"grad_norm": 1.4309169783530684,
"learning_rate": 4.6565761724579e-05,
"loss": 0.9452,
"step": 405
},
{
"epoch": 2.516228748068006,
"grad_norm": 1.4369198232445761,
"learning_rate": 4.6394486013534265e-05,
"loss": 0.9274,
"step": 406
},
{
"epoch": 2.5224111282843893,
"grad_norm": 1.3376962293350674,
"learning_rate": 4.6223089902188336e-05,
"loss": 0.9321,
"step": 407
},
{
"epoch": 2.528593508500773,
"grad_norm": 1.3175000668496903,
"learning_rate": 4.605157661771904e-05,
"loss": 0.9385,
"step": 408
},
{
"epoch": 2.534775888717156,
"grad_norm": 1.6617422309422225,
"learning_rate": 4.5879949389510406e-05,
"loss": 0.9301,
"step": 409
},
{
"epoch": 2.5409582689335393,
"grad_norm": 1.3172794288702236,
"learning_rate": 4.570821144909193e-05,
"loss": 0.9423,
"step": 410
},
{
"epoch": 2.547140649149923,
"grad_norm": 1.6218116986054838,
"learning_rate": 4.553636603007761e-05,
"loss": 0.9309,
"step": 411
},
{
"epoch": 2.553323029366306,
"grad_norm": 1.282648230450059,
"learning_rate": 4.536441636810521e-05,
"loss": 0.9393,
"step": 412
},
{
"epoch": 2.5595054095826892,
"grad_norm": 1.7085653102331866,
"learning_rate": 4.519236570077523e-05,
"loss": 0.9387,
"step": 413
},
{
"epoch": 2.565687789799073,
"grad_norm": 1.181853985248212,
"learning_rate": 4.502021726758994e-05,
"loss": 0.9329,
"step": 414
},
{
"epoch": 2.571870170015456,
"grad_norm": 1.7534552716011547,
"learning_rate": 4.484797430989248e-05,
"loss": 0.9412,
"step": 415
},
{
"epoch": 2.578052550231839,
"grad_norm": 1.3292585743678849,
"learning_rate": 4.467564007080574e-05,
"loss": 0.9399,
"step": 416
},
{
"epoch": 2.584234930448223,
"grad_norm": 1.7630942395608926,
"learning_rate": 4.4503217795171326e-05,
"loss": 0.9303,
"step": 417
},
{
"epoch": 2.590417310664606,
"grad_norm": 1.5547379842247044,
"learning_rate": 4.433071072948846e-05,
"loss": 0.9404,
"step": 418
},
{
"epoch": 2.596599690880989,
"grad_norm": 1.55358461949987,
"learning_rate": 4.415812212185287e-05,
"loss": 0.9479,
"step": 419
},
{
"epoch": 2.6027820710973724,
"grad_norm": 1.3519799459619244,
"learning_rate": 4.398545522189562e-05,
"loss": 0.9381,
"step": 420
},
{
"epoch": 2.6089644513137555,
"grad_norm": 1.4063429925697228,
"learning_rate": 4.38127132807219e-05,
"loss": 0.9308,
"step": 421
},
{
"epoch": 2.615146831530139,
"grad_norm": 1.1100048648257177,
"learning_rate": 4.363989955084988e-05,
"loss": 0.9316,
"step": 422
},
{
"epoch": 2.6213292117465223,
"grad_norm": 1.5473918787261554,
"learning_rate": 4.3467017286149355e-05,
"loss": 0.9288,
"step": 423
},
{
"epoch": 2.6275115919629055,
"grad_norm": 1.1413947086858753,
"learning_rate": 4.32940697417806e-05,
"loss": 0.9281,
"step": 424
},
{
"epoch": 2.633693972179289,
"grad_norm": 1.4822907840670647,
"learning_rate": 4.3121060174133006e-05,
"loss": 0.9391,
"step": 425
},
{
"epoch": 2.6398763523956723,
"grad_norm": 1.1893158134844697,
"learning_rate": 4.294799184076379e-05,
"loss": 0.9358,
"step": 426
},
{
"epoch": 2.6460587326120555,
"grad_norm": 1.395540540534361,
"learning_rate": 4.277486800033664e-05,
"loss": 0.9344,
"step": 427
},
{
"epoch": 2.652241112828439,
"grad_norm": 1.1389047154561869,
"learning_rate": 4.260169191256041e-05,
"loss": 0.9121,
"step": 428
},
{
"epoch": 2.6584234930448223,
"grad_norm": 1.349515321074191,
"learning_rate": 4.242846683812768e-05,
"loss": 0.9424,
"step": 429
},
{
"epoch": 2.6646058732612055,
"grad_norm": 1.2170924594039827,
"learning_rate": 4.225519603865338e-05,
"loss": 0.9285,
"step": 430
},
{
"epoch": 2.670788253477589,
"grad_norm": 1.2637766040441283,
"learning_rate": 4.208188277661341e-05,
"loss": 0.9489,
"step": 431
},
{
"epoch": 2.6769706336939723,
"grad_norm": 0.9722140941647698,
"learning_rate": 4.190853031528317e-05,
"loss": 0.9042,
"step": 432
},
{
"epoch": 2.6831530139103554,
"grad_norm": 1.3212458332351424,
"learning_rate": 4.173514191867615e-05,
"loss": 0.9294,
"step": 433
},
{
"epoch": 2.689335394126739,
"grad_norm": 1.0216097339005061,
"learning_rate": 4.156172085148244e-05,
"loss": 0.9365,
"step": 434
},
{
"epoch": 2.6955177743431222,
"grad_norm": 1.4931636832422879,
"learning_rate": 4.138827037900728e-05,
"loss": 0.9453,
"step": 435
},
{
"epoch": 2.7017001545595054,
"grad_norm": 1.5106660907983307,
"learning_rate": 4.1214793767109607e-05,
"loss": 0.9488,
"step": 436
},
{
"epoch": 2.7078825347758886,
"grad_norm": 0.8429680654036087,
"learning_rate": 4.104129428214048e-05,
"loss": 0.9286,
"step": 437
},
{
"epoch": 2.7140649149922718,
"grad_norm": 1.1888732405027307,
"learning_rate": 4.0867775190881673e-05,
"loss": 0.9389,
"step": 438
},
{
"epoch": 2.7202472952086554,
"grad_norm": 0.8249720034944886,
"learning_rate": 4.0694239760484084e-05,
"loss": 0.9236,
"step": 439
},
{
"epoch": 2.7264296754250386,
"grad_norm": 0.8112023719901648,
"learning_rate": 4.05206912584063e-05,
"loss": 0.93,
"step": 440
},
{
"epoch": 2.7326120556414217,
"grad_norm": 0.8700099921141801,
"learning_rate": 4.0347132952352976e-05,
"loss": 0.9274,
"step": 441
},
{
"epoch": 2.7387944358578054,
"grad_norm": 0.7451536808284058,
"learning_rate": 4.0173568110213444e-05,
"loss": 0.9444,
"step": 442
},
{
"epoch": 2.7449768160741885,
"grad_norm": 0.7730600029414318,
"learning_rate": 4e-05,
"loss": 0.9264,
"step": 443
},
{
"epoch": 2.7511591962905717,
"grad_norm": 0.64992428377988,
"learning_rate": 3.9826431889786576e-05,
"loss": 0.9228,
"step": 444
},
{
"epoch": 2.7573415765069553,
"grad_norm": 0.6092471954250116,
"learning_rate": 3.965286704764702e-05,
"loss": 0.9386,
"step": 445
},
{
"epoch": 2.7635239567233385,
"grad_norm": 0.5829783913867378,
"learning_rate": 3.947930874159371e-05,
"loss": 0.9466,
"step": 446
},
{
"epoch": 2.7697063369397217,
"grad_norm": 0.5791113943292933,
"learning_rate": 3.930576023951593e-05,
"loss": 0.9315,
"step": 447
},
{
"epoch": 2.7758887171561053,
"grad_norm": 0.496929341909277,
"learning_rate": 3.9132224809118347e-05,
"loss": 0.9433,
"step": 448
},
{
"epoch": 2.7820710973724885,
"grad_norm": 0.5814134399140919,
"learning_rate": 3.8958705717859526e-05,
"loss": 0.94,
"step": 449
},
{
"epoch": 2.7882534775888717,
"grad_norm": 0.49572568964279784,
"learning_rate": 3.878520623289041e-05,
"loss": 0.9288,
"step": 450
},
{
"epoch": 2.7944358578052553,
"grad_norm": 0.5284855150274684,
"learning_rate": 3.8611729620992726e-05,
"loss": 0.9333,
"step": 451
},
{
"epoch": 2.8006182380216385,
"grad_norm": 0.516369342375662,
"learning_rate": 3.8438279148517575e-05,
"loss": 0.9484,
"step": 452
},
{
"epoch": 2.8068006182380216,
"grad_norm": 0.46332381283153445,
"learning_rate": 3.826485808132386e-05,
"loss": 0.9394,
"step": 453
},
{
"epoch": 2.812982998454405,
"grad_norm": 0.5353899030659641,
"learning_rate": 3.8091469684716845e-05,
"loss": 0.93,
"step": 454
},
{
"epoch": 2.819165378670788,
"grad_norm": 0.3932685579553986,
"learning_rate": 3.7918117223386615e-05,
"loss": 0.9368,
"step": 455
},
{
"epoch": 2.8253477588871716,
"grad_norm": 0.43586697192948876,
"learning_rate": 3.774480396134663e-05,
"loss": 0.9397,
"step": 456
},
{
"epoch": 2.8315301391035548,
"grad_norm": 0.40361521367835107,
"learning_rate": 3.7571533161872334e-05,
"loss": 0.9213,
"step": 457
},
{
"epoch": 2.837712519319938,
"grad_norm": 0.4604217029229173,
"learning_rate": 3.7398308087439603e-05,
"loss": 0.9268,
"step": 458
},
{
"epoch": 2.8438948995363216,
"grad_norm": 0.538514910957252,
"learning_rate": 3.722513199966336e-05,
"loss": 0.9455,
"step": 459
},
{
"epoch": 2.8500772797527048,
"grad_norm": 0.6470702786089004,
"learning_rate": 3.705200815923622e-05,
"loss": 0.9123,
"step": 460
},
{
"epoch": 2.856259659969088,
"grad_norm": 0.5008393569354235,
"learning_rate": 3.6878939825867014e-05,
"loss": 0.9382,
"step": 461
},
{
"epoch": 2.8624420401854715,
"grad_norm": 0.41610970566343775,
"learning_rate": 3.670593025821942e-05,
"loss": 0.9387,
"step": 462
},
{
"epoch": 2.8686244204018547,
"grad_norm": 0.3582328984320768,
"learning_rate": 3.653298271385065e-05,
"loss": 0.9365,
"step": 463
},
{
"epoch": 2.874806800618238,
"grad_norm": 0.36268739794043653,
"learning_rate": 3.6360100449150135e-05,
"loss": 0.9398,
"step": 464
},
{
"epoch": 2.8809891808346215,
"grad_norm": 0.31637193895013355,
"learning_rate": 3.618728671927811e-05,
"loss": 0.9468,
"step": 465
},
{
"epoch": 2.8871715610510047,
"grad_norm": 0.37689814936512106,
"learning_rate": 3.6014544778104394e-05,
"loss": 0.9499,
"step": 466
},
{
"epoch": 2.893353941267388,
"grad_norm": 0.4779224408619651,
"learning_rate": 3.584187787814714e-05,
"loss": 0.9396,
"step": 467
},
{
"epoch": 2.8995363214837715,
"grad_norm": 0.509229566930791,
"learning_rate": 3.566928927051155e-05,
"loss": 0.9291,
"step": 468
},
{
"epoch": 2.9057187017001547,
"grad_norm": 0.48474577075424746,
"learning_rate": 3.5496782204828694e-05,
"loss": 0.9203,
"step": 469
},
{
"epoch": 2.911901081916538,
"grad_norm": 0.3971864637298685,
"learning_rate": 3.5324359929194274e-05,
"loss": 0.9394,
"step": 470
},
{
"epoch": 2.918083462132921,
"grad_norm": 0.3053259719312419,
"learning_rate": 3.5152025690107524e-05,
"loss": 0.9268,
"step": 471
},
{
"epoch": 2.9242658423493046,
"grad_norm": 0.29216270961639657,
"learning_rate": 3.497978273241007e-05,
"loss": 0.9148,
"step": 472
},
{
"epoch": 2.930448222565688,
"grad_norm": 0.2654490613569246,
"learning_rate": 3.480763429922479e-05,
"loss": 0.9379,
"step": 473
},
{
"epoch": 2.936630602782071,
"grad_norm": 0.38722618089775157,
"learning_rate": 3.463558363189479e-05,
"loss": 0.9277,
"step": 474
},
{
"epoch": 2.942812982998454,
"grad_norm": 0.3877069119339608,
"learning_rate": 3.4463633969922395e-05,
"loss": 0.9165,
"step": 475
},
{
"epoch": 2.948995363214838,
"grad_norm": 0.49582780013936295,
"learning_rate": 3.429178855090809e-05,
"loss": 0.9395,
"step": 476
},
{
"epoch": 2.955177743431221,
"grad_norm": 0.48453229307728,
"learning_rate": 3.412005061048959e-05,
"loss": 0.9184,
"step": 477
},
{
"epoch": 2.961360123647604,
"grad_norm": 0.372292891013131,
"learning_rate": 3.394842338228097e-05,
"loss": 0.9257,
"step": 478
},
{
"epoch": 2.9675425038639878,
"grad_norm": 0.21481675893348937,
"learning_rate": 3.377691009781168e-05,
"loss": 0.9342,
"step": 479
},
{
"epoch": 2.973724884080371,
"grad_norm": 0.2791079849765998,
"learning_rate": 3.360551398646575e-05,
"loss": 0.927,
"step": 480
},
{
"epoch": 2.979907264296754,
"grad_norm": 0.31864441303786156,
"learning_rate": 3.3434238275421e-05,
"loss": 0.9364,
"step": 481
},
{
"epoch": 2.9860896445131377,
"grad_norm": 0.3426078238120316,
"learning_rate": 3.326308618958828e-05,
"loss": 0.9237,
"step": 482
},
{
"epoch": 2.992272024729521,
"grad_norm": 0.33872000941997643,
"learning_rate": 3.309206095155069e-05,
"loss": 0.9331,
"step": 483
},
{
"epoch": 3.001545595054096,
"grad_norm": 0.41746066648889707,
"learning_rate": 3.292116578150298e-05,
"loss": 1.632,
"step": 484
},
{
"epoch": 3.007727975270479,
"grad_norm": 0.479690534350614,
"learning_rate": 3.2750403897190856e-05,
"loss": 0.8917,
"step": 485
},
{
"epoch": 3.0139103554868623,
"grad_norm": 0.37954910983497075,
"learning_rate": 3.2579778513850405e-05,
"loss": 0.8802,
"step": 486
},
{
"epoch": 3.020092735703246,
"grad_norm": 0.43731886585463775,
"learning_rate": 3.240929284414762e-05,
"loss": 0.9013,
"step": 487
},
{
"epoch": 3.026275115919629,
"grad_norm": 0.3851011426541529,
"learning_rate": 3.223895009811777e-05,
"loss": 0.876,
"step": 488
},
{
"epoch": 3.0324574961360122,
"grad_norm": 0.321671225557479,
"learning_rate": 3.2068753483105165e-05,
"loss": 0.8888,
"step": 489
},
{
"epoch": 3.038639876352396,
"grad_norm": 0.4076120056603458,
"learning_rate": 3.189870620370256e-05,
"loss": 0.9015,
"step": 490
},
{
"epoch": 3.044822256568779,
"grad_norm": 0.32597913024373476,
"learning_rate": 3.1728811461690904e-05,
"loss": 0.8858,
"step": 491
},
{
"epoch": 3.051004636785162,
"grad_norm": 0.39721200533205675,
"learning_rate": 3.155907245597912e-05,
"loss": 0.8971,
"step": 492
},
{
"epoch": 3.0571870170015454,
"grad_norm": 0.33689618227170504,
"learning_rate": 3.138949238254375e-05,
"loss": 0.8913,
"step": 493
},
{
"epoch": 3.063369397217929,
"grad_norm": 0.31095628858697094,
"learning_rate": 3.122007443436885e-05,
"loss": 0.8996,
"step": 494
},
{
"epoch": 3.069551777434312,
"grad_norm": 0.4028486236419519,
"learning_rate": 3.105082180138585e-05,
"loss": 0.8889,
"step": 495
},
{
"epoch": 3.0757341576506954,
"grad_norm": 0.484359324518245,
"learning_rate": 3.0881737670413534e-05,
"loss": 0.8964,
"step": 496
},
{
"epoch": 3.081916537867079,
"grad_norm": 0.38584852659317825,
"learning_rate": 3.071282522509798e-05,
"loss": 0.8794,
"step": 497
},
{
"epoch": 3.088098918083462,
"grad_norm": 0.3397498441090358,
"learning_rate": 3.054408764585261e-05,
"loss": 0.8957,
"step": 498
},
{
"epoch": 3.0942812982998453,
"grad_norm": 0.31333063472846795,
"learning_rate": 3.037552810979836e-05,
"loss": 0.8786,
"step": 499
},
{
"epoch": 3.100463678516229,
"grad_norm": 0.3649292765310615,
"learning_rate": 3.020714979070385e-05,
"loss": 0.8884,
"step": 500
},
{
"epoch": 3.106646058732612,
"grad_norm": 0.4200346160735661,
"learning_rate": 3.0038955858925556e-05,
"loss": 0.8846,
"step": 501
},
{
"epoch": 3.1128284389489953,
"grad_norm": 0.28488029439459195,
"learning_rate": 2.9870949481348214e-05,
"loss": 0.8942,
"step": 502
},
{
"epoch": 3.1190108191653785,
"grad_norm": 0.2697608832955852,
"learning_rate": 2.9703133821325125e-05,
"loss": 0.8915,
"step": 503
},
{
"epoch": 3.125193199381762,
"grad_norm": 0.3357046563951315,
"learning_rate": 2.953551203861863e-05,
"loss": 0.8885,
"step": 504
},
{
"epoch": 3.1313755795981453,
"grad_norm": 0.3200692987355479,
"learning_rate": 2.9368087289340575e-05,
"loss": 0.8844,
"step": 505
},
{
"epoch": 3.1375579598145285,
"grad_norm": 0.24036855461462353,
"learning_rate": 2.9200862725892897e-05,
"loss": 0.8929,
"step": 506
},
{
"epoch": 3.143740340030912,
"grad_norm": 0.238395091168019,
"learning_rate": 2.903384149690834e-05,
"loss": 0.8874,
"step": 507
},
{
"epoch": 3.1499227202472952,
"grad_norm": 0.2401017173847325,
"learning_rate": 2.8867026747191054e-05,
"loss": 0.8876,
"step": 508
},
{
"epoch": 3.1561051004636784,
"grad_norm": 0.2799086754124938,
"learning_rate": 2.8700421617657446e-05,
"loss": 0.8958,
"step": 509
},
{
"epoch": 3.162287480680062,
"grad_norm": 0.2780648545074759,
"learning_rate": 2.853402924527708e-05,
"loss": 0.8952,
"step": 510
},
{
"epoch": 3.1684698608964452,
"grad_norm": 0.23416871362717123,
"learning_rate": 2.8367852763013508e-05,
"loss": 0.8783,
"step": 511
},
{
"epoch": 3.1746522411128284,
"grad_norm": 0.19517664014067915,
"learning_rate": 2.8201895299765358e-05,
"loss": 0.8885,
"step": 512
},
{
"epoch": 3.1808346213292116,
"grad_norm": 0.24055029383355403,
"learning_rate": 2.803615998030738e-05,
"loss": 0.8781,
"step": 513
},
{
"epoch": 3.187017001545595,
"grad_norm": 0.21251629293077243,
"learning_rate": 2.787064992523169e-05,
"loss": 0.896,
"step": 514
},
{
"epoch": 3.1931993817619784,
"grad_norm": 0.2205317155524029,
"learning_rate": 2.7705368250888892e-05,
"loss": 0.8977,
"step": 515
},
{
"epoch": 3.1993817619783615,
"grad_norm": 0.2710843453756844,
"learning_rate": 2.7540318069329468e-05,
"loss": 0.8809,
"step": 516
},
{
"epoch": 3.205564142194745,
"grad_norm": 0.2134694304897921,
"learning_rate": 2.7375502488245218e-05,
"loss": 0.8852,
"step": 517
},
{
"epoch": 3.2117465224111283,
"grad_norm": 0.21525338899378574,
"learning_rate": 2.7210924610910658e-05,
"loss": 0.8843,
"step": 518
},
{
"epoch": 3.2179289026275115,
"grad_norm": 0.2257874439917892,
"learning_rate": 2.7046587536124656e-05,
"loss": 0.8912,
"step": 519
},
{
"epoch": 3.2241112828438947,
"grad_norm": 0.2146891709570046,
"learning_rate": 2.6882494358152055e-05,
"loss": 0.8902,
"step": 520
},
{
"epoch": 3.2302936630602783,
"grad_norm": 0.19521290168739072,
"learning_rate": 2.6718648166665445e-05,
"loss": 0.8931,
"step": 521
},
{
"epoch": 3.2364760432766615,
"grad_norm": 0.2375023715085163,
"learning_rate": 2.6555052046686943e-05,
"loss": 0.8799,
"step": 522
},
{
"epoch": 3.2426584234930447,
"grad_norm": 0.2440992341795168,
"learning_rate": 2.6391709078530106e-05,
"loss": 0.9015,
"step": 523
},
{
"epoch": 3.2488408037094283,
"grad_norm": 0.2660799604690821,
"learning_rate": 2.6228622337742016e-05,
"loss": 0.8869,
"step": 524
},
{
"epoch": 3.2550231839258115,
"grad_norm": 0.1898820986247372,
"learning_rate": 2.606579489504526e-05,
"loss": 0.8921,
"step": 525
},
{
"epoch": 3.2612055641421946,
"grad_norm": 0.21206019174841687,
"learning_rate": 2.590322981628016e-05,
"loss": 0.8969,
"step": 526
},
{
"epoch": 3.2673879443585783,
"grad_norm": 0.21516060193400413,
"learning_rate": 2.5740930162347053e-05,
"loss": 0.8839,
"step": 527
},
{
"epoch": 3.2735703245749614,
"grad_norm": 0.19044808648502842,
"learning_rate": 2.5578898989148692e-05,
"loss": 0.8927,
"step": 528
},
{
"epoch": 3.2797527047913446,
"grad_norm": 0.22486834698093985,
"learning_rate": 2.5417139347532625e-05,
"loss": 0.8889,
"step": 529
},
{
"epoch": 3.285935085007728,
"grad_norm": 0.19216048198447702,
"learning_rate": 2.525565428323379e-05,
"loss": 0.8893,
"step": 530
},
{
"epoch": 3.2921174652241114,
"grad_norm": 0.20965862072042166,
"learning_rate": 2.509444683681722e-05,
"loss": 0.8989,
"step": 531
},
{
"epoch": 3.2982998454404946,
"grad_norm": 0.1815566372511265,
"learning_rate": 2.49335200436207e-05,
"loss": 0.8795,
"step": 532
},
{
"epoch": 3.3044822256568778,
"grad_norm": 0.23072409231554872,
"learning_rate": 2.4772876933697658e-05,
"loss": 0.909,
"step": 533
},
{
"epoch": 3.3106646058732614,
"grad_norm": 0.2299841285498698,
"learning_rate": 2.461252053176015e-05,
"loss": 0.8966,
"step": 534
},
{
"epoch": 3.3168469860896446,
"grad_norm": 0.23678203939185868,
"learning_rate": 2.4452453857121846e-05,
"loss": 0.904,
"step": 535
},
{
"epoch": 3.3230293663060277,
"grad_norm": 0.23345278141855083,
"learning_rate": 2.429267992364123e-05,
"loss": 0.8709,
"step": 536
},
{
"epoch": 3.329211746522411,
"grad_norm": 0.21351104494670123,
"learning_rate": 2.413320173966481e-05,
"loss": 0.886,
"step": 537
},
{
"epoch": 3.3353941267387945,
"grad_norm": 0.24843505340058486,
"learning_rate": 2.3974022307970488e-05,
"loss": 0.8994,
"step": 538
},
{
"epoch": 3.3415765069551777,
"grad_norm": 0.2144602081955269,
"learning_rate": 2.3815144625711063e-05,
"loss": 0.8928,
"step": 539
},
{
"epoch": 3.347758887171561,
"grad_norm": 0.214440559424067,
"learning_rate": 2.3656571684357736e-05,
"loss": 0.8952,
"step": 540
},
{
"epoch": 3.3539412673879445,
"grad_norm": 0.1870653320452973,
"learning_rate": 2.34983064696438e-05,
"loss": 0.8789,
"step": 541
},
{
"epoch": 3.3601236476043277,
"grad_norm": 0.1816470397576087,
"learning_rate": 2.3340351961508495e-05,
"loss": 0.8904,
"step": 542
},
{
"epoch": 3.366306027820711,
"grad_norm": 0.23629374681914123,
"learning_rate": 2.318271113404075e-05,
"loss": 0.8853,
"step": 543
},
{
"epoch": 3.3724884080370945,
"grad_norm": 0.2136915907110259,
"learning_rate": 2.3025386955423342e-05,
"loss": 0.9049,
"step": 544
},
{
"epoch": 3.3786707882534777,
"grad_norm": 0.22758131014283642,
"learning_rate": 2.2868382387876905e-05,
"loss": 0.8876,
"step": 545
},
{
"epoch": 3.384853168469861,
"grad_norm": 0.2365037683934222,
"learning_rate": 2.2711700387604223e-05,
"loss": 0.9025,
"step": 546
},
{
"epoch": 3.391035548686244,
"grad_norm": 0.21359052590913488,
"learning_rate": 2.255534390473451e-05,
"loss": 0.8939,
"step": 547
},
{
"epoch": 3.3972179289026276,
"grad_norm": 0.21656702849393653,
"learning_rate": 2.2399315883267853e-05,
"loss": 0.8987,
"step": 548
},
{
"epoch": 3.403400309119011,
"grad_norm": 0.225661122723241,
"learning_rate": 2.2243619261019918e-05,
"loss": 0.8893,
"step": 549
},
{
"epoch": 3.409582689335394,
"grad_norm": 0.18874144934647666,
"learning_rate": 2.2088256969566413e-05,
"loss": 0.9041,
"step": 550
},
{
"epoch": 3.4157650695517776,
"grad_norm": 0.28668375357552045,
"learning_rate": 2.1933231934188104e-05,
"loss": 0.8855,
"step": 551
},
{
"epoch": 3.421947449768161,
"grad_norm": 0.2531027429361992,
"learning_rate": 2.1778547073815554e-05,
"loss": 0.8954,
"step": 552
},
{
"epoch": 3.428129829984544,
"grad_norm": 0.21767635407476482,
"learning_rate": 2.1624205300974346e-05,
"loss": 0.8951,
"step": 553
},
{
"epoch": 3.434312210200927,
"grad_norm": 0.18527077734942415,
"learning_rate": 2.1470209521730064e-05,
"loss": 0.891,
"step": 554
},
{
"epoch": 3.4404945904173108,
"grad_norm": 0.26428285410878966,
"learning_rate": 2.131656263563369e-05,
"loss": 0.9102,
"step": 555
},
{
"epoch": 3.446676970633694,
"grad_norm": 0.17077454909439538,
"learning_rate": 2.1163267535666994e-05,
"loss": 0.9022,
"step": 556
},
{
"epoch": 3.452859350850077,
"grad_norm": 0.2136026248917033,
"learning_rate": 2.1010327108188037e-05,
"loss": 0.8797,
"step": 557
},
{
"epoch": 3.4590417310664607,
"grad_norm": 0.23170046494717933,
"learning_rate": 2.0857744232876797e-05,
"loss": 0.9014,
"step": 558
},
{
"epoch": 3.465224111282844,
"grad_norm": 0.18466452017435814,
"learning_rate": 2.070552178268102e-05,
"loss": 0.893,
"step": 559
},
{
"epoch": 3.471406491499227,
"grad_norm": 0.2938322911655651,
"learning_rate": 2.0553662623762105e-05,
"loss": 0.8872,
"step": 560
},
{
"epoch": 3.4775888717156107,
"grad_norm": 0.28795550547667914,
"learning_rate": 2.040216961544109e-05,
"loss": 0.8975,
"step": 561
},
{
"epoch": 3.483771251931994,
"grad_norm": 0.22847154183739538,
"learning_rate": 2.0251045610144865e-05,
"loss": 0.8918,
"step": 562
},
{
"epoch": 3.489953632148377,
"grad_norm": 0.28080426895854366,
"learning_rate": 2.0100293453352446e-05,
"loss": 0.8703,
"step": 563
},
{
"epoch": 3.4961360123647607,
"grad_norm": 0.23972543857227632,
"learning_rate": 1.9949915983541405e-05,
"loss": 0.8794,
"step": 564
},
{
"epoch": 3.502318392581144,
"grad_norm": 0.2159429902592839,
"learning_rate": 1.9799916032134384e-05,
"loss": 0.8844,
"step": 565
},
{
"epoch": 3.508500772797527,
"grad_norm": 0.2126223807781731,
"learning_rate": 1.9650296423445865e-05,
"loss": 0.8934,
"step": 566
},
{
"epoch": 3.51468315301391,
"grad_norm": 0.25136395675418194,
"learning_rate": 1.9501059974628923e-05,
"loss": 0.8951,
"step": 567
},
{
"epoch": 3.5208655332302934,
"grad_norm": 0.2422901671285048,
"learning_rate": 1.9352209495622215e-05,
"loss": 0.8931,
"step": 568
},
{
"epoch": 3.527047913446677,
"grad_norm": 0.22399150155898304,
"learning_rate": 1.920374778909702e-05,
"loss": 0.894,
"step": 569
},
{
"epoch": 3.53323029366306,
"grad_norm": 0.17997096245249658,
"learning_rate": 1.9055677650404598e-05,
"loss": 0.8854,
"step": 570
},
{
"epoch": 3.5394126738794434,
"grad_norm": 0.20091228641905334,
"learning_rate": 1.890800186752336e-05,
"loss": 0.8912,
"step": 571
},
{
"epoch": 3.545595054095827,
"grad_norm": 0.1713636230002705,
"learning_rate": 1.8760723221006577e-05,
"loss": 0.8984,
"step": 572
},
{
"epoch": 3.55177743431221,
"grad_norm": 0.22444993402898378,
"learning_rate": 1.861384448392984e-05,
"loss": 0.8957,
"step": 573
},
{
"epoch": 3.5579598145285933,
"grad_norm": 0.1622926494483365,
"learning_rate": 1.8467368421839045e-05,
"loss": 0.8882,
"step": 574
},
{
"epoch": 3.564142194744977,
"grad_norm": 0.22561783403087973,
"learning_rate": 1.832129779269811e-05,
"loss": 0.8837,
"step": 575
},
{
"epoch": 3.57032457496136,
"grad_norm": 0.1776156141854532,
"learning_rate": 1.817563534683722e-05,
"loss": 0.8705,
"step": 576
},
{
"epoch": 3.5765069551777433,
"grad_norm": 0.1905281137121667,
"learning_rate": 1.8030383826900935e-05,
"loss": 0.8814,
"step": 577
},
{
"epoch": 3.582689335394127,
"grad_norm": 0.2044502851986332,
"learning_rate": 1.7885545967796606e-05,
"loss": 0.8981,
"step": 578
},
{
"epoch": 3.58887171561051,
"grad_norm": 0.1784191323277939,
"learning_rate": 1.7741124496642862e-05,
"loss": 0.8901,
"step": 579
},
{
"epoch": 3.5950540958268933,
"grad_norm": 0.18594801839576822,
"learning_rate": 1.7597122132718202e-05,
"loss": 0.8666,
"step": 580
},
{
"epoch": 3.601236476043277,
"grad_norm": 0.18365311459244033,
"learning_rate": 1.7453541587409963e-05,
"loss": 0.8936,
"step": 581
},
{
"epoch": 3.60741885625966,
"grad_norm": 0.20713153169279955,
"learning_rate": 1.731038556416304e-05,
"loss": 0.8898,
"step": 582
},
{
"epoch": 3.6136012364760433,
"grad_norm": 0.20905582740873777,
"learning_rate": 1.716765675842919e-05,
"loss": 0.8905,
"step": 583
},
{
"epoch": 3.6197836166924264,
"grad_norm": 0.18415135918656184,
"learning_rate": 1.70253578576161e-05,
"loss": 0.8945,
"step": 584
},
{
"epoch": 3.62596599690881,
"grad_norm": 0.2042255922116097,
"learning_rate": 1.688349154103699e-05,
"loss": 0.8839,
"step": 585
},
{
"epoch": 3.6321483771251932,
"grad_norm": 0.17189580999683968,
"learning_rate": 1.6742060479859944e-05,
"loss": 0.8747,
"step": 586
},
{
"epoch": 3.6383307573415764,
"grad_norm": 0.21696578872996222,
"learning_rate": 1.6601067337057783e-05,
"loss": 0.8863,
"step": 587
},
{
"epoch": 3.6445131375579596,
"grad_norm": 0.20566434805583705,
"learning_rate": 1.6460514767357855e-05,
"loss": 0.8978,
"step": 588
},
{
"epoch": 3.650695517774343,
"grad_norm": 0.212105435407208,
"learning_rate": 1.6320405417192066e-05,
"loss": 0.8796,
"step": 589
},
{
"epoch": 3.6568778979907264,
"grad_norm": 0.21704798077180454,
"learning_rate": 1.618074192464699e-05,
"loss": 0.8962,
"step": 590
},
{
"epoch": 3.6630602782071096,
"grad_norm": 0.26219775112073423,
"learning_rate": 1.6041526919414318e-05,
"loss": 0.8965,
"step": 591
},
{
"epoch": 3.669242658423493,
"grad_norm": 0.18361447307055298,
"learning_rate": 1.5902763022741247e-05,
"loss": 0.8912,
"step": 592
},
{
"epoch": 3.6754250386398764,
"grad_norm": 0.2767924709588361,
"learning_rate": 1.576445284738116e-05,
"loss": 0.8974,
"step": 593
},
{
"epoch": 3.6816074188562595,
"grad_norm": 0.21044903902705353,
"learning_rate": 1.5626598997544422e-05,
"loss": 0.8932,
"step": 594
},
{
"epoch": 3.687789799072643,
"grad_norm": 0.22141335816275728,
"learning_rate": 1.548920406884935e-05,
"loss": 0.8966,
"step": 595
},
{
"epoch": 3.6939721792890263,
"grad_norm": 0.19199822889292517,
"learning_rate": 1.5352270648273363e-05,
"loss": 0.8927,
"step": 596
},
{
"epoch": 3.7001545595054095,
"grad_norm": 0.2256809667548217,
"learning_rate": 1.5215801314104202e-05,
"loss": 0.8917,
"step": 597
},
{
"epoch": 3.706336939721793,
"grad_norm": 0.23515679935616451,
"learning_rate": 1.5079798635891468e-05,
"loss": 0.8907,
"step": 598
},
{
"epoch": 3.7125193199381763,
"grad_norm": 0.1640102854982691,
"learning_rate": 1.4944265174398203e-05,
"loss": 0.9016,
"step": 599
},
{
"epoch": 3.7187017001545595,
"grad_norm": 0.19789240249181025,
"learning_rate": 1.4809203481552681e-05,
"loss": 0.8811,
"step": 600
},
{
"epoch": 3.7248840803709427,
"grad_norm": 0.17774590981378566,
"learning_rate": 1.4674616100400303e-05,
"loss": 0.8811,
"step": 601
},
{
"epoch": 3.7310664605873263,
"grad_norm": 0.19736844331190076,
"learning_rate": 1.4540505565055858e-05,
"loss": 0.8996,
"step": 602
},
{
"epoch": 3.7372488408037094,
"grad_norm": 0.17893905096693596,
"learning_rate": 1.4406874400655615e-05,
"loss": 0.9037,
"step": 603
},
{
"epoch": 3.7434312210200926,
"grad_norm": 0.19684792739449095,
"learning_rate": 1.427372512330996e-05,
"loss": 0.8851,
"step": 604
},
{
"epoch": 3.749613601236476,
"grad_norm": 0.2512744015419969,
"learning_rate": 1.4141060240055859e-05,
"loss": 0.8954,
"step": 605
},
{
"epoch": 3.7557959814528594,
"grad_norm": 0.17423327717089102,
"learning_rate": 1.400888224880983e-05,
"loss": 0.8808,
"step": 606
},
{
"epoch": 3.7619783616692426,
"grad_norm": 0.1819843673558297,
"learning_rate": 1.3877193638320718e-05,
"loss": 0.8982,
"step": 607
},
{
"epoch": 3.7681607418856258,
"grad_norm": 0.1693796580052696,
"learning_rate": 1.3745996888122992e-05,
"loss": 0.8701,
"step": 608
},
{
"epoch": 3.7743431221020094,
"grad_norm": 0.1821658263170253,
"learning_rate": 1.361529446848998e-05,
"loss": 0.8843,
"step": 609
},
{
"epoch": 3.7805255023183926,
"grad_norm": 0.24650364015406534,
"learning_rate": 1.3485088840387369e-05,
"loss": 0.8966,
"step": 610
},
{
"epoch": 3.7867078825347757,
"grad_norm": 0.18771305026658164,
"learning_rate": 1.3355382455426892e-05,
"loss": 0.8846,
"step": 611
},
{
"epoch": 3.7928902627511594,
"grad_norm": 0.18610630863899502,
"learning_rate": 1.3226177755820087e-05,
"loss": 0.8846,
"step": 612
},
{
"epoch": 3.7990726429675425,
"grad_norm": 0.17549326513432761,
"learning_rate": 1.3097477174332482e-05,
"loss": 0.8855,
"step": 613
},
{
"epoch": 3.8052550231839257,
"grad_norm": 0.17546252086320724,
"learning_rate": 1.296928313423758e-05,
"loss": 0.894,
"step": 614
},
{
"epoch": 3.8114374034003093,
"grad_norm": 0.1904786805615447,
"learning_rate": 1.2841598049271395e-05,
"loss": 0.8847,
"step": 615
},
{
"epoch": 3.8176197836166925,
"grad_norm": 0.16824287711277486,
"learning_rate": 1.271442432358688e-05,
"loss": 0.882,
"step": 616
},
{
"epoch": 3.8238021638330757,
"grad_norm": 0.15191088188497087,
"learning_rate": 1.2587764351708813e-05,
"loss": 0.897,
"step": 617
},
{
"epoch": 3.8299845440494593,
"grad_norm": 0.17144785228852993,
"learning_rate": 1.2461620518488533e-05,
"loss": 0.8947,
"step": 618
},
{
"epoch": 3.8361669242658425,
"grad_norm": 0.1981987256042829,
"learning_rate": 1.2335995199059175e-05,
"loss": 0.8888,
"step": 619
},
{
"epoch": 3.8423493044822257,
"grad_norm": 0.1664686018609899,
"learning_rate": 1.2210890758790877e-05,
"loss": 0.8865,
"step": 620
},
{
"epoch": 3.848531684698609,
"grad_norm": 0.17952380029164736,
"learning_rate": 1.2086309553246282e-05,
"loss": 0.8735,
"step": 621
},
{
"epoch": 3.854714064914992,
"grad_norm": 0.1654528501349644,
"learning_rate": 1.1962253928136129e-05,
"loss": 0.8975,
"step": 622
},
{
"epoch": 3.8608964451313756,
"grad_norm": 0.19932349446317263,
"learning_rate": 1.1838726219275163e-05,
"loss": 0.8994,
"step": 623
},
{
"epoch": 3.867078825347759,
"grad_norm": 0.16515103220193778,
"learning_rate": 1.1715728752538103e-05,
"loss": 0.8795,
"step": 624
},
{
"epoch": 3.873261205564142,
"grad_norm": 0.15802093965187888,
"learning_rate": 1.1593263843815854e-05,
"loss": 0.8846,
"step": 625
},
{
"epoch": 3.8794435857805256,
"grad_norm": 0.18002633059902745,
"learning_rate": 1.1471333798971922e-05,
"loss": 0.8813,
"step": 626
},
{
"epoch": 3.885625965996909,
"grad_norm": 0.16043447191960278,
"learning_rate": 1.1349940913798978e-05,
"loss": 0.88,
"step": 627
},
{
"epoch": 3.891808346213292,
"grad_norm": 0.21327907033825158,
"learning_rate": 1.1229087473975641e-05,
"loss": 0.8961,
"step": 628
},
{
"epoch": 3.8979907264296756,
"grad_norm": 0.15650552928778205,
"learning_rate": 1.1108775755023422e-05,
"loss": 0.8748,
"step": 629
},
{
"epoch": 3.9041731066460588,
"grad_norm": 0.18123273261524603,
"learning_rate": 1.0989008022263906e-05,
"loss": 0.8935,
"step": 630
},
{
"epoch": 3.910355486862442,
"grad_norm": 0.15260920437276415,
"learning_rate": 1.08697865307761e-05,
"loss": 0.8852,
"step": 631
},
{
"epoch": 3.9165378670788256,
"grad_norm": 0.165453780873266,
"learning_rate": 1.0751113525353957e-05,
"loss": 0.9049,
"step": 632
},
{
"epoch": 3.9227202472952087,
"grad_norm": 0.13033209339725224,
"learning_rate": 1.0632991240464068e-05,
"loss": 0.8931,
"step": 633
},
{
"epoch": 3.928902627511592,
"grad_norm": 0.16631041812626945,
"learning_rate": 1.0515421900203714e-05,
"loss": 0.888,
"step": 634
},
{
"epoch": 3.9350850077279755,
"grad_norm": 0.16953390165836668,
"learning_rate": 1.0398407718258836e-05,
"loss": 0.8984,
"step": 635
},
{
"epoch": 3.9412673879443587,
"grad_norm": 0.14284684056392535,
"learning_rate": 1.028195089786248e-05,
"loss": 0.9026,
"step": 636
},
{
"epoch": 3.947449768160742,
"grad_norm": 0.16855244368923294,
"learning_rate": 1.0166053631753204e-05,
"loss": 0.8875,
"step": 637
},
{
"epoch": 3.953632148377125,
"grad_norm": 0.14590639079902812,
"learning_rate": 1.0050718102133916e-05,
"loss": 0.8929,
"step": 638
},
{
"epoch": 3.9598145285935082,
"grad_norm": 0.14384779730638278,
"learning_rate": 9.935946480630658e-06,
"loss": 0.8825,
"step": 639
},
{
"epoch": 3.965996908809892,
"grad_norm": 0.1534372289046035,
"learning_rate": 9.82174092825181e-06,
"loss": 0.8982,
"step": 640
},
{
"epoch": 3.972179289026275,
"grad_norm": 0.15609090465249761,
"learning_rate": 9.708103595347352e-06,
"loss": 0.8939,
"step": 641
},
{
"epoch": 3.978361669242658,
"grad_norm": 0.1403869008769113,
"learning_rate": 9.595036621568398e-06,
"loss": 0.8843,
"step": 642
},
{
"epoch": 3.984544049459042,
"grad_norm": 0.14678540378586016,
"learning_rate": 9.482542135826916e-06,
"loss": 0.8987,
"step": 643
},
{
"epoch": 3.990726429675425,
"grad_norm": 0.1875416299280276,
"learning_rate": 9.370622256255571e-06,
"loss": 0.8774,
"step": 644
},
{
"epoch": 3.996908809891808,
"grad_norm": 0.27186472008456647,
"learning_rate": 9.259279090167994e-06,
"loss": 1.5557,
"step": 645
},
{
"epoch": 4.006182380216384,
"grad_norm": 0.20982320372706692,
"learning_rate": 9.148514734018917e-06,
"loss": 0.87,
"step": 646
},
{
"epoch": 4.012364760432766,
"grad_norm": 0.17703070468066032,
"learning_rate": 9.038331273364869e-06,
"loss": 0.8649,
"step": 647
},
{
"epoch": 4.01854714064915,
"grad_norm": 0.16674753883298887,
"learning_rate": 8.928730782824755e-06,
"loss": 0.8577,
"step": 648
},
{
"epoch": 4.024729520865534,
"grad_norm": 0.16839923962797768,
"learning_rate": 8.819715326040965e-06,
"loss": 0.8571,
"step": 649
},
{
"epoch": 4.030911901081916,
"grad_norm": 0.220038624506249,
"learning_rate": 8.711286955640332e-06,
"loss": 0.8609,
"step": 650
},
{
"epoch": 4.0370942812983,
"grad_norm": 0.16790422765840488,
"learning_rate": 8.603447713195634e-06,
"loss": 0.8566,
"step": 651
},
{
"epoch": 4.043276661514684,
"grad_norm": 0.18920607343960402,
"learning_rate": 8.496199629187072e-06,
"loss": 0.8608,
"step": 652
},
{
"epoch": 4.049459041731066,
"grad_norm": 0.18186173348032736,
"learning_rate": 8.389544722964076e-06,
"loss": 0.8675,
"step": 653
},
{
"epoch": 4.05564142194745,
"grad_norm": 0.17348169700628485,
"learning_rate": 8.283485002707233e-06,
"loss": 0.8689,
"step": 654
},
{
"epoch": 4.061823802163833,
"grad_norm": 0.17798870193571564,
"learning_rate": 8.178022465390549e-06,
"loss": 0.853,
"step": 655
},
{
"epoch": 4.068006182380216,
"grad_norm": 0.18943450530940084,
"learning_rate": 8.073159096743799e-06,
"loss": 0.8645,
"step": 656
},
{
"epoch": 4.0741885625966,
"grad_norm": 0.1602196742298431,
"learning_rate": 7.968896871215147e-06,
"loss": 0.8517,
"step": 657
},
{
"epoch": 4.080370942812983,
"grad_norm": 0.16728654932655415,
"learning_rate": 7.865237751933965e-06,
"loss": 0.858,
"step": 658
},
{
"epoch": 4.086553323029366,
"grad_norm": 0.17345900037449719,
"learning_rate": 7.762183690673887e-06,
"loss": 0.8552,
"step": 659
},
{
"epoch": 4.09273570324575,
"grad_norm": 0.16849302168626767,
"learning_rate": 7.659736627816064e-06,
"loss": 0.8559,
"step": 660
},
{
"epoch": 4.098918083462133,
"grad_norm": 0.15814956982575118,
"learning_rate": 7.557898492312561e-06,
"loss": 0.8527,
"step": 661
},
{
"epoch": 4.105100463678516,
"grad_norm": 0.15201462004812408,
"learning_rate": 7.456671201650145e-06,
"loss": 0.8551,
"step": 662
},
{
"epoch": 4.1112828438949,
"grad_norm": 0.13755795678813534,
"learning_rate": 7.356056661814106e-06,
"loss": 0.8557,
"step": 663
},
{
"epoch": 4.117465224111283,
"grad_norm": 0.15959755820070046,
"learning_rate": 7.256056767252402e-06,
"loss": 0.8539,
"step": 664
},
{
"epoch": 4.123647604327666,
"grad_norm": 0.1511791942647596,
"learning_rate": 7.156673400839933e-06,
"loss": 0.853,
"step": 665
},
{
"epoch": 4.12982998454405,
"grad_norm": 0.14354560331309524,
"learning_rate": 7.057908433843205e-06,
"loss": 0.8464,
"step": 666
},
{
"epoch": 4.1360123647604325,
"grad_norm": 0.1365948998014824,
"learning_rate": 6.959763725884956e-06,
"loss": 0.86,
"step": 667
},
{
"epoch": 4.142194744976816,
"grad_norm": 0.1745107521168772,
"learning_rate": 6.862241124909262e-06,
"loss": 0.8559,
"step": 668
},
{
"epoch": 4.1483771251932,
"grad_norm": 0.16007671239764104,
"learning_rate": 6.765342467146632e-06,
"loss": 0.8547,
"step": 669
},
{
"epoch": 4.1545595054095825,
"grad_norm": 0.14588280210721616,
"learning_rate": 6.66906957707957e-06,
"loss": 0.8563,
"step": 670
},
{
"epoch": 4.160741885625966,
"grad_norm": 0.13534918455746356,
"learning_rate": 6.5734242674080705e-06,
"loss": 0.8536,
"step": 671
},
{
"epoch": 4.166924265842349,
"grad_norm": 0.16010646794448433,
"learning_rate": 6.478408339015603e-06,
"loss": 0.8624,
"step": 672
},
{
"epoch": 4.1731066460587325,
"grad_norm": 0.22075894460772627,
"learning_rate": 6.384023580935155e-06,
"loss": 0.8588,
"step": 673
},
{
"epoch": 4.179289026275116,
"grad_norm": 0.17179022848077144,
"learning_rate": 6.290271770315541e-06,
"loss": 0.8497,
"step": 674
},
{
"epoch": 4.185471406491499,
"grad_norm": 0.14586862929974034,
"learning_rate": 6.1971546723879864e-06,
"loss": 0.859,
"step": 675
},
{
"epoch": 4.1916537867078825,
"grad_norm": 0.1629793349005496,
"learning_rate": 6.104674040432797e-06,
"loss": 0.8649,
"step": 676
},
{
"epoch": 4.197836166924266,
"grad_norm": 0.15123853435524942,
"learning_rate": 6.012831615746484e-06,
"loss": 0.8536,
"step": 677
},
{
"epoch": 4.204018547140649,
"grad_norm": 0.14020729925440908,
"learning_rate": 5.9216291276088435e-06,
"loss": 0.8633,
"step": 678
},
{
"epoch": 4.210200927357032,
"grad_norm": 0.1342216269142737,
"learning_rate": 5.831068293250481e-06,
"loss": 0.863,
"step": 679
},
{
"epoch": 4.216383307573416,
"grad_norm": 0.16252430886540703,
"learning_rate": 5.74115081782042e-06,
"loss": 0.8623,
"step": 680
},
{
"epoch": 4.222565687789799,
"grad_norm": 0.16349108022579686,
"learning_rate": 5.651878394354091e-06,
"loss": 0.8582,
"step": 681
},
{
"epoch": 4.228748068006182,
"grad_norm": 0.14338160366251965,
"learning_rate": 5.563252703741322e-06,
"loss": 0.8565,
"step": 682
},
{
"epoch": 4.234930448222566,
"grad_norm": 0.12199876601554722,
"learning_rate": 5.475275414694787e-06,
"loss": 0.8553,
"step": 683
},
{
"epoch": 4.241112828438949,
"grad_norm": 0.13894410880991992,
"learning_rate": 5.387948183718563e-06,
"loss": 0.8586,
"step": 684
},
{
"epoch": 4.247295208655332,
"grad_norm": 0.14581402965942716,
"learning_rate": 5.30127265507693e-06,
"loss": 0.8639,
"step": 685
},
{
"epoch": 4.253477588871716,
"grad_norm": 0.12813655780527142,
"learning_rate": 5.215250460763397e-06,
"loss": 0.845,
"step": 686
},
{
"epoch": 4.259659969088099,
"grad_norm": 0.1303992841804699,
"learning_rate": 5.129883220470007e-06,
"loss": 0.8509,
"step": 687
},
{
"epoch": 4.265842349304482,
"grad_norm": 0.1334156527459534,
"learning_rate": 5.045172541556831e-06,
"loss": 0.8522,
"step": 688
},
{
"epoch": 4.272024729520865,
"grad_norm": 0.12398532810622111,
"learning_rate": 4.961120019021684e-06,
"loss": 0.8647,
"step": 689
},
{
"epoch": 4.278207109737249,
"grad_norm": 0.12909526271671973,
"learning_rate": 4.877727235470113e-06,
"loss": 0.8586,
"step": 690
},
{
"epoch": 4.284389489953632,
"grad_norm": 0.128229338131557,
"learning_rate": 4.794995761085593e-06,
"loss": 0.8606,
"step": 691
},
{
"epoch": 4.290571870170015,
"grad_norm": 0.12394373173993793,
"learning_rate": 4.712927153599967e-06,
"loss": 0.8561,
"step": 692
},
{
"epoch": 4.296754250386399,
"grad_norm": 0.12226322639491397,
"learning_rate": 4.631522958264083e-06,
"loss": 0.8526,
"step": 693
},
{
"epoch": 4.302936630602782,
"grad_norm": 0.11913527209349921,
"learning_rate": 4.550784707818756e-06,
"loss": 0.8556,
"step": 694
},
{
"epoch": 4.309119010819165,
"grad_norm": 0.1459462503792828,
"learning_rate": 4.47071392246587e-06,
"loss": 0.8616,
"step": 695
},
{
"epoch": 4.315301391035549,
"grad_norm": 0.13037715558952065,
"learning_rate": 4.3913121098397675e-06,
"loss": 0.8595,
"step": 696
},
{
"epoch": 4.321483771251932,
"grad_norm": 0.11600080031698429,
"learning_rate": 4.312580764978825e-06,
"loss": 0.8472,
"step": 697
},
{
"epoch": 4.327666151468315,
"grad_norm": 0.12978365860295593,
"learning_rate": 4.234521370297398e-06,
"loss": 0.8652,
"step": 698
},
{
"epoch": 4.333848531684699,
"grad_norm": 0.13293173492859417,
"learning_rate": 4.157135395557786e-06,
"loss": 0.862,
"step": 699
},
{
"epoch": 4.340030911901082,
"grad_norm": 0.1159570605469207,
"learning_rate": 4.080424297842656e-06,
"loss": 0.8626,
"step": 700
},
{
"epoch": 4.346213292117465,
"grad_norm": 0.12754643596592866,
"learning_rate": 4.004389521527543e-06,
"loss": 0.8599,
"step": 701
},
{
"epoch": 4.352395672333849,
"grad_norm": 0.12442724249604767,
"learning_rate": 3.929032498253729e-06,
"loss": 0.8643,
"step": 702
},
{
"epoch": 4.358578052550232,
"grad_norm": 0.12301901657925675,
"learning_rate": 3.8543546469011904e-06,
"loss": 0.8684,
"step": 703
},
{
"epoch": 4.364760432766615,
"grad_norm": 0.11793079664110967,
"learning_rate": 3.780357373561958e-06,
"loss": 0.8554,
"step": 704
},
{
"epoch": 4.370942812982999,
"grad_norm": 0.11575437385032925,
"learning_rate": 3.7070420715136133e-06,
"loss": 0.8649,
"step": 705
},
{
"epoch": 4.377125193199381,
"grad_norm": 0.12884152614107805,
"learning_rate": 3.634410121193059e-06,
"loss": 0.8549,
"step": 706
},
{
"epoch": 4.383307573415765,
"grad_norm": 0.13882711090422645,
"learning_rate": 3.562462890170526e-06,
"loss": 0.848,
"step": 707
},
{
"epoch": 4.3894899536321486,
"grad_norm": 0.13110201135261,
"learning_rate": 3.4912017331238057e-06,
"loss": 0.8634,
"step": 708
},
{
"epoch": 4.395672333848531,
"grad_norm": 0.1192918433930348,
"learning_rate": 3.420627991812788e-06,
"loss": 0.8595,
"step": 709
},
{
"epoch": 4.401854714064915,
"grad_norm": 0.12360082779438417,
"learning_rate": 3.3507429950541527e-06,
"loss": 0.8526,
"step": 710
},
{
"epoch": 4.4080370942812985,
"grad_norm": 0.12305496739557657,
"learning_rate": 3.281548058696373e-06,
"loss": 0.8583,
"step": 711
},
{
"epoch": 4.414219474497681,
"grad_norm": 0.11759210587278442,
"learning_rate": 3.2130444855949406e-06,
"loss": 0.8671,
"step": 712
},
{
"epoch": 4.420401854714065,
"grad_norm": 0.12180525882355446,
"learning_rate": 3.145233565587824e-06,
"loss": 0.8616,
"step": 713
},
{
"epoch": 4.4265842349304485,
"grad_norm": 0.11198832899126798,
"learning_rate": 3.078116575471173e-06,
"loss": 0.8566,
"step": 714
},
{
"epoch": 4.432766615146831,
"grad_norm": 0.11314209890394392,
"learning_rate": 3.0116947789753028e-06,
"loss": 0.8581,
"step": 715
},
{
"epoch": 4.438948995363215,
"grad_norm": 0.11662833912400648,
"learning_rate": 2.9459694267408977e-06,
"loss": 0.8646,
"step": 716
},
{
"epoch": 4.4451313755795985,
"grad_norm": 0.11975312911303863,
"learning_rate": 2.8809417562954435e-06,
"loss": 0.8636,
"step": 717
},
{
"epoch": 4.451313755795981,
"grad_norm": 0.11940652053609146,
"learning_rate": 2.8166129920299278e-06,
"loss": 0.8474,
"step": 718
},
{
"epoch": 4.457496136012365,
"grad_norm": 0.11252115918580587,
"learning_rate": 2.752984345175809e-06,
"loss": 0.848,
"step": 719
},
{
"epoch": 4.4636785162287484,
"grad_norm": 0.10311476898498252,
"learning_rate": 2.690057013782195e-06,
"loss": 0.8599,
"step": 720
},
{
"epoch": 4.469860896445131,
"grad_norm": 0.11653845815565568,
"learning_rate": 2.6278321826932818e-06,
"loss": 0.8585,
"step": 721
},
{
"epoch": 4.476043276661515,
"grad_norm": 0.1098739721478331,
"learning_rate": 2.566311023526056e-06,
"loss": 0.8704,
"step": 722
},
{
"epoch": 4.4822256568778975,
"grad_norm": 0.11465020117380761,
"learning_rate": 2.5054946946482208e-06,
"loss": 0.8619,
"step": 723
},
{
"epoch": 4.488408037094281,
"grad_norm": 0.11796720870498818,
"learning_rate": 2.445384341156389e-06,
"loss": 0.8498,
"step": 724
},
{
"epoch": 4.494590417310665,
"grad_norm": 0.1166183962860216,
"learning_rate": 2.3859810948545414e-06,
"loss": 0.8567,
"step": 725
},
{
"epoch": 4.500772797527048,
"grad_norm": 0.11682256683745675,
"learning_rate": 2.3272860742326798e-06,
"loss": 0.8556,
"step": 726
},
{
"epoch": 4.506955177743431,
"grad_norm": 0.11434906715119132,
"learning_rate": 2.269300384445812e-06,
"loss": 0.8533,
"step": 727
},
{
"epoch": 4.513137557959815,
"grad_norm": 0.10735947434712247,
"learning_rate": 2.2120251172931082e-06,
"loss": 0.856,
"step": 728
},
{
"epoch": 4.5193199381761975,
"grad_norm": 0.10434949655119075,
"learning_rate": 2.15546135119733e-06,
"loss": 0.8583,
"step": 729
},
{
"epoch": 4.525502318392581,
"grad_norm": 0.11530305446016027,
"learning_rate": 2.0996101511846056e-06,
"loss": 0.8522,
"step": 730
},
{
"epoch": 4.531684698608965,
"grad_norm": 0.11525723139251923,
"learning_rate": 2.0444725688642685e-06,
"loss": 0.8537,
"step": 731
},
{
"epoch": 4.5378670788253475,
"grad_norm": 0.10816720987898426,
"learning_rate": 1.9900496424091375e-06,
"loss": 0.8552,
"step": 732
},
{
"epoch": 4.544049459041731,
"grad_norm": 0.10384961737476203,
"learning_rate": 1.9363423965359195e-06,
"loss": 0.8562,
"step": 733
},
{
"epoch": 4.550231839258115,
"grad_norm": 0.11336995410635603,
"learning_rate": 1.883351842485972e-06,
"loss": 0.8506,
"step": 734
},
{
"epoch": 4.556414219474497,
"grad_norm": 0.1092135826099414,
"learning_rate": 1.8310789780061887e-06,
"loss": 0.8589,
"step": 735
},
{
"epoch": 4.562596599690881,
"grad_norm": 0.10993015362867578,
"learning_rate": 1.7795247873302735e-06,
"loss": 0.8568,
"step": 736
},
{
"epoch": 4.568778979907265,
"grad_norm": 0.10666335057459857,
"learning_rate": 1.728690241160189e-06,
"loss": 0.8549,
"step": 737
},
{
"epoch": 4.574961360123647,
"grad_norm": 0.10922147280384506,
"learning_rate": 1.6785762966478715e-06,
"loss": 0.8616,
"step": 738
},
{
"epoch": 4.581143740340031,
"grad_norm": 0.10394964664273364,
"learning_rate": 1.6291838973772068e-06,
"loss": 0.85,
"step": 739
},
{
"epoch": 4.587326120556414,
"grad_norm": 0.10207519688840014,
"learning_rate": 1.5805139733462827e-06,
"loss": 0.8542,
"step": 740
},
{
"epoch": 4.593508500772797,
"grad_norm": 0.11068267423445127,
"learning_rate": 1.532567440949868e-06,
"loss": 0.863,
"step": 741
},
{
"epoch": 4.599690880989181,
"grad_norm": 0.10290817052985828,
"learning_rate": 1.4853452029621518e-06,
"loss": 0.8582,
"step": 742
},
{
"epoch": 4.605873261205565,
"grad_norm": 0.10564697594520893,
"learning_rate": 1.4388481485197558e-06,
"loss": 0.8666,
"step": 743
},
{
"epoch": 4.612055641421947,
"grad_norm": 0.11348466915609945,
"learning_rate": 1.3930771531049847e-06,
"loss": 0.8543,
"step": 744
},
{
"epoch": 4.618238021638331,
"grad_norm": 0.10419274079331056,
"learning_rate": 1.3480330785293494e-06,
"loss": 0.8605,
"step": 745
},
{
"epoch": 4.624420401854714,
"grad_norm": 0.10717824788246437,
"learning_rate": 1.3037167729173273e-06,
"loss": 0.8602,
"step": 746
},
{
"epoch": 4.630602782071097,
"grad_norm": 0.11005975127499093,
"learning_rate": 1.2601290706904102e-06,
"loss": 0.8612,
"step": 747
},
{
"epoch": 4.636785162287481,
"grad_norm": 0.11366201823377237,
"learning_rate": 1.2172707925513838e-06,
"loss": 0.8692,
"step": 748
},
{
"epoch": 4.642967542503864,
"grad_norm": 0.10308899025027428,
"learning_rate": 1.1751427454688735e-06,
"loss": 0.8739,
"step": 749
},
{
"epoch": 4.649149922720247,
"grad_norm": 0.10280506406459125,
"learning_rate": 1.1337457226621518e-06,
"loss": 0.8507,
"step": 750
},
{
"epoch": 4.655332302936631,
"grad_norm": 0.10187117757230453,
"learning_rate": 1.0930805035862125e-06,
"loss": 0.8688,
"step": 751
},
{
"epoch": 4.661514683153014,
"grad_norm": 0.10669750138958087,
"learning_rate": 1.0531478539170713e-06,
"loss": 0.8576,
"step": 752
},
{
"epoch": 4.667697063369397,
"grad_norm": 0.1101955728854345,
"learning_rate": 1.0139485255373826e-06,
"loss": 0.8586,
"step": 753
},
{
"epoch": 4.673879443585781,
"grad_norm": 0.1012943029959148,
"learning_rate": 9.75483256522236e-07,
"loss": 0.8732,
"step": 754
},
{
"epoch": 4.680061823802164,
"grad_norm": 0.10388171869974447,
"learning_rate": 9.377527711253198e-07,
"loss": 0.8501,
"step": 755
},
{
"epoch": 4.686244204018547,
"grad_norm": 0.10442678790389989,
"learning_rate": 9.007577797652245e-07,
"loss": 0.8606,
"step": 756
},
{
"epoch": 4.69242658423493,
"grad_norm": 0.11370097685694985,
"learning_rate": 8.644989790121072e-07,
"loss": 0.8687,
"step": 757
},
{
"epoch": 4.698608964451314,
"grad_norm": 0.10057604736658246,
"learning_rate": 8.289770515745599e-07,
"loss": 0.8623,
"step": 758
},
{
"epoch": 4.704791344667697,
"grad_norm": 0.10924197746964438,
"learning_rate": 7.941926662867528e-07,
"loss": 0.8663,
"step": 759
},
{
"epoch": 4.710973724884081,
"grad_norm": 0.10139852002580828,
"learning_rate": 7.60146478095849e-07,
"loss": 0.8533,
"step": 760
},
{
"epoch": 4.717156105100464,
"grad_norm": 0.10196842257139534,
"learning_rate": 7.268391280496589e-07,
"loss": 0.8545,
"step": 761
},
{
"epoch": 4.723338485316847,
"grad_norm": 0.09982941082955533,
"learning_rate": 6.94271243284601e-07,
"loss": 0.8593,
"step": 762
},
{
"epoch": 4.72952086553323,
"grad_norm": 0.1057680391701117,
"learning_rate": 6.624434370138532e-07,
"loss": 0.8637,
"step": 763
},
{
"epoch": 4.7357032457496135,
"grad_norm": 0.09876242839351415,
"learning_rate": 6.313563085158425e-07,
"loss": 0.8527,
"step": 764
},
{
"epoch": 4.741885625965997,
"grad_norm": 0.1029622819478206,
"learning_rate": 6.010104431229202e-07,
"loss": 0.8512,
"step": 765
},
{
"epoch": 4.74806800618238,
"grad_norm": 0.1017976721996803,
"learning_rate": 5.714064122103935e-07,
"loss": 0.8535,
"step": 766
},
{
"epoch": 4.7542503863987635,
"grad_norm": 0.10123837826010473,
"learning_rate": 5.425447731857248e-07,
"loss": 0.8563,
"step": 767
},
{
"epoch": 4.760432766615147,
"grad_norm": 0.10371103096875577,
"learning_rate": 5.144260694780512e-07,
"loss": 0.8558,
"step": 768
},
{
"epoch": 4.76661514683153,
"grad_norm": 0.10142002208178304,
"learning_rate": 4.870508305279531e-07,
"loss": 0.8574,
"step": 769
},
{
"epoch": 4.7727975270479135,
"grad_norm": 0.10174156389331632,
"learning_rate": 4.604195717774973e-07,
"loss": 0.8678,
"step": 770
},
{
"epoch": 4.778979907264297,
"grad_norm": 0.10158475219119292,
"learning_rate": 4.3453279466049383e-07,
"loss": 0.8652,
"step": 771
},
{
"epoch": 4.78516228748068,
"grad_norm": 0.1037484611064643,
"learning_rate": 4.0939098659309895e-07,
"loss": 0.8489,
"step": 772
},
{
"epoch": 4.7913446676970635,
"grad_norm": 0.10162493935552491,
"learning_rate": 3.849946209646138e-07,
"loss": 0.87,
"step": 773
},
{
"epoch": 4.797527047913446,
"grad_norm": 0.10099579812699412,
"learning_rate": 3.6134415712857587e-07,
"loss": 0.861,
"step": 774
},
{
"epoch": 4.80370942812983,
"grad_norm": 0.10320302386148672,
"learning_rate": 3.384400403941124e-07,
"loss": 0.8506,
"step": 775
},
{
"epoch": 4.809891808346213,
"grad_norm": 0.10386437696890127,
"learning_rate": 3.1628270201754743e-07,
"loss": 0.8498,
"step": 776
},
{
"epoch": 4.816074188562597,
"grad_norm": 0.09844682031008335,
"learning_rate": 2.948725591942925e-07,
"loss": 0.8537,
"step": 777
},
{
"epoch": 4.82225656877898,
"grad_norm": 0.10232392208050557,
"learning_rate": 2.742100150509819e-07,
"loss": 0.8668,
"step": 778
},
{
"epoch": 4.828438948995363,
"grad_norm": 0.09744721925527271,
"learning_rate": 2.542954586378921e-07,
"loss": 0.8657,
"step": 779
},
{
"epoch": 4.834621329211746,
"grad_norm": 0.10125314166249126,
"learning_rate": 2.351292649216097e-07,
"loss": 0.8631,
"step": 780
},
{
"epoch": 4.84080370942813,
"grad_norm": 0.09816843886551711,
"learning_rate": 2.167117947779751e-07,
"loss": 0.8657,
"step": 781
},
{
"epoch": 4.846986089644513,
"grad_norm": 0.10038235663832168,
"learning_rate": 1.990433949852788e-07,
"loss": 0.8666,
"step": 782
},
{
"epoch": 4.853168469860896,
"grad_norm": 0.09799020116935744,
"learning_rate": 1.821243982177423e-07,
"loss": 0.8635,
"step": 783
},
{
"epoch": 4.85935085007728,
"grad_norm": 0.10142283764104856,
"learning_rate": 1.6595512303925199e-07,
"loss": 0.8496,
"step": 784
},
{
"epoch": 4.865533230293663,
"grad_norm": 0.10199837045642687,
"learning_rate": 1.5053587389735502e-07,
"loss": 0.8764,
"step": 785
},
{
"epoch": 4.871715610510046,
"grad_norm": 0.10337397487378153,
"learning_rate": 1.358669411175395e-07,
"loss": 0.8548,
"step": 786
},
{
"epoch": 4.87789799072643,
"grad_norm": 0.09793020007211001,
"learning_rate": 1.2194860089774995e-07,
"loss": 0.8653,
"step": 787
},
{
"epoch": 4.884080370942813,
"grad_norm": 0.09878155037542055,
"learning_rate": 1.0878111530320478e-07,
"loss": 0.8623,
"step": 788
},
{
"epoch": 4.890262751159196,
"grad_norm": 0.10147729772509538,
"learning_rate": 9.636473226144916e-08,
"loss": 0.8614,
"step": 789
},
{
"epoch": 4.89644513137558,
"grad_norm": 0.09737787354641762,
"learning_rate": 8.469968555769648e-08,
"loss": 0.8647,
"step": 790
},
{
"epoch": 4.902627511591962,
"grad_norm": 0.09933111257188475,
"learning_rate": 7.37861948304186e-08,
"loss": 0.8482,
"step": 791
},
{
"epoch": 4.908809891808346,
"grad_norm": 0.09569900244500511,
"learning_rate": 6.362446556720691e-08,
"loss": 0.851,
"step": 792
},
{
"epoch": 4.91499227202473,
"grad_norm": 0.1031116966509426,
"learning_rate": 5.421468910092209e-08,
"loss": 0.8663,
"step": 793
},
{
"epoch": 4.921174652241113,
"grad_norm": 0.11035260250280782,
"learning_rate": 4.555704260607474e-08,
"loss": 0.8488,
"step": 794
},
{
"epoch": 4.927357032457496,
"grad_norm": 0.09621566038424581,
"learning_rate": 3.765168909548589e-08,
"loss": 0.8606,
"step": 795
},
{
"epoch": 4.93353941267388,
"grad_norm": 0.10155877681943523,
"learning_rate": 3.049877741723606e-08,
"loss": 0.8639,
"step": 796
},
{
"epoch": 4.939721792890262,
"grad_norm": 0.10009225600988104,
"learning_rate": 2.4098442251849762e-08,
"loss": 0.8559,
"step": 797
},
{
"epoch": 4.945904173106646,
"grad_norm": 0.10209104156755659,
"learning_rate": 1.8450804109759745e-08,
"loss": 0.8421,
"step": 798
},
{
"epoch": 4.95208655332303,
"grad_norm": 0.10570200426208885,
"learning_rate": 1.3555969329037688e-08,
"loss": 0.8565,
"step": 799
},
{
"epoch": 4.958268933539412,
"grad_norm": 0.09810578254731156,
"learning_rate": 9.41403007340025e-09,
"loss": 0.8638,
"step": 800
},
{
"epoch": 4.964451313755796,
"grad_norm": 0.1002862128277217,
"learning_rate": 6.025064330463792e-09,
"loss": 0.8553,
"step": 801
},
{
"epoch": 4.97063369397218,
"grad_norm": 0.09978110120418406,
"learning_rate": 3.3891359102877773e-09,
"loss": 0.8633,
"step": 802
},
{
"epoch": 4.976816074188562,
"grad_norm": 0.09867622221883632,
"learning_rate": 1.506294444153511e-09,
"loss": 0.8764,
"step": 803
},
{
"epoch": 4.982998454404946,
"grad_norm": 0.10760151986299173,
"learning_rate": 3.7657538364932467e-10,
"loss": 0.8589,
"step": 804
},
{
"epoch": 4.9891808346213296,
"grad_norm": 0.10304725530595252,
"learning_rate": 0.0,
"loss": 0.8693,
"step": 805
},
{
"epoch": 4.9891808346213296,
"step": 805,
"total_flos": 1.657473506467224e+19,
"train_loss": 0.9660356465333737,
"train_runtime": 42578.0212,
"train_samples_per_second": 9.723,
"train_steps_per_second": 0.019
}
],
"logging_steps": 1.0,
"max_steps": 805,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.657473506467224e+19,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}