marianna13's picture
Upload folder using huggingface_hub
8e186b7 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0,
"eval_steps": 500,
"global_step": 624,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.004807692307692308,
"grad_norm": 6.990724563598633,
"learning_rate": 1.5873015873015874e-07,
"loss": 1.0623,
"step": 1
},
{
"epoch": 0.009615384615384616,
"grad_norm": 6.919309139251709,
"learning_rate": 3.174603174603175e-07,
"loss": 1.0437,
"step": 2
},
{
"epoch": 0.014423076923076924,
"grad_norm": 6.894330978393555,
"learning_rate": 4.7619047619047623e-07,
"loss": 1.047,
"step": 3
},
{
"epoch": 0.019230769230769232,
"grad_norm": 6.92576265335083,
"learning_rate": 6.34920634920635e-07,
"loss": 1.042,
"step": 4
},
{
"epoch": 0.02403846153846154,
"grad_norm": 6.855586051940918,
"learning_rate": 7.936507936507937e-07,
"loss": 1.0384,
"step": 5
},
{
"epoch": 0.028846153846153848,
"grad_norm": 6.824655532836914,
"learning_rate": 9.523809523809525e-07,
"loss": 1.0474,
"step": 6
},
{
"epoch": 0.03365384615384615,
"grad_norm": 6.360400199890137,
"learning_rate": 1.111111111111111e-06,
"loss": 1.0354,
"step": 7
},
{
"epoch": 0.038461538461538464,
"grad_norm": 6.237353324890137,
"learning_rate": 1.26984126984127e-06,
"loss": 1.035,
"step": 8
},
{
"epoch": 0.04326923076923077,
"grad_norm": 4.962989330291748,
"learning_rate": 1.4285714285714286e-06,
"loss": 0.986,
"step": 9
},
{
"epoch": 0.04807692307692308,
"grad_norm": 4.90160608291626,
"learning_rate": 1.5873015873015873e-06,
"loss": 0.9925,
"step": 10
},
{
"epoch": 0.052884615384615384,
"grad_norm": 4.697347640991211,
"learning_rate": 1.746031746031746e-06,
"loss": 0.9938,
"step": 11
},
{
"epoch": 0.057692307692307696,
"grad_norm": 2.8192811012268066,
"learning_rate": 1.904761904761905e-06,
"loss": 0.9378,
"step": 12
},
{
"epoch": 0.0625,
"grad_norm": 2.878005027770996,
"learning_rate": 2.0634920634920634e-06,
"loss": 0.9478,
"step": 13
},
{
"epoch": 0.0673076923076923,
"grad_norm": 2.691462993621826,
"learning_rate": 2.222222222222222e-06,
"loss": 0.9575,
"step": 14
},
{
"epoch": 0.07211538461538461,
"grad_norm": 2.599451780319214,
"learning_rate": 2.380952380952381e-06,
"loss": 0.9319,
"step": 15
},
{
"epoch": 0.07692307692307693,
"grad_norm": 3.288738489151001,
"learning_rate": 2.53968253968254e-06,
"loss": 0.9198,
"step": 16
},
{
"epoch": 0.08173076923076923,
"grad_norm": 3.5585649013519287,
"learning_rate": 2.6984126984126986e-06,
"loss": 0.8854,
"step": 17
},
{
"epoch": 0.08653846153846154,
"grad_norm": 3.675098419189453,
"learning_rate": 2.8571428571428573e-06,
"loss": 0.9028,
"step": 18
},
{
"epoch": 0.09134615384615384,
"grad_norm": 3.505143642425537,
"learning_rate": 3.015873015873016e-06,
"loss": 0.9001,
"step": 19
},
{
"epoch": 0.09615384615384616,
"grad_norm": 3.0340216159820557,
"learning_rate": 3.1746031746031746e-06,
"loss": 0.8766,
"step": 20
},
{
"epoch": 0.10096153846153846,
"grad_norm": 2.239123821258545,
"learning_rate": 3.3333333333333333e-06,
"loss": 0.8773,
"step": 21
},
{
"epoch": 0.10576923076923077,
"grad_norm": 1.5766472816467285,
"learning_rate": 3.492063492063492e-06,
"loss": 0.8311,
"step": 22
},
{
"epoch": 0.11057692307692307,
"grad_norm": 1.567060112953186,
"learning_rate": 3.6507936507936507e-06,
"loss": 0.8303,
"step": 23
},
{
"epoch": 0.11538461538461539,
"grad_norm": 1.373935580253601,
"learning_rate": 3.80952380952381e-06,
"loss": 0.8011,
"step": 24
},
{
"epoch": 0.1201923076923077,
"grad_norm": 1.1732724905014038,
"learning_rate": 3.968253968253968e-06,
"loss": 0.7878,
"step": 25
},
{
"epoch": 0.125,
"grad_norm": 1.0626819133758545,
"learning_rate": 4.126984126984127e-06,
"loss": 0.78,
"step": 26
},
{
"epoch": 0.12980769230769232,
"grad_norm": 1.0781354904174805,
"learning_rate": 4.2857142857142855e-06,
"loss": 0.7852,
"step": 27
},
{
"epoch": 0.1346153846153846,
"grad_norm": 1.006543517112732,
"learning_rate": 4.444444444444444e-06,
"loss": 0.7954,
"step": 28
},
{
"epoch": 0.13942307692307693,
"grad_norm": 0.9152565598487854,
"learning_rate": 4.603174603174604e-06,
"loss": 0.7573,
"step": 29
},
{
"epoch": 0.14423076923076922,
"grad_norm": 0.8509926199913025,
"learning_rate": 4.761904761904762e-06,
"loss": 0.7572,
"step": 30
},
{
"epoch": 0.14903846153846154,
"grad_norm": 0.8057727217674255,
"learning_rate": 4.920634920634921e-06,
"loss": 0.7536,
"step": 31
},
{
"epoch": 0.15384615384615385,
"grad_norm": 0.8267956376075745,
"learning_rate": 5.07936507936508e-06,
"loss": 0.7441,
"step": 32
},
{
"epoch": 0.15865384615384615,
"grad_norm": 0.8982202410697937,
"learning_rate": 5.2380952380952384e-06,
"loss": 0.7398,
"step": 33
},
{
"epoch": 0.16346153846153846,
"grad_norm": 0.7668877840042114,
"learning_rate": 5.396825396825397e-06,
"loss": 0.7356,
"step": 34
},
{
"epoch": 0.16826923076923078,
"grad_norm": 0.775622546672821,
"learning_rate": 5.555555555555557e-06,
"loss": 0.722,
"step": 35
},
{
"epoch": 0.17307692307692307,
"grad_norm": 0.7095115780830383,
"learning_rate": 5.7142857142857145e-06,
"loss": 0.7061,
"step": 36
},
{
"epoch": 0.1778846153846154,
"grad_norm": 0.7553266882896423,
"learning_rate": 5.873015873015874e-06,
"loss": 0.7194,
"step": 37
},
{
"epoch": 0.18269230769230768,
"grad_norm": 0.7487539052963257,
"learning_rate": 6.031746031746032e-06,
"loss": 0.7101,
"step": 38
},
{
"epoch": 0.1875,
"grad_norm": 0.7720729112625122,
"learning_rate": 6.1904761904761914e-06,
"loss": 0.7186,
"step": 39
},
{
"epoch": 0.19230769230769232,
"grad_norm": 0.7665765881538391,
"learning_rate": 6.349206349206349e-06,
"loss": 0.7262,
"step": 40
},
{
"epoch": 0.1971153846153846,
"grad_norm": 0.8309563398361206,
"learning_rate": 6.507936507936509e-06,
"loss": 0.7232,
"step": 41
},
{
"epoch": 0.20192307692307693,
"grad_norm": 0.8082911372184753,
"learning_rate": 6.666666666666667e-06,
"loss": 0.7006,
"step": 42
},
{
"epoch": 0.20673076923076922,
"grad_norm": 0.6936203837394714,
"learning_rate": 6.825396825396826e-06,
"loss": 0.7036,
"step": 43
},
{
"epoch": 0.21153846153846154,
"grad_norm": 0.8248268365859985,
"learning_rate": 6.984126984126984e-06,
"loss": 0.7022,
"step": 44
},
{
"epoch": 0.21634615384615385,
"grad_norm": 0.7197801470756531,
"learning_rate": 7.1428571428571436e-06,
"loss": 0.6632,
"step": 45
},
{
"epoch": 0.22115384615384615,
"grad_norm": 0.5920101404190063,
"learning_rate": 7.301587301587301e-06,
"loss": 0.682,
"step": 46
},
{
"epoch": 0.22596153846153846,
"grad_norm": 0.6113859415054321,
"learning_rate": 7.460317460317461e-06,
"loss": 0.6929,
"step": 47
},
{
"epoch": 0.23076923076923078,
"grad_norm": 0.6709004640579224,
"learning_rate": 7.61904761904762e-06,
"loss": 0.6728,
"step": 48
},
{
"epoch": 0.23557692307692307,
"grad_norm": 0.6471027731895447,
"learning_rate": 7.77777777777778e-06,
"loss": 0.6722,
"step": 49
},
{
"epoch": 0.2403846153846154,
"grad_norm": 0.5812222361564636,
"learning_rate": 7.936507936507936e-06,
"loss": 0.6787,
"step": 50
},
{
"epoch": 0.24519230769230768,
"grad_norm": 0.7254865765571594,
"learning_rate": 8.095238095238097e-06,
"loss": 0.6809,
"step": 51
},
{
"epoch": 0.25,
"grad_norm": 0.5452060103416443,
"learning_rate": 8.253968253968254e-06,
"loss": 0.6658,
"step": 52
},
{
"epoch": 0.2548076923076923,
"grad_norm": 0.6729876399040222,
"learning_rate": 8.412698412698414e-06,
"loss": 0.6777,
"step": 53
},
{
"epoch": 0.25961538461538464,
"grad_norm": 0.5397999286651611,
"learning_rate": 8.571428571428571e-06,
"loss": 0.6713,
"step": 54
},
{
"epoch": 0.2644230769230769,
"grad_norm": 0.6134487390518188,
"learning_rate": 8.730158730158731e-06,
"loss": 0.6668,
"step": 55
},
{
"epoch": 0.2692307692307692,
"grad_norm": 0.5382984280586243,
"learning_rate": 8.888888888888888e-06,
"loss": 0.6697,
"step": 56
},
{
"epoch": 0.27403846153846156,
"grad_norm": 0.6409310102462769,
"learning_rate": 9.047619047619049e-06,
"loss": 0.6725,
"step": 57
},
{
"epoch": 0.27884615384615385,
"grad_norm": 0.4940313398838043,
"learning_rate": 9.206349206349207e-06,
"loss": 0.6738,
"step": 58
},
{
"epoch": 0.28365384615384615,
"grad_norm": 0.5982186198234558,
"learning_rate": 9.365079365079366e-06,
"loss": 0.6625,
"step": 59
},
{
"epoch": 0.28846153846153844,
"grad_norm": 0.6377073526382446,
"learning_rate": 9.523809523809525e-06,
"loss": 0.6605,
"step": 60
},
{
"epoch": 0.2932692307692308,
"grad_norm": 0.6730812191963196,
"learning_rate": 9.682539682539683e-06,
"loss": 0.6589,
"step": 61
},
{
"epoch": 0.2980769230769231,
"grad_norm": 0.7052279710769653,
"learning_rate": 9.841269841269842e-06,
"loss": 0.6504,
"step": 62
},
{
"epoch": 0.30288461538461536,
"grad_norm": 0.5276140570640564,
"learning_rate": 1e-05,
"loss": 0.6623,
"step": 63
},
{
"epoch": 0.3076923076923077,
"grad_norm": 0.7044535279273987,
"learning_rate": 9.99992160057155e-06,
"loss": 0.6383,
"step": 64
},
{
"epoch": 0.3125,
"grad_norm": 0.7368455529212952,
"learning_rate": 9.999686404744782e-06,
"loss": 0.6447,
"step": 65
},
{
"epoch": 0.3173076923076923,
"grad_norm": 0.7093029618263245,
"learning_rate": 9.999294419895389e-06,
"loss": 0.6673,
"step": 66
},
{
"epoch": 0.32211538461538464,
"grad_norm": 0.7653289437294006,
"learning_rate": 9.998745658315924e-06,
"loss": 0.6437,
"step": 67
},
{
"epoch": 0.3269230769230769,
"grad_norm": 0.6811711192131042,
"learning_rate": 9.998040137215423e-06,
"loss": 0.656,
"step": 68
},
{
"epoch": 0.3317307692307692,
"grad_norm": 0.6195499897003174,
"learning_rate": 9.99717787871887e-06,
"loss": 0.6473,
"step": 69
},
{
"epoch": 0.33653846153846156,
"grad_norm": 0.7382338047027588,
"learning_rate": 9.99615890986649e-06,
"loss": 0.6484,
"step": 70
},
{
"epoch": 0.34134615384615385,
"grad_norm": 0.7722101807594299,
"learning_rate": 9.994983262612916e-06,
"loss": 0.6508,
"step": 71
},
{
"epoch": 0.34615384615384615,
"grad_norm": 0.7530032396316528,
"learning_rate": 9.993650973826177e-06,
"loss": 0.6346,
"step": 72
},
{
"epoch": 0.35096153846153844,
"grad_norm": 0.7273198962211609,
"learning_rate": 9.992162085286543e-06,
"loss": 0.6456,
"step": 73
},
{
"epoch": 0.3557692307692308,
"grad_norm": 0.7155181765556335,
"learning_rate": 9.990516643685222e-06,
"loss": 0.6553,
"step": 74
},
{
"epoch": 0.3605769230769231,
"grad_norm": 0.7894483804702759,
"learning_rate": 9.988714700622882e-06,
"loss": 0.648,
"step": 75
},
{
"epoch": 0.36538461538461536,
"grad_norm": 0.6426685452461243,
"learning_rate": 9.986756312608048e-06,
"loss": 0.6498,
"step": 76
},
{
"epoch": 0.3701923076923077,
"grad_norm": 0.6765972375869751,
"learning_rate": 9.98464154105532e-06,
"loss": 0.6478,
"step": 77
},
{
"epoch": 0.375,
"grad_norm": 0.6547264456748962,
"learning_rate": 9.982370452283451e-06,
"loss": 0.628,
"step": 78
},
{
"epoch": 0.3798076923076923,
"grad_norm": 0.6896611452102661,
"learning_rate": 9.979943117513265e-06,
"loss": 0.6204,
"step": 79
},
{
"epoch": 0.38461538461538464,
"grad_norm": 0.7491808533668518,
"learning_rate": 9.977359612865424e-06,
"loss": 0.6376,
"step": 80
},
{
"epoch": 0.3894230769230769,
"grad_norm": 0.6002772450447083,
"learning_rate": 9.974620019358046e-06,
"loss": 0.6378,
"step": 81
},
{
"epoch": 0.3942307692307692,
"grad_norm": 0.7602054476737976,
"learning_rate": 9.971724422904154e-06,
"loss": 0.643,
"step": 82
},
{
"epoch": 0.39903846153846156,
"grad_norm": 0.5507690906524658,
"learning_rate": 9.968672914308995e-06,
"loss": 0.6332,
"step": 83
},
{
"epoch": 0.40384615384615385,
"grad_norm": 0.6860342025756836,
"learning_rate": 9.965465589267176e-06,
"loss": 0.6489,
"step": 84
},
{
"epoch": 0.40865384615384615,
"grad_norm": 0.5931215286254883,
"learning_rate": 9.96210254835968e-06,
"loss": 0.6425,
"step": 85
},
{
"epoch": 0.41346153846153844,
"grad_norm": 0.5528920292854309,
"learning_rate": 9.9585838970507e-06,
"loss": 0.6361,
"step": 86
},
{
"epoch": 0.4182692307692308,
"grad_norm": 0.6603521704673767,
"learning_rate": 9.954909745684339e-06,
"loss": 0.63,
"step": 87
},
{
"epoch": 0.4230769230769231,
"grad_norm": 0.6493014693260193,
"learning_rate": 9.951080209481138e-06,
"loss": 0.6187,
"step": 88
},
{
"epoch": 0.42788461538461536,
"grad_norm": 0.5607807040214539,
"learning_rate": 9.947095408534483e-06,
"loss": 0.6459,
"step": 89
},
{
"epoch": 0.4326923076923077,
"grad_norm": 0.6842950582504272,
"learning_rate": 9.94295546780682e-06,
"loss": 0.6315,
"step": 90
},
{
"epoch": 0.4375,
"grad_norm": 0.7010789513587952,
"learning_rate": 9.93866051712574e-06,
"loss": 0.6353,
"step": 91
},
{
"epoch": 0.4423076923076923,
"grad_norm": 0.573545515537262,
"learning_rate": 9.934210691179918e-06,
"loss": 0.6149,
"step": 92
},
{
"epoch": 0.44711538461538464,
"grad_norm": 0.5375520586967468,
"learning_rate": 9.929606129514875e-06,
"loss": 0.6227,
"step": 93
},
{
"epoch": 0.4519230769230769,
"grad_norm": 0.6587491035461426,
"learning_rate": 9.924846976528618e-06,
"loss": 0.6297,
"step": 94
},
{
"epoch": 0.4567307692307692,
"grad_norm": 0.6161583662033081,
"learning_rate": 9.919933381467088e-06,
"loss": 0.6421,
"step": 95
},
{
"epoch": 0.46153846153846156,
"grad_norm": 0.5526056885719299,
"learning_rate": 9.91486549841951e-06,
"loss": 0.6506,
"step": 96
},
{
"epoch": 0.46634615384615385,
"grad_norm": 0.6603134870529175,
"learning_rate": 9.909643486313533e-06,
"loss": 0.639,
"step": 97
},
{
"epoch": 0.47115384615384615,
"grad_norm": 0.5784550309181213,
"learning_rate": 9.904267508910269e-06,
"loss": 0.6307,
"step": 98
},
{
"epoch": 0.47596153846153844,
"grad_norm": 0.5557315945625305,
"learning_rate": 9.898737734799134e-06,
"loss": 0.6163,
"step": 99
},
{
"epoch": 0.4807692307692308,
"grad_norm": 0.6521722078323364,
"learning_rate": 9.89305433739258e-06,
"loss": 0.6352,
"step": 100
},
{
"epoch": 0.4855769230769231,
"grad_norm": 0.6033216118812561,
"learning_rate": 9.887217494920655e-06,
"loss": 0.6253,
"step": 101
},
{
"epoch": 0.49038461538461536,
"grad_norm": 0.5049378871917725,
"learning_rate": 9.881227390425404e-06,
"loss": 0.6315,
"step": 102
},
{
"epoch": 0.4951923076923077,
"grad_norm": 0.5810129046440125,
"learning_rate": 9.875084211755127e-06,
"loss": 0.6303,
"step": 103
},
{
"epoch": 0.5,
"grad_norm": 0.5270701050758362,
"learning_rate": 9.868788151558513e-06,
"loss": 0.6275,
"step": 104
},
{
"epoch": 0.5048076923076923,
"grad_norm": 0.5791574120521545,
"learning_rate": 9.862339407278564e-06,
"loss": 0.6364,
"step": 105
},
{
"epoch": 0.5096153846153846,
"grad_norm": 0.5724871158599854,
"learning_rate": 9.855738181146427e-06,
"loss": 0.6222,
"step": 106
},
{
"epoch": 0.5144230769230769,
"grad_norm": 0.5011044144630432,
"learning_rate": 9.848984680175049e-06,
"loss": 0.6216,
"step": 107
},
{
"epoch": 0.5192307692307693,
"grad_norm": 0.5154832601547241,
"learning_rate": 9.84207911615267e-06,
"loss": 0.6224,
"step": 108
},
{
"epoch": 0.5240384615384616,
"grad_norm": 0.5105087757110596,
"learning_rate": 9.835021705636201e-06,
"loss": 0.6259,
"step": 109
},
{
"epoch": 0.5288461538461539,
"grad_norm": 0.66715407371521,
"learning_rate": 9.827812669944423e-06,
"loss": 0.6245,
"step": 110
},
{
"epoch": 0.5336538461538461,
"grad_norm": 0.5680528879165649,
"learning_rate": 9.82045223515105e-06,
"loss": 0.6264,
"step": 111
},
{
"epoch": 0.5384615384615384,
"grad_norm": 0.6406980156898499,
"learning_rate": 9.812940632077629e-06,
"loss": 0.6216,
"step": 112
},
{
"epoch": 0.5432692307692307,
"grad_norm": 0.7648453116416931,
"learning_rate": 9.805278096286318e-06,
"loss": 0.6225,
"step": 113
},
{
"epoch": 0.5480769230769231,
"grad_norm": 0.6609866619110107,
"learning_rate": 9.797464868072489e-06,
"loss": 0.6414,
"step": 114
},
{
"epoch": 0.5528846153846154,
"grad_norm": 0.6097372174263,
"learning_rate": 9.789501192457188e-06,
"loss": 0.6301,
"step": 115
},
{
"epoch": 0.5576923076923077,
"grad_norm": 0.6364119648933411,
"learning_rate": 9.781387319179465e-06,
"loss": 0.6302,
"step": 116
},
{
"epoch": 0.5625,
"grad_norm": 0.5458793640136719,
"learning_rate": 9.773123502688532e-06,
"loss": 0.6119,
"step": 117
},
{
"epoch": 0.5673076923076923,
"grad_norm": 0.7278120517730713,
"learning_rate": 9.764710002135784e-06,
"loss": 0.6298,
"step": 118
},
{
"epoch": 0.5721153846153846,
"grad_norm": 0.5531219244003296,
"learning_rate": 9.756147081366673e-06,
"loss": 0.6174,
"step": 119
},
{
"epoch": 0.5769230769230769,
"grad_norm": 0.6038543581962585,
"learning_rate": 9.747435008912438e-06,
"loss": 0.6192,
"step": 120
},
{
"epoch": 0.5817307692307693,
"grad_norm": 0.8131069540977478,
"learning_rate": 9.73857405798168e-06,
"loss": 0.6198,
"step": 121
},
{
"epoch": 0.5865384615384616,
"grad_norm": 0.5778982043266296,
"learning_rate": 9.729564506451791e-06,
"loss": 0.6205,
"step": 122
},
{
"epoch": 0.5913461538461539,
"grad_norm": 0.842856764793396,
"learning_rate": 9.720406636860252e-06,
"loss": 0.6289,
"step": 123
},
{
"epoch": 0.5961538461538461,
"grad_norm": 0.6863536238670349,
"learning_rate": 9.711100736395758e-06,
"loss": 0.6168,
"step": 124
},
{
"epoch": 0.6009615384615384,
"grad_norm": 0.6230315566062927,
"learning_rate": 9.70164709688922e-06,
"loss": 0.6125,
"step": 125
},
{
"epoch": 0.6057692307692307,
"grad_norm": 0.788438618183136,
"learning_rate": 9.69204601480461e-06,
"loss": 0.6296,
"step": 126
},
{
"epoch": 0.6105769230769231,
"grad_norm": 0.6000380516052246,
"learning_rate": 9.682297791229668e-06,
"loss": 0.6296,
"step": 127
},
{
"epoch": 0.6153846153846154,
"grad_norm": 0.73893141746521,
"learning_rate": 9.67240273186646e-06,
"loss": 0.6083,
"step": 128
},
{
"epoch": 0.6201923076923077,
"grad_norm": 0.6741968393325806,
"learning_rate": 9.66236114702178e-06,
"loss": 0.6141,
"step": 129
},
{
"epoch": 0.625,
"grad_norm": 0.7334549427032471,
"learning_rate": 9.652173351597435e-06,
"loss": 0.6052,
"step": 130
},
{
"epoch": 0.6298076923076923,
"grad_norm": 0.6906297206878662,
"learning_rate": 9.641839665080363e-06,
"loss": 0.6196,
"step": 131
},
{
"epoch": 0.6346153846153846,
"grad_norm": 0.7519117593765259,
"learning_rate": 9.631360411532609e-06,
"loss": 0.6209,
"step": 132
},
{
"epoch": 0.6394230769230769,
"grad_norm": 0.6346924304962158,
"learning_rate": 9.620735919581168e-06,
"loss": 0.624,
"step": 133
},
{
"epoch": 0.6442307692307693,
"grad_norm": 0.6513266563415527,
"learning_rate": 9.609966522407678e-06,
"loss": 0.6294,
"step": 134
},
{
"epoch": 0.6490384615384616,
"grad_norm": 0.719011664390564,
"learning_rate": 9.599052557737973e-06,
"loss": 0.6323,
"step": 135
},
{
"epoch": 0.6538461538461539,
"grad_norm": 0.5536561012268066,
"learning_rate": 9.58799436783149e-06,
"loss": 0.6081,
"step": 136
},
{
"epoch": 0.6586538461538461,
"grad_norm": 0.6664816737174988,
"learning_rate": 9.576792299470537e-06,
"loss": 0.6234,
"step": 137
},
{
"epoch": 0.6634615384615384,
"grad_norm": 0.6060488820075989,
"learning_rate": 9.565446703949417e-06,
"loss": 0.609,
"step": 138
},
{
"epoch": 0.6682692307692307,
"grad_norm": 0.7735944390296936,
"learning_rate": 9.55395793706341e-06,
"loss": 0.609,
"step": 139
},
{
"epoch": 0.6730769230769231,
"grad_norm": 0.5819251537322998,
"learning_rate": 9.542326359097619e-06,
"loss": 0.6116,
"step": 140
},
{
"epoch": 0.6778846153846154,
"grad_norm": 0.8322558403015137,
"learning_rate": 9.530552334815672e-06,
"loss": 0.6071,
"step": 141
},
{
"epoch": 0.6826923076923077,
"grad_norm": 0.744692862033844,
"learning_rate": 9.518636233448276e-06,
"loss": 0.5932,
"step": 142
},
{
"epoch": 0.6875,
"grad_norm": 0.9699648022651672,
"learning_rate": 9.506578428681648e-06,
"loss": 0.6119,
"step": 143
},
{
"epoch": 0.6923076923076923,
"grad_norm": 0.621521532535553,
"learning_rate": 9.494379298645788e-06,
"loss": 0.6121,
"step": 144
},
{
"epoch": 0.6971153846153846,
"grad_norm": 1.0709632635116577,
"learning_rate": 9.482039225902623e-06,
"loss": 0.6101,
"step": 145
},
{
"epoch": 0.7019230769230769,
"grad_norm": 0.5138935446739197,
"learning_rate": 9.469558597434018e-06,
"loss": 0.6283,
"step": 146
},
{
"epoch": 0.7067307692307693,
"grad_norm": 0.8398244976997375,
"learning_rate": 9.456937804629623e-06,
"loss": 0.6,
"step": 147
},
{
"epoch": 0.7115384615384616,
"grad_norm": 0.7468331456184387,
"learning_rate": 9.444177243274619e-06,
"loss": 0.6124,
"step": 148
},
{
"epoch": 0.7163461538461539,
"grad_norm": 0.7130693793296814,
"learning_rate": 9.43127731353729e-06,
"loss": 0.6195,
"step": 149
},
{
"epoch": 0.7211538461538461,
"grad_norm": 0.766042172908783,
"learning_rate": 9.418238419956484e-06,
"loss": 0.6052,
"step": 150
},
{
"epoch": 0.7259615384615384,
"grad_norm": 0.8716906309127808,
"learning_rate": 9.405060971428924e-06,
"loss": 0.6209,
"step": 151
},
{
"epoch": 0.7307692307692307,
"grad_norm": 0.6835882663726807,
"learning_rate": 9.391745381196382e-06,
"loss": 0.6174,
"step": 152
},
{
"epoch": 0.7355769230769231,
"grad_norm": 0.7342336177825928,
"learning_rate": 9.378292066832723e-06,
"loss": 0.6243,
"step": 153
},
{
"epoch": 0.7403846153846154,
"grad_norm": 0.6497460007667542,
"learning_rate": 9.364701450230813e-06,
"loss": 0.6313,
"step": 154
},
{
"epoch": 0.7451923076923077,
"grad_norm": 0.6471742391586304,
"learning_rate": 9.350973957589278e-06,
"loss": 0.6124,
"step": 155
},
{
"epoch": 0.75,
"grad_norm": 0.6364895105361938,
"learning_rate": 9.33711001939915e-06,
"loss": 0.614,
"step": 156
},
{
"epoch": 0.7548076923076923,
"grad_norm": 0.6573180556297302,
"learning_rate": 9.32311007043036e-06,
"loss": 0.6309,
"step": 157
},
{
"epoch": 0.7596153846153846,
"grad_norm": 0.6328399777412415,
"learning_rate": 9.30897454971811e-06,
"loss": 0.6101,
"step": 158
},
{
"epoch": 0.7644230769230769,
"grad_norm": 0.6596732139587402,
"learning_rate": 9.294703900549096e-06,
"loss": 0.6244,
"step": 159
},
{
"epoch": 0.7692307692307693,
"grad_norm": 0.6779056787490845,
"learning_rate": 9.280298570447612e-06,
"loss": 0.6095,
"step": 160
},
{
"epoch": 0.7740384615384616,
"grad_norm": 0.6027641892433167,
"learning_rate": 9.265759011161519e-06,
"loss": 0.5958,
"step": 161
},
{
"epoch": 0.7788461538461539,
"grad_norm": 0.7004779577255249,
"learning_rate": 9.251085678648072e-06,
"loss": 0.6221,
"step": 162
},
{
"epoch": 0.7836538461538461,
"grad_norm": 0.7348073124885559,
"learning_rate": 9.236279033059622e-06,
"loss": 0.6226,
"step": 163
},
{
"epoch": 0.7884615384615384,
"grad_norm": 0.5637420415878296,
"learning_rate": 9.221339538729191e-06,
"loss": 0.601,
"step": 164
},
{
"epoch": 0.7932692307692307,
"grad_norm": 0.8568174839019775,
"learning_rate": 9.206267664155906e-06,
"loss": 0.6081,
"step": 165
},
{
"epoch": 0.7980769230769231,
"grad_norm": 0.7954565286636353,
"learning_rate": 9.191063881990308e-06,
"loss": 0.6032,
"step": 166
},
{
"epoch": 0.8028846153846154,
"grad_norm": 0.6611636877059937,
"learning_rate": 9.17572866901953e-06,
"loss": 0.6101,
"step": 167
},
{
"epoch": 0.8076923076923077,
"grad_norm": 0.9244260787963867,
"learning_rate": 9.160262506152343e-06,
"loss": 0.6015,
"step": 168
},
{
"epoch": 0.8125,
"grad_norm": 0.6074010133743286,
"learning_rate": 9.14466587840408e-06,
"loss": 0.6106,
"step": 169
},
{
"epoch": 0.8173076923076923,
"grad_norm": 0.6845681071281433,
"learning_rate": 9.12893927488142e-06,
"loss": 0.6012,
"step": 170
},
{
"epoch": 0.8221153846153846,
"grad_norm": 0.5826325416564941,
"learning_rate": 9.113083188767057e-06,
"loss": 0.6137,
"step": 171
},
{
"epoch": 0.8269230769230769,
"grad_norm": 0.5240793228149414,
"learning_rate": 9.097098117304223e-06,
"loss": 0.6131,
"step": 172
},
{
"epoch": 0.8317307692307693,
"grad_norm": 0.5843315124511719,
"learning_rate": 9.08098456178111e-06,
"loss": 0.6072,
"step": 173
},
{
"epoch": 0.8365384615384616,
"grad_norm": 0.5391135215759277,
"learning_rate": 9.064743027515127e-06,
"loss": 0.6068,
"step": 174
},
{
"epoch": 0.8413461538461539,
"grad_norm": 0.6380318403244019,
"learning_rate": 9.048374023837086e-06,
"loss": 0.6035,
"step": 175
},
{
"epoch": 0.8461538461538461,
"grad_norm": 0.5520167350769043,
"learning_rate": 9.03187806407519e-06,
"loss": 0.6131,
"step": 176
},
{
"epoch": 0.8509615384615384,
"grad_norm": 0.5716068744659424,
"learning_rate": 9.015255665538972e-06,
"loss": 0.6138,
"step": 177
},
{
"epoch": 0.8557692307692307,
"grad_norm": 0.598839282989502,
"learning_rate": 8.998507349503048e-06,
"loss": 0.5959,
"step": 178
},
{
"epoch": 0.8605769230769231,
"grad_norm": 0.6445143818855286,
"learning_rate": 8.981633641190779e-06,
"loss": 0.6076,
"step": 179
},
{
"epoch": 0.8653846153846154,
"grad_norm": 0.6508511304855347,
"learning_rate": 8.964635069757803e-06,
"loss": 0.6021,
"step": 180
},
{
"epoch": 0.8701923076923077,
"grad_norm": 0.6267576813697815,
"learning_rate": 8.94751216827543e-06,
"loss": 0.5993,
"step": 181
},
{
"epoch": 0.875,
"grad_norm": 0.6657792925834656,
"learning_rate": 8.930265473713939e-06,
"loss": 0.601,
"step": 182
},
{
"epoch": 0.8798076923076923,
"grad_norm": 0.6445341110229492,
"learning_rate": 8.912895526925726e-06,
"loss": 0.6018,
"step": 183
},
{
"epoch": 0.8846153846153846,
"grad_norm": 0.506033718585968,
"learning_rate": 8.895402872628352e-06,
"loss": 0.6029,
"step": 184
},
{
"epoch": 0.8894230769230769,
"grad_norm": 0.6721712350845337,
"learning_rate": 8.87778805938746e-06,
"loss": 0.5957,
"step": 185
},
{
"epoch": 0.8942307692307693,
"grad_norm": 0.5404098033905029,
"learning_rate": 8.86005163959956e-06,
"loss": 0.601,
"step": 186
},
{
"epoch": 0.8990384615384616,
"grad_norm": 0.6862823367118835,
"learning_rate": 8.842194169474727e-06,
"loss": 0.5961,
"step": 187
},
{
"epoch": 0.9038461538461539,
"grad_norm": 0.6001689434051514,
"learning_rate": 8.824216209019139e-06,
"loss": 0.6023,
"step": 188
},
{
"epoch": 0.9086538461538461,
"grad_norm": 0.5968487858772278,
"learning_rate": 8.806118322017525e-06,
"loss": 0.6087,
"step": 189
},
{
"epoch": 0.9134615384615384,
"grad_norm": 0.5125237703323364,
"learning_rate": 8.787901076015487e-06,
"loss": 0.5929,
"step": 190
},
{
"epoch": 0.9182692307692307,
"grad_norm": 0.7270171046257019,
"learning_rate": 8.769565042301692e-06,
"loss": 0.6209,
"step": 191
},
{
"epoch": 0.9230769230769231,
"grad_norm": 0.5782671570777893,
"learning_rate": 8.751110795889966e-06,
"loss": 0.5954,
"step": 192
},
{
"epoch": 0.9278846153846154,
"grad_norm": 0.5898118615150452,
"learning_rate": 8.732538915501257e-06,
"loss": 0.5993,
"step": 193
},
{
"epoch": 0.9326923076923077,
"grad_norm": 0.5937283635139465,
"learning_rate": 8.71384998354549e-06,
"loss": 0.614,
"step": 194
},
{
"epoch": 0.9375,
"grad_norm": 0.6111398935317993,
"learning_rate": 8.695044586103297e-06,
"loss": 0.605,
"step": 195
},
{
"epoch": 0.9423076923076923,
"grad_norm": 0.592433750629425,
"learning_rate": 8.676123312907641e-06,
"loss": 0.6137,
"step": 196
},
{
"epoch": 0.9471153846153846,
"grad_norm": 0.579230546951294,
"learning_rate": 8.657086757325328e-06,
"loss": 0.5918,
"step": 197
},
{
"epoch": 0.9519230769230769,
"grad_norm": 0.7127505540847778,
"learning_rate": 8.637935516338384e-06,
"loss": 0.6034,
"step": 198
},
{
"epoch": 0.9567307692307693,
"grad_norm": 0.6220248341560364,
"learning_rate": 8.61867019052535e-06,
"loss": 0.6043,
"step": 199
},
{
"epoch": 0.9615384615384616,
"grad_norm": 0.523986279964447,
"learning_rate": 8.599291384042442e-06,
"loss": 0.5896,
"step": 200
},
{
"epoch": 0.9663461538461539,
"grad_norm": 0.7416500449180603,
"learning_rate": 8.579799704604597e-06,
"loss": 0.5966,
"step": 201
},
{
"epoch": 0.9711538461538461,
"grad_norm": 0.6234787106513977,
"learning_rate": 8.560195763466428e-06,
"loss": 0.6022,
"step": 202
},
{
"epoch": 0.9759615384615384,
"grad_norm": 0.608817458152771,
"learning_rate": 8.540480175403045e-06,
"loss": 0.6001,
"step": 203
},
{
"epoch": 0.9807692307692307,
"grad_norm": 0.5969523191452026,
"learning_rate": 8.520653558690785e-06,
"loss": 0.6143,
"step": 204
},
{
"epoch": 0.9855769230769231,
"grad_norm": 0.6564033031463623,
"learning_rate": 8.500716535087815e-06,
"loss": 0.5862,
"step": 205
},
{
"epoch": 0.9903846153846154,
"grad_norm": 0.5331311821937561,
"learning_rate": 8.480669729814635e-06,
"loss": 0.6073,
"step": 206
},
{
"epoch": 0.9951923076923077,
"grad_norm": 0.5960302948951721,
"learning_rate": 8.460513771534475e-06,
"loss": 0.5938,
"step": 207
},
{
"epoch": 1.0,
"grad_norm": 0.5379106402397156,
"learning_rate": 8.440249292333583e-06,
"loss": 0.6125,
"step": 208
},
{
"epoch": 1.0048076923076923,
"grad_norm": 0.6571820378303528,
"learning_rate": 8.41987692770139e-06,
"loss": 0.5833,
"step": 209
},
{
"epoch": 1.0096153846153846,
"grad_norm": 0.5532863736152649,
"learning_rate": 8.399397316510596e-06,
"loss": 0.5801,
"step": 210
},
{
"epoch": 1.0144230769230769,
"grad_norm": 0.5147875547409058,
"learning_rate": 8.378811100997122e-06,
"loss": 0.5696,
"step": 211
},
{
"epoch": 1.0192307692307692,
"grad_norm": 0.5540103316307068,
"learning_rate": 8.358118926739984e-06,
"loss": 0.578,
"step": 212
},
{
"epoch": 1.0240384615384615,
"grad_norm": 0.5395882725715637,
"learning_rate": 8.337321442641036e-06,
"loss": 0.5836,
"step": 213
},
{
"epoch": 1.0288461538461537,
"grad_norm": 0.5106703639030457,
"learning_rate": 8.316419300904622e-06,
"loss": 0.5531,
"step": 214
},
{
"epoch": 1.0336538461538463,
"grad_norm": 0.570037841796875,
"learning_rate": 8.295413157017127e-06,
"loss": 0.5619,
"step": 215
},
{
"epoch": 1.0384615384615385,
"grad_norm": 0.5976242423057556,
"learning_rate": 8.274303669726427e-06,
"loss": 0.559,
"step": 216
},
{
"epoch": 1.0432692307692308,
"grad_norm": 0.5582316517829895,
"learning_rate": 8.25309150102121e-06,
"loss": 0.5656,
"step": 217
},
{
"epoch": 1.0480769230769231,
"grad_norm": 0.535399854183197,
"learning_rate": 8.231777316110245e-06,
"loss": 0.5519,
"step": 218
},
{
"epoch": 1.0528846153846154,
"grad_norm": 0.6076732277870178,
"learning_rate": 8.210361783401491e-06,
"loss": 0.5601,
"step": 219
},
{
"epoch": 1.0576923076923077,
"grad_norm": 0.5462629199028015,
"learning_rate": 8.188845574481162e-06,
"loss": 0.5777,
"step": 220
},
{
"epoch": 1.0625,
"grad_norm": 0.6692743301391602,
"learning_rate": 8.167229364092648e-06,
"loss": 0.5874,
"step": 221
},
{
"epoch": 1.0673076923076923,
"grad_norm": 0.5311057567596436,
"learning_rate": 8.145513830115367e-06,
"loss": 0.5846,
"step": 222
},
{
"epoch": 1.0721153846153846,
"grad_norm": 0.5106709599494934,
"learning_rate": 8.1236996535435e-06,
"loss": 0.5643,
"step": 223
},
{
"epoch": 1.0769230769230769,
"grad_norm": 0.5355270504951477,
"learning_rate": 8.101787518464634e-06,
"loss": 0.5621,
"step": 224
},
{
"epoch": 1.0817307692307692,
"grad_norm": 0.5128294825553894,
"learning_rate": 8.079778112038318e-06,
"loss": 0.5695,
"step": 225
},
{
"epoch": 1.0865384615384615,
"grad_norm": 0.5155346989631653,
"learning_rate": 8.057672124474508e-06,
"loss": 0.5589,
"step": 226
},
{
"epoch": 1.0913461538461537,
"grad_norm": 0.6576012372970581,
"learning_rate": 8.035470249011916e-06,
"loss": 0.5612,
"step": 227
},
{
"epoch": 1.0961538461538463,
"grad_norm": 0.5001193284988403,
"learning_rate": 8.013173181896283e-06,
"loss": 0.5701,
"step": 228
},
{
"epoch": 1.1009615384615385,
"grad_norm": 0.6337225437164307,
"learning_rate": 7.990781622358535e-06,
"loss": 0.558,
"step": 229
},
{
"epoch": 1.1057692307692308,
"grad_norm": 0.622628390789032,
"learning_rate": 7.968296272592862e-06,
"loss": 0.5489,
"step": 230
},
{
"epoch": 1.1105769230769231,
"grad_norm": 0.5139206051826477,
"learning_rate": 7.945717837734688e-06,
"loss": 0.544,
"step": 231
},
{
"epoch": 1.1153846153846154,
"grad_norm": 0.4990007281303406,
"learning_rate": 7.923047025838573e-06,
"loss": 0.5771,
"step": 232
},
{
"epoch": 1.1201923076923077,
"grad_norm": 0.43576645851135254,
"learning_rate": 7.900284547855992e-06,
"loss": 0.5668,
"step": 233
},
{
"epoch": 1.125,
"grad_norm": 0.4898163378238678,
"learning_rate": 7.87743111761305e-06,
"loss": 0.5659,
"step": 234
},
{
"epoch": 1.1298076923076923,
"grad_norm": 0.44757646322250366,
"learning_rate": 7.8544874517881e-06,
"loss": 0.5384,
"step": 235
},
{
"epoch": 1.1346153846153846,
"grad_norm": 0.5017294883728027,
"learning_rate": 7.831454269889251e-06,
"loss": 0.5636,
"step": 236
},
{
"epoch": 1.1394230769230769,
"grad_norm": 0.5470489263534546,
"learning_rate": 7.808332294231824e-06,
"loss": 0.5653,
"step": 237
},
{
"epoch": 1.1442307692307692,
"grad_norm": 0.5034846067428589,
"learning_rate": 7.785122249915688e-06,
"loss": 0.5591,
"step": 238
},
{
"epoch": 1.1490384615384615,
"grad_norm": 0.5333671569824219,
"learning_rate": 7.76182486480253e-06,
"loss": 0.5473,
"step": 239
},
{
"epoch": 1.1538461538461537,
"grad_norm": 0.5511570572853088,
"learning_rate": 7.738440869493018e-06,
"loss": 0.5551,
"step": 240
},
{
"epoch": 1.1586538461538463,
"grad_norm": 0.4971109628677368,
"learning_rate": 7.714970997303898e-06,
"loss": 0.5444,
"step": 241
},
{
"epoch": 1.1634615384615385,
"grad_norm": 0.5407963991165161,
"learning_rate": 7.691415984244998e-06,
"loss": 0.5681,
"step": 242
},
{
"epoch": 1.1682692307692308,
"grad_norm": 0.4940282106399536,
"learning_rate": 7.667776568996143e-06,
"loss": 0.5734,
"step": 243
},
{
"epoch": 1.1730769230769231,
"grad_norm": 0.46702679991722107,
"learning_rate": 7.64405349288399e-06,
"loss": 0.5527,
"step": 244
},
{
"epoch": 1.1778846153846154,
"grad_norm": 0.49713486433029175,
"learning_rate": 7.62024749985878e-06,
"loss": 0.5618,
"step": 245
},
{
"epoch": 1.1826923076923077,
"grad_norm": 0.5031074285507202,
"learning_rate": 7.596359336471015e-06,
"loss": 0.546,
"step": 246
},
{
"epoch": 1.1875,
"grad_norm": 0.5209451913833618,
"learning_rate": 7.572389751848037e-06,
"loss": 0.5599,
"step": 247
},
{
"epoch": 1.1923076923076923,
"grad_norm": 0.5612902045249939,
"learning_rate": 7.548339497670538e-06,
"loss": 0.5537,
"step": 248
},
{
"epoch": 1.1971153846153846,
"grad_norm": 0.595409631729126,
"learning_rate": 7.524209328148995e-06,
"loss": 0.5488,
"step": 249
},
{
"epoch": 1.2019230769230769,
"grad_norm": 0.5032594203948975,
"learning_rate": 7.500000000000001e-06,
"loss": 0.5835,
"step": 250
},
{
"epoch": 1.2067307692307692,
"grad_norm": 0.5198405385017395,
"learning_rate": 7.4757122724225575e-06,
"loss": 0.5783,
"step": 251
},
{
"epoch": 1.2115384615384615,
"grad_norm": 0.552277147769928,
"learning_rate": 7.451346907074245e-06,
"loss": 0.5609,
"step": 252
},
{
"epoch": 1.2163461538461537,
"grad_norm": 0.554760217666626,
"learning_rate": 7.426904668047352e-06,
"loss": 0.5678,
"step": 253
},
{
"epoch": 1.2211538461538463,
"grad_norm": 0.49330809712409973,
"learning_rate": 7.40238632184491e-06,
"loss": 0.5779,
"step": 254
},
{
"epoch": 1.2259615384615385,
"grad_norm": 0.6228109002113342,
"learning_rate": 7.377792637356644e-06,
"loss": 0.5496,
"step": 255
},
{
"epoch": 1.2307692307692308,
"grad_norm": 0.4928394854068756,
"learning_rate": 7.35312438583488e-06,
"loss": 0.5928,
"step": 256
},
{
"epoch": 1.2355769230769231,
"grad_norm": 0.5220669507980347,
"learning_rate": 7.3283823408703466e-06,
"loss": 0.5657,
"step": 257
},
{
"epoch": 1.2403846153846154,
"grad_norm": 0.5418768525123596,
"learning_rate": 7.303567278367918e-06,
"loss": 0.5493,
"step": 258
},
{
"epoch": 1.2451923076923077,
"grad_norm": 0.6059012413024902,
"learning_rate": 7.278679976522279e-06,
"loss": 0.5781,
"step": 259
},
{
"epoch": 1.25,
"grad_norm": 0.5703076720237732,
"learning_rate": 7.253721215793528e-06,
"loss": 0.5766,
"step": 260
},
{
"epoch": 1.2548076923076923,
"grad_norm": 0.7210759520530701,
"learning_rate": 7.2286917788826926e-06,
"loss": 0.5951,
"step": 261
},
{
"epoch": 1.2596153846153846,
"grad_norm": 0.49044269323349,
"learning_rate": 7.203592450707193e-06,
"loss": 0.5625,
"step": 262
},
{
"epoch": 1.2644230769230769,
"grad_norm": 0.5832895636558533,
"learning_rate": 7.178424018376224e-06,
"loss": 0.5594,
"step": 263
},
{
"epoch": 1.2692307692307692,
"grad_norm": 0.5973503589630127,
"learning_rate": 7.153187271166071e-06,
"loss": 0.5629,
"step": 264
},
{
"epoch": 1.2740384615384617,
"grad_norm": 0.5439810752868652,
"learning_rate": 7.127883000495353e-06,
"loss": 0.5548,
"step": 265
},
{
"epoch": 1.2788461538461537,
"grad_norm": 0.5828291177749634,
"learning_rate": 7.102511999900213e-06,
"loss": 0.5533,
"step": 266
},
{
"epoch": 1.2836538461538463,
"grad_norm": 0.601917028427124,
"learning_rate": 7.0770750650094335e-06,
"loss": 0.5673,
"step": 267
},
{
"epoch": 1.2884615384615383,
"grad_norm": 0.6008758544921875,
"learning_rate": 7.051572993519474e-06,
"loss": 0.557,
"step": 268
},
{
"epoch": 1.2932692307692308,
"grad_norm": 0.49223729968070984,
"learning_rate": 7.026006585169467e-06,
"loss": 0.5623,
"step": 269
},
{
"epoch": 1.2980769230769231,
"grad_norm": 0.6068645119667053,
"learning_rate": 7.0003766417161335e-06,
"loss": 0.5706,
"step": 270
},
{
"epoch": 1.3028846153846154,
"grad_norm": 0.5903732776641846,
"learning_rate": 6.974683966908642e-06,
"loss": 0.5625,
"step": 271
},
{
"epoch": 1.3076923076923077,
"grad_norm": 0.550208568572998,
"learning_rate": 6.948929366463397e-06,
"loss": 0.5873,
"step": 272
},
{
"epoch": 1.3125,
"grad_norm": 0.4981112480163574,
"learning_rate": 6.923113648038784e-06,
"loss": 0.5586,
"step": 273
},
{
"epoch": 1.3173076923076923,
"grad_norm": 0.48777687549591064,
"learning_rate": 6.897237621209831e-06,
"loss": 0.5635,
"step": 274
},
{
"epoch": 1.3221153846153846,
"grad_norm": 0.540177047252655,
"learning_rate": 6.87130209744282e-06,
"loss": 0.5619,
"step": 275
},
{
"epoch": 1.3269230769230769,
"grad_norm": 0.49552926421165466,
"learning_rate": 6.845307890069851e-06,
"loss": 0.5506,
"step": 276
},
{
"epoch": 1.3317307692307692,
"grad_norm": 0.6340379118919373,
"learning_rate": 6.8192558142633215e-06,
"loss": 0.5794,
"step": 277
},
{
"epoch": 1.3365384615384617,
"grad_norm": 0.4628427028656006,
"learning_rate": 6.7931466870103735e-06,
"loss": 0.5692,
"step": 278
},
{
"epoch": 1.3413461538461537,
"grad_norm": 0.5693328380584717,
"learning_rate": 6.766981327087271e-06,
"loss": 0.5617,
"step": 279
},
{
"epoch": 1.3461538461538463,
"grad_norm": 0.564184308052063,
"learning_rate": 6.740760555033715e-06,
"loss": 0.5501,
"step": 280
},
{
"epoch": 1.3509615384615383,
"grad_norm": 0.4685192406177521,
"learning_rate": 6.714485193127126e-06,
"loss": 0.5469,
"step": 281
},
{
"epoch": 1.3557692307692308,
"grad_norm": 0.5547739863395691,
"learning_rate": 6.688156065356845e-06,
"loss": 0.5787,
"step": 282
},
{
"epoch": 1.3605769230769231,
"grad_norm": 0.5170398354530334,
"learning_rate": 6.6617739973982985e-06,
"loss": 0.5526,
"step": 283
},
{
"epoch": 1.3653846153846154,
"grad_norm": 0.5516389608383179,
"learning_rate": 6.635339816587109e-06,
"loss": 0.5769,
"step": 284
},
{
"epoch": 1.3701923076923077,
"grad_norm": 0.6596104502677917,
"learning_rate": 6.60885435189314e-06,
"loss": 0.5687,
"step": 285
},
{
"epoch": 1.375,
"grad_norm": 0.5153890252113342,
"learning_rate": 6.582318433894513e-06,
"loss": 0.5623,
"step": 286
},
{
"epoch": 1.3798076923076923,
"grad_norm": 0.7115325927734375,
"learning_rate": 6.555732894751548e-06,
"loss": 0.5588,
"step": 287
},
{
"epoch": 1.3846153846153846,
"grad_norm": 0.5545114278793335,
"learning_rate": 6.529098568180672e-06,
"loss": 0.5605,
"step": 288
},
{
"epoch": 1.3894230769230769,
"grad_norm": 0.617314338684082,
"learning_rate": 6.502416289428282e-06,
"loss": 0.5768,
"step": 289
},
{
"epoch": 1.3942307692307692,
"grad_norm": 0.6576439142227173,
"learning_rate": 6.475686895244534e-06,
"loss": 0.5544,
"step": 290
},
{
"epoch": 1.3990384615384617,
"grad_norm": 0.5532417893409729,
"learning_rate": 6.448911223857124e-06,
"loss": 0.5457,
"step": 291
},
{
"epoch": 1.4038461538461537,
"grad_norm": 0.7050262689590454,
"learning_rate": 6.422090114944982e-06,
"loss": 0.5703,
"step": 292
},
{
"epoch": 1.4086538461538463,
"grad_norm": 0.5105507969856262,
"learning_rate": 6.3952244096119535e-06,
"loss": 0.5502,
"step": 293
},
{
"epoch": 1.4134615384615383,
"grad_norm": 0.6280075907707214,
"learning_rate": 6.368314950360416e-06,
"loss": 0.572,
"step": 294
},
{
"epoch": 1.4182692307692308,
"grad_norm": 0.5413298606872559,
"learning_rate": 6.341362581064856e-06,
"loss": 0.5599,
"step": 295
},
{
"epoch": 1.4230769230769231,
"grad_norm": 0.5309041738510132,
"learning_rate": 6.314368146945418e-06,
"loss": 0.5496,
"step": 296
},
{
"epoch": 1.4278846153846154,
"grad_norm": 0.585471510887146,
"learning_rate": 6.28733249454138e-06,
"loss": 0.5519,
"step": 297
},
{
"epoch": 1.4326923076923077,
"grad_norm": 0.4619200825691223,
"learning_rate": 6.260256471684622e-06,
"loss": 0.562,
"step": 298
},
{
"epoch": 1.4375,
"grad_norm": 0.568500816822052,
"learning_rate": 6.233140927473033e-06,
"loss": 0.5404,
"step": 299
},
{
"epoch": 1.4423076923076923,
"grad_norm": 0.5464410781860352,
"learning_rate": 6.205986712243876e-06,
"loss": 0.5724,
"step": 300
},
{
"epoch": 1.4471153846153846,
"grad_norm": 0.5567378401756287,
"learning_rate": 6.178794677547138e-06,
"loss": 0.5841,
"step": 301
},
{
"epoch": 1.4519230769230769,
"grad_norm": 0.5256141424179077,
"learning_rate": 6.151565676118805e-06,
"loss": 0.556,
"step": 302
},
{
"epoch": 1.4567307692307692,
"grad_norm": 0.5142210721969604,
"learning_rate": 6.124300561854139e-06,
"loss": 0.5667,
"step": 303
},
{
"epoch": 1.4615384615384617,
"grad_norm": 0.5487565398216248,
"learning_rate": 6.097000189780893e-06,
"loss": 0.5557,
"step": 304
},
{
"epoch": 1.4663461538461537,
"grad_norm": 0.5075297951698303,
"learning_rate": 6.0696654160324875e-06,
"loss": 0.5737,
"step": 305
},
{
"epoch": 1.4711538461538463,
"grad_norm": 0.5169580578804016,
"learning_rate": 6.042297097821184e-06,
"loss": 0.5487,
"step": 306
},
{
"epoch": 1.4759615384615383,
"grad_norm": 0.5064544677734375,
"learning_rate": 6.014896093411181e-06,
"loss": 0.5471,
"step": 307
},
{
"epoch": 1.4807692307692308,
"grad_norm": 0.4726231098175049,
"learning_rate": 5.987463262091715e-06,
"loss": 0.544,
"step": 308
},
{
"epoch": 1.4855769230769231,
"grad_norm": 0.5868953466415405,
"learning_rate": 5.959999464150101e-06,
"loss": 0.5579,
"step": 309
},
{
"epoch": 1.4903846153846154,
"grad_norm": 0.5180251598358154,
"learning_rate": 5.932505560844766e-06,
"loss": 0.5547,
"step": 310
},
{
"epoch": 1.4951923076923077,
"grad_norm": 0.5364190936088562,
"learning_rate": 5.904982414378233e-06,
"loss": 0.5838,
"step": 311
},
{
"epoch": 1.5,
"grad_norm": 0.5316298007965088,
"learning_rate": 5.877430887870081e-06,
"loss": 0.5821,
"step": 312
},
{
"epoch": 1.5048076923076923,
"grad_norm": 0.4852922558784485,
"learning_rate": 5.849851845329884e-06,
"loss": 0.5718,
"step": 313
},
{
"epoch": 1.5096153846153846,
"grad_norm": 0.4950406551361084,
"learning_rate": 5.822246151630109e-06,
"loss": 0.5475,
"step": 314
},
{
"epoch": 1.5144230769230769,
"grad_norm": 0.5721108317375183,
"learning_rate": 5.794614672479e-06,
"loss": 0.5682,
"step": 315
},
{
"epoch": 1.5192307692307692,
"grad_norm": 0.43430256843566895,
"learning_rate": 5.766958274393428e-06,
"loss": 0.5723,
"step": 316
},
{
"epoch": 1.5240384615384617,
"grad_norm": 0.5114768147468567,
"learning_rate": 5.739277824671711e-06,
"loss": 0.5724,
"step": 317
},
{
"epoch": 1.5288461538461537,
"grad_norm": 0.5058174133300781,
"learning_rate": 5.711574191366427e-06,
"loss": 0.5636,
"step": 318
},
{
"epoch": 1.5336538461538463,
"grad_norm": 0.4865401089191437,
"learning_rate": 5.683848243257181e-06,
"loss": 0.5444,
"step": 319
},
{
"epoch": 1.5384615384615383,
"grad_norm": 0.5734021067619324,
"learning_rate": 5.656100849823366e-06,
"loss": 0.5703,
"step": 320
},
{
"epoch": 1.5432692307692308,
"grad_norm": 0.4867899715900421,
"learning_rate": 5.628332881216899e-06,
"loss": 0.5728,
"step": 321
},
{
"epoch": 1.5480769230769231,
"grad_norm": 0.46812713146209717,
"learning_rate": 5.600545208234927e-06,
"loss": 0.5624,
"step": 322
},
{
"epoch": 1.5528846153846154,
"grad_norm": 0.5345550775527954,
"learning_rate": 5.57273870229252e-06,
"loss": 0.5487,
"step": 323
},
{
"epoch": 1.5576923076923077,
"grad_norm": 0.45916569232940674,
"learning_rate": 5.544914235395347e-06,
"loss": 0.551,
"step": 324
},
{
"epoch": 1.5625,
"grad_norm": 0.4594832956790924,
"learning_rate": 5.517072680112332e-06,
"loss": 0.5619,
"step": 325
},
{
"epoch": 1.5673076923076923,
"grad_norm": 0.5421361923217773,
"learning_rate": 5.4892149095482815e-06,
"loss": 0.5509,
"step": 326
},
{
"epoch": 1.5721153846153846,
"grad_norm": 0.4083757996559143,
"learning_rate": 5.46134179731651e-06,
"loss": 0.567,
"step": 327
},
{
"epoch": 1.5769230769230769,
"grad_norm": 0.4842531085014343,
"learning_rate": 5.4334542175114495e-06,
"loss": 0.5572,
"step": 328
},
{
"epoch": 1.5817307692307692,
"grad_norm": 0.5053375363349915,
"learning_rate": 5.40555304468122e-06,
"loss": 0.5653,
"step": 329
},
{
"epoch": 1.5865384615384617,
"grad_norm": 0.5320326685905457,
"learning_rate": 5.377639153800229e-06,
"loss": 0.5595,
"step": 330
},
{
"epoch": 1.5913461538461537,
"grad_norm": 0.502543568611145,
"learning_rate": 5.34971342024171e-06,
"loss": 0.572,
"step": 331
},
{
"epoch": 1.5961538461538463,
"grad_norm": 0.47758862376213074,
"learning_rate": 5.321776719750283e-06,
"loss": 0.5742,
"step": 332
},
{
"epoch": 1.6009615384615383,
"grad_norm": 0.49216747283935547,
"learning_rate": 5.29382992841449e-06,
"loss": 0.5618,
"step": 333
},
{
"epoch": 1.6057692307692308,
"grad_norm": 0.5157171487808228,
"learning_rate": 5.265873922639315e-06,
"loss": 0.5513,
"step": 334
},
{
"epoch": 1.6105769230769231,
"grad_norm": 0.48715752363204956,
"learning_rate": 5.237909579118713e-06,
"loss": 0.5626,
"step": 335
},
{
"epoch": 1.6153846153846154,
"grad_norm": 0.48291143774986267,
"learning_rate": 5.209937774808098e-06,
"loss": 0.559,
"step": 336
},
{
"epoch": 1.6201923076923077,
"grad_norm": 0.5301973819732666,
"learning_rate": 5.181959386896862e-06,
"loss": 0.5612,
"step": 337
},
{
"epoch": 1.625,
"grad_norm": 0.4594566226005554,
"learning_rate": 5.153975292780852e-06,
"loss": 0.5637,
"step": 338
},
{
"epoch": 1.6298076923076923,
"grad_norm": 0.4743099510669708,
"learning_rate": 5.125986370034862e-06,
"loss": 0.568,
"step": 339
},
{
"epoch": 1.6346153846153846,
"grad_norm": 0.46242398023605347,
"learning_rate": 5.097993496385112e-06,
"loss": 0.5612,
"step": 340
},
{
"epoch": 1.6394230769230769,
"grad_norm": 0.45778384804725647,
"learning_rate": 5.069997549681718e-06,
"loss": 0.5608,
"step": 341
},
{
"epoch": 1.6442307692307692,
"grad_norm": 0.47530800104141235,
"learning_rate": 5.041999407871168e-06,
"loss": 0.5687,
"step": 342
},
{
"epoch": 1.6490384615384617,
"grad_norm": 0.5340310335159302,
"learning_rate": 5.01399994896879e-06,
"loss": 0.5452,
"step": 343
},
{
"epoch": 1.6538461538461537,
"grad_norm": 0.43904221057891846,
"learning_rate": 4.986000051031212e-06,
"loss": 0.5535,
"step": 344
},
{
"epoch": 1.6586538461538463,
"grad_norm": 0.49031099677085876,
"learning_rate": 4.958000592128834e-06,
"loss": 0.534,
"step": 345
},
{
"epoch": 1.6634615384615383,
"grad_norm": 0.4739511013031006,
"learning_rate": 4.930002450318282e-06,
"loss": 0.5489,
"step": 346
},
{
"epoch": 1.6682692307692308,
"grad_norm": 0.4124240577220917,
"learning_rate": 4.9020065036148885e-06,
"loss": 0.5575,
"step": 347
},
{
"epoch": 1.6730769230769231,
"grad_norm": 0.4636637270450592,
"learning_rate": 4.874013629965138e-06,
"loss": 0.5614,
"step": 348
},
{
"epoch": 1.6778846153846154,
"grad_norm": 0.4664590060710907,
"learning_rate": 4.846024707219149e-06,
"loss": 0.5743,
"step": 349
},
{
"epoch": 1.6826923076923077,
"grad_norm": 0.41771262884140015,
"learning_rate": 4.818040613103139e-06,
"loss": 0.5555,
"step": 350
},
{
"epoch": 1.6875,
"grad_norm": 0.46983399987220764,
"learning_rate": 4.790062225191902e-06,
"loss": 0.5587,
"step": 351
},
{
"epoch": 1.6923076923076923,
"grad_norm": 0.4576813578605652,
"learning_rate": 4.762090420881289e-06,
"loss": 0.5577,
"step": 352
},
{
"epoch": 1.6971153846153846,
"grad_norm": 0.4811893403530121,
"learning_rate": 4.734126077360685e-06,
"loss": 0.5537,
"step": 353
},
{
"epoch": 1.7019230769230769,
"grad_norm": 0.4063519239425659,
"learning_rate": 4.706170071585513e-06,
"loss": 0.5662,
"step": 354
},
{
"epoch": 1.7067307692307692,
"grad_norm": 0.4614051580429077,
"learning_rate": 4.678223280249718e-06,
"loss": 0.5503,
"step": 355
},
{
"epoch": 1.7115384615384617,
"grad_norm": 0.43803513050079346,
"learning_rate": 4.650286579758291e-06,
"loss": 0.5652,
"step": 356
},
{
"epoch": 1.7163461538461537,
"grad_norm": 0.42544105648994446,
"learning_rate": 4.622360846199772e-06,
"loss": 0.5431,
"step": 357
},
{
"epoch": 1.7211538461538463,
"grad_norm": 0.39839082956314087,
"learning_rate": 4.594446955318781e-06,
"loss": 0.555,
"step": 358
},
{
"epoch": 1.7259615384615383,
"grad_norm": 0.38914695382118225,
"learning_rate": 4.566545782488554e-06,
"loss": 0.5623,
"step": 359
},
{
"epoch": 1.7307692307692308,
"grad_norm": 0.410859078168869,
"learning_rate": 4.53865820268349e-06,
"loss": 0.5626,
"step": 360
},
{
"epoch": 1.7355769230769231,
"grad_norm": 0.4289568364620209,
"learning_rate": 4.510785090451719e-06,
"loss": 0.5659,
"step": 361
},
{
"epoch": 1.7403846153846154,
"grad_norm": 0.4343273639678955,
"learning_rate": 4.482927319887669e-06,
"loss": 0.5543,
"step": 362
},
{
"epoch": 1.7451923076923077,
"grad_norm": 0.41860419511795044,
"learning_rate": 4.455085764604653e-06,
"loss": 0.5501,
"step": 363
},
{
"epoch": 1.75,
"grad_norm": 0.4254235625267029,
"learning_rate": 4.427261297707482e-06,
"loss": 0.5478,
"step": 364
},
{
"epoch": 1.7548076923076923,
"grad_norm": 0.420207142829895,
"learning_rate": 4.399454791765076e-06,
"loss": 0.5683,
"step": 365
},
{
"epoch": 1.7596153846153846,
"grad_norm": 0.4237710237503052,
"learning_rate": 4.371667118783101e-06,
"loss": 0.5551,
"step": 366
},
{
"epoch": 1.7644230769230769,
"grad_norm": 0.42474254965782166,
"learning_rate": 4.343899150176635e-06,
"loss": 0.5508,
"step": 367
},
{
"epoch": 1.7692307692307692,
"grad_norm": 0.4262840449810028,
"learning_rate": 4.316151756742821e-06,
"loss": 0.5545,
"step": 368
},
{
"epoch": 1.7740384615384617,
"grad_norm": 0.4713822603225708,
"learning_rate": 4.2884258086335755e-06,
"loss": 0.5632,
"step": 369
},
{
"epoch": 1.7788461538461537,
"grad_norm": 0.465378999710083,
"learning_rate": 4.26072217532829e-06,
"loss": 0.5706,
"step": 370
},
{
"epoch": 1.7836538461538463,
"grad_norm": 0.4259505271911621,
"learning_rate": 4.233041725606573e-06,
"loss": 0.5653,
"step": 371
},
{
"epoch": 1.7884615384615383,
"grad_norm": 0.4873887300491333,
"learning_rate": 4.205385327521002e-06,
"loss": 0.5547,
"step": 372
},
{
"epoch": 1.7932692307692308,
"grad_norm": 0.5444599390029907,
"learning_rate": 4.177753848369892e-06,
"loss": 0.5635,
"step": 373
},
{
"epoch": 1.7980769230769231,
"grad_norm": 0.4417326748371124,
"learning_rate": 4.1501481546701185e-06,
"loss": 0.5542,
"step": 374
},
{
"epoch": 1.8028846153846154,
"grad_norm": 0.42065057158470154,
"learning_rate": 4.12256911212992e-06,
"loss": 0.5712,
"step": 375
},
{
"epoch": 1.8076923076923077,
"grad_norm": 0.4802209436893463,
"learning_rate": 4.095017585621767e-06,
"loss": 0.5762,
"step": 376
},
{
"epoch": 1.8125,
"grad_norm": 0.47133877873420715,
"learning_rate": 4.067494439155236e-06,
"loss": 0.5706,
"step": 377
},
{
"epoch": 1.8173076923076923,
"grad_norm": 0.42150482535362244,
"learning_rate": 4.0400005358499e-06,
"loss": 0.5549,
"step": 378
},
{
"epoch": 1.8221153846153846,
"grad_norm": 0.48820599913597107,
"learning_rate": 4.012536737908288e-06,
"loss": 0.5646,
"step": 379
},
{
"epoch": 1.8269230769230769,
"grad_norm": 0.46589240431785583,
"learning_rate": 3.985103906588821e-06,
"loss": 0.5521,
"step": 380
},
{
"epoch": 1.8317307692307692,
"grad_norm": 0.43386465311050415,
"learning_rate": 3.957702902178816e-06,
"loss": 0.5603,
"step": 381
},
{
"epoch": 1.8365384615384617,
"grad_norm": 0.438164085149765,
"learning_rate": 3.930334583967514e-06,
"loss": 0.5547,
"step": 382
},
{
"epoch": 1.8413461538461537,
"grad_norm": 0.43723925948143005,
"learning_rate": 3.902999810219109e-06,
"loss": 0.5277,
"step": 383
},
{
"epoch": 1.8461538461538463,
"grad_norm": 0.4327050745487213,
"learning_rate": 3.875699438145862e-06,
"loss": 0.5426,
"step": 384
},
{
"epoch": 1.8509615384615383,
"grad_norm": 0.4460081160068512,
"learning_rate": 3.8484343238811976e-06,
"loss": 0.5478,
"step": 385
},
{
"epoch": 1.8557692307692308,
"grad_norm": 0.40871721506118774,
"learning_rate": 3.821205322452863e-06,
"loss": 0.5674,
"step": 386
},
{
"epoch": 1.8605769230769231,
"grad_norm": 0.39245539903640747,
"learning_rate": 3.794013287756125e-06,
"loss": 0.5587,
"step": 387
},
{
"epoch": 1.8653846153846154,
"grad_norm": 0.4072204828262329,
"learning_rate": 3.766859072526969e-06,
"loss": 0.5503,
"step": 388
},
{
"epoch": 1.8701923076923077,
"grad_norm": 0.4407781958580017,
"learning_rate": 3.7397435283153795e-06,
"loss": 0.5612,
"step": 389
},
{
"epoch": 1.875,
"grad_norm": 0.4392929673194885,
"learning_rate": 3.712667505458622e-06,
"loss": 0.5543,
"step": 390
},
{
"epoch": 1.8798076923076923,
"grad_norm": 0.4082246720790863,
"learning_rate": 3.685631853054583e-06,
"loss": 0.5384,
"step": 391
},
{
"epoch": 1.8846153846153846,
"grad_norm": 0.40940192341804504,
"learning_rate": 3.658637418935146e-06,
"loss": 0.5586,
"step": 392
},
{
"epoch": 1.8894230769230769,
"grad_norm": 0.4302951395511627,
"learning_rate": 3.6316850496395863e-06,
"loss": 0.5541,
"step": 393
},
{
"epoch": 1.8942307692307692,
"grad_norm": 0.472651869058609,
"learning_rate": 3.6047755903880478e-06,
"loss": 0.5523,
"step": 394
},
{
"epoch": 1.8990384615384617,
"grad_norm": 0.4774600863456726,
"learning_rate": 3.577909885055019e-06,
"loss": 0.5551,
"step": 395
},
{
"epoch": 1.9038461538461537,
"grad_norm": 0.40797290205955505,
"learning_rate": 3.5510887761428764e-06,
"loss": 0.5675,
"step": 396
},
{
"epoch": 1.9086538461538463,
"grad_norm": 0.4044354557991028,
"learning_rate": 3.524313104755468e-06,
"loss": 0.5804,
"step": 397
},
{
"epoch": 1.9134615384615383,
"grad_norm": 0.48422014713287354,
"learning_rate": 3.4975837105717203e-06,
"loss": 0.5551,
"step": 398
},
{
"epoch": 1.9182692307692308,
"grad_norm": 0.4386497735977173,
"learning_rate": 3.4709014318193298e-06,
"loss": 0.5637,
"step": 399
},
{
"epoch": 1.9230769230769231,
"grad_norm": 0.41975468397140503,
"learning_rate": 3.4442671052484545e-06,
"loss": 0.5714,
"step": 400
},
{
"epoch": 1.9278846153846154,
"grad_norm": 0.37439560890197754,
"learning_rate": 3.4176815661054884e-06,
"loss": 0.5422,
"step": 401
},
{
"epoch": 1.9326923076923077,
"grad_norm": 0.40816617012023926,
"learning_rate": 3.3911456481068613e-06,
"loss": 0.5438,
"step": 402
},
{
"epoch": 1.9375,
"grad_norm": 0.47742047905921936,
"learning_rate": 3.3646601834128924e-06,
"loss": 0.5584,
"step": 403
},
{
"epoch": 1.9423076923076923,
"grad_norm": 0.44102102518081665,
"learning_rate": 3.3382260026017027e-06,
"loss": 0.5481,
"step": 404
},
{
"epoch": 1.9471153846153846,
"grad_norm": 0.39150795340538025,
"learning_rate": 3.311843934643157e-06,
"loss": 0.5545,
"step": 405
},
{
"epoch": 1.9519230769230769,
"grad_norm": 0.4709937572479248,
"learning_rate": 3.2855148068728753e-06,
"loss": 0.5654,
"step": 406
},
{
"epoch": 1.9567307692307692,
"grad_norm": 0.4761818051338196,
"learning_rate": 3.2592394449662867e-06,
"loss": 0.5343,
"step": 407
},
{
"epoch": 1.9615384615384617,
"grad_norm": 0.44496825337409973,
"learning_rate": 3.233018672912731e-06,
"loss": 0.5527,
"step": 408
},
{
"epoch": 1.9663461538461537,
"grad_norm": 0.5426850318908691,
"learning_rate": 3.2068533129896273e-06,
"loss": 0.5473,
"step": 409
},
{
"epoch": 1.9711538461538463,
"grad_norm": 0.499978244304657,
"learning_rate": 3.1807441857366798e-06,
"loss": 0.5656,
"step": 410
},
{
"epoch": 1.9759615384615383,
"grad_norm": 0.41915661096572876,
"learning_rate": 3.1546921099301507e-06,
"loss": 0.5797,
"step": 411
},
{
"epoch": 1.9807692307692308,
"grad_norm": 0.4491116404533386,
"learning_rate": 3.1286979025571817e-06,
"loss": 0.5662,
"step": 412
},
{
"epoch": 1.9855769230769231,
"grad_norm": 0.42887210845947266,
"learning_rate": 3.1027623787901706e-06,
"loss": 0.5593,
"step": 413
},
{
"epoch": 1.9903846153846154,
"grad_norm": 0.445030152797699,
"learning_rate": 3.076886351961217e-06,
"loss": 0.547,
"step": 414
},
{
"epoch": 1.9951923076923077,
"grad_norm": 0.47931942343711853,
"learning_rate": 3.0510706335366034e-06,
"loss": 0.5684,
"step": 415
},
{
"epoch": 2.0,
"grad_norm": 0.4522424638271332,
"learning_rate": 3.02531603309136e-06,
"loss": 0.5519,
"step": 416
},
{
"epoch": 2.0048076923076925,
"grad_norm": 0.42797061800956726,
"learning_rate": 2.9996233582838686e-06,
"loss": 0.5399,
"step": 417
},
{
"epoch": 2.0096153846153846,
"grad_norm": 0.46265673637390137,
"learning_rate": 2.973993414830534e-06,
"loss": 0.5101,
"step": 418
},
{
"epoch": 2.014423076923077,
"grad_norm": 0.46875470876693726,
"learning_rate": 2.948427006480528e-06,
"loss": 0.5288,
"step": 419
},
{
"epoch": 2.019230769230769,
"grad_norm": 0.3783641755580902,
"learning_rate": 2.9229249349905686e-06,
"loss": 0.5254,
"step": 420
},
{
"epoch": 2.0240384615384617,
"grad_norm": 0.4915563464164734,
"learning_rate": 2.897488000099788e-06,
"loss": 0.5348,
"step": 421
},
{
"epoch": 2.0288461538461537,
"grad_norm": 0.46728116273880005,
"learning_rate": 2.8721169995046503e-06,
"loss": 0.5388,
"step": 422
},
{
"epoch": 2.0336538461538463,
"grad_norm": 0.4580628275871277,
"learning_rate": 2.846812728833931e-06,
"loss": 0.5255,
"step": 423
},
{
"epoch": 2.0384615384615383,
"grad_norm": 0.42225995659828186,
"learning_rate": 2.8215759816237748e-06,
"loss": 0.5206,
"step": 424
},
{
"epoch": 2.043269230769231,
"grad_norm": 0.4477675259113312,
"learning_rate": 2.796407549292809e-06,
"loss": 0.503,
"step": 425
},
{
"epoch": 2.048076923076923,
"grad_norm": 0.4319400489330292,
"learning_rate": 2.771308221117309e-06,
"loss": 0.5187,
"step": 426
},
{
"epoch": 2.0528846153846154,
"grad_norm": 0.4639637768268585,
"learning_rate": 2.7462787842064753e-06,
"loss": 0.5327,
"step": 427
},
{
"epoch": 2.0576923076923075,
"grad_norm": 0.3952256739139557,
"learning_rate": 2.7213200234777215e-06,
"loss": 0.5132,
"step": 428
},
{
"epoch": 2.0625,
"grad_norm": 0.38108691573143005,
"learning_rate": 2.696432721632082e-06,
"loss": 0.5536,
"step": 429
},
{
"epoch": 2.0673076923076925,
"grad_norm": 0.3831665515899658,
"learning_rate": 2.671617659129655e-06,
"loss": 0.5069,
"step": 430
},
{
"epoch": 2.0721153846153846,
"grad_norm": 0.38643160462379456,
"learning_rate": 2.646875614165121e-06,
"loss": 0.5222,
"step": 431
},
{
"epoch": 2.076923076923077,
"grad_norm": 0.40774914622306824,
"learning_rate": 2.6222073626433587e-06,
"loss": 0.5208,
"step": 432
},
{
"epoch": 2.081730769230769,
"grad_norm": 0.40251675248146057,
"learning_rate": 2.597613678155092e-06,
"loss": 0.513,
"step": 433
},
{
"epoch": 2.0865384615384617,
"grad_norm": 0.38830575346946716,
"learning_rate": 2.573095331952646e-06,
"loss": 0.5445,
"step": 434
},
{
"epoch": 2.0913461538461537,
"grad_norm": 0.4013608694076538,
"learning_rate": 2.5486530929257574e-06,
"loss": 0.5285,
"step": 435
},
{
"epoch": 2.0961538461538463,
"grad_norm": 0.3629082441329956,
"learning_rate": 2.5242877275774446e-06,
"loss": 0.532,
"step": 436
},
{
"epoch": 2.1009615384615383,
"grad_norm": 0.40173375606536865,
"learning_rate": 2.5000000000000015e-06,
"loss": 0.529,
"step": 437
},
{
"epoch": 2.105769230769231,
"grad_norm": 0.3868687152862549,
"learning_rate": 2.475790671851007e-06,
"loss": 0.5278,
"step": 438
},
{
"epoch": 2.110576923076923,
"grad_norm": 0.41604334115982056,
"learning_rate": 2.4516605023294626e-06,
"loss": 0.5284,
"step": 439
},
{
"epoch": 2.1153846153846154,
"grad_norm": 0.43302401900291443,
"learning_rate": 2.4276102481519655e-06,
"loss": 0.5286,
"step": 440
},
{
"epoch": 2.1201923076923075,
"grad_norm": 0.36202147603034973,
"learning_rate": 2.403640663528986e-06,
"loss": 0.5415,
"step": 441
},
{
"epoch": 2.125,
"grad_norm": 0.36743971705436707,
"learning_rate": 2.379752500141222e-06,
"loss": 0.5429,
"step": 442
},
{
"epoch": 2.1298076923076925,
"grad_norm": 0.4273446202278137,
"learning_rate": 2.355946507116012e-06,
"loss": 0.5329,
"step": 443
},
{
"epoch": 2.1346153846153846,
"grad_norm": 0.4194796681404114,
"learning_rate": 2.332223431003859e-06,
"loss": 0.547,
"step": 444
},
{
"epoch": 2.139423076923077,
"grad_norm": 0.3628264367580414,
"learning_rate": 2.3085840157550036e-06,
"loss": 0.5254,
"step": 445
},
{
"epoch": 2.144230769230769,
"grad_norm": 0.3705046772956848,
"learning_rate": 2.2850290026961032e-06,
"loss": 0.5359,
"step": 446
},
{
"epoch": 2.1490384615384617,
"grad_norm": 0.39662665128707886,
"learning_rate": 2.2615591305069846e-06,
"loss": 0.5121,
"step": 447
},
{
"epoch": 2.1538461538461537,
"grad_norm": 0.3908136785030365,
"learning_rate": 2.238175135197471e-06,
"loss": 0.5438,
"step": 448
},
{
"epoch": 2.1586538461538463,
"grad_norm": 0.36550912261009216,
"learning_rate": 2.2148777500843125e-06,
"loss": 0.5169,
"step": 449
},
{
"epoch": 2.1634615384615383,
"grad_norm": 0.39665716886520386,
"learning_rate": 2.1916677057681786e-06,
"loss": 0.4957,
"step": 450
},
{
"epoch": 2.168269230769231,
"grad_norm": 0.3526061475276947,
"learning_rate": 2.1685457301107506e-06,
"loss": 0.5355,
"step": 451
},
{
"epoch": 2.173076923076923,
"grad_norm": 0.3616856336593628,
"learning_rate": 2.145512548211902e-06,
"loss": 0.5136,
"step": 452
},
{
"epoch": 2.1778846153846154,
"grad_norm": 0.36202114820480347,
"learning_rate": 2.1225688823869494e-06,
"loss": 0.5107,
"step": 453
},
{
"epoch": 2.1826923076923075,
"grad_norm": 0.38488027453422546,
"learning_rate": 2.09971545214401e-06,
"loss": 0.5279,
"step": 454
},
{
"epoch": 2.1875,
"grad_norm": 0.38552579283714294,
"learning_rate": 2.0769529741614297e-06,
"loss": 0.5216,
"step": 455
},
{
"epoch": 2.1923076923076925,
"grad_norm": 0.40065112709999084,
"learning_rate": 2.054282162265313e-06,
"loss": 0.5092,
"step": 456
},
{
"epoch": 2.1971153846153846,
"grad_norm": 0.3517535924911499,
"learning_rate": 2.0317037274071412e-06,
"loss": 0.5175,
"step": 457
},
{
"epoch": 2.201923076923077,
"grad_norm": 0.3569267988204956,
"learning_rate": 2.009218377641466e-06,
"loss": 0.5361,
"step": 458
},
{
"epoch": 2.206730769230769,
"grad_norm": 0.3930048942565918,
"learning_rate": 1.9868268181037186e-06,
"loss": 0.5252,
"step": 459
},
{
"epoch": 2.2115384615384617,
"grad_norm": 0.37880074977874756,
"learning_rate": 1.964529750988086e-06,
"loss": 0.5251,
"step": 460
},
{
"epoch": 2.2163461538461537,
"grad_norm": 0.38495028018951416,
"learning_rate": 1.9423278755254933e-06,
"loss": 0.532,
"step": 461
},
{
"epoch": 2.2211538461538463,
"grad_norm": 0.3525002598762512,
"learning_rate": 1.9202218879616824e-06,
"loss": 0.5215,
"step": 462
},
{
"epoch": 2.2259615384615383,
"grad_norm": 0.3603316843509674,
"learning_rate": 1.8982124815353665e-06,
"loss": 0.5221,
"step": 463
},
{
"epoch": 2.230769230769231,
"grad_norm": 0.35701784491539,
"learning_rate": 1.8763003464565022e-06,
"loss": 0.5117,
"step": 464
},
{
"epoch": 2.235576923076923,
"grad_norm": 0.3503670394420624,
"learning_rate": 1.854486169884635e-06,
"loss": 0.5457,
"step": 465
},
{
"epoch": 2.2403846153846154,
"grad_norm": 0.3608977198600769,
"learning_rate": 1.8327706359073526e-06,
"loss": 0.5061,
"step": 466
},
{
"epoch": 2.2451923076923075,
"grad_norm": 0.3715517222881317,
"learning_rate": 1.8111544255188402e-06,
"loss": 0.5429,
"step": 467
},
{
"epoch": 2.25,
"grad_norm": 0.393916517496109,
"learning_rate": 1.7896382165985094e-06,
"loss": 0.5272,
"step": 468
},
{
"epoch": 2.2548076923076925,
"grad_norm": 0.36094385385513306,
"learning_rate": 1.768222683889757e-06,
"loss": 0.5299,
"step": 469
},
{
"epoch": 2.2596153846153846,
"grad_norm": 0.3601950407028198,
"learning_rate": 1.746908498978791e-06,
"loss": 0.5251,
"step": 470
},
{
"epoch": 2.264423076923077,
"grad_norm": 0.3479694426059723,
"learning_rate": 1.7256963302735752e-06,
"loss": 0.5398,
"step": 471
},
{
"epoch": 2.269230769230769,
"grad_norm": 0.3505624532699585,
"learning_rate": 1.7045868429828745e-06,
"loss": 0.4987,
"step": 472
},
{
"epoch": 2.2740384615384617,
"grad_norm": 0.3886091411113739,
"learning_rate": 1.6835806990953802e-06,
"loss": 0.5267,
"step": 473
},
{
"epoch": 2.2788461538461537,
"grad_norm": 0.37276989221572876,
"learning_rate": 1.6626785573589667e-06,
"loss": 0.5153,
"step": 474
},
{
"epoch": 2.2836538461538463,
"grad_norm": 0.3589704930782318,
"learning_rate": 1.6418810732600177e-06,
"loss": 0.513,
"step": 475
},
{
"epoch": 2.2884615384615383,
"grad_norm": 0.3904787302017212,
"learning_rate": 1.6211888990028785e-06,
"loss": 0.5202,
"step": 476
},
{
"epoch": 2.293269230769231,
"grad_norm": 0.3644522726535797,
"learning_rate": 1.6006026834894068e-06,
"loss": 0.521,
"step": 477
},
{
"epoch": 2.298076923076923,
"grad_norm": 0.3736514151096344,
"learning_rate": 1.5801230722986104e-06,
"loss": 0.5056,
"step": 478
},
{
"epoch": 2.3028846153846154,
"grad_norm": 0.35614803433418274,
"learning_rate": 1.5597507076664187e-06,
"loss": 0.5307,
"step": 479
},
{
"epoch": 2.3076923076923075,
"grad_norm": 0.35307708382606506,
"learning_rate": 1.5394862284655266e-06,
"loss": 0.5279,
"step": 480
},
{
"epoch": 2.3125,
"grad_norm": 0.37877628207206726,
"learning_rate": 1.5193302701853674e-06,
"loss": 0.5286,
"step": 481
},
{
"epoch": 2.3173076923076925,
"grad_norm": 0.35032036900520325,
"learning_rate": 1.499283464912188e-06,
"loss": 0.5401,
"step": 482
},
{
"epoch": 2.3221153846153846,
"grad_norm": 0.3594701886177063,
"learning_rate": 1.4793464413092161e-06,
"loss": 0.5239,
"step": 483
},
{
"epoch": 2.326923076923077,
"grad_norm": 0.35576295852661133,
"learning_rate": 1.459519824596956e-06,
"loss": 0.5325,
"step": 484
},
{
"epoch": 2.331730769230769,
"grad_norm": 0.38037362694740295,
"learning_rate": 1.4398042365335745e-06,
"loss": 0.5181,
"step": 485
},
{
"epoch": 2.3365384615384617,
"grad_norm": 0.3730925917625427,
"learning_rate": 1.4202002953954042e-06,
"loss": 0.5153,
"step": 486
},
{
"epoch": 2.3413461538461537,
"grad_norm": 0.3507210612297058,
"learning_rate": 1.4007086159575595e-06,
"loss": 0.5119,
"step": 487
},
{
"epoch": 2.3461538461538463,
"grad_norm": 0.3383399248123169,
"learning_rate": 1.3813298094746491e-06,
"loss": 0.5458,
"step": 488
},
{
"epoch": 2.3509615384615383,
"grad_norm": 0.34395211935043335,
"learning_rate": 1.362064483661617e-06,
"loss": 0.5318,
"step": 489
},
{
"epoch": 2.355769230769231,
"grad_norm": 0.3459123373031616,
"learning_rate": 1.3429132426746743e-06,
"loss": 0.5194,
"step": 490
},
{
"epoch": 2.360576923076923,
"grad_norm": 0.38007426261901855,
"learning_rate": 1.3238766870923592e-06,
"loss": 0.5311,
"step": 491
},
{
"epoch": 2.3653846153846154,
"grad_norm": 0.3795618414878845,
"learning_rate": 1.3049554138967052e-06,
"loss": 0.5362,
"step": 492
},
{
"epoch": 2.3701923076923075,
"grad_norm": 0.3444887101650238,
"learning_rate": 1.286150016454511e-06,
"loss": 0.5263,
"step": 493
},
{
"epoch": 2.375,
"grad_norm": 0.39823612570762634,
"learning_rate": 1.267461084498744e-06,
"loss": 0.5171,
"step": 494
},
{
"epoch": 2.3798076923076925,
"grad_norm": 0.3500751852989197,
"learning_rate": 1.2488892041100364e-06,
"loss": 0.5099,
"step": 495
},
{
"epoch": 2.3846153846153846,
"grad_norm": 0.34816107153892517,
"learning_rate": 1.2304349576983094e-06,
"loss": 0.5343,
"step": 496
},
{
"epoch": 2.389423076923077,
"grad_norm": 0.3754243552684784,
"learning_rate": 1.2120989239845149e-06,
"loss": 0.5265,
"step": 497
},
{
"epoch": 2.394230769230769,
"grad_norm": 0.32973939180374146,
"learning_rate": 1.1938816779824753e-06,
"loss": 0.5437,
"step": 498
},
{
"epoch": 2.3990384615384617,
"grad_norm": 0.37367406487464905,
"learning_rate": 1.1757837909808628e-06,
"loss": 0.5212,
"step": 499
},
{
"epoch": 2.4038461538461537,
"grad_norm": 0.34787219762802124,
"learning_rate": 1.157805830525275e-06,
"loss": 0.5225,
"step": 500
},
{
"epoch": 2.4086538461538463,
"grad_norm": 0.3542671799659729,
"learning_rate": 1.1399483604004403e-06,
"loss": 0.5301,
"step": 501
},
{
"epoch": 2.4134615384615383,
"grad_norm": 0.36626604199409485,
"learning_rate": 1.1222119406125426e-06,
"loss": 0.5234,
"step": 502
},
{
"epoch": 2.418269230769231,
"grad_norm": 0.3557561933994293,
"learning_rate": 1.1045971273716476e-06,
"loss": 0.5316,
"step": 503
},
{
"epoch": 2.423076923076923,
"grad_norm": 0.3680206835269928,
"learning_rate": 1.0871044730742752e-06,
"loss": 0.5261,
"step": 504
},
{
"epoch": 2.4278846153846154,
"grad_norm": 0.39090457558631897,
"learning_rate": 1.0697345262860638e-06,
"loss": 0.5279,
"step": 505
},
{
"epoch": 2.4326923076923075,
"grad_norm": 0.33717867732048035,
"learning_rate": 1.0524878317245713e-06,
"loss": 0.5218,
"step": 506
},
{
"epoch": 2.4375,
"grad_norm": 0.357800155878067,
"learning_rate": 1.0353649302421982e-06,
"loss": 0.5227,
"step": 507
},
{
"epoch": 2.4423076923076925,
"grad_norm": 0.34802302718162537,
"learning_rate": 1.0183663588092214e-06,
"loss": 0.5566,
"step": 508
},
{
"epoch": 2.4471153846153846,
"grad_norm": 0.3453739583492279,
"learning_rate": 1.0014926504969535e-06,
"loss": 0.5259,
"step": 509
},
{
"epoch": 2.451923076923077,
"grad_norm": 0.3763096034526825,
"learning_rate": 9.847443344610296e-07,
"loss": 0.5201,
"step": 510
},
{
"epoch": 2.456730769230769,
"grad_norm": 0.3687525689601898,
"learning_rate": 9.681219359248106e-07,
"loss": 0.5224,
"step": 511
},
{
"epoch": 2.4615384615384617,
"grad_norm": 0.3717356026172638,
"learning_rate": 9.516259761629148e-07,
"loss": 0.5145,
"step": 512
},
{
"epoch": 2.4663461538461537,
"grad_norm": 0.36599400639533997,
"learning_rate": 9.352569724848715e-07,
"loss": 0.5086,
"step": 513
},
{
"epoch": 2.4711538461538463,
"grad_norm": 0.33455225825309753,
"learning_rate": 9.190154382188921e-07,
"loss": 0.5148,
"step": 514
},
{
"epoch": 2.4759615384615383,
"grad_norm": 0.3153638243675232,
"learning_rate": 9.029018826957775e-07,
"loss": 0.5286,
"step": 515
},
{
"epoch": 2.480769230769231,
"grad_norm": 0.3458288609981537,
"learning_rate": 8.86916811232944e-07,
"loss": 0.5337,
"step": 516
},
{
"epoch": 2.485576923076923,
"grad_norm": 0.33229777216911316,
"learning_rate": 8.710607251185799e-07,
"loss": 0.5443,
"step": 517
},
{
"epoch": 2.4903846153846154,
"grad_norm": 0.3534226417541504,
"learning_rate": 8.553341215959215e-07,
"loss": 0.5391,
"step": 518
},
{
"epoch": 2.4951923076923075,
"grad_norm": 0.3250564634799957,
"learning_rate": 8.397374938476594e-07,
"loss": 0.5482,
"step": 519
},
{
"epoch": 2.5,
"grad_norm": 0.3450680375099182,
"learning_rate": 8.242713309804729e-07,
"loss": 0.5215,
"step": 520
},
{
"epoch": 2.5048076923076925,
"grad_norm": 0.3476369380950928,
"learning_rate": 8.089361180096927e-07,
"loss": 0.5308,
"step": 521
},
{
"epoch": 2.5096153846153846,
"grad_norm": 0.35948845744132996,
"learning_rate": 7.937323358440935e-07,
"loss": 0.5019,
"step": 522
},
{
"epoch": 2.5144230769230766,
"grad_norm": 0.3242650628089905,
"learning_rate": 7.786604612708093e-07,
"loss": 0.5317,
"step": 523
},
{
"epoch": 2.519230769230769,
"grad_norm": 0.3232422173023224,
"learning_rate": 7.637209669403789e-07,
"loss": 0.5165,
"step": 524
},
{
"epoch": 2.5240384615384617,
"grad_norm": 0.32376691699028015,
"learning_rate": 7.489143213519301e-07,
"loss": 0.5199,
"step": 525
},
{
"epoch": 2.5288461538461537,
"grad_norm": 0.3350430130958557,
"learning_rate": 7.342409888384816e-07,
"loss": 0.5087,
"step": 526
},
{
"epoch": 2.5336538461538463,
"grad_norm": 0.3153380751609802,
"learning_rate": 7.197014295523879e-07,
"loss": 0.5377,
"step": 527
},
{
"epoch": 2.5384615384615383,
"grad_norm": 0.36297881603240967,
"learning_rate": 7.052960994509056e-07,
"loss": 0.5238,
"step": 528
},
{
"epoch": 2.543269230769231,
"grad_norm": 0.34906795620918274,
"learning_rate": 6.910254502818914e-07,
"loss": 0.5119,
"step": 529
},
{
"epoch": 2.5480769230769234,
"grad_norm": 0.3447514474391937,
"learning_rate": 6.768899295696413e-07,
"loss": 0.5261,
"step": 530
},
{
"epoch": 2.5528846153846154,
"grad_norm": 0.3420708477497101,
"learning_rate": 6.628899806008515e-07,
"loss": 0.5201,
"step": 531
},
{
"epoch": 2.5576923076923075,
"grad_norm": 0.35410743951797485,
"learning_rate": 6.490260424107231e-07,
"loss": 0.5136,
"step": 532
},
{
"epoch": 2.5625,
"grad_norm": 0.3542673885822296,
"learning_rate": 6.352985497691883e-07,
"loss": 0.5212,
"step": 533
},
{
"epoch": 2.5673076923076925,
"grad_norm": 0.35093072056770325,
"learning_rate": 6.217079331672777e-07,
"loss": 0.5121,
"step": 534
},
{
"epoch": 2.5721153846153846,
"grad_norm": 0.33335307240486145,
"learning_rate": 6.082546188036204e-07,
"loss": 0.5007,
"step": 535
},
{
"epoch": 2.5769230769230766,
"grad_norm": 0.3272242546081543,
"learning_rate": 5.949390285710777e-07,
"loss": 0.5443,
"step": 536
},
{
"epoch": 2.581730769230769,
"grad_norm": 0.34555184841156006,
"learning_rate": 5.817615800435167e-07,
"loss": 0.5255,
"step": 537
},
{
"epoch": 2.5865384615384617,
"grad_norm": 0.3510143458843231,
"learning_rate": 5.687226864627115e-07,
"loss": 0.5238,
"step": 538
},
{
"epoch": 2.5913461538461537,
"grad_norm": 0.35968631505966187,
"learning_rate": 5.558227567253832e-07,
"loss": 0.5186,
"step": 539
},
{
"epoch": 2.5961538461538463,
"grad_norm": 0.35358065366744995,
"learning_rate": 5.430621953703785e-07,
"loss": 0.5132,
"step": 540
},
{
"epoch": 2.6009615384615383,
"grad_norm": 0.34705501794815063,
"learning_rate": 5.304414025659832e-07,
"loss": 0.5418,
"step": 541
},
{
"epoch": 2.605769230769231,
"grad_norm": 0.34814560413360596,
"learning_rate": 5.179607740973764e-07,
"loss": 0.5217,
"step": 542
},
{
"epoch": 2.6105769230769234,
"grad_norm": 0.32552292943000793,
"learning_rate": 5.056207013542131e-07,
"loss": 0.5224,
"step": 543
},
{
"epoch": 2.6153846153846154,
"grad_norm": 0.3452615737915039,
"learning_rate": 4.934215713183527e-07,
"loss": 0.5226,
"step": 544
},
{
"epoch": 2.6201923076923075,
"grad_norm": 0.3437098264694214,
"learning_rate": 4.813637665517251e-07,
"loss": 0.5258,
"step": 545
},
{
"epoch": 2.625,
"grad_norm": 0.36558979749679565,
"learning_rate": 4.6944766518432936e-07,
"loss": 0.5137,
"step": 546
},
{
"epoch": 2.6298076923076925,
"grad_norm": 0.33913153409957886,
"learning_rate": 4.576736409023813e-07,
"loss": 0.508,
"step": 547
},
{
"epoch": 2.6346153846153846,
"grad_norm": 0.35266298055648804,
"learning_rate": 4.460420629365919e-07,
"loss": 0.5167,
"step": 548
},
{
"epoch": 2.6394230769230766,
"grad_norm": 0.3532564640045166,
"learning_rate": 4.3455329605058436e-07,
"loss": 0.5145,
"step": 549
},
{
"epoch": 2.644230769230769,
"grad_norm": 0.3402523994445801,
"learning_rate": 4.232077005294638e-07,
"loss": 0.5237,
"step": 550
},
{
"epoch": 2.6490384615384617,
"grad_norm": 0.3120058476924896,
"learning_rate": 4.120056321685101e-07,
"loss": 0.562,
"step": 551
},
{
"epoch": 2.6538461538461537,
"grad_norm": 0.33833712339401245,
"learning_rate": 4.009474422620269e-07,
"loss": 0.5257,
"step": 552
},
{
"epoch": 2.6586538461538463,
"grad_norm": 0.333404004573822,
"learning_rate": 3.900334775923237e-07,
"loss": 0.5115,
"step": 553
},
{
"epoch": 2.6634615384615383,
"grad_norm": 0.32522115111351013,
"learning_rate": 3.7926408041883355e-07,
"loss": 0.5286,
"step": 554
},
{
"epoch": 2.668269230769231,
"grad_norm": 0.3394847810268402,
"learning_rate": 3.6863958846739213e-07,
"loss": 0.5195,
"step": 555
},
{
"epoch": 2.6730769230769234,
"grad_norm": 0.3137637674808502,
"learning_rate": 3.581603349196372e-07,
"loss": 0.5339,
"step": 556
},
{
"epoch": 2.6778846153846154,
"grad_norm": 0.35012125968933105,
"learning_rate": 3.4782664840256387e-07,
"loss": 0.5054,
"step": 557
},
{
"epoch": 2.6826923076923075,
"grad_norm": 0.3344334065914154,
"learning_rate": 3.3763885297822153e-07,
"loss": 0.5303,
"step": 558
},
{
"epoch": 2.6875,
"grad_norm": 0.3342319428920746,
"learning_rate": 3.275972681335421e-07,
"loss": 0.512,
"step": 559
},
{
"epoch": 2.6923076923076925,
"grad_norm": 0.32932013273239136,
"learning_rate": 3.1770220877033243e-07,
"loss": 0.5354,
"step": 560
},
{
"epoch": 2.6971153846153846,
"grad_norm": 0.34048202633857727,
"learning_rate": 3.0795398519539113e-07,
"loss": 0.5216,
"step": 561
},
{
"epoch": 2.7019230769230766,
"grad_norm": 0.3443983495235443,
"learning_rate": 2.9835290311078123e-07,
"loss": 0.5131,
"step": 562
},
{
"epoch": 2.706730769230769,
"grad_norm": 0.34546706080436707,
"learning_rate": 2.888992636042437e-07,
"loss": 0.4995,
"step": 563
},
{
"epoch": 2.7115384615384617,
"grad_norm": 0.32637590169906616,
"learning_rate": 2.7959336313974847e-07,
"loss": 0.524,
"step": 564
},
{
"epoch": 2.7163461538461537,
"grad_norm": 0.32179778814315796,
"learning_rate": 2.704354935482095e-07,
"loss": 0.5369,
"step": 565
},
{
"epoch": 2.7211538461538463,
"grad_norm": 0.3366025388240814,
"learning_rate": 2.6142594201832183e-07,
"loss": 0.5021,
"step": 566
},
{
"epoch": 2.7259615384615383,
"grad_norm": 0.3441389501094818,
"learning_rate": 2.525649910875627e-07,
"loss": 0.522,
"step": 567
},
{
"epoch": 2.730769230769231,
"grad_norm": 0.3519365191459656,
"learning_rate": 2.438529186333288e-07,
"loss": 0.5234,
"step": 568
},
{
"epoch": 2.7355769230769234,
"grad_norm": 0.330476313829422,
"learning_rate": 2.3528999786421758e-07,
"loss": 0.5309,
"step": 569
},
{
"epoch": 2.7403846153846154,
"grad_norm": 0.3406146466732025,
"learning_rate": 2.2687649731146844e-07,
"loss": 0.5279,
"step": 570
},
{
"epoch": 2.7451923076923075,
"grad_norm": 0.3351764678955078,
"learning_rate": 2.1861268082053466e-07,
"loss": 0.5275,
"step": 571
},
{
"epoch": 2.75,
"grad_norm": 0.37317317724227905,
"learning_rate": 2.104988075428127e-07,
"loss": 0.5285,
"step": 572
},
{
"epoch": 2.7548076923076925,
"grad_norm": 0.3595113158226013,
"learning_rate": 2.0253513192751374e-07,
"loss": 0.5287,
"step": 573
},
{
"epoch": 2.7596153846153846,
"grad_norm": 0.33269202709198,
"learning_rate": 1.947219037136827e-07,
"loss": 0.5151,
"step": 574
},
{
"epoch": 2.7644230769230766,
"grad_norm": 0.341153085231781,
"learning_rate": 1.8705936792237255e-07,
"loss": 0.5243,
"step": 575
},
{
"epoch": 2.769230769230769,
"grad_norm": 0.3603745102882385,
"learning_rate": 1.7954776484895188e-07,
"loss": 0.5228,
"step": 576
},
{
"epoch": 2.7740384615384617,
"grad_norm": 0.3396835923194885,
"learning_rate": 1.7218733005557707e-07,
"loss": 0.5073,
"step": 577
},
{
"epoch": 2.7788461538461537,
"grad_norm": 0.3207492530345917,
"learning_rate": 1.6497829436380009e-07,
"loss": 0.5266,
"step": 578
},
{
"epoch": 2.7836538461538463,
"grad_norm": 0.3595437705516815,
"learning_rate": 1.5792088384733174e-07,
"loss": 0.5045,
"step": 579
},
{
"epoch": 2.7884615384615383,
"grad_norm": 0.3366645574569702,
"learning_rate": 1.510153198249531e-07,
"loss": 0.5434,
"step": 580
},
{
"epoch": 2.793269230769231,
"grad_norm": 0.3330222964286804,
"learning_rate": 1.4426181885357215e-07,
"loss": 0.5246,
"step": 581
},
{
"epoch": 2.7980769230769234,
"grad_norm": 0.3313714861869812,
"learning_rate": 1.376605927214364e-07,
"loss": 0.5196,
"step": 582
},
{
"epoch": 2.8028846153846154,
"grad_norm": 0.3222265839576721,
"learning_rate": 1.312118484414876e-07,
"loss": 0.5153,
"step": 583
},
{
"epoch": 2.8076923076923075,
"grad_norm": 0.3020241856575012,
"learning_rate": 1.2491578824487204e-07,
"loss": 0.5489,
"step": 584
},
{
"epoch": 2.8125,
"grad_norm": 0.3178803026676178,
"learning_rate": 1.1877260957459835e-07,
"loss": 0.5352,
"step": 585
},
{
"epoch": 2.8173076923076925,
"grad_norm": 0.3120872676372528,
"learning_rate": 1.1278250507934518e-07,
"loss": 0.5294,
"step": 586
},
{
"epoch": 2.8221153846153846,
"grad_norm": 0.3329028785228729,
"learning_rate": 1.0694566260742001e-07,
"loss": 0.5307,
"step": 587
},
{
"epoch": 2.8269230769230766,
"grad_norm": 0.3385623097419739,
"learning_rate": 1.0126226520086823e-07,
"loss": 0.524,
"step": 588
},
{
"epoch": 2.831730769230769,
"grad_norm": 0.3274289071559906,
"learning_rate": 9.573249108973281e-08,
"loss": 0.5196,
"step": 589
},
{
"epoch": 2.8365384615384617,
"grad_norm": 0.31717026233673096,
"learning_rate": 9.035651368646647e-08,
"loss": 0.5163,
"step": 590
},
{
"epoch": 2.8413461538461537,
"grad_norm": 0.3221368193626404,
"learning_rate": 8.513450158049109e-08,
"loss": 0.5256,
"step": 591
},
{
"epoch": 2.8461538461538463,
"grad_norm": 0.3442428708076477,
"learning_rate": 8.006661853291298e-08,
"loss": 0.5197,
"step": 592
},
{
"epoch": 2.8509615384615383,
"grad_norm": 0.33662882447242737,
"learning_rate": 7.515302347138486e-08,
"loss": 0.5188,
"step": 593
},
{
"epoch": 2.855769230769231,
"grad_norm": 0.33708903193473816,
"learning_rate": 7.03938704851248e-08,
"loss": 0.5268,
"step": 594
},
{
"epoch": 2.8605769230769234,
"grad_norm": 0.3152363896369934,
"learning_rate": 6.578930882008283e-08,
"loss": 0.5295,
"step": 595
},
{
"epoch": 2.8653846153846154,
"grad_norm": 0.32401999831199646,
"learning_rate": 6.133948287426028e-08,
"loss": 0.511,
"step": 596
},
{
"epoch": 2.8701923076923075,
"grad_norm": 0.32977667450904846,
"learning_rate": 5.704453219318118e-08,
"loss": 0.514,
"step": 597
},
{
"epoch": 2.875,
"grad_norm": 0.32670727372169495,
"learning_rate": 5.2904591465516855e-08,
"loss": 0.514,
"step": 598
},
{
"epoch": 2.8798076923076925,
"grad_norm": 0.32648882269859314,
"learning_rate": 4.891979051886153e-08,
"loss": 0.5181,
"step": 599
},
{
"epoch": 2.8846153846153846,
"grad_norm": 0.324532687664032,
"learning_rate": 4.509025431566283e-08,
"loss": 0.5221,
"step": 600
},
{
"epoch": 2.8894230769230766,
"grad_norm": 0.31154200434684753,
"learning_rate": 4.141610294930043e-08,
"loss": 0.5212,
"step": 601
},
{
"epoch": 2.894230769230769,
"grad_norm": 0.3022911250591278,
"learning_rate": 3.7897451640321326e-08,
"loss": 0.5227,
"step": 602
},
{
"epoch": 2.8990384615384617,
"grad_norm": 0.3582252860069275,
"learning_rate": 3.4534410732825485e-08,
"loss": 0.5237,
"step": 603
},
{
"epoch": 2.9038461538461537,
"grad_norm": 0.3342374265193939,
"learning_rate": 3.1327085691006954e-08,
"loss": 0.5086,
"step": 604
},
{
"epoch": 2.9086538461538463,
"grad_norm": 0.35545840859413147,
"learning_rate": 2.8275577095846495e-08,
"loss": 0.5204,
"step": 605
},
{
"epoch": 2.9134615384615383,
"grad_norm": 0.37066027522087097,
"learning_rate": 2.5379980641955792e-08,
"loss": 0.519,
"step": 606
},
{
"epoch": 2.918269230769231,
"grad_norm": 0.3329092860221863,
"learning_rate": 2.264038713457706e-08,
"loss": 0.5424,
"step": 607
},
{
"epoch": 2.9230769230769234,
"grad_norm": 0.34132009744644165,
"learning_rate": 2.0056882486736982e-08,
"loss": 0.5209,
"step": 608
},
{
"epoch": 2.9278846153846154,
"grad_norm": 0.32534360885620117,
"learning_rate": 1.762954771655001e-08,
"loss": 0.5298,
"step": 609
},
{
"epoch": 2.9326923076923075,
"grad_norm": 0.3154315948486328,
"learning_rate": 1.5358458944680356e-08,
"loss": 0.5075,
"step": 610
},
{
"epoch": 2.9375,
"grad_norm": 0.3233172595500946,
"learning_rate": 1.3243687391952809e-08,
"loss": 0.5264,
"step": 611
},
{
"epoch": 2.9423076923076925,
"grad_norm": 0.3091405928134918,
"learning_rate": 1.1285299377118974e-08,
"loss": 0.5431,
"step": 612
},
{
"epoch": 2.9471153846153846,
"grad_norm": 0.32396966218948364,
"learning_rate": 9.48335631477948e-09,
"loss": 0.5235,
"step": 613
},
{
"epoch": 2.9519230769230766,
"grad_norm": 0.3247375786304474,
"learning_rate": 7.837914713457184e-09,
"loss": 0.53,
"step": 614
},
{
"epoch": 2.956730769230769,
"grad_norm": 0.31406140327453613,
"learning_rate": 6.349026173824713e-09,
"loss": 0.517,
"step": 615
},
{
"epoch": 2.9615384615384617,
"grad_norm": 0.3556351661682129,
"learning_rate": 5.016737387085191e-09,
"loss": 0.5218,
"step": 616
},
{
"epoch": 2.9663461538461537,
"grad_norm": 0.34563007950782776,
"learning_rate": 3.841090133511749e-09,
"loss": 0.5199,
"step": 617
},
{
"epoch": 2.9711538461538463,
"grad_norm": 0.33256563544273376,
"learning_rate": 2.8221212811324616e-09,
"loss": 0.511,
"step": 618
},
{
"epoch": 2.9759615384615383,
"grad_norm": 0.32485440373420715,
"learning_rate": 1.959862784577937e-09,
"loss": 0.5198,
"step": 619
},
{
"epoch": 2.980769230769231,
"grad_norm": 0.30697503685951233,
"learning_rate": 1.2543416840771206e-09,
"loss": 0.5484,
"step": 620
},
{
"epoch": 2.9855769230769234,
"grad_norm": 0.3317164182662964,
"learning_rate": 7.055801046113031e-10,
"loss": 0.5138,
"step": 621
},
{
"epoch": 2.9903846153846154,
"grad_norm": 0.3181401789188385,
"learning_rate": 3.1359525521801326e-10,
"loss": 0.5243,
"step": 622
},
{
"epoch": 2.9951923076923075,
"grad_norm": 0.32515648007392883,
"learning_rate": 7.839942845144777e-11,
"loss": 0.5262,
"step": 623
},
{
"epoch": 3.0,
"grad_norm": 0.3372482359409332,
"learning_rate": 0.0,
"loss": 0.5101,
"step": 624
},
{
"epoch": 3.0,
"step": 624,
"total_flos": 1149901297156096.0,
"train_loss": 0.5865381705837373,
"train_runtime": 8382.2551,
"train_samples_per_second": 7.124,
"train_steps_per_second": 0.074
}
],
"logging_steps": 1,
"max_steps": 624,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1149901297156096.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}