amburger66's picture
LoRA fine-tune on RobotSmith task03 - correlation fixed
1d7080b verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.018306636155606407,
"eval_steps": 50,
"global_step": 1000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 1.8306636155606407e-05,
"grad_norm": 6.084787368774414,
"learning_rate": 0.0,
"loss": 0.6392,
"step": 1
},
{
"epoch": 3.6613272311212814e-05,
"grad_norm": 8.000160217285156,
"learning_rate": 2.0000000000000002e-07,
"loss": 0.8375,
"step": 2
},
{
"epoch": 5.491990846681922e-05,
"grad_norm": 9.977635383605957,
"learning_rate": 4.0000000000000003e-07,
"loss": 0.4734,
"step": 3
},
{
"epoch": 7.322654462242563e-05,
"grad_norm": 6.60877799987793,
"learning_rate": 6.000000000000001e-07,
"loss": 0.7562,
"step": 4
},
{
"epoch": 9.153318077803204e-05,
"grad_norm": 10.706132888793945,
"learning_rate": 8.000000000000001e-07,
"loss": 1.0019,
"step": 5
},
{
"epoch": 0.00010983981693363844,
"grad_norm": 10.783683776855469,
"learning_rate": 1.0000000000000002e-06,
"loss": 0.9186,
"step": 6
},
{
"epoch": 0.00012814645308924485,
"grad_norm": 9.012995719909668,
"learning_rate": 1.2000000000000002e-06,
"loss": 1.029,
"step": 7
},
{
"epoch": 0.00014645308924485126,
"grad_norm": 2.700565814971924,
"learning_rate": 1.4000000000000001e-06,
"loss": 0.4689,
"step": 8
},
{
"epoch": 0.00016475972540045766,
"grad_norm": 7.6247029304504395,
"learning_rate": 1.6000000000000001e-06,
"loss": 0.6074,
"step": 9
},
{
"epoch": 0.00018306636155606407,
"grad_norm": 6.389184951782227,
"learning_rate": 1.8000000000000001e-06,
"loss": 0.8861,
"step": 10
},
{
"epoch": 0.00020137299771167048,
"grad_norm": 6.685202121734619,
"learning_rate": 2.0000000000000003e-06,
"loss": 0.5527,
"step": 11
},
{
"epoch": 0.00021967963386727689,
"grad_norm": 6.647484302520752,
"learning_rate": 2.2e-06,
"loss": 0.5459,
"step": 12
},
{
"epoch": 0.0002379862700228833,
"grad_norm": 7.842504978179932,
"learning_rate": 2.4000000000000003e-06,
"loss": 0.7269,
"step": 13
},
{
"epoch": 0.0002562929061784897,
"grad_norm": 9.710944175720215,
"learning_rate": 2.6e-06,
"loss": 0.7054,
"step": 14
},
{
"epoch": 0.00027459954233409613,
"grad_norm": 10.485865592956543,
"learning_rate": 2.8000000000000003e-06,
"loss": 1.1884,
"step": 15
},
{
"epoch": 0.0002929061784897025,
"grad_norm": 3.510791301727295,
"learning_rate": 3e-06,
"loss": 0.5241,
"step": 16
},
{
"epoch": 0.00031121281464530895,
"grad_norm": 11.97901439666748,
"learning_rate": 3.2000000000000003e-06,
"loss": 1.0192,
"step": 17
},
{
"epoch": 0.00032951945080091533,
"grad_norm": 8.022102355957031,
"learning_rate": 3.4000000000000005e-06,
"loss": 0.5731,
"step": 18
},
{
"epoch": 0.00034782608695652176,
"grad_norm": 7.800467014312744,
"learning_rate": 3.6000000000000003e-06,
"loss": 0.5588,
"step": 19
},
{
"epoch": 0.00036613272311212814,
"grad_norm": 5.90647029876709,
"learning_rate": 3.8000000000000005e-06,
"loss": 0.4773,
"step": 20
},
{
"epoch": 0.0003844393592677346,
"grad_norm": 10.05663776397705,
"learning_rate": 4.000000000000001e-06,
"loss": 0.8649,
"step": 21
},
{
"epoch": 0.00040274599542334096,
"grad_norm": 3.0191540718078613,
"learning_rate": 4.2000000000000004e-06,
"loss": 0.2516,
"step": 22
},
{
"epoch": 0.0004210526315789474,
"grad_norm": 8.017966270446777,
"learning_rate": 4.4e-06,
"loss": 0.7442,
"step": 23
},
{
"epoch": 0.00043935926773455377,
"grad_norm": 6.663602352142334,
"learning_rate": 4.600000000000001e-06,
"loss": 0.6345,
"step": 24
},
{
"epoch": 0.0004576659038901602,
"grad_norm": 8.327048301696777,
"learning_rate": 4.800000000000001e-06,
"loss": 0.6079,
"step": 25
},
{
"epoch": 0.0004759725400457666,
"grad_norm": 15.23943042755127,
"learning_rate": 5e-06,
"loss": 0.9292,
"step": 26
},
{
"epoch": 0.000494279176201373,
"grad_norm": 11.373315811157227,
"learning_rate": 5.2e-06,
"loss": 0.589,
"step": 27
},
{
"epoch": 0.0005125858123569794,
"grad_norm": 9.903393745422363,
"learning_rate": 5.400000000000001e-06,
"loss": 0.4872,
"step": 28
},
{
"epoch": 0.0005308924485125858,
"grad_norm": 17.918140411376953,
"learning_rate": 5.600000000000001e-06,
"loss": 1.1316,
"step": 29
},
{
"epoch": 0.0005491990846681923,
"grad_norm": 7.308037281036377,
"learning_rate": 5.8e-06,
"loss": 0.389,
"step": 30
},
{
"epoch": 0.0005675057208237986,
"grad_norm": 8.060295104980469,
"learning_rate": 6e-06,
"loss": 0.582,
"step": 31
},
{
"epoch": 0.000585812356979405,
"grad_norm": 8.332834243774414,
"learning_rate": 6.200000000000001e-06,
"loss": 1.1902,
"step": 32
},
{
"epoch": 0.0006041189931350115,
"grad_norm": 8.524596214294434,
"learning_rate": 6.4000000000000006e-06,
"loss": 0.3101,
"step": 33
},
{
"epoch": 0.0006224256292906179,
"grad_norm": 10.255339622497559,
"learning_rate": 6.600000000000001e-06,
"loss": 0.5759,
"step": 34
},
{
"epoch": 0.0006407322654462242,
"grad_norm": 12.656609535217285,
"learning_rate": 6.800000000000001e-06,
"loss": 1.0338,
"step": 35
},
{
"epoch": 0.0006590389016018307,
"grad_norm": 6.707568645477295,
"learning_rate": 7e-06,
"loss": 0.3607,
"step": 36
},
{
"epoch": 0.0006773455377574371,
"grad_norm": 6.263491153717041,
"learning_rate": 7.2000000000000005e-06,
"loss": 0.4405,
"step": 37
},
{
"epoch": 0.0006956521739130435,
"grad_norm": 5.70526647567749,
"learning_rate": 7.4e-06,
"loss": 0.5682,
"step": 38
},
{
"epoch": 0.0007139588100686499,
"grad_norm": 4.024957180023193,
"learning_rate": 7.600000000000001e-06,
"loss": 0.5816,
"step": 39
},
{
"epoch": 0.0007322654462242563,
"grad_norm": 8.361418724060059,
"learning_rate": 7.800000000000002e-06,
"loss": 0.8628,
"step": 40
},
{
"epoch": 0.0007505720823798627,
"grad_norm": 4.543780326843262,
"learning_rate": 8.000000000000001e-06,
"loss": 0.3442,
"step": 41
},
{
"epoch": 0.0007688787185354692,
"grad_norm": 8.164483070373535,
"learning_rate": 8.2e-06,
"loss": 0.5429,
"step": 42
},
{
"epoch": 0.0007871853546910755,
"grad_norm": 7.965429306030273,
"learning_rate": 8.400000000000001e-06,
"loss": 0.7367,
"step": 43
},
{
"epoch": 0.0008054919908466819,
"grad_norm": 16.735258102416992,
"learning_rate": 8.6e-06,
"loss": 0.9958,
"step": 44
},
{
"epoch": 0.0008237986270022883,
"grad_norm": 8.915084838867188,
"learning_rate": 8.8e-06,
"loss": 0.4859,
"step": 45
},
{
"epoch": 0.0008421052631578948,
"grad_norm": 9.113476753234863,
"learning_rate": 9e-06,
"loss": 1.1163,
"step": 46
},
{
"epoch": 0.0008604118993135011,
"grad_norm": 11.264951705932617,
"learning_rate": 9.200000000000002e-06,
"loss": 0.7696,
"step": 47
},
{
"epoch": 0.0008787185354691075,
"grad_norm": 11.313631057739258,
"learning_rate": 9.4e-06,
"loss": 1.0535,
"step": 48
},
{
"epoch": 0.000897025171624714,
"grad_norm": 2.9443583488464355,
"learning_rate": 9.600000000000001e-06,
"loss": 0.2185,
"step": 49
},
{
"epoch": 0.0009153318077803204,
"grad_norm": 5.448572635650635,
"learning_rate": 9.800000000000001e-06,
"loss": 0.3536,
"step": 50
},
{
"epoch": 0.0009336384439359267,
"grad_norm": 7.4222731590271,
"learning_rate": 1e-05,
"loss": 0.5806,
"step": 51
},
{
"epoch": 0.0009519450800915332,
"grad_norm": 6.155364990234375,
"learning_rate": 1.02e-05,
"loss": 0.8657,
"step": 52
},
{
"epoch": 0.0009702517162471396,
"grad_norm": 7.7926249504089355,
"learning_rate": 1.04e-05,
"loss": 0.8655,
"step": 53
},
{
"epoch": 0.000988558352402746,
"grad_norm": 7.443905830383301,
"learning_rate": 1.0600000000000002e-05,
"loss": 0.701,
"step": 54
},
{
"epoch": 0.0010068649885583525,
"grad_norm": 5.242746353149414,
"learning_rate": 1.0800000000000002e-05,
"loss": 0.3073,
"step": 55
},
{
"epoch": 0.0010251716247139588,
"grad_norm": 3.663163661956787,
"learning_rate": 1.1000000000000001e-05,
"loss": 0.4435,
"step": 56
},
{
"epoch": 0.0010434782608695651,
"grad_norm": 6.215541839599609,
"learning_rate": 1.1200000000000001e-05,
"loss": 0.4789,
"step": 57
},
{
"epoch": 0.0010617848970251717,
"grad_norm": 8.457674026489258,
"learning_rate": 1.14e-05,
"loss": 0.9246,
"step": 58
},
{
"epoch": 0.001080091533180778,
"grad_norm": 5.378246307373047,
"learning_rate": 1.16e-05,
"loss": 0.3759,
"step": 59
},
{
"epoch": 0.0010983981693363845,
"grad_norm": 10.745940208435059,
"learning_rate": 1.18e-05,
"loss": 1.2856,
"step": 60
},
{
"epoch": 0.0011167048054919909,
"grad_norm": 2.361469030380249,
"learning_rate": 1.2e-05,
"loss": 0.2053,
"step": 61
},
{
"epoch": 0.0011350114416475972,
"grad_norm": 7.341982841491699,
"learning_rate": 1.22e-05,
"loss": 0.5777,
"step": 62
},
{
"epoch": 0.0011533180778032037,
"grad_norm": 1.9843751192092896,
"learning_rate": 1.2400000000000002e-05,
"loss": 0.1818,
"step": 63
},
{
"epoch": 0.00117162471395881,
"grad_norm": 7.640516757965088,
"learning_rate": 1.2600000000000001e-05,
"loss": 1.0094,
"step": 64
},
{
"epoch": 0.0011899313501144164,
"grad_norm": 15.899582862854004,
"learning_rate": 1.2800000000000001e-05,
"loss": 1.5402,
"step": 65
},
{
"epoch": 0.001208237986270023,
"grad_norm": 6.037946701049805,
"learning_rate": 1.3000000000000001e-05,
"loss": 0.5179,
"step": 66
},
{
"epoch": 0.0012265446224256293,
"grad_norm": 1.6879180669784546,
"learning_rate": 1.3200000000000002e-05,
"loss": 0.2435,
"step": 67
},
{
"epoch": 0.0012448512585812358,
"grad_norm": 3.539372682571411,
"learning_rate": 1.3400000000000002e-05,
"loss": 0.2219,
"step": 68
},
{
"epoch": 0.0012631578947368421,
"grad_norm": 9.577892303466797,
"learning_rate": 1.3600000000000002e-05,
"loss": 1.0921,
"step": 69
},
{
"epoch": 0.0012814645308924484,
"grad_norm": 7.512180328369141,
"learning_rate": 1.38e-05,
"loss": 1.1851,
"step": 70
},
{
"epoch": 0.001299771167048055,
"grad_norm": 6.8162841796875,
"learning_rate": 1.4e-05,
"loss": 0.5342,
"step": 71
},
{
"epoch": 0.0013180778032036613,
"grad_norm": 13.964672088623047,
"learning_rate": 1.4200000000000001e-05,
"loss": 0.8404,
"step": 72
},
{
"epoch": 0.0013363844393592676,
"grad_norm": 10.541035652160645,
"learning_rate": 1.4400000000000001e-05,
"loss": 0.9331,
"step": 73
},
{
"epoch": 0.0013546910755148742,
"grad_norm": 5.94256067276001,
"learning_rate": 1.46e-05,
"loss": 0.7509,
"step": 74
},
{
"epoch": 0.0013729977116704805,
"grad_norm": 6.224687099456787,
"learning_rate": 1.48e-05,
"loss": 0.5732,
"step": 75
},
{
"epoch": 0.001391304347826087,
"grad_norm": 6.311083793640137,
"learning_rate": 1.5000000000000002e-05,
"loss": 0.6101,
"step": 76
},
{
"epoch": 0.0014096109839816934,
"grad_norm": 5.685784339904785,
"learning_rate": 1.5200000000000002e-05,
"loss": 0.6369,
"step": 77
},
{
"epoch": 0.0014279176201372997,
"grad_norm": 4.5122504234313965,
"learning_rate": 1.54e-05,
"loss": 0.5075,
"step": 78
},
{
"epoch": 0.0014462242562929062,
"grad_norm": 4.641362190246582,
"learning_rate": 1.5600000000000003e-05,
"loss": 0.3565,
"step": 79
},
{
"epoch": 0.0014645308924485126,
"grad_norm": 6.5127177238464355,
"learning_rate": 1.58e-05,
"loss": 0.9192,
"step": 80
},
{
"epoch": 0.001482837528604119,
"grad_norm": 4.107303619384766,
"learning_rate": 1.6000000000000003e-05,
"loss": 0.458,
"step": 81
},
{
"epoch": 0.0015011441647597254,
"grad_norm": 5.691149711608887,
"learning_rate": 1.62e-05,
"loss": 0.4345,
"step": 82
},
{
"epoch": 0.0015194508009153318,
"grad_norm": 4.126266002655029,
"learning_rate": 1.64e-05,
"loss": 0.3686,
"step": 83
},
{
"epoch": 0.0015377574370709383,
"grad_norm": 5.336889266967773,
"learning_rate": 1.66e-05,
"loss": 0.3863,
"step": 84
},
{
"epoch": 0.0015560640732265446,
"grad_norm": 9.846256256103516,
"learning_rate": 1.6800000000000002e-05,
"loss": 0.8348,
"step": 85
},
{
"epoch": 0.001574370709382151,
"grad_norm": 3.1803081035614014,
"learning_rate": 1.7e-05,
"loss": 0.5672,
"step": 86
},
{
"epoch": 0.0015926773455377575,
"grad_norm": 4.788447380065918,
"learning_rate": 1.72e-05,
"loss": 0.3226,
"step": 87
},
{
"epoch": 0.0016109839816933638,
"grad_norm": 4.775529861450195,
"learning_rate": 1.7400000000000003e-05,
"loss": 0.7328,
"step": 88
},
{
"epoch": 0.0016292906178489702,
"grad_norm": 4.884747505187988,
"learning_rate": 1.76e-05,
"loss": 0.5481,
"step": 89
},
{
"epoch": 0.0016475972540045767,
"grad_norm": 11.710714340209961,
"learning_rate": 1.7800000000000002e-05,
"loss": 0.7645,
"step": 90
},
{
"epoch": 0.001665903890160183,
"grad_norm": 7.733964443206787,
"learning_rate": 1.8e-05,
"loss": 0.6914,
"step": 91
},
{
"epoch": 0.0016842105263157896,
"grad_norm": 3.1426684856414795,
"learning_rate": 1.8200000000000002e-05,
"loss": 0.214,
"step": 92
},
{
"epoch": 0.001702517162471396,
"grad_norm": 8.66888427734375,
"learning_rate": 1.8400000000000003e-05,
"loss": 0.7952,
"step": 93
},
{
"epoch": 0.0017208237986270022,
"grad_norm": 4.978145599365234,
"learning_rate": 1.86e-05,
"loss": 0.7386,
"step": 94
},
{
"epoch": 0.0017391304347826088,
"grad_norm": 9.449872970581055,
"learning_rate": 1.88e-05,
"loss": 0.7716,
"step": 95
},
{
"epoch": 0.001757437070938215,
"grad_norm": 5.055934429168701,
"learning_rate": 1.9e-05,
"loss": 0.6527,
"step": 96
},
{
"epoch": 0.0017757437070938214,
"grad_norm": 3.6194591522216797,
"learning_rate": 1.9200000000000003e-05,
"loss": 0.424,
"step": 97
},
{
"epoch": 0.001794050343249428,
"grad_norm": 5.86330509185791,
"learning_rate": 1.94e-05,
"loss": 0.4895,
"step": 98
},
{
"epoch": 0.0018123569794050343,
"grad_norm": 5.012547492980957,
"learning_rate": 1.9600000000000002e-05,
"loss": 0.3257,
"step": 99
},
{
"epoch": 0.0018306636155606408,
"grad_norm": 6.429784297943115,
"learning_rate": 1.98e-05,
"loss": 0.4644,
"step": 100
},
{
"epoch": 0.0018489702517162471,
"grad_norm": 7.005829811096191,
"learning_rate": 2e-05,
"loss": 0.3813,
"step": 101
},
{
"epoch": 0.0018672768878718535,
"grad_norm": 5.1893415451049805,
"learning_rate": 1.9999939076577906e-05,
"loss": 0.484,
"step": 102
},
{
"epoch": 0.00188558352402746,
"grad_norm": 5.120594501495361,
"learning_rate": 1.9999756307053947e-05,
"loss": 0.4228,
"step": 103
},
{
"epoch": 0.0019038901601830663,
"grad_norm": 4.1509904861450195,
"learning_rate": 1.9999451693655125e-05,
"loss": 0.3841,
"step": 104
},
{
"epoch": 0.0019221967963386727,
"grad_norm": 6.078530788421631,
"learning_rate": 1.9999025240093045e-05,
"loss": 0.509,
"step": 105
},
{
"epoch": 0.0019405034324942792,
"grad_norm": 4.366968631744385,
"learning_rate": 1.9998476951563914e-05,
"loss": 0.4813,
"step": 106
},
{
"epoch": 0.0019588100686498858,
"grad_norm": 3.9347426891326904,
"learning_rate": 1.9997806834748455e-05,
"loss": 0.3545,
"step": 107
},
{
"epoch": 0.001977116704805492,
"grad_norm": 8.536993026733398,
"learning_rate": 1.9997014897811834e-05,
"loss": 0.7704,
"step": 108
},
{
"epoch": 0.0019954233409610984,
"grad_norm": 4.600913047790527,
"learning_rate": 1.9996101150403543e-05,
"loss": 0.4502,
"step": 109
},
{
"epoch": 0.002013729977116705,
"grad_norm": 10.658208847045898,
"learning_rate": 1.9995065603657317e-05,
"loss": 1.076,
"step": 110
},
{
"epoch": 0.002032036613272311,
"grad_norm": 5.670320510864258,
"learning_rate": 1.999390827019096e-05,
"loss": 0.7519,
"step": 111
},
{
"epoch": 0.0020503432494279176,
"grad_norm": 7.06428337097168,
"learning_rate": 1.999262916410621e-05,
"loss": 0.4423,
"step": 112
},
{
"epoch": 0.002068649885583524,
"grad_norm": 5.1104254722595215,
"learning_rate": 1.9991228300988586e-05,
"loss": 0.4446,
"step": 113
},
{
"epoch": 0.0020869565217391303,
"grad_norm": 3.8655929565429688,
"learning_rate": 1.998970569790715e-05,
"loss": 0.3284,
"step": 114
},
{
"epoch": 0.002105263157894737,
"grad_norm": 2.9013712406158447,
"learning_rate": 1.9988061373414342e-05,
"loss": 0.3442,
"step": 115
},
{
"epoch": 0.0021235697940503433,
"grad_norm": 13.76830768585205,
"learning_rate": 1.9986295347545738e-05,
"loss": 1.2942,
"step": 116
},
{
"epoch": 0.0021418764302059494,
"grad_norm": 2.8040542602539062,
"learning_rate": 1.9984407641819812e-05,
"loss": 0.1902,
"step": 117
},
{
"epoch": 0.002160183066361556,
"grad_norm": 5.416018009185791,
"learning_rate": 1.9982398279237657e-05,
"loss": 0.5331,
"step": 118
},
{
"epoch": 0.0021784897025171625,
"grad_norm": 4.10576057434082,
"learning_rate": 1.9980267284282718e-05,
"loss": 0.352,
"step": 119
},
{
"epoch": 0.002196796338672769,
"grad_norm": 3.7827014923095703,
"learning_rate": 1.9978014682920503e-05,
"loss": 0.3332,
"step": 120
},
{
"epoch": 0.002215102974828375,
"grad_norm": 5.295076847076416,
"learning_rate": 1.9975640502598243e-05,
"loss": 0.2993,
"step": 121
},
{
"epoch": 0.0022334096109839817,
"grad_norm": 5.779493808746338,
"learning_rate": 1.997314477224458e-05,
"loss": 0.8626,
"step": 122
},
{
"epoch": 0.0022517162471395883,
"grad_norm": 6.701866626739502,
"learning_rate": 1.9970527522269204e-05,
"loss": 0.4652,
"step": 123
},
{
"epoch": 0.0022700228832951944,
"grad_norm": 11.795001983642578,
"learning_rate": 1.9967788784562474e-05,
"loss": 0.4393,
"step": 124
},
{
"epoch": 0.002288329519450801,
"grad_norm": 3.3595681190490723,
"learning_rate": 1.9964928592495046e-05,
"loss": 0.2178,
"step": 125
},
{
"epoch": 0.0023066361556064075,
"grad_norm": 5.417965412139893,
"learning_rate": 1.9961946980917457e-05,
"loss": 0.6563,
"step": 126
},
{
"epoch": 0.0023249427917620136,
"grad_norm": 9.861664772033691,
"learning_rate": 1.9958843986159705e-05,
"loss": 0.9203,
"step": 127
},
{
"epoch": 0.00234324942791762,
"grad_norm": 6.221032619476318,
"learning_rate": 1.99556196460308e-05,
"loss": 0.7037,
"step": 128
},
{
"epoch": 0.0023615560640732267,
"grad_norm": 2.6676557064056396,
"learning_rate": 1.9952273999818312e-05,
"loss": 0.1846,
"step": 129
},
{
"epoch": 0.0023798627002288328,
"grad_norm": 4.812030792236328,
"learning_rate": 1.9948807088287884e-05,
"loss": 0.3774,
"step": 130
},
{
"epoch": 0.0023981693363844393,
"grad_norm": 9.71884822845459,
"learning_rate": 1.9945218953682736e-05,
"loss": 0.4416,
"step": 131
},
{
"epoch": 0.002416475972540046,
"grad_norm": 5.498316287994385,
"learning_rate": 1.9941509639723155e-05,
"loss": 0.7989,
"step": 132
},
{
"epoch": 0.002434782608695652,
"grad_norm": 2.3858511447906494,
"learning_rate": 1.9937679191605964e-05,
"loss": 0.1593,
"step": 133
},
{
"epoch": 0.0024530892448512585,
"grad_norm": 11.50888442993164,
"learning_rate": 1.9933727656003964e-05,
"loss": 0.5981,
"step": 134
},
{
"epoch": 0.002471395881006865,
"grad_norm": 6.000802516937256,
"learning_rate": 1.992965508106537e-05,
"loss": 0.5936,
"step": 135
},
{
"epoch": 0.0024897025171624716,
"grad_norm": 5.783041477203369,
"learning_rate": 1.9925461516413224e-05,
"loss": 0.5644,
"step": 136
},
{
"epoch": 0.0025080091533180777,
"grad_norm": 4.8185954093933105,
"learning_rate": 1.9921147013144782e-05,
"loss": 0.5223,
"step": 137
},
{
"epoch": 0.0025263157894736842,
"grad_norm": 5.774040222167969,
"learning_rate": 1.9916711623830904e-05,
"loss": 0.4565,
"step": 138
},
{
"epoch": 0.0025446224256292908,
"grad_norm": 7.319972515106201,
"learning_rate": 1.991215540251542e-05,
"loss": 0.6701,
"step": 139
},
{
"epoch": 0.002562929061784897,
"grad_norm": 7.840810775756836,
"learning_rate": 1.9907478404714438e-05,
"loss": 0.7881,
"step": 140
},
{
"epoch": 0.0025812356979405034,
"grad_norm": 3.3393847942352295,
"learning_rate": 1.9902680687415704e-05,
"loss": 0.3798,
"step": 141
},
{
"epoch": 0.00259954233409611,
"grad_norm": 7.432023048400879,
"learning_rate": 1.989776230907789e-05,
"loss": 0.4875,
"step": 142
},
{
"epoch": 0.002617848970251716,
"grad_norm": 4.963670253753662,
"learning_rate": 1.9892723329629885e-05,
"loss": 0.439,
"step": 143
},
{
"epoch": 0.0026361556064073226,
"grad_norm": 5.046631813049316,
"learning_rate": 1.988756381047006e-05,
"loss": 0.5672,
"step": 144
},
{
"epoch": 0.002654462242562929,
"grad_norm": 7.650580883026123,
"learning_rate": 1.988228381446553e-05,
"loss": 0.7363,
"step": 145
},
{
"epoch": 0.0026727688787185353,
"grad_norm": 7.517553806304932,
"learning_rate": 1.9876883405951378e-05,
"loss": 0.707,
"step": 146
},
{
"epoch": 0.002691075514874142,
"grad_norm": 2.3692057132720947,
"learning_rate": 1.987136265072988e-05,
"loss": 0.2329,
"step": 147
},
{
"epoch": 0.0027093821510297484,
"grad_norm": 3.4414758682250977,
"learning_rate": 1.9865721616069695e-05,
"loss": 0.3901,
"step": 148
},
{
"epoch": 0.0027276887871853545,
"grad_norm": 4.910306930541992,
"learning_rate": 1.985996037070505e-05,
"loss": 0.5497,
"step": 149
},
{
"epoch": 0.002745995423340961,
"grad_norm": 6.384312629699707,
"learning_rate": 1.9854078984834904e-05,
"loss": 0.481,
"step": 150
},
{
"epoch": 0.0027643020594965676,
"grad_norm": 3.9407761096954346,
"learning_rate": 1.9848077530122083e-05,
"loss": 0.2972,
"step": 151
},
{
"epoch": 0.002782608695652174,
"grad_norm": 5.0933122634887695,
"learning_rate": 1.984195607969242e-05,
"loss": 0.293,
"step": 152
},
{
"epoch": 0.00280091533180778,
"grad_norm": 12.269235610961914,
"learning_rate": 1.983571470813386e-05,
"loss": 0.7898,
"step": 153
},
{
"epoch": 0.0028192219679633868,
"grad_norm": 7.075507640838623,
"learning_rate": 1.9829353491495545e-05,
"loss": 0.7744,
"step": 154
},
{
"epoch": 0.0028375286041189933,
"grad_norm": 6.747109413146973,
"learning_rate": 1.982287250728689e-05,
"loss": 0.8134,
"step": 155
},
{
"epoch": 0.0028558352402745994,
"grad_norm": 6.466359615325928,
"learning_rate": 1.9816271834476642e-05,
"loss": 0.379,
"step": 156
},
{
"epoch": 0.002874141876430206,
"grad_norm": 4.720534801483154,
"learning_rate": 1.9809551553491918e-05,
"loss": 0.3114,
"step": 157
},
{
"epoch": 0.0028924485125858125,
"grad_norm": 6.6202921867370605,
"learning_rate": 1.9802711746217222e-05,
"loss": 0.4371,
"step": 158
},
{
"epoch": 0.0029107551487414186,
"grad_norm": 1.6486483812332153,
"learning_rate": 1.979575249599344e-05,
"loss": 0.0489,
"step": 159
},
{
"epoch": 0.002929061784897025,
"grad_norm": 7.626447677612305,
"learning_rate": 1.9788673887616852e-05,
"loss": 0.4217,
"step": 160
},
{
"epoch": 0.0029473684210526317,
"grad_norm": 4.698169231414795,
"learning_rate": 1.9781476007338058e-05,
"loss": 0.264,
"step": 161
},
{
"epoch": 0.002965675057208238,
"grad_norm": 8.18157958984375,
"learning_rate": 1.9774158942860962e-05,
"loss": 0.6062,
"step": 162
},
{
"epoch": 0.0029839816933638443,
"grad_norm": 1.4313340187072754,
"learning_rate": 1.9766722783341682e-05,
"loss": 0.1536,
"step": 163
},
{
"epoch": 0.003002288329519451,
"grad_norm": 5.4740705490112305,
"learning_rate": 1.9759167619387474e-05,
"loss": 0.4488,
"step": 164
},
{
"epoch": 0.0030205949656750574,
"grad_norm": 8.551219940185547,
"learning_rate": 1.9751493543055634e-05,
"loss": 0.4957,
"step": 165
},
{
"epoch": 0.0030389016018306635,
"grad_norm": 6.109156131744385,
"learning_rate": 1.9743700647852356e-05,
"loss": 0.5566,
"step": 166
},
{
"epoch": 0.00305720823798627,
"grad_norm": 8.344240188598633,
"learning_rate": 1.9735789028731603e-05,
"loss": 0.8809,
"step": 167
},
{
"epoch": 0.0030755148741418766,
"grad_norm": 7.053730010986328,
"learning_rate": 1.972775878209397e-05,
"loss": 0.6096,
"step": 168
},
{
"epoch": 0.0030938215102974827,
"grad_norm": 8.59719181060791,
"learning_rate": 1.9719610005785466e-05,
"loss": 0.4776,
"step": 169
},
{
"epoch": 0.0031121281464530893,
"grad_norm": 7.1130571365356445,
"learning_rate": 1.971134279909636e-05,
"loss": 0.8,
"step": 170
},
{
"epoch": 0.003130434782608696,
"grad_norm": 5.579831123352051,
"learning_rate": 1.9702957262759964e-05,
"loss": 0.4925,
"step": 171
},
{
"epoch": 0.003148741418764302,
"grad_norm": 3.0703647136688232,
"learning_rate": 1.9694453498951392e-05,
"loss": 0.2191,
"step": 172
},
{
"epoch": 0.0031670480549199085,
"grad_norm": 2.900186538696289,
"learning_rate": 1.9685831611286312e-05,
"loss": 0.3114,
"step": 173
},
{
"epoch": 0.003185354691075515,
"grad_norm": 1.6251899003982544,
"learning_rate": 1.9677091704819714e-05,
"loss": 0.2184,
"step": 174
},
{
"epoch": 0.003203661327231121,
"grad_norm": 8.224297523498535,
"learning_rate": 1.9668233886044597e-05,
"loss": 0.6323,
"step": 175
},
{
"epoch": 0.0032219679633867277,
"grad_norm": 7.209324836730957,
"learning_rate": 1.9659258262890683e-05,
"loss": 0.6906,
"step": 176
},
{
"epoch": 0.003240274599542334,
"grad_norm": 4.86000919342041,
"learning_rate": 1.9650164944723116e-05,
"loss": 0.3577,
"step": 177
},
{
"epoch": 0.0032585812356979403,
"grad_norm": 2.0504913330078125,
"learning_rate": 1.96409540423411e-05,
"loss": 0.3458,
"step": 178
},
{
"epoch": 0.003276887871853547,
"grad_norm": 3.175110101699829,
"learning_rate": 1.9631625667976584e-05,
"loss": 0.244,
"step": 179
},
{
"epoch": 0.0032951945080091534,
"grad_norm": 4.946536540985107,
"learning_rate": 1.9622179935292855e-05,
"loss": 0.4006,
"step": 180
},
{
"epoch": 0.00331350114416476,
"grad_norm": 5.009103298187256,
"learning_rate": 1.961261695938319e-05,
"loss": 0.2098,
"step": 181
},
{
"epoch": 0.003331807780320366,
"grad_norm": 4.852262020111084,
"learning_rate": 1.9602936856769432e-05,
"loss": 0.3542,
"step": 182
},
{
"epoch": 0.0033501144164759726,
"grad_norm": 3.4230353832244873,
"learning_rate": 1.9593139745400575e-05,
"loss": 0.3436,
"step": 183
},
{
"epoch": 0.003368421052631579,
"grad_norm": 4.419967174530029,
"learning_rate": 1.9583225744651334e-05,
"loss": 0.2836,
"step": 184
},
{
"epoch": 0.0033867276887871852,
"grad_norm": 8.700972557067871,
"learning_rate": 1.9573194975320672e-05,
"loss": 0.6588,
"step": 185
},
{
"epoch": 0.003405034324942792,
"grad_norm": 8.335065841674805,
"learning_rate": 1.9563047559630356e-05,
"loss": 0.6679,
"step": 186
},
{
"epoch": 0.0034233409610983983,
"grad_norm": 10.574755668640137,
"learning_rate": 1.9552783621223437e-05,
"loss": 0.6755,
"step": 187
},
{
"epoch": 0.0034416475972540044,
"grad_norm": 4.3693671226501465,
"learning_rate": 1.954240328516277e-05,
"loss": 0.2874,
"step": 188
},
{
"epoch": 0.003459954233409611,
"grad_norm": 3.7697315216064453,
"learning_rate": 1.9531906677929472e-05,
"loss": 0.2167,
"step": 189
},
{
"epoch": 0.0034782608695652175,
"grad_norm": 9.218770027160645,
"learning_rate": 1.9521293927421388e-05,
"loss": 0.3588,
"step": 190
},
{
"epoch": 0.0034965675057208236,
"grad_norm": 8.993297576904297,
"learning_rate": 1.9510565162951538e-05,
"loss": 0.7196,
"step": 191
},
{
"epoch": 0.00351487414187643,
"grad_norm": 5.189692974090576,
"learning_rate": 1.9499720515246524e-05,
"loss": 0.4028,
"step": 192
},
{
"epoch": 0.0035331807780320367,
"grad_norm": 2.8236076831817627,
"learning_rate": 1.9488760116444966e-05,
"loss": 0.337,
"step": 193
},
{
"epoch": 0.003551487414187643,
"grad_norm": 9.882134437561035,
"learning_rate": 1.947768410009586e-05,
"loss": 0.8133,
"step": 194
},
{
"epoch": 0.0035697940503432494,
"grad_norm": 6.428056716918945,
"learning_rate": 1.9466492601156964e-05,
"loss": 0.7583,
"step": 195
},
{
"epoch": 0.003588100686498856,
"grad_norm": 9.280491828918457,
"learning_rate": 1.945518575599317e-05,
"loss": 0.4404,
"step": 196
},
{
"epoch": 0.0036064073226544625,
"grad_norm": 6.298871040344238,
"learning_rate": 1.944376370237481e-05,
"loss": 0.3666,
"step": 197
},
{
"epoch": 0.0036247139588100686,
"grad_norm": 6.506961822509766,
"learning_rate": 1.943222657947601e-05,
"loss": 0.2982,
"step": 198
},
{
"epoch": 0.003643020594965675,
"grad_norm": 4.006402015686035,
"learning_rate": 1.942057452787297e-05,
"loss": 0.3572,
"step": 199
},
{
"epoch": 0.0036613272311212816,
"grad_norm": 3.9380829334259033,
"learning_rate": 1.9408807689542257e-05,
"loss": 0.2622,
"step": 200
},
{
"epoch": 0.0036796338672768878,
"grad_norm": 3.6081972122192383,
"learning_rate": 1.9396926207859085e-05,
"loss": 0.3038,
"step": 201
},
{
"epoch": 0.0036979405034324943,
"grad_norm": 4.289360046386719,
"learning_rate": 1.938493022759556e-05,
"loss": 0.2393,
"step": 202
},
{
"epoch": 0.003716247139588101,
"grad_norm": 5.4397358894348145,
"learning_rate": 1.937281989491892e-05,
"loss": 0.416,
"step": 203
},
{
"epoch": 0.003734553775743707,
"grad_norm": 2.6657931804656982,
"learning_rate": 1.9360595357389735e-05,
"loss": 0.1847,
"step": 204
},
{
"epoch": 0.0037528604118993135,
"grad_norm": 10.061616897583008,
"learning_rate": 1.9348256763960146e-05,
"loss": 1.0184,
"step": 205
},
{
"epoch": 0.00377116704805492,
"grad_norm": 5.163506031036377,
"learning_rate": 1.9335804264972018e-05,
"loss": 0.4527,
"step": 206
},
{
"epoch": 0.003789473684210526,
"grad_norm": 8.941624641418457,
"learning_rate": 1.9323238012155125e-05,
"loss": 0.4174,
"step": 207
},
{
"epoch": 0.0038077803203661327,
"grad_norm": 10.84901237487793,
"learning_rate": 1.9310558158625286e-05,
"loss": 0.7763,
"step": 208
},
{
"epoch": 0.0038260869565217392,
"grad_norm": 6.312254428863525,
"learning_rate": 1.9297764858882516e-05,
"loss": 0.5921,
"step": 209
},
{
"epoch": 0.0038443935926773453,
"grad_norm": 2.949817419052124,
"learning_rate": 1.9284858268809135e-05,
"loss": 0.1387,
"step": 210
},
{
"epoch": 0.003862700228832952,
"grad_norm": 8.410788536071777,
"learning_rate": 1.9271838545667876e-05,
"loss": 0.6722,
"step": 211
},
{
"epoch": 0.0038810068649885584,
"grad_norm": 14.2056303024292,
"learning_rate": 1.925870584809995e-05,
"loss": 1.3406,
"step": 212
},
{
"epoch": 0.003899313501144165,
"grad_norm": 6.280990123748779,
"learning_rate": 1.9245460336123136e-05,
"loss": 0.2529,
"step": 213
},
{
"epoch": 0.0039176201372997715,
"grad_norm": 3.3927037715911865,
"learning_rate": 1.923210217112981e-05,
"loss": 0.181,
"step": 214
},
{
"epoch": 0.003935926773455377,
"grad_norm": 4.4074811935424805,
"learning_rate": 1.9218631515885007e-05,
"loss": 0.3954,
"step": 215
},
{
"epoch": 0.003954233409610984,
"grad_norm": 4.823078155517578,
"learning_rate": 1.9205048534524405e-05,
"loss": 0.5657,
"step": 216
},
{
"epoch": 0.00397254004576659,
"grad_norm": 5.334933757781982,
"learning_rate": 1.9191353392552346e-05,
"loss": 0.4124,
"step": 217
},
{
"epoch": 0.003990846681922197,
"grad_norm": 7.765951633453369,
"learning_rate": 1.9177546256839814e-05,
"loss": 0.4361,
"step": 218
},
{
"epoch": 0.004009153318077803,
"grad_norm": 5.867908000946045,
"learning_rate": 1.9163627295622397e-05,
"loss": 0.3579,
"step": 219
},
{
"epoch": 0.00402745995423341,
"grad_norm": 6.852147102355957,
"learning_rate": 1.914959667849825e-05,
"loss": 0.7059,
"step": 220
},
{
"epoch": 0.0040457665903890164,
"grad_norm": 5.715799808502197,
"learning_rate": 1.913545457642601e-05,
"loss": 0.6624,
"step": 221
},
{
"epoch": 0.004064073226544622,
"grad_norm": 4.819512367248535,
"learning_rate": 1.9121201161722732e-05,
"loss": 0.3062,
"step": 222
},
{
"epoch": 0.004082379862700229,
"grad_norm": 5.430914878845215,
"learning_rate": 1.910683660806177e-05,
"loss": 0.664,
"step": 223
},
{
"epoch": 0.004100686498855835,
"grad_norm": 16.048826217651367,
"learning_rate": 1.9092361090470688e-05,
"loss": 0.9425,
"step": 224
},
{
"epoch": 0.004118993135011442,
"grad_norm": 10.175516128540039,
"learning_rate": 1.907777478532909e-05,
"loss": 0.4463,
"step": 225
},
{
"epoch": 0.004137299771167048,
"grad_norm": 10.81609058380127,
"learning_rate": 1.9063077870366504e-05,
"loss": 0.5496,
"step": 226
},
{
"epoch": 0.004155606407322655,
"grad_norm": 3.3909196853637695,
"learning_rate": 1.9048270524660197e-05,
"loss": 0.2642,
"step": 227
},
{
"epoch": 0.0041739130434782605,
"grad_norm": 7.220968723297119,
"learning_rate": 1.903335292863301e-05,
"loss": 0.3606,
"step": 228
},
{
"epoch": 0.004192219679633867,
"grad_norm": 2.8659698963165283,
"learning_rate": 1.901832526405114e-05,
"loss": 0.2331,
"step": 229
},
{
"epoch": 0.004210526315789474,
"grad_norm": 8.508347511291504,
"learning_rate": 1.9003187714021936e-05,
"loss": 0.5547,
"step": 230
},
{
"epoch": 0.00422883295194508,
"grad_norm": 2.9436657428741455,
"learning_rate": 1.8987940462991673e-05,
"loss": 0.1906,
"step": 231
},
{
"epoch": 0.004247139588100687,
"grad_norm": 8.40645980834961,
"learning_rate": 1.8972583696743284e-05,
"loss": 0.5893,
"step": 232
},
{
"epoch": 0.004265446224256293,
"grad_norm": 4.141918182373047,
"learning_rate": 1.895711760239413e-05,
"loss": 0.5183,
"step": 233
},
{
"epoch": 0.004283752860411899,
"grad_norm": 7.489543914794922,
"learning_rate": 1.8941542368393683e-05,
"loss": 0.4938,
"step": 234
},
{
"epoch": 0.0043020594965675054,
"grad_norm": 3.483055591583252,
"learning_rate": 1.892585818452126e-05,
"loss": 0.2687,
"step": 235
},
{
"epoch": 0.004320366132723112,
"grad_norm": 5.23473596572876,
"learning_rate": 1.891006524188368e-05,
"loss": 0.2855,
"step": 236
},
{
"epoch": 0.0043386727688787185,
"grad_norm": 4.449317455291748,
"learning_rate": 1.889416373291298e-05,
"loss": 0.3082,
"step": 237
},
{
"epoch": 0.004356979405034325,
"grad_norm": 3.1681277751922607,
"learning_rate": 1.8878153851364013e-05,
"loss": 0.2741,
"step": 238
},
{
"epoch": 0.004375286041189932,
"grad_norm": 10.63996410369873,
"learning_rate": 1.8862035792312148e-05,
"loss": 0.6235,
"step": 239
},
{
"epoch": 0.004393592677345538,
"grad_norm": 5.231956958770752,
"learning_rate": 1.884580975215084e-05,
"loss": 0.4739,
"step": 240
},
{
"epoch": 0.004411899313501144,
"grad_norm": 3.019294023513794,
"learning_rate": 1.8829475928589272e-05,
"loss": 0.4061,
"step": 241
},
{
"epoch": 0.00443020594965675,
"grad_norm": 5.6815595626831055,
"learning_rate": 1.8813034520649923e-05,
"loss": 0.4274,
"step": 242
},
{
"epoch": 0.004448512585812357,
"grad_norm": 5.3184123039245605,
"learning_rate": 1.879648572866617e-05,
"loss": 0.3172,
"step": 243
},
{
"epoch": 0.0044668192219679635,
"grad_norm": 1.4854844808578491,
"learning_rate": 1.8779829754279806e-05,
"loss": 0.1407,
"step": 244
},
{
"epoch": 0.00448512585812357,
"grad_norm": 9.262789726257324,
"learning_rate": 1.8763066800438638e-05,
"loss": 0.7686,
"step": 245
},
{
"epoch": 0.0045034324942791765,
"grad_norm": 6.8137359619140625,
"learning_rate": 1.874619707139396e-05,
"loss": 0.3486,
"step": 246
},
{
"epoch": 0.004521739130434782,
"grad_norm": 11.356719017028809,
"learning_rate": 1.8729220772698096e-05,
"loss": 0.4303,
"step": 247
},
{
"epoch": 0.004540045766590389,
"grad_norm": 5.79303503036499,
"learning_rate": 1.8712138111201898e-05,
"loss": 0.3068,
"step": 248
},
{
"epoch": 0.004558352402745995,
"grad_norm": 1.9887948036193848,
"learning_rate": 1.869494929505219e-05,
"loss": 0.1456,
"step": 249
},
{
"epoch": 0.004576659038901602,
"grad_norm": 10.916764259338379,
"learning_rate": 1.8677654533689287e-05,
"loss": 1.1486,
"step": 250
},
{
"epoch": 0.004594965675057208,
"grad_norm": 4.801982879638672,
"learning_rate": 1.866025403784439e-05,
"loss": 0.2377,
"step": 251
},
{
"epoch": 0.004613272311212815,
"grad_norm": 5.9616498947143555,
"learning_rate": 1.864274801953705e-05,
"loss": 0.213,
"step": 252
},
{
"epoch": 0.0046315789473684215,
"grad_norm": 4.0821099281311035,
"learning_rate": 1.8625136692072577e-05,
"loss": 0.2355,
"step": 253
},
{
"epoch": 0.004649885583524027,
"grad_norm": 3.7081425189971924,
"learning_rate": 1.860742027003944e-05,
"loss": 0.2571,
"step": 254
},
{
"epoch": 0.004668192219679634,
"grad_norm": 8.445603370666504,
"learning_rate": 1.8589598969306646e-05,
"loss": 0.6145,
"step": 255
},
{
"epoch": 0.00468649885583524,
"grad_norm": 11.145668029785156,
"learning_rate": 1.8571673007021124e-05,
"loss": 0.7036,
"step": 256
},
{
"epoch": 0.004704805491990847,
"grad_norm": 3.8045926094055176,
"learning_rate": 1.855364260160507e-05,
"loss": 0.3276,
"step": 257
},
{
"epoch": 0.004723112128146453,
"grad_norm": 8.324919700622559,
"learning_rate": 1.8535507972753275e-05,
"loss": 0.9585,
"step": 258
},
{
"epoch": 0.00474141876430206,
"grad_norm": 9.286887168884277,
"learning_rate": 1.851726934143048e-05,
"loss": 0.4821,
"step": 259
},
{
"epoch": 0.0047597254004576655,
"grad_norm": 1.0730000734329224,
"learning_rate": 1.849892692986864e-05,
"loss": 0.1082,
"step": 260
},
{
"epoch": 0.004778032036613272,
"grad_norm": 4.952276229858398,
"learning_rate": 1.848048096156426e-05,
"loss": 0.2672,
"step": 261
},
{
"epoch": 0.004796338672768879,
"grad_norm": 4.847801208496094,
"learning_rate": 1.8461931661275642e-05,
"loss": 0.4634,
"step": 262
},
{
"epoch": 0.004814645308924485,
"grad_norm": 2.5741231441497803,
"learning_rate": 1.8443279255020153e-05,
"loss": 0.209,
"step": 263
},
{
"epoch": 0.004832951945080092,
"grad_norm": 7.57988977432251,
"learning_rate": 1.842452397007148e-05,
"loss": 0.4493,
"step": 264
},
{
"epoch": 0.004851258581235698,
"grad_norm": 7.51473331451416,
"learning_rate": 1.8405666034956842e-05,
"loss": 0.6737,
"step": 265
},
{
"epoch": 0.004869565217391304,
"grad_norm": 13.536398887634277,
"learning_rate": 1.8386705679454243e-05,
"loss": 0.8438,
"step": 266
},
{
"epoch": 0.0048878718535469105,
"grad_norm": 3.2198519706726074,
"learning_rate": 1.836764313458962e-05,
"loss": 0.251,
"step": 267
},
{
"epoch": 0.004906178489702517,
"grad_norm": 5.983921527862549,
"learning_rate": 1.8348478632634067e-05,
"loss": 0.3383,
"step": 268
},
{
"epoch": 0.0049244851258581235,
"grad_norm": 10.424074172973633,
"learning_rate": 1.8329212407100996e-05,
"loss": 1.0936,
"step": 269
},
{
"epoch": 0.00494279176201373,
"grad_norm": 6.141103267669678,
"learning_rate": 1.8309844692743283e-05,
"loss": 0.5605,
"step": 270
},
{
"epoch": 0.004961098398169337,
"grad_norm": 2.830514430999756,
"learning_rate": 1.8290375725550417e-05,
"loss": 0.1908,
"step": 271
},
{
"epoch": 0.004979405034324943,
"grad_norm": 6.214183330535889,
"learning_rate": 1.827080574274562e-05,
"loss": 0.6819,
"step": 272
},
{
"epoch": 0.004997711670480549,
"grad_norm": 7.017670631408691,
"learning_rate": 1.8251134982782952e-05,
"loss": 0.5617,
"step": 273
},
{
"epoch": 0.005016018306636155,
"grad_norm": 15.681363105773926,
"learning_rate": 1.8231363685344422e-05,
"loss": 0.8103,
"step": 274
},
{
"epoch": 0.005034324942791762,
"grad_norm": 7.182861328125,
"learning_rate": 1.821149209133704e-05,
"loss": 0.4888,
"step": 275
},
{
"epoch": 0.0050526315789473685,
"grad_norm": 10.65903377532959,
"learning_rate": 1.819152044288992e-05,
"loss": 0.7242,
"step": 276
},
{
"epoch": 0.005070938215102975,
"grad_norm": 4.960131645202637,
"learning_rate": 1.8171448983351284e-05,
"loss": 0.4659,
"step": 277
},
{
"epoch": 0.0050892448512585816,
"grad_norm": 2.2972419261932373,
"learning_rate": 1.815127795728554e-05,
"loss": 0.1826,
"step": 278
},
{
"epoch": 0.005107551487414187,
"grad_norm": 5.904301166534424,
"learning_rate": 1.8131007610470278e-05,
"loss": 0.3366,
"step": 279
},
{
"epoch": 0.005125858123569794,
"grad_norm": 5.723453998565674,
"learning_rate": 1.8110638189893267e-05,
"loss": 0.3315,
"step": 280
},
{
"epoch": 0.0051441647597254,
"grad_norm": 5.037336349487305,
"learning_rate": 1.8090169943749477e-05,
"loss": 0.3074,
"step": 281
},
{
"epoch": 0.005162471395881007,
"grad_norm": 7.443374156951904,
"learning_rate": 1.806960312143802e-05,
"loss": 0.4225,
"step": 282
},
{
"epoch": 0.005180778032036613,
"grad_norm": 4.295177459716797,
"learning_rate": 1.804893797355914e-05,
"loss": 0.3837,
"step": 283
},
{
"epoch": 0.00519908466819222,
"grad_norm": 3.3948891162872314,
"learning_rate": 1.8028174751911147e-05,
"loss": 0.349,
"step": 284
},
{
"epoch": 0.0052173913043478265,
"grad_norm": 6.52895975112915,
"learning_rate": 1.8007313709487334e-05,
"loss": 0.4161,
"step": 285
},
{
"epoch": 0.005235697940503432,
"grad_norm": 8.563170433044434,
"learning_rate": 1.798635510047293e-05,
"loss": 0.5702,
"step": 286
},
{
"epoch": 0.005254004576659039,
"grad_norm": 3.7967607975006104,
"learning_rate": 1.7965299180241963e-05,
"loss": 0.1922,
"step": 287
},
{
"epoch": 0.005272311212814645,
"grad_norm": 5.493015766143799,
"learning_rate": 1.7944146205354182e-05,
"loss": 0.6328,
"step": 288
},
{
"epoch": 0.005290617848970252,
"grad_norm": 10.71493148803711,
"learning_rate": 1.792289643355191e-05,
"loss": 0.6749,
"step": 289
},
{
"epoch": 0.005308924485125858,
"grad_norm": 7.881803512573242,
"learning_rate": 1.7901550123756906e-05,
"loss": 0.4414,
"step": 290
},
{
"epoch": 0.005327231121281465,
"grad_norm": 3.409003734588623,
"learning_rate": 1.788010753606722e-05,
"loss": 0.217,
"step": 291
},
{
"epoch": 0.0053455377574370706,
"grad_norm": 5.169612884521484,
"learning_rate": 1.785856893175402e-05,
"loss": 0.4037,
"step": 292
},
{
"epoch": 0.005363844393592677,
"grad_norm": 5.614249229431152,
"learning_rate": 1.78369345732584e-05,
"loss": 0.3583,
"step": 293
},
{
"epoch": 0.005382151029748284,
"grad_norm": 5.413997173309326,
"learning_rate": 1.781520472418819e-05,
"loss": 0.4758,
"step": 294
},
{
"epoch": 0.00540045766590389,
"grad_norm": 5.485922813415527,
"learning_rate": 1.7793379649314743e-05,
"loss": 0.4584,
"step": 295
},
{
"epoch": 0.005418764302059497,
"grad_norm": 6.144253253936768,
"learning_rate": 1.777145961456971e-05,
"loss": 0.5489,
"step": 296
},
{
"epoch": 0.005437070938215103,
"grad_norm": 4.56196403503418,
"learning_rate": 1.7749444887041797e-05,
"loss": 0.4218,
"step": 297
},
{
"epoch": 0.005455377574370709,
"grad_norm": 12.976069450378418,
"learning_rate": 1.7727335734973512e-05,
"loss": 0.519,
"step": 298
},
{
"epoch": 0.0054736842105263155,
"grad_norm": 10.657815933227539,
"learning_rate": 1.7705132427757895e-05,
"loss": 0.8982,
"step": 299
},
{
"epoch": 0.005491990846681922,
"grad_norm": 3.892559766769409,
"learning_rate": 1.7682835235935236e-05,
"loss": 0.2004,
"step": 300
},
{
"epoch": 0.005510297482837529,
"grad_norm": 6.767181396484375,
"learning_rate": 1.766044443118978e-05,
"loss": 0.5441,
"step": 301
},
{
"epoch": 0.005528604118993135,
"grad_norm": 5.241102695465088,
"learning_rate": 1.7637960286346423e-05,
"loss": 0.3085,
"step": 302
},
{
"epoch": 0.005546910755148742,
"grad_norm": 4.85935640335083,
"learning_rate": 1.761538307536737e-05,
"loss": 0.4702,
"step": 303
},
{
"epoch": 0.005565217391304348,
"grad_norm": 8.737496376037598,
"learning_rate": 1.759271307334881e-05,
"loss": 0.7425,
"step": 304
},
{
"epoch": 0.005583524027459954,
"grad_norm": 10.740922927856445,
"learning_rate": 1.7569950556517566e-05,
"loss": 0.9976,
"step": 305
},
{
"epoch": 0.00560183066361556,
"grad_norm": 7.7551422119140625,
"learning_rate": 1.7547095802227723e-05,
"loss": 0.5183,
"step": 306
},
{
"epoch": 0.005620137299771167,
"grad_norm": 6.327582836151123,
"learning_rate": 1.7524149088957244e-05,
"loss": 0.6134,
"step": 307
},
{
"epoch": 0.0056384439359267735,
"grad_norm": 10.561834335327148,
"learning_rate": 1.7501110696304598e-05,
"loss": 0.6976,
"step": 308
},
{
"epoch": 0.00565675057208238,
"grad_norm": 4.514830112457275,
"learning_rate": 1.747798090498532e-05,
"loss": 0.3118,
"step": 309
},
{
"epoch": 0.005675057208237987,
"grad_norm": 3.9753475189208984,
"learning_rate": 1.7454759996828622e-05,
"loss": 0.3003,
"step": 310
},
{
"epoch": 0.005693363844393592,
"grad_norm": 7.0595011711120605,
"learning_rate": 1.7431448254773943e-05,
"loss": 0.5337,
"step": 311
},
{
"epoch": 0.005711670480549199,
"grad_norm": 4.571330547332764,
"learning_rate": 1.74080459628675e-05,
"loss": 0.4606,
"step": 312
},
{
"epoch": 0.005729977116704805,
"grad_norm": 7.209574222564697,
"learning_rate": 1.7384553406258842e-05,
"loss": 0.4431,
"step": 313
},
{
"epoch": 0.005748283752860412,
"grad_norm": 3.1496009826660156,
"learning_rate": 1.7360970871197347e-05,
"loss": 0.288,
"step": 314
},
{
"epoch": 0.0057665903890160184,
"grad_norm": 1.1498371362686157,
"learning_rate": 1.7337298645028764e-05,
"loss": 0.0904,
"step": 315
},
{
"epoch": 0.005784897025171625,
"grad_norm": 4.189731121063232,
"learning_rate": 1.7313537016191706e-05,
"loss": 0.2142,
"step": 316
},
{
"epoch": 0.0058032036613272315,
"grad_norm": 3.495272636413574,
"learning_rate": 1.7289686274214116e-05,
"loss": 0.2809,
"step": 317
},
{
"epoch": 0.005821510297482837,
"grad_norm": 11.465441703796387,
"learning_rate": 1.7265746709709762e-05,
"loss": 0.9229,
"step": 318
},
{
"epoch": 0.005839816933638444,
"grad_norm": 3.298283338546753,
"learning_rate": 1.7241718614374678e-05,
"loss": 0.2105,
"step": 319
},
{
"epoch": 0.00585812356979405,
"grad_norm": 4.148833274841309,
"learning_rate": 1.7217602280983622e-05,
"loss": 0.2279,
"step": 320
},
{
"epoch": 0.005876430205949657,
"grad_norm": 2.6801838874816895,
"learning_rate": 1.7193398003386514e-05,
"loss": 0.2207,
"step": 321
},
{
"epoch": 0.005894736842105263,
"grad_norm": 5.371076583862305,
"learning_rate": 1.716910607650483e-05,
"loss": 0.6822,
"step": 322
},
{
"epoch": 0.00591304347826087,
"grad_norm": 5.155342102050781,
"learning_rate": 1.7144726796328034e-05,
"loss": 0.6256,
"step": 323
},
{
"epoch": 0.005931350114416476,
"grad_norm": 5.462609767913818,
"learning_rate": 1.712026045990997e-05,
"loss": 0.3712,
"step": 324
},
{
"epoch": 0.005949656750572082,
"grad_norm": 2.920656204223633,
"learning_rate": 1.709570736536521e-05,
"loss": 0.2011,
"step": 325
},
{
"epoch": 0.005967963386727689,
"grad_norm": 6.90954065322876,
"learning_rate": 1.7071067811865477e-05,
"loss": 0.3713,
"step": 326
},
{
"epoch": 0.005986270022883295,
"grad_norm": 4.551551342010498,
"learning_rate": 1.7046342099635948e-05,
"loss": 0.3052,
"step": 327
},
{
"epoch": 0.006004576659038902,
"grad_norm": 4.894721031188965,
"learning_rate": 1.7021530529951627e-05,
"loss": 0.8055,
"step": 328
},
{
"epoch": 0.006022883295194508,
"grad_norm": 4.173544406890869,
"learning_rate": 1.6996633405133656e-05,
"loss": 0.4642,
"step": 329
},
{
"epoch": 0.006041189931350115,
"grad_norm": 5.222576141357422,
"learning_rate": 1.697165102854565e-05,
"loss": 0.3175,
"step": 330
},
{
"epoch": 0.0060594965675057205,
"grad_norm": 6.459454536437988,
"learning_rate": 1.6946583704589973e-05,
"loss": 0.5357,
"step": 331
},
{
"epoch": 0.006077803203661327,
"grad_norm": 3.6619327068328857,
"learning_rate": 1.692143173870407e-05,
"loss": 0.4273,
"step": 332
},
{
"epoch": 0.006096109839816934,
"grad_norm": 5.7568840980529785,
"learning_rate": 1.68961954373567e-05,
"loss": 0.4508,
"step": 333
},
{
"epoch": 0.00611441647597254,
"grad_norm": 7.910932540893555,
"learning_rate": 1.6870875108044233e-05,
"loss": 0.4166,
"step": 334
},
{
"epoch": 0.006132723112128147,
"grad_norm": 4.656468391418457,
"learning_rate": 1.684547105928689e-05,
"loss": 0.3099,
"step": 335
},
{
"epoch": 0.006151029748283753,
"grad_norm": 6.832411289215088,
"learning_rate": 1.6819983600624986e-05,
"loss": 0.4713,
"step": 336
},
{
"epoch": 0.006169336384439359,
"grad_norm": 3.675230026245117,
"learning_rate": 1.6794413042615168e-05,
"loss": 0.3749,
"step": 337
},
{
"epoch": 0.0061876430205949655,
"grad_norm": 14.497318267822266,
"learning_rate": 1.6768759696826608e-05,
"loss": 1.1848,
"step": 338
},
{
"epoch": 0.006205949656750572,
"grad_norm": 8.537349700927734,
"learning_rate": 1.6743023875837233e-05,
"loss": 0.4933,
"step": 339
},
{
"epoch": 0.0062242562929061785,
"grad_norm": 4.796643257141113,
"learning_rate": 1.6717205893229904e-05,
"loss": 0.4819,
"step": 340
},
{
"epoch": 0.006242562929061785,
"grad_norm": 5.899059772491455,
"learning_rate": 1.6691306063588583e-05,
"loss": 0.6742,
"step": 341
},
{
"epoch": 0.006260869565217392,
"grad_norm": 8.666280746459961,
"learning_rate": 1.6665324702494524e-05,
"loss": 0.5037,
"step": 342
},
{
"epoch": 0.006279176201372997,
"grad_norm": 14.649405479431152,
"learning_rate": 1.6639262126522417e-05,
"loss": 0.7326,
"step": 343
},
{
"epoch": 0.006297482837528604,
"grad_norm": 6.6728034019470215,
"learning_rate": 1.661311865323652e-05,
"loss": 0.4329,
"step": 344
},
{
"epoch": 0.00631578947368421,
"grad_norm": 3.9211013317108154,
"learning_rate": 1.6586894601186804e-05,
"loss": 0.4379,
"step": 345
},
{
"epoch": 0.006334096109839817,
"grad_norm": 5.018696308135986,
"learning_rate": 1.6560590289905074e-05,
"loss": 0.3585,
"step": 346
},
{
"epoch": 0.0063524027459954235,
"grad_norm": 10.415823936462402,
"learning_rate": 1.6534206039901057e-05,
"loss": 0.5712,
"step": 347
},
{
"epoch": 0.00637070938215103,
"grad_norm": 3.3000543117523193,
"learning_rate": 1.650774217265851e-05,
"loss": 0.262,
"step": 348
},
{
"epoch": 0.0063890160183066366,
"grad_norm": 5.603179454803467,
"learning_rate": 1.6481199010631312e-05,
"loss": 0.2821,
"step": 349
},
{
"epoch": 0.006407322654462242,
"grad_norm": 10.997979164123535,
"learning_rate": 1.645457687723951e-05,
"loss": 0.8257,
"step": 350
},
{
"epoch": 0.006425629290617849,
"grad_norm": 9.430671691894531,
"learning_rate": 1.6427876096865394e-05,
"loss": 0.6242,
"step": 351
},
{
"epoch": 0.006443935926773455,
"grad_norm": 4.974905014038086,
"learning_rate": 1.6401096994849558e-05,
"loss": 0.3398,
"step": 352
},
{
"epoch": 0.006462242562929062,
"grad_norm": 6.002130031585693,
"learning_rate": 1.63742398974869e-05,
"loss": 0.4058,
"step": 353
},
{
"epoch": 0.006480549199084668,
"grad_norm": 4.933552265167236,
"learning_rate": 1.6347305132022677e-05,
"loss": 0.4209,
"step": 354
},
{
"epoch": 0.006498855835240275,
"grad_norm": 4.84297513961792,
"learning_rate": 1.632029302664851e-05,
"loss": 0.3322,
"step": 355
},
{
"epoch": 0.006517162471395881,
"grad_norm": 5.379430294036865,
"learning_rate": 1.6293203910498375e-05,
"loss": 0.4546,
"step": 356
},
{
"epoch": 0.006535469107551487,
"grad_norm": 3.9145123958587646,
"learning_rate": 1.6266038113644605e-05,
"loss": 0.3207,
"step": 357
},
{
"epoch": 0.006553775743707094,
"grad_norm": 6.354018211364746,
"learning_rate": 1.6238795967093865e-05,
"loss": 0.4729,
"step": 358
},
{
"epoch": 0.0065720823798627,
"grad_norm": 9.218265533447266,
"learning_rate": 1.6211477802783105e-05,
"loss": 0.6496,
"step": 359
},
{
"epoch": 0.006590389016018307,
"grad_norm": 12.858467102050781,
"learning_rate": 1.6184083953575543e-05,
"loss": 0.4073,
"step": 360
},
{
"epoch": 0.006608695652173913,
"grad_norm": 8.530418395996094,
"learning_rate": 1.6156614753256583e-05,
"loss": 0.5826,
"step": 361
},
{
"epoch": 0.00662700228832952,
"grad_norm": 8.6444673538208,
"learning_rate": 1.6129070536529767e-05,
"loss": 0.6264,
"step": 362
},
{
"epoch": 0.0066453089244851255,
"grad_norm": 9.804875373840332,
"learning_rate": 1.610145163901268e-05,
"loss": 0.4984,
"step": 363
},
{
"epoch": 0.006663615560640732,
"grad_norm": 7.5225324630737305,
"learning_rate": 1.607375839723287e-05,
"loss": 0.4179,
"step": 364
},
{
"epoch": 0.006681922196796339,
"grad_norm": 6.881463527679443,
"learning_rate": 1.6045991148623752e-05,
"loss": 0.4956,
"step": 365
},
{
"epoch": 0.006700228832951945,
"grad_norm": 6.102224349975586,
"learning_rate": 1.6018150231520486e-05,
"loss": 0.4461,
"step": 366
},
{
"epoch": 0.006718535469107552,
"grad_norm": 11.278656005859375,
"learning_rate": 1.599023598515586e-05,
"loss": 0.8761,
"step": 367
},
{
"epoch": 0.006736842105263158,
"grad_norm": 5.988121032714844,
"learning_rate": 1.5962248749656158e-05,
"loss": 0.2178,
"step": 368
},
{
"epoch": 0.006755148741418764,
"grad_norm": 0.9311094880104065,
"learning_rate": 1.5934188866037017e-05,
"loss": 0.0839,
"step": 369
},
{
"epoch": 0.0067734553775743705,
"grad_norm": 5.130373001098633,
"learning_rate": 1.5906056676199256e-05,
"loss": 0.1672,
"step": 370
},
{
"epoch": 0.006791762013729977,
"grad_norm": 5.0459418296813965,
"learning_rate": 1.5877852522924733e-05,
"loss": 0.2407,
"step": 371
},
{
"epoch": 0.006810068649885584,
"grad_norm": 2.93607497215271,
"learning_rate": 1.584957674987216e-05,
"loss": 0.1925,
"step": 372
},
{
"epoch": 0.00682837528604119,
"grad_norm": 12.006101608276367,
"learning_rate": 1.5821229701572897e-05,
"loss": 0.8912,
"step": 373
},
{
"epoch": 0.006846681922196797,
"grad_norm": 15.027911186218262,
"learning_rate": 1.5792811723426787e-05,
"loss": 0.4139,
"step": 374
},
{
"epoch": 0.006864988558352402,
"grad_norm": 5.793044090270996,
"learning_rate": 1.5764323161697933e-05,
"loss": 0.2851,
"step": 375
},
{
"epoch": 0.006883295194508009,
"grad_norm": 8.51421070098877,
"learning_rate": 1.573576436351046e-05,
"loss": 0.5144,
"step": 376
},
{
"epoch": 0.006901601830663615,
"grad_norm": 9.018112182617188,
"learning_rate": 1.570713567684432e-05,
"loss": 0.4011,
"step": 377
},
{
"epoch": 0.006919908466819222,
"grad_norm": 12.444863319396973,
"learning_rate": 1.5678437450531014e-05,
"loss": 0.8046,
"step": 378
},
{
"epoch": 0.0069382151029748285,
"grad_norm": 7.43892765045166,
"learning_rate": 1.564967003424938e-05,
"loss": 0.4985,
"step": 379
},
{
"epoch": 0.006956521739130435,
"grad_norm": 3.2682807445526123,
"learning_rate": 1.5620833778521306e-05,
"loss": 0.3606,
"step": 380
},
{
"epoch": 0.006974828375286042,
"grad_norm": 6.130132675170898,
"learning_rate": 1.5591929034707468e-05,
"loss": 0.5154,
"step": 381
},
{
"epoch": 0.006993135011441647,
"grad_norm": 5.139265060424805,
"learning_rate": 1.556295615500305e-05,
"loss": 0.3442,
"step": 382
},
{
"epoch": 0.007011441647597254,
"grad_norm": 5.858025074005127,
"learning_rate": 1.553391549243344e-05,
"loss": 0.316,
"step": 383
},
{
"epoch": 0.00702974828375286,
"grad_norm": 5.114577293395996,
"learning_rate": 1.5504807400849957e-05,
"loss": 0.3712,
"step": 384
},
{
"epoch": 0.007048054919908467,
"grad_norm": 11.650129318237305,
"learning_rate": 1.5475632234925505e-05,
"loss": 0.2964,
"step": 385
},
{
"epoch": 0.007066361556064073,
"grad_norm": 4.760644435882568,
"learning_rate": 1.5446390350150272e-05,
"loss": 0.4069,
"step": 386
},
{
"epoch": 0.00708466819221968,
"grad_norm": 4.894204139709473,
"learning_rate": 1.54170821028274e-05,
"loss": 0.4782,
"step": 387
},
{
"epoch": 0.007102974828375286,
"grad_norm": 4.8495001792907715,
"learning_rate": 1.5387707850068633e-05,
"loss": 0.3845,
"step": 388
},
{
"epoch": 0.007121281464530892,
"grad_norm": 3.118293285369873,
"learning_rate": 1.5358267949789968e-05,
"loss": 0.17,
"step": 389
},
{
"epoch": 0.007139588100686499,
"grad_norm": 4.566890239715576,
"learning_rate": 1.53287627607073e-05,
"loss": 0.207,
"step": 390
},
{
"epoch": 0.007157894736842105,
"grad_norm": 9.8530855178833,
"learning_rate": 1.529919264233205e-05,
"loss": 0.5756,
"step": 391
},
{
"epoch": 0.007176201372997712,
"grad_norm": 3.8654870986938477,
"learning_rate": 1.5269557954966777e-05,
"loss": 0.2833,
"step": 392
},
{
"epoch": 0.007194508009153318,
"grad_norm": 8.076715469360352,
"learning_rate": 1.5239859059700794e-05,
"loss": 0.434,
"step": 393
},
{
"epoch": 0.007212814645308925,
"grad_norm": 2.7720115184783936,
"learning_rate": 1.5210096318405768e-05,
"loss": 0.1749,
"step": 394
},
{
"epoch": 0.007231121281464531,
"grad_norm": 3.135864496231079,
"learning_rate": 1.5180270093731305e-05,
"loss": 0.0707,
"step": 395
},
{
"epoch": 0.007249427917620137,
"grad_norm": 7.354649543762207,
"learning_rate": 1.5150380749100545e-05,
"loss": 0.5681,
"step": 396
},
{
"epoch": 0.007267734553775744,
"grad_norm": 5.514926433563232,
"learning_rate": 1.5120428648705716e-05,
"loss": 0.2812,
"step": 397
},
{
"epoch": 0.00728604118993135,
"grad_norm": 4.094933986663818,
"learning_rate": 1.5090414157503715e-05,
"loss": 0.189,
"step": 398
},
{
"epoch": 0.007304347826086957,
"grad_norm": 1.2968307733535767,
"learning_rate": 1.5060337641211637e-05,
"loss": 0.1388,
"step": 399
},
{
"epoch": 0.007322654462242563,
"grad_norm": 9.890887260437012,
"learning_rate": 1.5030199466302354e-05,
"loss": 0.4099,
"step": 400
},
{
"epoch": 0.007340961098398169,
"grad_norm": 7.428032875061035,
"learning_rate": 1.5000000000000002e-05,
"loss": 0.5361,
"step": 401
},
{
"epoch": 0.0073592677345537755,
"grad_norm": 6.782509803771973,
"learning_rate": 1.4969739610275556e-05,
"loss": 0.3347,
"step": 402
},
{
"epoch": 0.007377574370709382,
"grad_norm": 8.611628532409668,
"learning_rate": 1.493941866584231e-05,
"loss": 0.6473,
"step": 403
},
{
"epoch": 0.007395881006864989,
"grad_norm": 11.101713180541992,
"learning_rate": 1.490903753615141e-05,
"loss": 0.5101,
"step": 404
},
{
"epoch": 0.007414187643020595,
"grad_norm": 2.3792383670806885,
"learning_rate": 1.4878596591387329e-05,
"loss": 0.2176,
"step": 405
},
{
"epoch": 0.007432494279176202,
"grad_norm": 3.647657871246338,
"learning_rate": 1.4848096202463373e-05,
"loss": 0.1662,
"step": 406
},
{
"epoch": 0.007450800915331807,
"grad_norm": 2.6126086711883545,
"learning_rate": 1.4817536741017153e-05,
"loss": 0.223,
"step": 407
},
{
"epoch": 0.007469107551487414,
"grad_norm": 7.960146903991699,
"learning_rate": 1.478691857940607e-05,
"loss": 0.4825,
"step": 408
},
{
"epoch": 0.0074874141876430204,
"grad_norm": 2.9891467094421387,
"learning_rate": 1.4756242090702756e-05,
"loss": 0.2072,
"step": 409
},
{
"epoch": 0.007505720823798627,
"grad_norm": 7.48441743850708,
"learning_rate": 1.4725507648690542e-05,
"loss": 0.6046,
"step": 410
},
{
"epoch": 0.0075240274599542335,
"grad_norm": 4.050131320953369,
"learning_rate": 1.469471562785891e-05,
"loss": 0.343,
"step": 411
},
{
"epoch": 0.00754233409610984,
"grad_norm": 7.872916221618652,
"learning_rate": 1.4663866403398915e-05,
"loss": 0.6128,
"step": 412
},
{
"epoch": 0.007560640732265447,
"grad_norm": 4.682326793670654,
"learning_rate": 1.463296035119862e-05,
"loss": 0.3317,
"step": 413
},
{
"epoch": 0.007578947368421052,
"grad_norm": 4.932008266448975,
"learning_rate": 1.4601997847838518e-05,
"loss": 0.2332,
"step": 414
},
{
"epoch": 0.007597254004576659,
"grad_norm": 7.188138008117676,
"learning_rate": 1.4570979270586944e-05,
"loss": 0.5127,
"step": 415
},
{
"epoch": 0.007615560640732265,
"grad_norm": 6.72917366027832,
"learning_rate": 1.4539904997395468e-05,
"loss": 0.2865,
"step": 416
},
{
"epoch": 0.007633867276887872,
"grad_norm": 6.936765670776367,
"learning_rate": 1.4508775406894308e-05,
"loss": 0.4261,
"step": 417
},
{
"epoch": 0.0076521739130434785,
"grad_norm": 6.002128601074219,
"learning_rate": 1.4477590878387697e-05,
"loss": 0.3089,
"step": 418
},
{
"epoch": 0.007670480549199085,
"grad_norm": 3.699465036392212,
"learning_rate": 1.4446351791849276e-05,
"loss": 0.2833,
"step": 419
},
{
"epoch": 0.007688787185354691,
"grad_norm": 5.764729022979736,
"learning_rate": 1.4415058527917454e-05,
"loss": 0.3462,
"step": 420
},
{
"epoch": 0.007707093821510297,
"grad_norm": 2.2470004558563232,
"learning_rate": 1.4383711467890776e-05,
"loss": 0.1425,
"step": 421
},
{
"epoch": 0.007725400457665904,
"grad_norm": 3.7865355014801025,
"learning_rate": 1.4352310993723277e-05,
"loss": 0.1958,
"step": 422
},
{
"epoch": 0.00774370709382151,
"grad_norm": 5.835235595703125,
"learning_rate": 1.4320857488019826e-05,
"loss": 0.3167,
"step": 423
},
{
"epoch": 0.007762013729977117,
"grad_norm": 4.612227439880371,
"learning_rate": 1.4289351334031461e-05,
"loss": 0.259,
"step": 424
},
{
"epoch": 0.007780320366132723,
"grad_norm": 5.527163982391357,
"learning_rate": 1.4257792915650728e-05,
"loss": 0.3875,
"step": 425
},
{
"epoch": 0.00779862700228833,
"grad_norm": 1.5776690244674683,
"learning_rate": 1.4226182617406996e-05,
"loss": 0.1545,
"step": 426
},
{
"epoch": 0.007816933638443936,
"grad_norm": 5.22613525390625,
"learning_rate": 1.4194520824461773e-05,
"loss": 0.2102,
"step": 427
},
{
"epoch": 0.007835240274599543,
"grad_norm": 8.434508323669434,
"learning_rate": 1.4162807922604014e-05,
"loss": 0.4331,
"step": 428
},
{
"epoch": 0.00785354691075515,
"grad_norm": 6.714724063873291,
"learning_rate": 1.413104429824542e-05,
"loss": 0.4454,
"step": 429
},
{
"epoch": 0.007871853546910754,
"grad_norm": 5.015308856964111,
"learning_rate": 1.4099230338415728e-05,
"loss": 0.1637,
"step": 430
},
{
"epoch": 0.007890160183066361,
"grad_norm": 5.9333367347717285,
"learning_rate": 1.4067366430758004e-05,
"loss": 0.3313,
"step": 431
},
{
"epoch": 0.007908466819221967,
"grad_norm": 2.7909371852874756,
"learning_rate": 1.4035452963523903e-05,
"loss": 0.1595,
"step": 432
},
{
"epoch": 0.007926773455377574,
"grad_norm": 12.026054382324219,
"learning_rate": 1.4003490325568953e-05,
"loss": 0.5666,
"step": 433
},
{
"epoch": 0.00794508009153318,
"grad_norm": 9.2349271774292,
"learning_rate": 1.3971478906347806e-05,
"loss": 0.4163,
"step": 434
},
{
"epoch": 0.007963386727688787,
"grad_norm": 4.861546516418457,
"learning_rate": 1.3939419095909513e-05,
"loss": 0.2674,
"step": 435
},
{
"epoch": 0.007981693363844394,
"grad_norm": 6.877256393432617,
"learning_rate": 1.3907311284892737e-05,
"loss": 0.2162,
"step": 436
},
{
"epoch": 0.008,
"grad_norm": 9.61545467376709,
"learning_rate": 1.3875155864521031e-05,
"loss": 0.5273,
"step": 437
},
{
"epoch": 0.008018306636155607,
"grad_norm": 4.861330509185791,
"learning_rate": 1.3842953226598036e-05,
"loss": 0.1181,
"step": 438
},
{
"epoch": 0.008036613272311213,
"grad_norm": 12.819157600402832,
"learning_rate": 1.3810703763502744e-05,
"loss": 0.9099,
"step": 439
},
{
"epoch": 0.00805491990846682,
"grad_norm": 11.029340744018555,
"learning_rate": 1.3778407868184674e-05,
"loss": 0.5066,
"step": 440
},
{
"epoch": 0.008073226544622426,
"grad_norm": 12.776625633239746,
"learning_rate": 1.3746065934159123e-05,
"loss": 0.8391,
"step": 441
},
{
"epoch": 0.008091533180778033,
"grad_norm": 10.414067268371582,
"learning_rate": 1.371367835550235e-05,
"loss": 0.579,
"step": 442
},
{
"epoch": 0.008109839816933638,
"grad_norm": 2.4149062633514404,
"learning_rate": 1.3681245526846782e-05,
"loss": 0.2106,
"step": 443
},
{
"epoch": 0.008128146453089244,
"grad_norm": 3.4909825325012207,
"learning_rate": 1.3648767843376196e-05,
"loss": 0.2114,
"step": 444
},
{
"epoch": 0.00814645308924485,
"grad_norm": 2.9647722244262695,
"learning_rate": 1.3616245700820922e-05,
"loss": 0.2153,
"step": 445
},
{
"epoch": 0.008164759725400457,
"grad_norm": 2.9645230770111084,
"learning_rate": 1.3583679495453e-05,
"loss": 0.1417,
"step": 446
},
{
"epoch": 0.008183066361556064,
"grad_norm": 5.128696918487549,
"learning_rate": 1.3551069624081372e-05,
"loss": 0.2745,
"step": 447
},
{
"epoch": 0.00820137299771167,
"grad_norm": 8.987137794494629,
"learning_rate": 1.3518416484047018e-05,
"loss": 0.5339,
"step": 448
},
{
"epoch": 0.008219679633867277,
"grad_norm": 9.38960075378418,
"learning_rate": 1.3485720473218153e-05,
"loss": 0.4518,
"step": 449
},
{
"epoch": 0.008237986270022883,
"grad_norm": 2.744861364364624,
"learning_rate": 1.3452981989985347e-05,
"loss": 0.153,
"step": 450
},
{
"epoch": 0.00825629290617849,
"grad_norm": 6.3267998695373535,
"learning_rate": 1.342020143325669e-05,
"loss": 0.4077,
"step": 451
},
{
"epoch": 0.008274599542334097,
"grad_norm": 9.530548095703125,
"learning_rate": 1.3387379202452917e-05,
"loss": 0.5712,
"step": 452
},
{
"epoch": 0.008292906178489703,
"grad_norm": 1.2639689445495605,
"learning_rate": 1.3354515697502552e-05,
"loss": 0.0901,
"step": 453
},
{
"epoch": 0.00831121281464531,
"grad_norm": 10.240217208862305,
"learning_rate": 1.3321611318837033e-05,
"loss": 0.4039,
"step": 454
},
{
"epoch": 0.008329519450800914,
"grad_norm": 7.718742370605469,
"learning_rate": 1.3288666467385834e-05,
"loss": 0.5048,
"step": 455
},
{
"epoch": 0.008347826086956521,
"grad_norm": 1.411474347114563,
"learning_rate": 1.3255681544571568e-05,
"loss": 0.1291,
"step": 456
},
{
"epoch": 0.008366132723112128,
"grad_norm": 3.7180333137512207,
"learning_rate": 1.3222656952305113e-05,
"loss": 0.1669,
"step": 457
},
{
"epoch": 0.008384439359267734,
"grad_norm": 4.709652423858643,
"learning_rate": 1.3189593092980701e-05,
"loss": 0.2531,
"step": 458
},
{
"epoch": 0.00840274599542334,
"grad_norm": 2.9395062923431396,
"learning_rate": 1.3156490369471026e-05,
"loss": 0.1908,
"step": 459
},
{
"epoch": 0.008421052631578947,
"grad_norm": 1.6285467147827148,
"learning_rate": 1.3123349185122328e-05,
"loss": 0.1829,
"step": 460
},
{
"epoch": 0.008439359267734554,
"grad_norm": 2.359907627105713,
"learning_rate": 1.3090169943749475e-05,
"loss": 0.078,
"step": 461
},
{
"epoch": 0.00845766590389016,
"grad_norm": 5.721633434295654,
"learning_rate": 1.3056953049631059e-05,
"loss": 0.5493,
"step": 462
},
{
"epoch": 0.008475972540045767,
"grad_norm": 7.441493034362793,
"learning_rate": 1.3023698907504447e-05,
"loss": 0.3625,
"step": 463
},
{
"epoch": 0.008494279176201373,
"grad_norm": 8.423327445983887,
"learning_rate": 1.2990407922560869e-05,
"loss": 0.841,
"step": 464
},
{
"epoch": 0.00851258581235698,
"grad_norm": 13.417081832885742,
"learning_rate": 1.2957080500440469e-05,
"loss": 0.3607,
"step": 465
},
{
"epoch": 0.008530892448512586,
"grad_norm": 19.080902099609375,
"learning_rate": 1.2923717047227368e-05,
"loss": 0.8309,
"step": 466
},
{
"epoch": 0.008549199084668193,
"grad_norm": 5.821935176849365,
"learning_rate": 1.2890317969444716e-05,
"loss": 0.1997,
"step": 467
},
{
"epoch": 0.008567505720823798,
"grad_norm": 6.380805492401123,
"learning_rate": 1.2856883674049736e-05,
"loss": 0.3366,
"step": 468
},
{
"epoch": 0.008585812356979404,
"grad_norm": 4.085707664489746,
"learning_rate": 1.2823414568428767e-05,
"loss": 0.2064,
"step": 469
},
{
"epoch": 0.008604118993135011,
"grad_norm": 1.9464811086654663,
"learning_rate": 1.2789911060392295e-05,
"loss": 0.1908,
"step": 470
},
{
"epoch": 0.008622425629290617,
"grad_norm": 3.138880729675293,
"learning_rate": 1.2756373558169992e-05,
"loss": 0.1714,
"step": 471
},
{
"epoch": 0.008640732265446224,
"grad_norm": 4.658113479614258,
"learning_rate": 1.2722802470405744e-05,
"loss": 0.3097,
"step": 472
},
{
"epoch": 0.00865903890160183,
"grad_norm": 4.06643533706665,
"learning_rate": 1.2689198206152657e-05,
"loss": 0.2363,
"step": 473
},
{
"epoch": 0.008677345537757437,
"grad_norm": 7.818174839019775,
"learning_rate": 1.265556117486809e-05,
"loss": 0.3208,
"step": 474
},
{
"epoch": 0.008695652173913044,
"grad_norm": 5.261242866516113,
"learning_rate": 1.2621891786408648e-05,
"loss": 0.3357,
"step": 475
},
{
"epoch": 0.00871395881006865,
"grad_norm": 6.41402006149292,
"learning_rate": 1.2588190451025209e-05,
"loss": 0.2269,
"step": 476
},
{
"epoch": 0.008732265446224257,
"grad_norm": 3.909989833831787,
"learning_rate": 1.2554457579357906e-05,
"loss": 0.2456,
"step": 477
},
{
"epoch": 0.008750572082379863,
"grad_norm": 6.242023944854736,
"learning_rate": 1.252069358243114e-05,
"loss": 0.4592,
"step": 478
},
{
"epoch": 0.00876887871853547,
"grad_norm": 11.872090339660645,
"learning_rate": 1.2486898871648552e-05,
"loss": 0.6729,
"step": 479
},
{
"epoch": 0.008787185354691076,
"grad_norm": 5.960324287414551,
"learning_rate": 1.2453073858788027e-05,
"loss": 0.4166,
"step": 480
},
{
"epoch": 0.008805491990846681,
"grad_norm": 1.5860638618469238,
"learning_rate": 1.2419218955996677e-05,
"loss": 0.1463,
"step": 481
},
{
"epoch": 0.008823798627002288,
"grad_norm": 2.930913209915161,
"learning_rate": 1.238533457578581e-05,
"loss": 0.1741,
"step": 482
},
{
"epoch": 0.008842105263157894,
"grad_norm": 7.377974033355713,
"learning_rate": 1.23514211310259e-05,
"loss": 0.4284,
"step": 483
},
{
"epoch": 0.0088604118993135,
"grad_norm": 3.70469069480896,
"learning_rate": 1.2317479034941572e-05,
"loss": 0.2728,
"step": 484
},
{
"epoch": 0.008878718535469107,
"grad_norm": 5.633390426635742,
"learning_rate": 1.2283508701106559e-05,
"loss": 0.3421,
"step": 485
},
{
"epoch": 0.008897025171624714,
"grad_norm": 0.9088148474693298,
"learning_rate": 1.2249510543438652e-05,
"loss": 0.1071,
"step": 486
},
{
"epoch": 0.00891533180778032,
"grad_norm": 10.652883529663086,
"learning_rate": 1.2215484976194675e-05,
"loss": 0.4244,
"step": 487
},
{
"epoch": 0.008933638443935927,
"grad_norm": 5.040556907653809,
"learning_rate": 1.2181432413965428e-05,
"loss": 0.2866,
"step": 488
},
{
"epoch": 0.008951945080091533,
"grad_norm": 5.78645133972168,
"learning_rate": 1.2147353271670634e-05,
"loss": 0.2152,
"step": 489
},
{
"epoch": 0.00897025171624714,
"grad_norm": 8.46097183227539,
"learning_rate": 1.211324796455389e-05,
"loss": 0.6545,
"step": 490
},
{
"epoch": 0.008988558352402747,
"grad_norm": 5.180769443511963,
"learning_rate": 1.2079116908177592e-05,
"loss": 0.234,
"step": 491
},
{
"epoch": 0.009006864988558353,
"grad_norm": 5.95151424407959,
"learning_rate": 1.2044960518417902e-05,
"loss": 0.2579,
"step": 492
},
{
"epoch": 0.00902517162471396,
"grad_norm": 7.002593994140625,
"learning_rate": 1.2010779211459649e-05,
"loss": 0.5269,
"step": 493
},
{
"epoch": 0.009043478260869564,
"grad_norm": 7.671480178833008,
"learning_rate": 1.1976573403791263e-05,
"loss": 0.4812,
"step": 494
},
{
"epoch": 0.009061784897025171,
"grad_norm": 7.638554096221924,
"learning_rate": 1.194234351219972e-05,
"loss": 0.3599,
"step": 495
},
{
"epoch": 0.009080091533180778,
"grad_norm": 6.174886226654053,
"learning_rate": 1.190808995376545e-05,
"loss": 0.275,
"step": 496
},
{
"epoch": 0.009098398169336384,
"grad_norm": 3.23649525642395,
"learning_rate": 1.187381314585725e-05,
"loss": 0.179,
"step": 497
},
{
"epoch": 0.00911670480549199,
"grad_norm": 6.402664661407471,
"learning_rate": 1.1839513506127202e-05,
"loss": 0.3912,
"step": 498
},
{
"epoch": 0.009135011441647597,
"grad_norm": 7.799668788909912,
"learning_rate": 1.1805191452505602e-05,
"loss": 0.3725,
"step": 499
},
{
"epoch": 0.009153318077803204,
"grad_norm": 13.813634872436523,
"learning_rate": 1.1770847403195836e-05,
"loss": 0.7032,
"step": 500
},
{
"epoch": 0.00917162471395881,
"grad_norm": 2.5564143657684326,
"learning_rate": 1.1736481776669307e-05,
"loss": 0.129,
"step": 501
},
{
"epoch": 0.009189931350114417,
"grad_norm": 4.1047821044921875,
"learning_rate": 1.1702094991660326e-05,
"loss": 0.2844,
"step": 502
},
{
"epoch": 0.009208237986270023,
"grad_norm": 8.63683032989502,
"learning_rate": 1.1667687467161025e-05,
"loss": 0.4529,
"step": 503
},
{
"epoch": 0.00922654462242563,
"grad_norm": 5.256021976470947,
"learning_rate": 1.1633259622416224e-05,
"loss": 0.4106,
"step": 504
},
{
"epoch": 0.009244851258581236,
"grad_norm": 4.8041486740112305,
"learning_rate": 1.159881187691835e-05,
"loss": 0.2019,
"step": 505
},
{
"epoch": 0.009263157894736843,
"grad_norm": 5.618558406829834,
"learning_rate": 1.156434465040231e-05,
"loss": 0.3514,
"step": 506
},
{
"epoch": 0.009281464530892448,
"grad_norm": 3.8470354080200195,
"learning_rate": 1.1529858362840383e-05,
"loss": 0.2582,
"step": 507
},
{
"epoch": 0.009299771167048054,
"grad_norm": 4.10807991027832,
"learning_rate": 1.1495353434437098e-05,
"loss": 0.2775,
"step": 508
},
{
"epoch": 0.00931807780320366,
"grad_norm": 3.2404677867889404,
"learning_rate": 1.1460830285624119e-05,
"loss": 0.0828,
"step": 509
},
{
"epoch": 0.009336384439359267,
"grad_norm": 4.644530296325684,
"learning_rate": 1.1426289337055119e-05,
"loss": 0.2594,
"step": 510
},
{
"epoch": 0.009354691075514874,
"grad_norm": 4.381004810333252,
"learning_rate": 1.1391731009600655e-05,
"loss": 0.3301,
"step": 511
},
{
"epoch": 0.00937299771167048,
"grad_norm": 2.8095078468322754,
"learning_rate": 1.1357155724343046e-05,
"loss": 0.1854,
"step": 512
},
{
"epoch": 0.009391304347826087,
"grad_norm": 4.646561622619629,
"learning_rate": 1.1322563902571227e-05,
"loss": 0.2856,
"step": 513
},
{
"epoch": 0.009409610983981694,
"grad_norm": 2.435589075088501,
"learning_rate": 1.128795596577563e-05,
"loss": 0.158,
"step": 514
},
{
"epoch": 0.0094279176201373,
"grad_norm": 5.535358905792236,
"learning_rate": 1.1253332335643043e-05,
"loss": 0.4016,
"step": 515
},
{
"epoch": 0.009446224256292907,
"grad_norm": 2.8487491607666016,
"learning_rate": 1.1218693434051475e-05,
"loss": 0.1856,
"step": 516
},
{
"epoch": 0.009464530892448513,
"grad_norm": 6.440304756164551,
"learning_rate": 1.1184039683065014e-05,
"loss": 0.4992,
"step": 517
},
{
"epoch": 0.00948283752860412,
"grad_norm": 4.590803146362305,
"learning_rate": 1.1149371504928667e-05,
"loss": 0.3192,
"step": 518
},
{
"epoch": 0.009501144164759726,
"grad_norm": 3.818535566329956,
"learning_rate": 1.1114689322063255e-05,
"loss": 0.2371,
"step": 519
},
{
"epoch": 0.009519450800915331,
"grad_norm": 6.345479965209961,
"learning_rate": 1.1079993557060228e-05,
"loss": 0.3573,
"step": 520
},
{
"epoch": 0.009537757437070938,
"grad_norm": 13.15767765045166,
"learning_rate": 1.1045284632676535e-05,
"loss": 0.4145,
"step": 521
},
{
"epoch": 0.009556064073226544,
"grad_norm": 2.1148345470428467,
"learning_rate": 1.1010562971829464e-05,
"loss": 0.1284,
"step": 522
},
{
"epoch": 0.00957437070938215,
"grad_norm": 5.396875858306885,
"learning_rate": 1.0975828997591496e-05,
"loss": 0.6029,
"step": 523
},
{
"epoch": 0.009592677345537757,
"grad_norm": 5.095512866973877,
"learning_rate": 1.0941083133185146e-05,
"loss": 0.316,
"step": 524
},
{
"epoch": 0.009610983981693364,
"grad_norm": 8.121919631958008,
"learning_rate": 1.0906325801977804e-05,
"loss": 0.3049,
"step": 525
},
{
"epoch": 0.00962929061784897,
"grad_norm": 3.969857931137085,
"learning_rate": 1.0871557427476585e-05,
"loss": 0.1074,
"step": 526
},
{
"epoch": 0.009647597254004577,
"grad_norm": 12.454261779785156,
"learning_rate": 1.083677843332316e-05,
"loss": 0.6658,
"step": 527
},
{
"epoch": 0.009665903890160183,
"grad_norm": 4.157962799072266,
"learning_rate": 1.0801989243288588e-05,
"loss": 0.343,
"step": 528
},
{
"epoch": 0.00968421052631579,
"grad_norm": 0.6788906455039978,
"learning_rate": 1.0767190281268187e-05,
"loss": 0.0834,
"step": 529
},
{
"epoch": 0.009702517162471396,
"grad_norm": 7.104922294616699,
"learning_rate": 1.0732381971276318e-05,
"loss": 0.4511,
"step": 530
},
{
"epoch": 0.009720823798627003,
"grad_norm": 10.119832038879395,
"learning_rate": 1.0697564737441254e-05,
"loss": 0.4547,
"step": 531
},
{
"epoch": 0.009739130434782608,
"grad_norm": 1.6746511459350586,
"learning_rate": 1.0662739004000005e-05,
"loss": 0.1375,
"step": 532
},
{
"epoch": 0.009757437070938214,
"grad_norm": 9.788320541381836,
"learning_rate": 1.0627905195293135e-05,
"loss": 0.4845,
"step": 533
},
{
"epoch": 0.009775743707093821,
"grad_norm": 9.758012771606445,
"learning_rate": 1.0593063735759619e-05,
"loss": 0.4237,
"step": 534
},
{
"epoch": 0.009794050343249427,
"grad_norm": 2.4389655590057373,
"learning_rate": 1.055821504993164e-05,
"loss": 0.1618,
"step": 535
},
{
"epoch": 0.009812356979405034,
"grad_norm": 10.49396800994873,
"learning_rate": 1.0523359562429441e-05,
"loss": 0.5892,
"step": 536
},
{
"epoch": 0.00983066361556064,
"grad_norm": 5.466527462005615,
"learning_rate": 1.0488497697956134e-05,
"loss": 0.2774,
"step": 537
},
{
"epoch": 0.009848970251716247,
"grad_norm": 2.714876890182495,
"learning_rate": 1.0453629881292537e-05,
"loss": 0.1092,
"step": 538
},
{
"epoch": 0.009867276887871854,
"grad_norm": 4.077674865722656,
"learning_rate": 1.0418756537291996e-05,
"loss": 0.2116,
"step": 539
},
{
"epoch": 0.00988558352402746,
"grad_norm": 3.0669403076171875,
"learning_rate": 1.03838780908752e-05,
"loss": 0.1422,
"step": 540
},
{
"epoch": 0.009903890160183067,
"grad_norm": 8.089181900024414,
"learning_rate": 1.0348994967025012e-05,
"loss": 0.4507,
"step": 541
},
{
"epoch": 0.009922196796338673,
"grad_norm": 3.917191982269287,
"learning_rate": 1.0314107590781284e-05,
"loss": 0.2293,
"step": 542
},
{
"epoch": 0.00994050343249428,
"grad_norm": 3.5208070278167725,
"learning_rate": 1.0279216387235691e-05,
"loss": 0.201,
"step": 543
},
{
"epoch": 0.009958810068649886,
"grad_norm": 11.041540145874023,
"learning_rate": 1.0244321781526533e-05,
"loss": 0.4846,
"step": 544
},
{
"epoch": 0.009977116704805491,
"grad_norm": 5.149035930633545,
"learning_rate": 1.0209424198833571e-05,
"loss": 0.2859,
"step": 545
},
{
"epoch": 0.009995423340961098,
"grad_norm": 4.3648762702941895,
"learning_rate": 1.0174524064372837e-05,
"loss": 0.3187,
"step": 546
},
{
"epoch": 0.010013729977116704,
"grad_norm": 14.32425594329834,
"learning_rate": 1.0139621803391454e-05,
"loss": 0.7906,
"step": 547
},
{
"epoch": 0.01003203661327231,
"grad_norm": 7.4310455322265625,
"learning_rate": 1.010471784116246e-05,
"loss": 0.3219,
"step": 548
},
{
"epoch": 0.010050343249427917,
"grad_norm": 4.170273780822754,
"learning_rate": 1.0069812602979617e-05,
"loss": 0.23,
"step": 549
},
{
"epoch": 0.010068649885583524,
"grad_norm": 11.858202934265137,
"learning_rate": 1.0034906514152239e-05,
"loss": 0.4112,
"step": 550
},
{
"epoch": 0.01008695652173913,
"grad_norm": 7.049015998840332,
"learning_rate": 1e-05,
"loss": 0.2449,
"step": 551
},
{
"epoch": 0.010105263157894737,
"grad_norm": 8.323487281799316,
"learning_rate": 9.965093485847766e-06,
"loss": 0.3886,
"step": 552
},
{
"epoch": 0.010123569794050344,
"grad_norm": 1.819872498512268,
"learning_rate": 9.930187397020385e-06,
"loss": 0.0831,
"step": 553
},
{
"epoch": 0.01014187643020595,
"grad_norm": 8.343167304992676,
"learning_rate": 9.895282158837545e-06,
"loss": 0.7165,
"step": 554
},
{
"epoch": 0.010160183066361557,
"grad_norm": 6.083837985992432,
"learning_rate": 9.860378196608549e-06,
"loss": 0.3716,
"step": 555
},
{
"epoch": 0.010178489702517163,
"grad_norm": 5.233552932739258,
"learning_rate": 9.825475935627165e-06,
"loss": 0.3486,
"step": 556
},
{
"epoch": 0.01019679633867277,
"grad_norm": 2.339454174041748,
"learning_rate": 9.790575801166432e-06,
"loss": 0.1367,
"step": 557
},
{
"epoch": 0.010215102974828374,
"grad_norm": 5.374737739562988,
"learning_rate": 9.75567821847347e-06,
"loss": 0.4209,
"step": 558
},
{
"epoch": 0.010233409610983981,
"grad_norm": 8.515985488891602,
"learning_rate": 9.720783612764314e-06,
"loss": 0.3302,
"step": 559
},
{
"epoch": 0.010251716247139588,
"grad_norm": 9.269789695739746,
"learning_rate": 9.685892409218718e-06,
"loss": 0.2903,
"step": 560
},
{
"epoch": 0.010270022883295194,
"grad_norm": 3.935189723968506,
"learning_rate": 9.651005032974994e-06,
"loss": 0.3017,
"step": 561
},
{
"epoch": 0.0102883295194508,
"grad_norm": 10.850844383239746,
"learning_rate": 9.616121909124801e-06,
"loss": 0.4161,
"step": 562
},
{
"epoch": 0.010306636155606407,
"grad_norm": 9.301300048828125,
"learning_rate": 9.581243462708007e-06,
"loss": 0.5199,
"step": 563
},
{
"epoch": 0.010324942791762014,
"grad_norm": 5.9267191886901855,
"learning_rate": 9.546370118707463e-06,
"loss": 0.2799,
"step": 564
},
{
"epoch": 0.01034324942791762,
"grad_norm": 9.080803871154785,
"learning_rate": 9.511502302043867e-06,
"loss": 0.7154,
"step": 565
},
{
"epoch": 0.010361556064073227,
"grad_norm": 7.055878162384033,
"learning_rate": 9.476640437570562e-06,
"loss": 0.304,
"step": 566
},
{
"epoch": 0.010379862700228833,
"grad_norm": 2.7219274044036865,
"learning_rate": 9.441784950068362e-06,
"loss": 0.1827,
"step": 567
},
{
"epoch": 0.01039816933638444,
"grad_norm": 2.3669350147247314,
"learning_rate": 9.406936264240386e-06,
"loss": 0.1686,
"step": 568
},
{
"epoch": 0.010416475972540046,
"grad_norm": 4.634760856628418,
"learning_rate": 9.372094804706867e-06,
"loss": 0.2163,
"step": 569
},
{
"epoch": 0.010434782608695653,
"grad_norm": 1.3255306482315063,
"learning_rate": 9.337260996000002e-06,
"loss": 0.0883,
"step": 570
},
{
"epoch": 0.010453089244851258,
"grad_norm": 2.8958237171173096,
"learning_rate": 9.302435262558748e-06,
"loss": 0.1593,
"step": 571
},
{
"epoch": 0.010471395881006864,
"grad_norm": 8.792821884155273,
"learning_rate": 9.267618028723687e-06,
"loss": 0.4316,
"step": 572
},
{
"epoch": 0.010489702517162471,
"grad_norm": 11.037833213806152,
"learning_rate": 9.232809718731815e-06,
"loss": 0.6522,
"step": 573
},
{
"epoch": 0.010508009153318077,
"grad_norm": 2.741342067718506,
"learning_rate": 9.198010756711413e-06,
"loss": 0.1564,
"step": 574
},
{
"epoch": 0.010526315789473684,
"grad_norm": 5.826968193054199,
"learning_rate": 9.163221566676847e-06,
"loss": 0.294,
"step": 575
},
{
"epoch": 0.01054462242562929,
"grad_norm": 2.7347750663757324,
"learning_rate": 9.128442572523418e-06,
"loss": 0.1361,
"step": 576
},
{
"epoch": 0.010562929061784897,
"grad_norm": 8.53330135345459,
"learning_rate": 9.093674198022201e-06,
"loss": 0.275,
"step": 577
},
{
"epoch": 0.010581235697940504,
"grad_norm": 3.5783607959747314,
"learning_rate": 9.058916866814857e-06,
"loss": 0.2126,
"step": 578
},
{
"epoch": 0.01059954233409611,
"grad_norm": 12.933378219604492,
"learning_rate": 9.024171002408507e-06,
"loss": 0.5151,
"step": 579
},
{
"epoch": 0.010617848970251717,
"grad_norm": 5.769047737121582,
"learning_rate": 8.989437028170537e-06,
"loss": 0.2029,
"step": 580
},
{
"epoch": 0.010636155606407323,
"grad_norm": 10.376677513122559,
"learning_rate": 8.954715367323468e-06,
"loss": 0.5334,
"step": 581
},
{
"epoch": 0.01065446224256293,
"grad_norm": 11.327596664428711,
"learning_rate": 8.920006442939772e-06,
"loss": 0.3964,
"step": 582
},
{
"epoch": 0.010672768878718536,
"grad_norm": 7.816294193267822,
"learning_rate": 8.885310677936746e-06,
"loss": 0.3265,
"step": 583
},
{
"epoch": 0.010691075514874141,
"grad_norm": 14.427319526672363,
"learning_rate": 8.850628495071336e-06,
"loss": 0.6769,
"step": 584
},
{
"epoch": 0.010709382151029748,
"grad_norm": 2.7052078247070312,
"learning_rate": 8.815960316934991e-06,
"loss": 0.1255,
"step": 585
},
{
"epoch": 0.010727688787185354,
"grad_norm": 7.072470188140869,
"learning_rate": 8.781306565948528e-06,
"loss": 0.4344,
"step": 586
},
{
"epoch": 0.01074599542334096,
"grad_norm": 15.953747749328613,
"learning_rate": 8.746667664356957e-06,
"loss": 0.6664,
"step": 587
},
{
"epoch": 0.010764302059496567,
"grad_norm": 4.023654937744141,
"learning_rate": 8.712044034224374e-06,
"loss": 0.1642,
"step": 588
},
{
"epoch": 0.010782608695652174,
"grad_norm": 9.880155563354492,
"learning_rate": 8.677436097428775e-06,
"loss": 0.5635,
"step": 589
},
{
"epoch": 0.01080091533180778,
"grad_norm": 7.572893142700195,
"learning_rate": 8.642844275656957e-06,
"loss": 0.6194,
"step": 590
},
{
"epoch": 0.010819221967963387,
"grad_norm": 4.445793151855469,
"learning_rate": 8.60826899039935e-06,
"loss": 0.2096,
"step": 591
},
{
"epoch": 0.010837528604118993,
"grad_norm": 9.583978652954102,
"learning_rate": 8.573710662944884e-06,
"loss": 0.2836,
"step": 592
},
{
"epoch": 0.0108558352402746,
"grad_norm": 8.179230690002441,
"learning_rate": 8.539169714375885e-06,
"loss": 0.3744,
"step": 593
},
{
"epoch": 0.010874141876430207,
"grad_norm": 10.7511568069458,
"learning_rate": 8.504646565562907e-06,
"loss": 0.5062,
"step": 594
},
{
"epoch": 0.010892448512585813,
"grad_norm": 2.8209664821624756,
"learning_rate": 8.47014163715962e-06,
"loss": 0.1292,
"step": 595
},
{
"epoch": 0.010910755148741418,
"grad_norm": 5.161312580108643,
"learning_rate": 8.43565534959769e-06,
"loss": 0.2225,
"step": 596
},
{
"epoch": 0.010929061784897024,
"grad_norm": 6.31545877456665,
"learning_rate": 8.401188123081653e-06,
"loss": 0.4825,
"step": 597
},
{
"epoch": 0.010947368421052631,
"grad_norm": 7.428010940551758,
"learning_rate": 8.366740377583781e-06,
"loss": 0.28,
"step": 598
},
{
"epoch": 0.010965675057208238,
"grad_norm": 10.632474899291992,
"learning_rate": 8.332312532838978e-06,
"loss": 0.4979,
"step": 599
},
{
"epoch": 0.010983981693363844,
"grad_norm": 5.0757880210876465,
"learning_rate": 8.297905008339677e-06,
"loss": 0.4114,
"step": 600
},
{
"epoch": 0.01100228832951945,
"grad_norm": 10.615399360656738,
"learning_rate": 8.263518223330698e-06,
"loss": 0.4028,
"step": 601
},
{
"epoch": 0.011020594965675057,
"grad_norm": 3.5888783931732178,
"learning_rate": 8.22915259680417e-06,
"loss": 0.1922,
"step": 602
},
{
"epoch": 0.011038901601830664,
"grad_norm": 7.211901664733887,
"learning_rate": 8.194808547494401e-06,
"loss": 0.4079,
"step": 603
},
{
"epoch": 0.01105720823798627,
"grad_norm": 2.416304111480713,
"learning_rate": 8.1604864938728e-06,
"loss": 0.1201,
"step": 604
},
{
"epoch": 0.011075514874141877,
"grad_norm": 4.646475791931152,
"learning_rate": 8.126186854142752e-06,
"loss": 0.2859,
"step": 605
},
{
"epoch": 0.011093821510297483,
"grad_norm": 3.045999765396118,
"learning_rate": 8.091910046234552e-06,
"loss": 0.1901,
"step": 606
},
{
"epoch": 0.01111212814645309,
"grad_norm": 0.9611418843269348,
"learning_rate": 8.057656487800283e-06,
"loss": 0.1063,
"step": 607
},
{
"epoch": 0.011130434782608696,
"grad_norm": 2.64760160446167,
"learning_rate": 8.023426596208739e-06,
"loss": 0.1395,
"step": 608
},
{
"epoch": 0.011148741418764301,
"grad_norm": 12.34020709991455,
"learning_rate": 7.989220788540356e-06,
"loss": 0.4971,
"step": 609
},
{
"epoch": 0.011167048054919908,
"grad_norm": 5.713735103607178,
"learning_rate": 7.955039481582098e-06,
"loss": 0.1714,
"step": 610
},
{
"epoch": 0.011185354691075514,
"grad_norm": 18.283546447753906,
"learning_rate": 7.92088309182241e-06,
"loss": 0.6558,
"step": 611
},
{
"epoch": 0.01120366132723112,
"grad_norm": 12.52721118927002,
"learning_rate": 7.886752035446116e-06,
"loss": 0.5603,
"step": 612
},
{
"epoch": 0.011221967963386727,
"grad_norm": 4.516887664794922,
"learning_rate": 7.852646728329368e-06,
"loss": 0.2777,
"step": 613
},
{
"epoch": 0.011240274599542334,
"grad_norm": 12.416860580444336,
"learning_rate": 7.818567586034578e-06,
"loss": 0.5139,
"step": 614
},
{
"epoch": 0.01125858123569794,
"grad_norm": 7.2105255126953125,
"learning_rate": 7.784515023805328e-06,
"loss": 0.3159,
"step": 615
},
{
"epoch": 0.011276887871853547,
"grad_norm": 5.7121100425720215,
"learning_rate": 7.750489456561351e-06,
"loss": 0.2603,
"step": 616
},
{
"epoch": 0.011295194508009154,
"grad_norm": 8.093098640441895,
"learning_rate": 7.716491298893443e-06,
"loss": 0.4499,
"step": 617
},
{
"epoch": 0.01131350114416476,
"grad_norm": 4.29197883605957,
"learning_rate": 7.68252096505843e-06,
"loss": 0.1985,
"step": 618
},
{
"epoch": 0.011331807780320367,
"grad_norm": 2.614251136779785,
"learning_rate": 7.6485788689741e-06,
"loss": 0.1525,
"step": 619
},
{
"epoch": 0.011350114416475973,
"grad_norm": 7.588433742523193,
"learning_rate": 7.6146654242141935e-06,
"loss": 0.5013,
"step": 620
},
{
"epoch": 0.01136842105263158,
"grad_norm": 7.058767318725586,
"learning_rate": 7.580781044003324e-06,
"loss": 0.2994,
"step": 621
},
{
"epoch": 0.011386727688787185,
"grad_norm": 4.412848949432373,
"learning_rate": 7.546926141211975e-06,
"loss": 0.2056,
"step": 622
},
{
"epoch": 0.011405034324942791,
"grad_norm": 7.570004463195801,
"learning_rate": 7.513101128351454e-06,
"loss": 0.3943,
"step": 623
},
{
"epoch": 0.011423340961098398,
"grad_norm": 4.8941521644592285,
"learning_rate": 7.4793064175688635e-06,
"loss": 0.2364,
"step": 624
},
{
"epoch": 0.011441647597254004,
"grad_norm": 7.014512538909912,
"learning_rate": 7.445542420642097e-06,
"loss": 0.3111,
"step": 625
},
{
"epoch": 0.01145995423340961,
"grad_norm": 2.618640661239624,
"learning_rate": 7.411809548974792e-06,
"loss": 0.1614,
"step": 626
},
{
"epoch": 0.011478260869565217,
"grad_norm": 5.4425740242004395,
"learning_rate": 7.378108213591355e-06,
"loss": 0.3619,
"step": 627
},
{
"epoch": 0.011496567505720824,
"grad_norm": 3.9418482780456543,
"learning_rate": 7.344438825131912e-06,
"loss": 0.1967,
"step": 628
},
{
"epoch": 0.01151487414187643,
"grad_norm": 5.235764026641846,
"learning_rate": 7.310801793847344e-06,
"loss": 0.2332,
"step": 629
},
{
"epoch": 0.011533180778032037,
"grad_norm": 3.303947687149048,
"learning_rate": 7.277197529594257e-06,
"loss": 0.1708,
"step": 630
},
{
"epoch": 0.011551487414187643,
"grad_norm": 13.982658386230469,
"learning_rate": 7.243626441830009e-06,
"loss": 0.62,
"step": 631
},
{
"epoch": 0.01156979405034325,
"grad_norm": 5.784631252288818,
"learning_rate": 7.210088939607709e-06,
"loss": 0.3401,
"step": 632
},
{
"epoch": 0.011588100686498857,
"grad_norm": 5.933520793914795,
"learning_rate": 7.176585431571235e-06,
"loss": 0.3589,
"step": 633
},
{
"epoch": 0.011606407322654463,
"grad_norm": 12.28591537475586,
"learning_rate": 7.143116325950266e-06,
"loss": 0.6409,
"step": 634
},
{
"epoch": 0.011624713958810068,
"grad_norm": 1.4337087869644165,
"learning_rate": 7.109682030555283e-06,
"loss": 0.0955,
"step": 635
},
{
"epoch": 0.011643020594965674,
"grad_norm": 3.1845862865448,
"learning_rate": 7.076282952772634e-06,
"loss": 0.2039,
"step": 636
},
{
"epoch": 0.011661327231121281,
"grad_norm": 3.719296455383301,
"learning_rate": 7.042919499559538e-06,
"loss": 0.1679,
"step": 637
},
{
"epoch": 0.011679633867276887,
"grad_norm": 8.414267539978027,
"learning_rate": 7.009592077439135e-06,
"loss": 0.4527,
"step": 638
},
{
"epoch": 0.011697940503432494,
"grad_norm": 4.962318420410156,
"learning_rate": 6.976301092495556e-06,
"loss": 0.257,
"step": 639
},
{
"epoch": 0.0117162471395881,
"grad_norm": 5.534420967102051,
"learning_rate": 6.943046950368944e-06,
"loss": 0.1673,
"step": 640
},
{
"epoch": 0.011734553775743707,
"grad_norm": 7.917168140411377,
"learning_rate": 6.909830056250527e-06,
"loss": 0.2011,
"step": 641
},
{
"epoch": 0.011752860411899314,
"grad_norm": 6.127618312835693,
"learning_rate": 6.876650814877675e-06,
"loss": 0.2565,
"step": 642
},
{
"epoch": 0.01177116704805492,
"grad_norm": 1.4675836563110352,
"learning_rate": 6.843509630528977e-06,
"loss": 0.0989,
"step": 643
},
{
"epoch": 0.011789473684210527,
"grad_norm": 10.80344009399414,
"learning_rate": 6.8104069070193e-06,
"loss": 0.5205,
"step": 644
},
{
"epoch": 0.011807780320366133,
"grad_norm": 5.06679630279541,
"learning_rate": 6.777343047694891e-06,
"loss": 0.2618,
"step": 645
},
{
"epoch": 0.01182608695652174,
"grad_norm": 18.99068832397461,
"learning_rate": 6.744318455428436e-06,
"loss": 1.1171,
"step": 646
},
{
"epoch": 0.011844393592677346,
"grad_norm": 4.126583576202393,
"learning_rate": 6.711333532614168e-06,
"loss": 0.2228,
"step": 647
},
{
"epoch": 0.011862700228832951,
"grad_norm": 2.62414288520813,
"learning_rate": 6.67838868116297e-06,
"loss": 0.1465,
"step": 648
},
{
"epoch": 0.011881006864988558,
"grad_norm": 14.556570053100586,
"learning_rate": 6.645484302497452e-06,
"loss": 0.4093,
"step": 649
},
{
"epoch": 0.011899313501144164,
"grad_norm": 3.685044288635254,
"learning_rate": 6.612620797547087e-06,
"loss": 0.2372,
"step": 650
},
{
"epoch": 0.01191762013729977,
"grad_norm": 10.447038650512695,
"learning_rate": 6.579798566743314e-06,
"loss": 0.3067,
"step": 651
},
{
"epoch": 0.011935926773455377,
"grad_norm": 7.927669525146484,
"learning_rate": 6.547018010014654e-06,
"loss": 0.5193,
"step": 652
},
{
"epoch": 0.011954233409610984,
"grad_norm": 18.05657386779785,
"learning_rate": 6.5142795267818505e-06,
"loss": 0.9131,
"step": 653
},
{
"epoch": 0.01197254004576659,
"grad_norm": 3.7571918964385986,
"learning_rate": 6.481583515952983e-06,
"loss": 0.2054,
"step": 654
},
{
"epoch": 0.011990846681922197,
"grad_norm": 5.364101409912109,
"learning_rate": 6.448930375918632e-06,
"loss": 0.1835,
"step": 655
},
{
"epoch": 0.012009153318077804,
"grad_norm": 9.139827728271484,
"learning_rate": 6.4163205045469975e-06,
"loss": 0.76,
"step": 656
},
{
"epoch": 0.01202745995423341,
"grad_norm": 1.704281210899353,
"learning_rate": 6.383754299179079e-06,
"loss": 0.1308,
"step": 657
},
{
"epoch": 0.012045766590389017,
"grad_norm": 7.968606472015381,
"learning_rate": 6.351232156623803e-06,
"loss": 0.6407,
"step": 658
},
{
"epoch": 0.012064073226544623,
"grad_norm": 17.709571838378906,
"learning_rate": 6.318754473153221e-06,
"loss": 0.9402,
"step": 659
},
{
"epoch": 0.01208237986270023,
"grad_norm": 11.887946128845215,
"learning_rate": 6.286321644497655e-06,
"loss": 0.5389,
"step": 660
},
{
"epoch": 0.012100686498855834,
"grad_norm": 4.999807357788086,
"learning_rate": 6.25393406584088e-06,
"loss": 0.3483,
"step": 661
},
{
"epoch": 0.012118993135011441,
"grad_norm": 7.6661906242370605,
"learning_rate": 6.22159213181533e-06,
"loss": 0.4436,
"step": 662
},
{
"epoch": 0.012137299771167048,
"grad_norm": 7.397403240203857,
"learning_rate": 6.18929623649726e-06,
"loss": 0.3315,
"step": 663
},
{
"epoch": 0.012155606407322654,
"grad_norm": 8.096484184265137,
"learning_rate": 6.157046773401964e-06,
"loss": 0.3889,
"step": 664
},
{
"epoch": 0.01217391304347826,
"grad_norm": 4.6776251792907715,
"learning_rate": 6.124844135478971e-06,
"loss": 0.2081,
"step": 665
},
{
"epoch": 0.012192219679633867,
"grad_norm": 8.530952453613281,
"learning_rate": 6.092688715107265e-06,
"loss": 0.4058,
"step": 666
},
{
"epoch": 0.012210526315789474,
"grad_norm": 9.915332794189453,
"learning_rate": 6.06058090409049e-06,
"loss": 0.4034,
"step": 667
},
{
"epoch": 0.01222883295194508,
"grad_norm": 16.072465896606445,
"learning_rate": 6.028521093652195e-06,
"loss": 0.5987,
"step": 668
},
{
"epoch": 0.012247139588100687,
"grad_norm": 4.9345479011535645,
"learning_rate": 5.996509674431053e-06,
"loss": 0.223,
"step": 669
},
{
"epoch": 0.012265446224256293,
"grad_norm": 4.367536544799805,
"learning_rate": 5.9645470364761e-06,
"loss": 0.357,
"step": 670
},
{
"epoch": 0.0122837528604119,
"grad_norm": 5.174431324005127,
"learning_rate": 5.932633569242e-06,
"loss": 0.3424,
"step": 671
},
{
"epoch": 0.012302059496567506,
"grad_norm": 5.893050193786621,
"learning_rate": 5.900769661584273e-06,
"loss": 0.2669,
"step": 672
},
{
"epoch": 0.012320366132723111,
"grad_norm": 10.575066566467285,
"learning_rate": 5.868955701754584e-06,
"loss": 0.3644,
"step": 673
},
{
"epoch": 0.012338672768878718,
"grad_norm": 5.934246063232422,
"learning_rate": 5.83719207739599e-06,
"loss": 0.3487,
"step": 674
},
{
"epoch": 0.012356979405034324,
"grad_norm": 2.32735538482666,
"learning_rate": 5.8054791755382286e-06,
"loss": 0.1629,
"step": 675
},
{
"epoch": 0.012375286041189931,
"grad_norm": 10.4105224609375,
"learning_rate": 5.773817382593008e-06,
"loss": 0.577,
"step": 676
},
{
"epoch": 0.012393592677345537,
"grad_norm": 8.903114318847656,
"learning_rate": 5.742207084349274e-06,
"loss": 0.5795,
"step": 677
},
{
"epoch": 0.012411899313501144,
"grad_norm": 1.7812707424163818,
"learning_rate": 5.710648665968543e-06,
"loss": 0.0989,
"step": 678
},
{
"epoch": 0.01243020594965675,
"grad_norm": 5.7843017578125,
"learning_rate": 5.679142511980176e-06,
"loss": 0.331,
"step": 679
},
{
"epoch": 0.012448512585812357,
"grad_norm": 10.94100570678711,
"learning_rate": 5.647689006276727e-06,
"loss": 0.6601,
"step": 680
},
{
"epoch": 0.012466819221967964,
"grad_norm": 3.161262273788452,
"learning_rate": 5.616288532109225e-06,
"loss": 0.1807,
"step": 681
},
{
"epoch": 0.01248512585812357,
"grad_norm": 5.56771183013916,
"learning_rate": 5.584941472082549e-06,
"loss": 0.3165,
"step": 682
},
{
"epoch": 0.012503432494279177,
"grad_norm": 2.321823835372925,
"learning_rate": 5.553648208150728e-06,
"loss": 0.0925,
"step": 683
},
{
"epoch": 0.012521739130434783,
"grad_norm": 5.972341537475586,
"learning_rate": 5.522409121612304e-06,
"loss": 0.301,
"step": 684
},
{
"epoch": 0.01254004576659039,
"grad_norm": 14.91727352142334,
"learning_rate": 5.491224593105695e-06,
"loss": 0.8664,
"step": 685
},
{
"epoch": 0.012558352402745995,
"grad_norm": 4.990755558013916,
"learning_rate": 5.460095002604533e-06,
"loss": 0.2396,
"step": 686
},
{
"epoch": 0.012576659038901601,
"grad_norm": 3.9700381755828857,
"learning_rate": 5.429020729413062e-06,
"loss": 0.2423,
"step": 687
},
{
"epoch": 0.012594965675057208,
"grad_norm": 19.159067153930664,
"learning_rate": 5.398002152161484e-06,
"loss": 0.8266,
"step": 688
},
{
"epoch": 0.012613272311212814,
"grad_norm": 6.696817398071289,
"learning_rate": 5.367039648801386e-06,
"loss": 0.3413,
"step": 689
},
{
"epoch": 0.01263157894736842,
"grad_norm": 7.948276042938232,
"learning_rate": 5.336133596601089e-06,
"loss": 0.2789,
"step": 690
},
{
"epoch": 0.012649885583524027,
"grad_norm": 4.675381660461426,
"learning_rate": 5.305284372141095e-06,
"loss": 0.2435,
"step": 691
},
{
"epoch": 0.012668192219679634,
"grad_norm": 8.05580997467041,
"learning_rate": 5.274492351309462e-06,
"loss": 0.3781,
"step": 692
},
{
"epoch": 0.01268649885583524,
"grad_norm": 8.211023330688477,
"learning_rate": 5.243757909297247e-06,
"loss": 0.1573,
"step": 693
},
{
"epoch": 0.012704805491990847,
"grad_norm": 2.2971174716949463,
"learning_rate": 5.213081420593933e-06,
"loss": 0.1092,
"step": 694
},
{
"epoch": 0.012723112128146453,
"grad_norm": 8.110382080078125,
"learning_rate": 5.1824632589828465e-06,
"loss": 0.3773,
"step": 695
},
{
"epoch": 0.01274141876430206,
"grad_norm": 4.399291038513184,
"learning_rate": 5.151903797536631e-06,
"loss": 0.1347,
"step": 696
},
{
"epoch": 0.012759725400457667,
"grad_norm": 6.359978199005127,
"learning_rate": 5.121403408612672e-06,
"loss": 0.1828,
"step": 697
},
{
"epoch": 0.012778032036613273,
"grad_norm": 3.2302865982055664,
"learning_rate": 5.090962463848592e-06,
"loss": 0.1979,
"step": 698
},
{
"epoch": 0.012796338672768878,
"grad_norm": 7.771028518676758,
"learning_rate": 5.060581334157693e-06,
"loss": 0.4509,
"step": 699
},
{
"epoch": 0.012814645308924484,
"grad_norm": 4.159087181091309,
"learning_rate": 5.030260389724447e-06,
"loss": 0.1944,
"step": 700
},
{
"epoch": 0.012832951945080091,
"grad_norm": 5.334461688995361,
"learning_rate": 5.000000000000003e-06,
"loss": 0.1792,
"step": 701
},
{
"epoch": 0.012851258581235698,
"grad_norm": 2.703159809112549,
"learning_rate": 4.96980053369765e-06,
"loss": 0.1473,
"step": 702
},
{
"epoch": 0.012869565217391304,
"grad_norm": 9.341408729553223,
"learning_rate": 4.939662358788364e-06,
"loss": 0.4148,
"step": 703
},
{
"epoch": 0.01288787185354691,
"grad_norm": 5.660780906677246,
"learning_rate": 4.909585842496287e-06,
"loss": 0.4013,
"step": 704
},
{
"epoch": 0.012906178489702517,
"grad_norm": 14.756714820861816,
"learning_rate": 4.879571351294287e-06,
"loss": 0.5132,
"step": 705
},
{
"epoch": 0.012924485125858124,
"grad_norm": 3.988741159439087,
"learning_rate": 4.849619250899458e-06,
"loss": 0.2555,
"step": 706
},
{
"epoch": 0.01294279176201373,
"grad_norm": 4.656280040740967,
"learning_rate": 4.8197299062687e-06,
"loss": 0.2938,
"step": 707
},
{
"epoch": 0.012961098398169337,
"grad_norm": 7.464823246002197,
"learning_rate": 4.78990368159424e-06,
"loss": 0.4899,
"step": 708
},
{
"epoch": 0.012979405034324943,
"grad_norm": 1.0329279899597168,
"learning_rate": 4.76014094029921e-06,
"loss": 0.0952,
"step": 709
},
{
"epoch": 0.01299771167048055,
"grad_norm": 4.358473300933838,
"learning_rate": 4.7304420450332244e-06,
"loss": 0.2474,
"step": 710
},
{
"epoch": 0.013016018306636156,
"grad_norm": 4.45465612411499,
"learning_rate": 4.700807357667953e-06,
"loss": 0.2223,
"step": 711
},
{
"epoch": 0.013034324942791761,
"grad_norm": 2.9852383136749268,
"learning_rate": 4.671237239292699e-06,
"loss": 0.2014,
"step": 712
},
{
"epoch": 0.013052631578947368,
"grad_norm": 8.365155220031738,
"learning_rate": 4.641732050210032e-06,
"loss": 0.4184,
"step": 713
},
{
"epoch": 0.013070938215102974,
"grad_norm": 6.666082382202148,
"learning_rate": 4.612292149931369e-06,
"loss": 0.3087,
"step": 714
},
{
"epoch": 0.01308924485125858,
"grad_norm": 3.9271891117095947,
"learning_rate": 4.582917897172603e-06,
"loss": 0.1685,
"step": 715
},
{
"epoch": 0.013107551487414187,
"grad_norm": 4.308122634887695,
"learning_rate": 4.5536096498497295e-06,
"loss": 0.1991,
"step": 716
},
{
"epoch": 0.013125858123569794,
"grad_norm": 4.309776306152344,
"learning_rate": 4.524367765074499e-06,
"loss": 0.263,
"step": 717
},
{
"epoch": 0.0131441647597254,
"grad_norm": 3.0838520526885986,
"learning_rate": 4.495192599150045e-06,
"loss": 0.1774,
"step": 718
},
{
"epoch": 0.013162471395881007,
"grad_norm": 7.4343695640563965,
"learning_rate": 4.46608450756656e-06,
"loss": 0.4462,
"step": 719
},
{
"epoch": 0.013180778032036614,
"grad_norm": 1.0408004522323608,
"learning_rate": 4.437043844996952e-06,
"loss": 0.0999,
"step": 720
},
{
"epoch": 0.01319908466819222,
"grad_norm": 0.9866322875022888,
"learning_rate": 4.408070965292534e-06,
"loss": 0.0948,
"step": 721
},
{
"epoch": 0.013217391304347827,
"grad_norm": 1.4886623620986938,
"learning_rate": 4.379166221478697e-06,
"loss": 0.1075,
"step": 722
},
{
"epoch": 0.013235697940503433,
"grad_norm": 14.501533508300781,
"learning_rate": 4.350329965750622e-06,
"loss": 0.4004,
"step": 723
},
{
"epoch": 0.01325400457665904,
"grad_norm": 5.454537391662598,
"learning_rate": 4.321562549468991e-06,
"loss": 0.219,
"step": 724
},
{
"epoch": 0.013272311212814645,
"grad_norm": 7.190273761749268,
"learning_rate": 4.292864323155684e-06,
"loss": 0.4322,
"step": 725
},
{
"epoch": 0.013290617848970251,
"grad_norm": 13.830406188964844,
"learning_rate": 4.264235636489542e-06,
"loss": 0.63,
"step": 726
},
{
"epoch": 0.013308924485125858,
"grad_norm": 2.6480069160461426,
"learning_rate": 4.235676838302069e-06,
"loss": 0.1278,
"step": 727
},
{
"epoch": 0.013327231121281464,
"grad_norm": 13.472162246704102,
"learning_rate": 4.207188276573214e-06,
"loss": 0.5058,
"step": 728
},
{
"epoch": 0.01334553775743707,
"grad_norm": 2.51118540763855,
"learning_rate": 4.178770298427107e-06,
"loss": 0.149,
"step": 729
},
{
"epoch": 0.013363844393592677,
"grad_norm": 2.3269636631011963,
"learning_rate": 4.150423250127846e-06,
"loss": 0.1129,
"step": 730
},
{
"epoch": 0.013382151029748284,
"grad_norm": 1.8809363842010498,
"learning_rate": 4.12214747707527e-06,
"loss": 0.1052,
"step": 731
},
{
"epoch": 0.01340045766590389,
"grad_norm": 2.9234514236450195,
"learning_rate": 4.093943323800746e-06,
"loss": 0.1489,
"step": 732
},
{
"epoch": 0.013418764302059497,
"grad_norm": 8.009096145629883,
"learning_rate": 4.065811133962987e-06,
"loss": 0.2956,
"step": 733
},
{
"epoch": 0.013437070938215103,
"grad_norm": 16.991966247558594,
"learning_rate": 4.037751250343841e-06,
"loss": 1.0203,
"step": 734
},
{
"epoch": 0.01345537757437071,
"grad_norm": 8.988302230834961,
"learning_rate": 4.009764014844143e-06,
"loss": 0.4323,
"step": 735
},
{
"epoch": 0.013473684210526317,
"grad_norm": 2.6384384632110596,
"learning_rate": 3.981849768479516e-06,
"loss": 0.1374,
"step": 736
},
{
"epoch": 0.013491990846681921,
"grad_norm": 6.374378681182861,
"learning_rate": 3.954008851376252e-06,
"loss": 0.2113,
"step": 737
},
{
"epoch": 0.013510297482837528,
"grad_norm": 8.83869457244873,
"learning_rate": 3.9262416027671354e-06,
"loss": 0.4818,
"step": 738
},
{
"epoch": 0.013528604118993134,
"grad_norm": 6.571460723876953,
"learning_rate": 3.898548360987325e-06,
"loss": 0.1552,
"step": 739
},
{
"epoch": 0.013546910755148741,
"grad_norm": 18.608381271362305,
"learning_rate": 3.8709294634702374e-06,
"loss": 0.3997,
"step": 740
},
{
"epoch": 0.013565217391304348,
"grad_norm": 9.389320373535156,
"learning_rate": 3.8433852467434175e-06,
"loss": 0.4528,
"step": 741
},
{
"epoch": 0.013583524027459954,
"grad_norm": 8.535320281982422,
"learning_rate": 3.81591604642446e-06,
"loss": 0.4345,
"step": 742
},
{
"epoch": 0.01360183066361556,
"grad_norm": 1.2252271175384521,
"learning_rate": 3.7885221972168974e-06,
"loss": 0.1047,
"step": 743
},
{
"epoch": 0.013620137299771167,
"grad_norm": 5.069352149963379,
"learning_rate": 3.7612040329061405e-06,
"loss": 0.0966,
"step": 744
},
{
"epoch": 0.013638443935926774,
"grad_norm": 9.090739250183105,
"learning_rate": 3.7339618863553983e-06,
"loss": 0.3199,
"step": 745
},
{
"epoch": 0.01365675057208238,
"grad_norm": 6.387343406677246,
"learning_rate": 3.7067960895016277e-06,
"loss": 0.2299,
"step": 746
},
{
"epoch": 0.013675057208237987,
"grad_norm": 7.78866720199585,
"learning_rate": 3.679706973351491e-06,
"loss": 0.4284,
"step": 747
},
{
"epoch": 0.013693363844393593,
"grad_norm": 5.307522773742676,
"learning_rate": 3.6526948679773256e-06,
"loss": 0.1915,
"step": 748
},
{
"epoch": 0.0137116704805492,
"grad_norm": 8.756507873535156,
"learning_rate": 3.625760102513103e-06,
"loss": 0.2633,
"step": 749
},
{
"epoch": 0.013729977116704805,
"grad_norm": 3.3090696334838867,
"learning_rate": 3.598903005150444e-06,
"loss": 0.1686,
"step": 750
},
{
"epoch": 0.013748283752860411,
"grad_norm": 11.242887496948242,
"learning_rate": 3.5721239031346067e-06,
"loss": 0.9433,
"step": 751
},
{
"epoch": 0.013766590389016018,
"grad_norm": 7.644121170043945,
"learning_rate": 3.545423122760493e-06,
"loss": 0.3955,
"step": 752
},
{
"epoch": 0.013784897025171624,
"grad_norm": 9.132620811462402,
"learning_rate": 3.5188009893686916e-06,
"loss": 0.5613,
"step": 753
},
{
"epoch": 0.01380320366132723,
"grad_norm": 8.751481056213379,
"learning_rate": 3.492257827341492e-06,
"loss": 0.5503,
"step": 754
},
{
"epoch": 0.013821510297482837,
"grad_norm": 7.459012985229492,
"learning_rate": 3.4657939600989453e-06,
"loss": 0.2748,
"step": 755
},
{
"epoch": 0.013839816933638444,
"grad_norm": 7.091245651245117,
"learning_rate": 3.4394097100949286e-06,
"loss": 0.3414,
"step": 756
},
{
"epoch": 0.01385812356979405,
"grad_norm": 6.701774597167969,
"learning_rate": 3.4131053988131947e-06,
"loss": 0.2112,
"step": 757
},
{
"epoch": 0.013876430205949657,
"grad_norm": 5.4992475509643555,
"learning_rate": 3.3868813467634833e-06,
"loss": 0.2098,
"step": 758
},
{
"epoch": 0.013894736842105264,
"grad_norm": 4.821681976318359,
"learning_rate": 3.360737873477584e-06,
"loss": 0.2729,
"step": 759
},
{
"epoch": 0.01391304347826087,
"grad_norm": 4.573626518249512,
"learning_rate": 3.3346752975054763e-06,
"loss": 0.3077,
"step": 760
},
{
"epoch": 0.013931350114416477,
"grad_norm": 5.970454692840576,
"learning_rate": 3.308693936411421e-06,
"loss": 0.2764,
"step": 761
},
{
"epoch": 0.013949656750572083,
"grad_norm": 15.272013664245605,
"learning_rate": 3.2827941067700996e-06,
"loss": 0.3641,
"step": 762
},
{
"epoch": 0.013967963386727688,
"grad_norm": 7.55459451675415,
"learning_rate": 3.2569761241627694e-06,
"loss": 0.1848,
"step": 763
},
{
"epoch": 0.013986270022883295,
"grad_norm": 4.483034133911133,
"learning_rate": 3.2312403031733943e-06,
"loss": 0.221,
"step": 764
},
{
"epoch": 0.014004576659038901,
"grad_norm": 4.3897624015808105,
"learning_rate": 3.2055869573848374e-06,
"loss": 0.175,
"step": 765
},
{
"epoch": 0.014022883295194508,
"grad_norm": 8.360304832458496,
"learning_rate": 3.1800163993750166e-06,
"loss": 0.3661,
"step": 766
},
{
"epoch": 0.014041189931350114,
"grad_norm": 5.351900100708008,
"learning_rate": 3.1545289407131128e-06,
"loss": 0.1164,
"step": 767
},
{
"epoch": 0.01405949656750572,
"grad_norm": 6.663723468780518,
"learning_rate": 3.1291248919557717e-06,
"loss": 0.1693,
"step": 768
},
{
"epoch": 0.014077803203661327,
"grad_norm": 2.0629947185516357,
"learning_rate": 3.103804562643302e-06,
"loss": 0.0827,
"step": 769
},
{
"epoch": 0.014096109839816934,
"grad_norm": 4.476418495178223,
"learning_rate": 3.0785682612959334e-06,
"loss": 0.238,
"step": 770
},
{
"epoch": 0.01411441647597254,
"grad_norm": 1.5098648071289062,
"learning_rate": 3.0534162954100264e-06,
"loss": 0.1068,
"step": 771
},
{
"epoch": 0.014132723112128147,
"grad_norm": 10.149811744689941,
"learning_rate": 3.028348971454356e-06,
"loss": 0.3284,
"step": 772
},
{
"epoch": 0.014151029748283753,
"grad_norm": 5.156161785125732,
"learning_rate": 3.003366594866345e-06,
"loss": 0.2241,
"step": 773
},
{
"epoch": 0.01416933638443936,
"grad_norm": 3.282696008682251,
"learning_rate": 2.978469470048376e-06,
"loss": 0.1718,
"step": 774
},
{
"epoch": 0.014187643020594966,
"grad_norm": 7.071142196655273,
"learning_rate": 2.953657900364053e-06,
"loss": 0.4103,
"step": 775
},
{
"epoch": 0.014205949656750571,
"grad_norm": 3.6657493114471436,
"learning_rate": 2.9289321881345257e-06,
"loss": 0.1982,
"step": 776
},
{
"epoch": 0.014224256292906178,
"grad_norm": 5.000716209411621,
"learning_rate": 2.9042926346347932e-06,
"loss": 0.193,
"step": 777
},
{
"epoch": 0.014242562929061784,
"grad_norm": 2.750865936279297,
"learning_rate": 2.8797395400900362e-06,
"loss": 0.1396,
"step": 778
},
{
"epoch": 0.014260869565217391,
"grad_norm": 8.415980339050293,
"learning_rate": 2.855273203671969e-06,
"loss": 0.5748,
"step": 779
},
{
"epoch": 0.014279176201372997,
"grad_norm": 4.887707710266113,
"learning_rate": 2.830893923495173e-06,
"loss": 0.2645,
"step": 780
},
{
"epoch": 0.014297482837528604,
"grad_norm": 4.593505382537842,
"learning_rate": 2.8066019966134907e-06,
"loss": 0.212,
"step": 781
},
{
"epoch": 0.01431578947368421,
"grad_norm": 4.513943672180176,
"learning_rate": 2.7823977190163788e-06,
"loss": 0.2007,
"step": 782
},
{
"epoch": 0.014334096109839817,
"grad_norm": 1.8813406229019165,
"learning_rate": 2.7582813856253276e-06,
"loss": 0.1123,
"step": 783
},
{
"epoch": 0.014352402745995424,
"grad_norm": 7.425732612609863,
"learning_rate": 2.7342532902902418e-06,
"loss": 0.2917,
"step": 784
},
{
"epoch": 0.01437070938215103,
"grad_norm": 3.4717419147491455,
"learning_rate": 2.7103137257858867e-06,
"loss": 0.1469,
"step": 785
},
{
"epoch": 0.014389016018306637,
"grad_norm": 3.6180458068847656,
"learning_rate": 2.6864629838082957e-06,
"loss": 0.1825,
"step": 786
},
{
"epoch": 0.014407322654462243,
"grad_norm": 11.820043563842773,
"learning_rate": 2.6627013549712355e-06,
"loss": 0.3843,
"step": 787
},
{
"epoch": 0.01442562929061785,
"grad_norm": 1.4070264101028442,
"learning_rate": 2.639029128802657e-06,
"loss": 0.1179,
"step": 788
},
{
"epoch": 0.014443935926773455,
"grad_norm": 3.2094695568084717,
"learning_rate": 2.615446593741161e-06,
"loss": 0.1432,
"step": 789
},
{
"epoch": 0.014462242562929061,
"grad_norm": 11.187945365905762,
"learning_rate": 2.5919540371325005e-06,
"loss": 0.5913,
"step": 790
},
{
"epoch": 0.014480549199084668,
"grad_norm": 8.291489601135254,
"learning_rate": 2.5685517452260566e-06,
"loss": 0.5073,
"step": 791
},
{
"epoch": 0.014498855835240274,
"grad_norm": 3.545529365539551,
"learning_rate": 2.5452400031713786e-06,
"loss": 0.1465,
"step": 792
},
{
"epoch": 0.01451716247139588,
"grad_norm": 7.315761566162109,
"learning_rate": 2.522019095014683e-06,
"loss": 0.624,
"step": 793
},
{
"epoch": 0.014535469107551487,
"grad_norm": 11.311168670654297,
"learning_rate": 2.4988893036954045e-06,
"loss": 0.5427,
"step": 794
},
{
"epoch": 0.014553775743707094,
"grad_norm": 7.208177089691162,
"learning_rate": 2.4758509110427576e-06,
"loss": 0.2099,
"step": 795
},
{
"epoch": 0.0145720823798627,
"grad_norm": 6.952854156494141,
"learning_rate": 2.45290419777228e-06,
"loss": 0.2803,
"step": 796
},
{
"epoch": 0.014590389016018307,
"grad_norm": 5.105114459991455,
"learning_rate": 2.4300494434824373e-06,
"loss": 0.2132,
"step": 797
},
{
"epoch": 0.014608695652173913,
"grad_norm": 5.196943283081055,
"learning_rate": 2.407286926651192e-06,
"loss": 0.2229,
"step": 798
},
{
"epoch": 0.01462700228832952,
"grad_norm": 5.220571041107178,
"learning_rate": 2.3846169246326345e-06,
"loss": 0.2873,
"step": 799
},
{
"epoch": 0.014645308924485127,
"grad_norm": 2.8920493125915527,
"learning_rate": 2.362039713653581e-06,
"loss": 0.1196,
"step": 800
},
{
"epoch": 0.014663615560640733,
"grad_norm": 6.959986686706543,
"learning_rate": 2.339555568810221e-06,
"loss": 0.2673,
"step": 801
},
{
"epoch": 0.014681922196796338,
"grad_norm": 19.750226974487305,
"learning_rate": 2.317164764064769e-06,
"loss": 0.2558,
"step": 802
},
{
"epoch": 0.014700228832951944,
"grad_norm": 4.109073638916016,
"learning_rate": 2.2948675722421086e-06,
"loss": 0.2048,
"step": 803
},
{
"epoch": 0.014718535469107551,
"grad_norm": 6.572871208190918,
"learning_rate": 2.27266426502649e-06,
"loss": 0.5073,
"step": 804
},
{
"epoch": 0.014736842105263158,
"grad_norm": 5.497032642364502,
"learning_rate": 2.2505551129582047e-06,
"loss": 0.2909,
"step": 805
},
{
"epoch": 0.014755148741418764,
"grad_norm": 4.798393726348877,
"learning_rate": 2.2285403854302912e-06,
"loss": 0.1645,
"step": 806
},
{
"epoch": 0.01477345537757437,
"grad_norm": 6.386361122131348,
"learning_rate": 2.206620350685257e-06,
"loss": 0.2036,
"step": 807
},
{
"epoch": 0.014791762013729977,
"grad_norm": 15.141334533691406,
"learning_rate": 2.1847952758118118e-06,
"loss": 0.6469,
"step": 808
},
{
"epoch": 0.014810068649885584,
"grad_norm": 2.4272351264953613,
"learning_rate": 2.163065426741603e-06,
"loss": 0.1063,
"step": 809
},
{
"epoch": 0.01482837528604119,
"grad_norm": 9.701894760131836,
"learning_rate": 2.1414310682459805e-06,
"loss": 0.7668,
"step": 810
},
{
"epoch": 0.014846681922196797,
"grad_norm": 2.3073835372924805,
"learning_rate": 2.119892463932781e-06,
"loss": 0.161,
"step": 811
},
{
"epoch": 0.014864988558352403,
"grad_norm": 8.390331268310547,
"learning_rate": 2.098449876243096e-06,
"loss": 0.3498,
"step": 812
},
{
"epoch": 0.01488329519450801,
"grad_norm": 5.491335868835449,
"learning_rate": 2.0771035664480944e-06,
"loss": 0.2787,
"step": 813
},
{
"epoch": 0.014901601830663615,
"grad_norm": 4.279118061065674,
"learning_rate": 2.0558537946458177e-06,
"loss": 0.1625,
"step": 814
},
{
"epoch": 0.014919908466819221,
"grad_norm": 7.896650791168213,
"learning_rate": 2.0347008197580376e-06,
"loss": 0.3681,
"step": 815
},
{
"epoch": 0.014938215102974828,
"grad_norm": 10.665939331054688,
"learning_rate": 2.013644899527074e-06,
"loss": 0.4568,
"step": 816
},
{
"epoch": 0.014956521739130434,
"grad_norm": 24.828960418701172,
"learning_rate": 1.9926862905126663e-06,
"loss": 1.5729,
"step": 817
},
{
"epoch": 0.014974828375286041,
"grad_norm": 12.46844482421875,
"learning_rate": 1.9718252480888567e-06,
"loss": 0.587,
"step": 818
},
{
"epoch": 0.014993135011441647,
"grad_norm": 6.524204730987549,
"learning_rate": 1.95106202644086e-06,
"loss": 0.2851,
"step": 819
},
{
"epoch": 0.015011441647597254,
"grad_norm": 9.998326301574707,
"learning_rate": 1.930396878561983e-06,
"loss": 0.5134,
"step": 820
},
{
"epoch": 0.01502974828375286,
"grad_norm": 15.49677848815918,
"learning_rate": 1.9098300562505266e-06,
"loss": 0.6058,
"step": 821
},
{
"epoch": 0.015048054919908467,
"grad_norm": 5.195704460144043,
"learning_rate": 1.8893618101067357e-06,
"loss": 0.2136,
"step": 822
},
{
"epoch": 0.015066361556064074,
"grad_norm": 7.91325044631958,
"learning_rate": 1.8689923895297247e-06,
"loss": 0.3135,
"step": 823
},
{
"epoch": 0.01508466819221968,
"grad_norm": 12.137787818908691,
"learning_rate": 1.848722042714457e-06,
"loss": 0.9225,
"step": 824
},
{
"epoch": 0.015102974828375287,
"grad_norm": 5.33144998550415,
"learning_rate": 1.8285510166487154e-06,
"loss": 0.2151,
"step": 825
},
{
"epoch": 0.015121281464530893,
"grad_norm": 6.352451324462891,
"learning_rate": 1.808479557110081e-06,
"loss": 0.2846,
"step": 826
},
{
"epoch": 0.015139588100686498,
"grad_norm": 5.770841598510742,
"learning_rate": 1.7885079086629598e-06,
"loss": 0.258,
"step": 827
},
{
"epoch": 0.015157894736842105,
"grad_norm": 9.67846393585205,
"learning_rate": 1.7686363146555807e-06,
"loss": 0.359,
"step": 828
},
{
"epoch": 0.015176201372997711,
"grad_norm": 1.321305751800537,
"learning_rate": 1.7488650172170496e-06,
"loss": 0.0735,
"step": 829
},
{
"epoch": 0.015194508009153318,
"grad_norm": 14.980170249938965,
"learning_rate": 1.7291942572543806e-06,
"loss": 0.3208,
"step": 830
},
{
"epoch": 0.015212814645308924,
"grad_norm": 12.814326286315918,
"learning_rate": 1.709624274449584e-06,
"loss": 0.5511,
"step": 831
},
{
"epoch": 0.01523112128146453,
"grad_norm": 9.458513259887695,
"learning_rate": 1.6901553072567189e-06,
"loss": 0.3438,
"step": 832
},
{
"epoch": 0.015249427917620137,
"grad_norm": 2.2716009616851807,
"learning_rate": 1.6707875928990059e-06,
"loss": 0.158,
"step": 833
},
{
"epoch": 0.015267734553775744,
"grad_norm": 4.668413162231445,
"learning_rate": 1.651521367365936e-06,
"loss": 0.2095,
"step": 834
},
{
"epoch": 0.01528604118993135,
"grad_norm": 6.753487586975098,
"learning_rate": 1.6323568654103838e-06,
"loss": 0.3238,
"step": 835
},
{
"epoch": 0.015304347826086957,
"grad_norm": 4.7871785163879395,
"learning_rate": 1.6132943205457607e-06,
"loss": 0.2238,
"step": 836
},
{
"epoch": 0.015322654462242563,
"grad_norm": 9.5857515335083,
"learning_rate": 1.5943339650431578e-06,
"loss": 0.3882,
"step": 837
},
{
"epoch": 0.01534096109839817,
"grad_norm": 5.032664775848389,
"learning_rate": 1.5754760299285255e-06,
"loss": 0.2316,
"step": 838
},
{
"epoch": 0.015359267734553777,
"grad_norm": 0.9537063837051392,
"learning_rate": 1.5567207449798517e-06,
"loss": 0.0631,
"step": 839
},
{
"epoch": 0.015377574370709381,
"grad_norm": 3.6486315727233887,
"learning_rate": 1.538068338724361e-06,
"loss": 0.1585,
"step": 840
},
{
"epoch": 0.015395881006864988,
"grad_norm": 7.9478840827941895,
"learning_rate": 1.5195190384357405e-06,
"loss": 0.3083,
"step": 841
},
{
"epoch": 0.015414187643020594,
"grad_norm": 4.013637542724609,
"learning_rate": 1.5010730701313626e-06,
"loss": 0.1468,
"step": 842
},
{
"epoch": 0.015432494279176201,
"grad_norm": 13.950557708740234,
"learning_rate": 1.4827306585695234e-06,
"loss": 0.4453,
"step": 843
},
{
"epoch": 0.015450800915331808,
"grad_norm": 8.558002471923828,
"learning_rate": 1.4644920272467245e-06,
"loss": 0.5517,
"step": 844
},
{
"epoch": 0.015469107551487414,
"grad_norm": 4.049571514129639,
"learning_rate": 1.446357398394934e-06,
"loss": 0.2165,
"step": 845
},
{
"epoch": 0.01548741418764302,
"grad_norm": 7.492833137512207,
"learning_rate": 1.4283269929788779e-06,
"loss": 0.3542,
"step": 846
},
{
"epoch": 0.015505720823798627,
"grad_norm": 4.005830764770508,
"learning_rate": 1.4104010306933558e-06,
"loss": 0.2093,
"step": 847
},
{
"epoch": 0.015524027459954234,
"grad_norm": 4.699175834655762,
"learning_rate": 1.3925797299605649e-06,
"loss": 0.1869,
"step": 848
},
{
"epoch": 0.01554233409610984,
"grad_norm": 7.27205228805542,
"learning_rate": 1.3748633079274254e-06,
"loss": 0.3142,
"step": 849
},
{
"epoch": 0.015560640732265447,
"grad_norm": 6.372773170471191,
"learning_rate": 1.3572519804629537e-06,
"loss": 0.2057,
"step": 850
},
{
"epoch": 0.015578947368421053,
"grad_norm": 6.957703590393066,
"learning_rate": 1.339745962155613e-06,
"loss": 0.2801,
"step": 851
},
{
"epoch": 0.01559725400457666,
"grad_norm": 7.599903106689453,
"learning_rate": 1.322345466310717e-06,
"loss": 0.2687,
"step": 852
},
{
"epoch": 0.015615560640732265,
"grad_norm": 5.099765300750732,
"learning_rate": 1.30505070494781e-06,
"loss": 0.2243,
"step": 853
},
{
"epoch": 0.015633867276887873,
"grad_norm": 8.895179748535156,
"learning_rate": 1.2878618887981064e-06,
"loss": 0.4707,
"step": 854
},
{
"epoch": 0.01565217391304348,
"grad_norm": 2.2955360412597656,
"learning_rate": 1.2707792273019049e-06,
"loss": 0.0989,
"step": 855
},
{
"epoch": 0.015670480549199086,
"grad_norm": 7.87874698638916,
"learning_rate": 1.2538029286060428e-06,
"loss": 0.376,
"step": 856
},
{
"epoch": 0.015688787185354693,
"grad_norm": 9.83598518371582,
"learning_rate": 1.2369331995613664e-06,
"loss": 0.6476,
"step": 857
},
{
"epoch": 0.0157070938215103,
"grad_norm": 2.0399458408355713,
"learning_rate": 1.2201702457201948e-06,
"loss": 0.1026,
"step": 858
},
{
"epoch": 0.015725400457665902,
"grad_norm": 6.7959442138671875,
"learning_rate": 1.2035142713338366e-06,
"loss": 0.1965,
"step": 859
},
{
"epoch": 0.01574370709382151,
"grad_norm": 9.003751754760742,
"learning_rate": 1.1869654793500784e-06,
"loss": 0.5292,
"step": 860
},
{
"epoch": 0.015762013729977115,
"grad_norm": 11.155455589294434,
"learning_rate": 1.1705240714107301e-06,
"loss": 0.5597,
"step": 861
},
{
"epoch": 0.015780320366132722,
"grad_norm": 9.739468574523926,
"learning_rate": 1.1541902478491607e-06,
"loss": 0.5616,
"step": 862
},
{
"epoch": 0.01579862700228833,
"grad_norm": 4.332462310791016,
"learning_rate": 1.1379642076878528e-06,
"loss": 0.1764,
"step": 863
},
{
"epoch": 0.015816933638443935,
"grad_norm": 4.867697715759277,
"learning_rate": 1.1218461486359878e-06,
"loss": 0.3082,
"step": 864
},
{
"epoch": 0.01583524027459954,
"grad_norm": 6.073899269104004,
"learning_rate": 1.1058362670870248e-06,
"loss": 0.3394,
"step": 865
},
{
"epoch": 0.015853546910755148,
"grad_norm": 15.319504737854004,
"learning_rate": 1.0899347581163222e-06,
"loss": 0.3632,
"step": 866
},
{
"epoch": 0.015871853546910755,
"grad_norm": 8.388349533081055,
"learning_rate": 1.0741418154787443e-06,
"loss": 0.477,
"step": 867
},
{
"epoch": 0.01589016018306636,
"grad_norm": 4.829792499542236,
"learning_rate": 1.058457631606319e-06,
"loss": 0.2643,
"step": 868
},
{
"epoch": 0.015908466819221968,
"grad_norm": 10.846988677978516,
"learning_rate": 1.042882397605871e-06,
"loss": 0.6206,
"step": 869
},
{
"epoch": 0.015926773455377574,
"grad_norm": 9.261481285095215,
"learning_rate": 1.0274163032567165e-06,
"loss": 0.3205,
"step": 870
},
{
"epoch": 0.01594508009153318,
"grad_norm": 1.8402290344238281,
"learning_rate": 1.012059537008332e-06,
"loss": 0.0896,
"step": 871
},
{
"epoch": 0.015963386727688787,
"grad_norm": 4.178070545196533,
"learning_rate": 9.968122859780648e-07,
"loss": 0.2115,
"step": 872
},
{
"epoch": 0.015981693363844394,
"grad_norm": 1.212082862854004,
"learning_rate": 9.816747359488632e-07,
"loss": 0.0796,
"step": 873
},
{
"epoch": 0.016,
"grad_norm": 17.429851531982422,
"learning_rate": 9.666470713669918e-07,
"loss": 0.6567,
"step": 874
},
{
"epoch": 0.016018306636155607,
"grad_norm": 3.6043081283569336,
"learning_rate": 9.517294753398066e-07,
"loss": 0.2011,
"step": 875
},
{
"epoch": 0.016036613272311213,
"grad_norm": 6.399504661560059,
"learning_rate": 9.369221296335007e-07,
"loss": 0.3302,
"step": 876
},
{
"epoch": 0.01605491990846682,
"grad_norm": 17.06818389892578,
"learning_rate": 9.222252146709143e-07,
"loss": 0.853,
"step": 877
},
{
"epoch": 0.016073226544622427,
"grad_norm": 2.9740846157073975,
"learning_rate": 9.076389095293148e-07,
"loss": 0.1675,
"step": 878
},
{
"epoch": 0.016091533180778033,
"grad_norm": 3.0257978439331055,
"learning_rate": 8.931633919382299e-07,
"loss": 0.2347,
"step": 879
},
{
"epoch": 0.01610983981693364,
"grad_norm": 1.34092378616333,
"learning_rate": 8.787988382772705e-07,
"loss": 0.118,
"step": 880
},
{
"epoch": 0.016128146453089246,
"grad_norm": 0.5438145995140076,
"learning_rate": 8.645454235739903e-07,
"loss": 0.0775,
"step": 881
},
{
"epoch": 0.016146453089244853,
"grad_norm": 6.302069187164307,
"learning_rate": 8.504033215017527e-07,
"loss": 0.3397,
"step": 882
},
{
"epoch": 0.01616475972540046,
"grad_norm": 11.530533790588379,
"learning_rate": 8.363727043776037e-07,
"loss": 0.3956,
"step": 883
},
{
"epoch": 0.016183066361556066,
"grad_norm": 12.510212898254395,
"learning_rate": 8.224537431601886e-07,
"loss": 0.2942,
"step": 884
},
{
"epoch": 0.01620137299771167,
"grad_norm": 10.052515983581543,
"learning_rate": 8.086466074476562e-07,
"loss": 0.2827,
"step": 885
},
{
"epoch": 0.016219679633867275,
"grad_norm": 2.31078839302063,
"learning_rate": 7.949514654755963e-07,
"loss": 0.1479,
"step": 886
},
{
"epoch": 0.016237986270022882,
"grad_norm": 16.138065338134766,
"learning_rate": 7.81368484114996e-07,
"loss": 0.4199,
"step": 887
},
{
"epoch": 0.01625629290617849,
"grad_norm": 3.326777458190918,
"learning_rate": 7.678978288701911e-07,
"loss": 0.1543,
"step": 888
},
{
"epoch": 0.016274599542334095,
"grad_norm": 5.338883876800537,
"learning_rate": 7.545396638768698e-07,
"loss": 0.1821,
"step": 889
},
{
"epoch": 0.0162929061784897,
"grad_norm": 7.2971014976501465,
"learning_rate": 7.412941519000527e-07,
"loss": 0.3375,
"step": 890
},
{
"epoch": 0.016311212814645308,
"grad_norm": 6.16159725189209,
"learning_rate": 7.281614543321269e-07,
"loss": 0.2027,
"step": 891
},
{
"epoch": 0.016329519450800915,
"grad_norm": 12.14189338684082,
"learning_rate": 7.151417311908648e-07,
"loss": 0.4996,
"step": 892
},
{
"epoch": 0.01634782608695652,
"grad_norm": 9.135406494140625,
"learning_rate": 7.022351411174866e-07,
"loss": 0.3862,
"step": 893
},
{
"epoch": 0.016366132723112128,
"grad_norm": 9.651875495910645,
"learning_rate": 6.894418413747183e-07,
"loss": 0.3808,
"step": 894
},
{
"epoch": 0.016384439359267734,
"grad_norm": 4.33256196975708,
"learning_rate": 6.767619878448783e-07,
"loss": 0.2232,
"step": 895
},
{
"epoch": 0.01640274599542334,
"grad_norm": 5.881313800811768,
"learning_rate": 6.641957350279838e-07,
"loss": 0.2579,
"step": 896
},
{
"epoch": 0.016421052631578947,
"grad_norm": 2.510111093521118,
"learning_rate": 6.517432360398556e-07,
"loss": 0.1578,
"step": 897
},
{
"epoch": 0.016439359267734554,
"grad_norm": 2.6253507137298584,
"learning_rate": 6.394046426102673e-07,
"loss": 0.1334,
"step": 898
},
{
"epoch": 0.01645766590389016,
"grad_norm": 5.544801712036133,
"learning_rate": 6.271801050810856e-07,
"loss": 0.3322,
"step": 899
},
{
"epoch": 0.016475972540045767,
"grad_norm": 5.784806251525879,
"learning_rate": 6.150697724044407e-07,
"loss": 0.3011,
"step": 900
},
{
"epoch": 0.016494279176201374,
"grad_norm": 10.115022659301758,
"learning_rate": 6.030737921409169e-07,
"loss": 0.4213,
"step": 901
},
{
"epoch": 0.01651258581235698,
"grad_norm": 13.211285591125488,
"learning_rate": 5.911923104577455e-07,
"loss": 0.2885,
"step": 902
},
{
"epoch": 0.016530892448512587,
"grad_norm": 4.184572696685791,
"learning_rate": 5.794254721270331e-07,
"loss": 0.2343,
"step": 903
},
{
"epoch": 0.016549199084668193,
"grad_norm": 6.290907859802246,
"learning_rate": 5.677734205239904e-07,
"loss": 0.259,
"step": 904
},
{
"epoch": 0.0165675057208238,
"grad_norm": 4.348118305206299,
"learning_rate": 5.562362976251901e-07,
"loss": 0.2195,
"step": 905
},
{
"epoch": 0.016585812356979406,
"grad_norm": 7.895175457000732,
"learning_rate": 5.448142440068316e-07,
"loss": 0.3582,
"step": 906
},
{
"epoch": 0.016604118993135013,
"grad_norm": 6.373378753662109,
"learning_rate": 5.335073988430373e-07,
"loss": 0.4875,
"step": 907
},
{
"epoch": 0.01662242562929062,
"grad_norm": 4.442923069000244,
"learning_rate": 5.223158999041444e-07,
"loss": 0.2706,
"step": 908
},
{
"epoch": 0.016640732265446226,
"grad_norm": 5.623570442199707,
"learning_rate": 5.112398835550348e-07,
"loss": 0.249,
"step": 909
},
{
"epoch": 0.01665903890160183,
"grad_norm": 5.596260070800781,
"learning_rate": 5.002794847534765e-07,
"loss": 0.3783,
"step": 910
},
{
"epoch": 0.016677345537757435,
"grad_norm": 4.682590484619141,
"learning_rate": 4.894348370484648e-07,
"loss": 0.2253,
"step": 911
},
{
"epoch": 0.016695652173913042,
"grad_norm": 8.983024597167969,
"learning_rate": 4.787060725786141e-07,
"loss": 0.3797,
"step": 912
},
{
"epoch": 0.01671395881006865,
"grad_norm": 1.1545554399490356,
"learning_rate": 4.6809332207053083e-07,
"loss": 0.1166,
"step": 913
},
{
"epoch": 0.016732265446224255,
"grad_norm": 5.555506229400635,
"learning_rate": 4.575967148372318e-07,
"loss": 0.2964,
"step": 914
},
{
"epoch": 0.01675057208237986,
"grad_norm": 2.3009276390075684,
"learning_rate": 4.4721637877656377e-07,
"loss": 0.1593,
"step": 915
},
{
"epoch": 0.016768878718535468,
"grad_norm": 4.247383117675781,
"learning_rate": 4.3695244036964567e-07,
"loss": 0.1959,
"step": 916
},
{
"epoch": 0.016787185354691075,
"grad_norm": 8.078374862670898,
"learning_rate": 4.268050246793276e-07,
"loss": 0.1937,
"step": 917
},
{
"epoch": 0.01680549199084668,
"grad_norm": 4.602707386016846,
"learning_rate": 4.167742553486676e-07,
"loss": 0.2737,
"step": 918
},
{
"epoch": 0.016823798627002288,
"grad_norm": 4.030495643615723,
"learning_rate": 4.068602545994249e-07,
"loss": 0.1949,
"step": 919
},
{
"epoch": 0.016842105263157894,
"grad_norm": 6.921680450439453,
"learning_rate": 3.9706314323056936e-07,
"loss": 0.3539,
"step": 920
},
{
"epoch": 0.0168604118993135,
"grad_norm": 4.294434070587158,
"learning_rate": 3.8738304061681107e-07,
"loss": 0.1642,
"step": 921
},
{
"epoch": 0.016878718535469107,
"grad_norm": 8.08474349975586,
"learning_rate": 3.7782006470714614e-07,
"loss": 0.531,
"step": 922
},
{
"epoch": 0.016897025171624714,
"grad_norm": 5.4197540283203125,
"learning_rate": 3.68374332023419e-07,
"loss": 0.1086,
"step": 923
},
{
"epoch": 0.01691533180778032,
"grad_norm": 10.48359203338623,
"learning_rate": 3.590459576589e-07,
"loss": 0.4211,
"step": 924
},
{
"epoch": 0.016933638443935927,
"grad_norm": 6.148892879486084,
"learning_rate": 3.498350552768859e-07,
"loss": 0.2935,
"step": 925
},
{
"epoch": 0.016951945080091534,
"grad_norm": 37.46620559692383,
"learning_rate": 3.4074173710931804e-07,
"loss": 0.4102,
"step": 926
},
{
"epoch": 0.01697025171624714,
"grad_norm": 8.583579063415527,
"learning_rate": 3.3176611395540625e-07,
"loss": 0.5772,
"step": 927
},
{
"epoch": 0.016988558352402747,
"grad_norm": 7.78873872756958,
"learning_rate": 3.2290829518028867e-07,
"loss": 0.4475,
"step": 928
},
{
"epoch": 0.017006864988558353,
"grad_norm": 1.913015365600586,
"learning_rate": 3.1416838871368925e-07,
"loss": 0.1075,
"step": 929
},
{
"epoch": 0.01702517162471396,
"grad_norm": 19.77604103088379,
"learning_rate": 3.0554650104861137e-07,
"loss": 0.8314,
"step": 930
},
{
"epoch": 0.017043478260869566,
"grad_norm": 6.767706394195557,
"learning_rate": 2.970427372400353e-07,
"loss": 0.2661,
"step": 931
},
{
"epoch": 0.017061784897025173,
"grad_norm": 6.047036647796631,
"learning_rate": 2.8865720090364037e-07,
"loss": 0.3102,
"step": 932
},
{
"epoch": 0.01708009153318078,
"grad_norm": 3.897958517074585,
"learning_rate": 2.8038999421453827e-07,
"loss": 0.2229,
"step": 933
},
{
"epoch": 0.017098398169336386,
"grad_norm": 14.799800872802734,
"learning_rate": 2.7224121790603517e-07,
"loss": 0.9037,
"step": 934
},
{
"epoch": 0.017116704805491992,
"grad_norm": 1.944534420967102,
"learning_rate": 2.6421097126839714e-07,
"loss": 0.0869,
"step": 935
},
{
"epoch": 0.017135011441647596,
"grad_norm": 6.719526290893555,
"learning_rate": 2.5629935214764866e-07,
"loss": 0.2525,
"step": 936
},
{
"epoch": 0.017153318077803202,
"grad_norm": 1.7623887062072754,
"learning_rate": 2.4850645694436736e-07,
"loss": 0.0721,
"step": 937
},
{
"epoch": 0.01717162471395881,
"grad_norm": 2.431997776031494,
"learning_rate": 2.4083238061252565e-07,
"loss": 0.1643,
"step": 938
},
{
"epoch": 0.017189931350114415,
"grad_norm": 11.728501319885254,
"learning_rate": 2.332772166583208e-07,
"loss": 0.234,
"step": 939
},
{
"epoch": 0.017208237986270022,
"grad_norm": 1.5342614650726318,
"learning_rate": 2.2584105713904126e-07,
"loss": 0.1473,
"step": 940
},
{
"epoch": 0.01722654462242563,
"grad_norm": 8.040061950683594,
"learning_rate": 2.1852399266194312e-07,
"loss": 0.3064,
"step": 941
},
{
"epoch": 0.017244851258581235,
"grad_norm": 2.8261466026306152,
"learning_rate": 2.1132611238315004e-07,
"loss": 0.2135,
"step": 942
},
{
"epoch": 0.01726315789473684,
"grad_norm": 5.933513164520264,
"learning_rate": 2.0424750400655947e-07,
"loss": 0.3271,
"step": 943
},
{
"epoch": 0.017281464530892448,
"grad_norm": 3.4956777095794678,
"learning_rate": 1.9728825378278248e-07,
"loss": 0.1417,
"step": 944
},
{
"epoch": 0.017299771167048054,
"grad_norm": 6.9485697746276855,
"learning_rate": 1.9044844650808468e-07,
"loss": 0.326,
"step": 945
},
{
"epoch": 0.01731807780320366,
"grad_norm": 9.41317081451416,
"learning_rate": 1.8372816552336025e-07,
"loss": 0.4231,
"step": 946
},
{
"epoch": 0.017336384439359268,
"grad_norm": 6.496486663818359,
"learning_rate": 1.7712749271311392e-07,
"loss": 0.2621,
"step": 947
},
{
"epoch": 0.017354691075514874,
"grad_norm": 9.458244323730469,
"learning_rate": 1.706465085044584e-07,
"loss": 0.3791,
"step": 948
},
{
"epoch": 0.01737299771167048,
"grad_norm": 1.7211376428604126,
"learning_rate": 1.6428529186614195e-07,
"loss": 0.1238,
"step": 949
},
{
"epoch": 0.017391304347826087,
"grad_norm": 2.092992067337036,
"learning_rate": 1.580439203075812e-07,
"loss": 0.1183,
"step": 950
},
{
"epoch": 0.017409610983981694,
"grad_norm": 7.563133716583252,
"learning_rate": 1.519224698779198e-07,
"loss": 0.2907,
"step": 951
},
{
"epoch": 0.0174279176201373,
"grad_norm": 13.335310935974121,
"learning_rate": 1.4592101516509916e-07,
"loss": 0.7671,
"step": 952
},
{
"epoch": 0.017446224256292907,
"grad_norm": 10.47811508178711,
"learning_rate": 1.400396292949513e-07,
"loss": 0.4623,
"step": 953
},
{
"epoch": 0.017464530892448513,
"grad_norm": 7.000487804412842,
"learning_rate": 1.3427838393030634e-07,
"loss": 0.4022,
"step": 954
},
{
"epoch": 0.01748283752860412,
"grad_norm": 6.856893062591553,
"learning_rate": 1.2863734927012094e-07,
"loss": 0.3277,
"step": 955
},
{
"epoch": 0.017501144164759726,
"grad_norm": 12.255921363830566,
"learning_rate": 1.231165940486234e-07,
"loss": 0.502,
"step": 956
},
{
"epoch": 0.017519450800915333,
"grad_norm": 24.51825714111328,
"learning_rate": 1.1771618553447217e-07,
"loss": 0.3786,
"step": 957
},
{
"epoch": 0.01753775743707094,
"grad_norm": 2.42156982421875,
"learning_rate": 1.1243618952994195e-07,
"loss": 0.1061,
"step": 958
},
{
"epoch": 0.017556064073226546,
"grad_norm": 19.74112319946289,
"learning_rate": 1.0727667037011668e-07,
"loss": 0.6959,
"step": 959
},
{
"epoch": 0.017574370709382153,
"grad_norm": 2.00630784034729,
"learning_rate": 1.0223769092211012e-07,
"loss": 0.1106,
"step": 960
},
{
"epoch": 0.01759267734553776,
"grad_norm": 7.67366361618042,
"learning_rate": 9.731931258429638e-08,
"loss": 0.173,
"step": 961
},
{
"epoch": 0.017610983981693362,
"grad_norm": 9.130098342895508,
"learning_rate": 9.252159528556404e-08,
"loss": 0.4265,
"step": 962
},
{
"epoch": 0.01762929061784897,
"grad_norm": 2.1006438732147217,
"learning_rate": 8.784459748458318e-08,
"loss": 0.0866,
"step": 963
},
{
"epoch": 0.017647597254004575,
"grad_norm": 9.933135032653809,
"learning_rate": 8.328837616909612e-08,
"loss": 0.2907,
"step": 964
},
{
"epoch": 0.017665903890160182,
"grad_norm": 6.022252559661865,
"learning_rate": 7.885298685522235e-08,
"loss": 0.3691,
"step": 965
},
{
"epoch": 0.01768421052631579,
"grad_norm": 3.950028419494629,
"learning_rate": 7.453848358678018e-08,
"loss": 0.138,
"step": 966
},
{
"epoch": 0.017702517162471395,
"grad_norm": 5.163845539093018,
"learning_rate": 7.034491893463059e-08,
"loss": 0.2737,
"step": 967
},
{
"epoch": 0.017720823798627,
"grad_norm": 7.640467166900635,
"learning_rate": 6.627234399603554e-08,
"loss": 0.3658,
"step": 968
},
{
"epoch": 0.017739130434782608,
"grad_norm": 6.5803985595703125,
"learning_rate": 6.232080839403631e-08,
"loss": 0.3093,
"step": 969
},
{
"epoch": 0.017757437070938215,
"grad_norm": 4.211984634399414,
"learning_rate": 5.849036027684607e-08,
"loss": 0.1708,
"step": 970
},
{
"epoch": 0.01777574370709382,
"grad_norm": 8.148463249206543,
"learning_rate": 5.4781046317267103e-08,
"loss": 0.3152,
"step": 971
},
{
"epoch": 0.017794050343249428,
"grad_norm": 4.271562099456787,
"learning_rate": 5.119291171211793e-08,
"loss": 0.1852,
"step": 972
},
{
"epoch": 0.017812356979405034,
"grad_norm": 8.320340156555176,
"learning_rate": 4.772600018168816e-08,
"loss": 0.3206,
"step": 973
},
{
"epoch": 0.01783066361556064,
"grad_norm": 9.28972053527832,
"learning_rate": 4.438035396920004e-08,
"loss": 0.3864,
"step": 974
},
{
"epoch": 0.017848970251716247,
"grad_norm": 12.746179580688477,
"learning_rate": 4.115601384029666e-08,
"loss": 0.694,
"step": 975
},
{
"epoch": 0.017867276887871854,
"grad_norm": 12.518298149108887,
"learning_rate": 3.805301908254455e-08,
"loss": 0.4285,
"step": 976
},
{
"epoch": 0.01788558352402746,
"grad_norm": 6.564645767211914,
"learning_rate": 3.50714075049563e-08,
"loss": 0.3002,
"step": 977
},
{
"epoch": 0.017903890160183067,
"grad_norm": 8.353341102600098,
"learning_rate": 3.22112154375287e-08,
"loss": 0.3428,
"step": 978
},
{
"epoch": 0.017922196796338673,
"grad_norm": 8.169654846191406,
"learning_rate": 2.947247773079753e-08,
"loss": 0.3041,
"step": 979
},
{
"epoch": 0.01794050343249428,
"grad_norm": 3.4119813442230225,
"learning_rate": 2.6855227755419046e-08,
"loss": 0.1635,
"step": 980
},
{
"epoch": 0.017958810068649887,
"grad_norm": 0.7701112627983093,
"learning_rate": 2.4359497401758026e-08,
"loss": 0.0799,
"step": 981
},
{
"epoch": 0.017977116704805493,
"grad_norm": 7.056750297546387,
"learning_rate": 2.1985317079500358e-08,
"loss": 0.2713,
"step": 982
},
{
"epoch": 0.0179954233409611,
"grad_norm": 13.480681419372559,
"learning_rate": 1.973271571728441e-08,
"loss": 0.5698,
"step": 983
},
{
"epoch": 0.018013729977116706,
"grad_norm": 2.141348123550415,
"learning_rate": 1.7601720762346895e-08,
"loss": 0.1175,
"step": 984
},
{
"epoch": 0.018032036613272313,
"grad_norm": 4.411203384399414,
"learning_rate": 1.5592358180189782e-08,
"loss": 0.1344,
"step": 985
},
{
"epoch": 0.01805034324942792,
"grad_norm": 9.80256462097168,
"learning_rate": 1.370465245426167e-08,
"loss": 0.3765,
"step": 986
},
{
"epoch": 0.018068649885583522,
"grad_norm": 20.578289031982422,
"learning_rate": 1.1938626585660252e-08,
"loss": 0.9648,
"step": 987
},
{
"epoch": 0.01808695652173913,
"grad_norm": 6.394954681396484,
"learning_rate": 1.0294302092853647e-08,
"loss": 0.2917,
"step": 988
},
{
"epoch": 0.018105263157894735,
"grad_norm": 2.968735933303833,
"learning_rate": 8.771699011416169e-09,
"loss": 0.1552,
"step": 989
},
{
"epoch": 0.018123569794050342,
"grad_norm": 9.075217247009277,
"learning_rate": 7.370835893788508e-09,
"loss": 0.5275,
"step": 990
},
{
"epoch": 0.01814187643020595,
"grad_norm": 0.4628898799419403,
"learning_rate": 6.091729809042379e-09,
"loss": 0.0548,
"step": 991
},
{
"epoch": 0.018160183066361555,
"grad_norm": 9.862579345703125,
"learning_rate": 4.9343963426840006e-09,
"loss": 0.4799,
"step": 992
},
{
"epoch": 0.01817848970251716,
"grad_norm": 0.9560385942459106,
"learning_rate": 3.898849596456477e-09,
"loss": 0.0928,
"step": 993
},
{
"epoch": 0.018196796338672768,
"grad_norm": 14.042981147766113,
"learning_rate": 2.9851021881688314e-09,
"loss": 0.4677,
"step": 994
},
{
"epoch": 0.018215102974828375,
"grad_norm": 1.057991862297058,
"learning_rate": 2.193165251545004e-09,
"loss": 0.112,
"step": 995
},
{
"epoch": 0.01823340961098398,
"grad_norm": 2.119772434234619,
"learning_rate": 1.5230484360873043e-09,
"loss": 0.105,
"step": 996
},
{
"epoch": 0.018251716247139588,
"grad_norm": 8.898747444152832,
"learning_rate": 9.74759906957612e-10,
"loss": 0.403,
"step": 997
},
{
"epoch": 0.018270022883295194,
"grad_norm": 4.880192756652832,
"learning_rate": 5.483063448785686e-10,
"loss": 0.2703,
"step": 998
},
{
"epoch": 0.0182883295194508,
"grad_norm": 1.7518311738967896,
"learning_rate": 2.436929460525317e-10,
"loss": 0.0996,
"step": 999
},
{
"epoch": 0.018306636155606407,
"grad_norm": 4.965724945068359,
"learning_rate": 6.092342209607083e-11,
"loss": 0.208,
"step": 1000
}
],
"logging_steps": 1,
"max_steps": 1000,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 200,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": false,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 0.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}