marianna13's picture
Upload folder using huggingface_hub
613e18e verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0,
"eval_steps": 500,
"global_step": 1359,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.002207505518763797,
"grad_norm": 2.7993483543395996,
"learning_rate": 7.352941176470589e-08,
"loss": 0.7615,
"step": 1
},
{
"epoch": 0.004415011037527594,
"grad_norm": 2.814159393310547,
"learning_rate": 1.4705882352941178e-07,
"loss": 0.7759,
"step": 2
},
{
"epoch": 0.006622516556291391,
"grad_norm": 2.932206392288208,
"learning_rate": 2.2058823529411768e-07,
"loss": 0.7897,
"step": 3
},
{
"epoch": 0.008830022075055188,
"grad_norm": 2.909721851348877,
"learning_rate": 2.9411764705882356e-07,
"loss": 0.7698,
"step": 4
},
{
"epoch": 0.011037527593818985,
"grad_norm": 2.855281114578247,
"learning_rate": 3.6764705882352943e-07,
"loss": 0.7673,
"step": 5
},
{
"epoch": 0.013245033112582781,
"grad_norm": 2.930614709854126,
"learning_rate": 4.4117647058823536e-07,
"loss": 0.7914,
"step": 6
},
{
"epoch": 0.01545253863134658,
"grad_norm": 2.8653368949890137,
"learning_rate": 5.147058823529412e-07,
"loss": 0.7713,
"step": 7
},
{
"epoch": 0.017660044150110375,
"grad_norm": 2.7474284172058105,
"learning_rate": 5.882352941176471e-07,
"loss": 0.7713,
"step": 8
},
{
"epoch": 0.019867549668874173,
"grad_norm": 2.6992416381835938,
"learning_rate": 6.61764705882353e-07,
"loss": 0.7549,
"step": 9
},
{
"epoch": 0.02207505518763797,
"grad_norm": 2.718104362487793,
"learning_rate": 7.352941176470589e-07,
"loss": 0.757,
"step": 10
},
{
"epoch": 0.024282560706401765,
"grad_norm": 2.670048952102661,
"learning_rate": 8.088235294117648e-07,
"loss": 0.7367,
"step": 11
},
{
"epoch": 0.026490066225165563,
"grad_norm": 2.205970287322998,
"learning_rate": 8.823529411764707e-07,
"loss": 0.7397,
"step": 12
},
{
"epoch": 0.02869757174392936,
"grad_norm": 2.2243752479553223,
"learning_rate": 9.558823529411764e-07,
"loss": 0.7565,
"step": 13
},
{
"epoch": 0.03090507726269316,
"grad_norm": 2.2006163597106934,
"learning_rate": 1.0294117647058825e-06,
"loss": 0.744,
"step": 14
},
{
"epoch": 0.033112582781456956,
"grad_norm": 2.0692009925842285,
"learning_rate": 1.1029411764705884e-06,
"loss": 0.7282,
"step": 15
},
{
"epoch": 0.03532008830022075,
"grad_norm": 1.755692958831787,
"learning_rate": 1.1764705882352942e-06,
"loss": 0.7027,
"step": 16
},
{
"epoch": 0.037527593818984545,
"grad_norm": 1.429651141166687,
"learning_rate": 1.25e-06,
"loss": 0.7028,
"step": 17
},
{
"epoch": 0.039735099337748346,
"grad_norm": 1.4149061441421509,
"learning_rate": 1.323529411764706e-06,
"loss": 0.7027,
"step": 18
},
{
"epoch": 0.04194260485651214,
"grad_norm": 1.4027491807937622,
"learning_rate": 1.3970588235294119e-06,
"loss": 0.7034,
"step": 19
},
{
"epoch": 0.04415011037527594,
"grad_norm": 1.3382784128189087,
"learning_rate": 1.4705882352941177e-06,
"loss": 0.6838,
"step": 20
},
{
"epoch": 0.046357615894039736,
"grad_norm": 1.2848469018936157,
"learning_rate": 1.5441176470588238e-06,
"loss": 0.6904,
"step": 21
},
{
"epoch": 0.04856512141280353,
"grad_norm": 1.0581680536270142,
"learning_rate": 1.6176470588235297e-06,
"loss": 0.6658,
"step": 22
},
{
"epoch": 0.05077262693156733,
"grad_norm": 1.0082190036773682,
"learning_rate": 1.6911764705882356e-06,
"loss": 0.653,
"step": 23
},
{
"epoch": 0.052980132450331126,
"grad_norm": 0.979928195476532,
"learning_rate": 1.7647058823529414e-06,
"loss": 0.6501,
"step": 24
},
{
"epoch": 0.05518763796909492,
"grad_norm": 0.9646239876747131,
"learning_rate": 1.8382352941176473e-06,
"loss": 0.6415,
"step": 25
},
{
"epoch": 0.05739514348785872,
"grad_norm": 0.8932761549949646,
"learning_rate": 1.9117647058823528e-06,
"loss": 0.6523,
"step": 26
},
{
"epoch": 0.059602649006622516,
"grad_norm": 0.8707468509674072,
"learning_rate": 1.985294117647059e-06,
"loss": 0.6434,
"step": 27
},
{
"epoch": 0.06181015452538632,
"grad_norm": 0.8291523456573486,
"learning_rate": 2.058823529411765e-06,
"loss": 0.6333,
"step": 28
},
{
"epoch": 0.0640176600441501,
"grad_norm": 0.7280705571174622,
"learning_rate": 2.132352941176471e-06,
"loss": 0.6198,
"step": 29
},
{
"epoch": 0.06622516556291391,
"grad_norm": 0.6107202172279358,
"learning_rate": 2.2058823529411767e-06,
"loss": 0.6057,
"step": 30
},
{
"epoch": 0.0684326710816777,
"grad_norm": 0.6679992079734802,
"learning_rate": 2.2794117647058826e-06,
"loss": 0.5983,
"step": 31
},
{
"epoch": 0.0706401766004415,
"grad_norm": 0.6695526242256165,
"learning_rate": 2.3529411764705885e-06,
"loss": 0.5946,
"step": 32
},
{
"epoch": 0.0728476821192053,
"grad_norm": 0.6435050964355469,
"learning_rate": 2.4264705882352943e-06,
"loss": 0.5975,
"step": 33
},
{
"epoch": 0.07505518763796909,
"grad_norm": 0.6036580204963684,
"learning_rate": 2.5e-06,
"loss": 0.592,
"step": 34
},
{
"epoch": 0.0772626931567329,
"grad_norm": 0.5135894417762756,
"learning_rate": 2.5735294117647057e-06,
"loss": 0.596,
"step": 35
},
{
"epoch": 0.07947019867549669,
"grad_norm": 0.4571033716201782,
"learning_rate": 2.647058823529412e-06,
"loss": 0.5764,
"step": 36
},
{
"epoch": 0.08167770419426049,
"grad_norm": 0.48609447479248047,
"learning_rate": 2.720588235294118e-06,
"loss": 0.5593,
"step": 37
},
{
"epoch": 0.08388520971302428,
"grad_norm": 0.49421966075897217,
"learning_rate": 2.7941176470588237e-06,
"loss": 0.5634,
"step": 38
},
{
"epoch": 0.08609271523178808,
"grad_norm": 0.48713281750679016,
"learning_rate": 2.8676470588235296e-06,
"loss": 0.5577,
"step": 39
},
{
"epoch": 0.08830022075055188,
"grad_norm": 0.42998674511909485,
"learning_rate": 2.9411764705882355e-06,
"loss": 0.5602,
"step": 40
},
{
"epoch": 0.09050772626931568,
"grad_norm": 0.39199528098106384,
"learning_rate": 3.0147058823529413e-06,
"loss": 0.5509,
"step": 41
},
{
"epoch": 0.09271523178807947,
"grad_norm": 0.3977169692516327,
"learning_rate": 3.0882352941176476e-06,
"loss": 0.5408,
"step": 42
},
{
"epoch": 0.09492273730684327,
"grad_norm": 0.3659592866897583,
"learning_rate": 3.161764705882353e-06,
"loss": 0.5468,
"step": 43
},
{
"epoch": 0.09713024282560706,
"grad_norm": 0.41232988238334656,
"learning_rate": 3.2352941176470594e-06,
"loss": 0.5312,
"step": 44
},
{
"epoch": 0.09933774834437085,
"grad_norm": 0.41818928718566895,
"learning_rate": 3.308823529411765e-06,
"loss": 0.542,
"step": 45
},
{
"epoch": 0.10154525386313466,
"grad_norm": 0.38174012303352356,
"learning_rate": 3.382352941176471e-06,
"loss": 0.5341,
"step": 46
},
{
"epoch": 0.10375275938189846,
"grad_norm": 0.3903500735759735,
"learning_rate": 3.4558823529411766e-06,
"loss": 0.5215,
"step": 47
},
{
"epoch": 0.10596026490066225,
"grad_norm": 0.3658589720726013,
"learning_rate": 3.529411764705883e-06,
"loss": 0.5277,
"step": 48
},
{
"epoch": 0.10816777041942605,
"grad_norm": 0.3160182237625122,
"learning_rate": 3.6029411764705883e-06,
"loss": 0.5193,
"step": 49
},
{
"epoch": 0.11037527593818984,
"grad_norm": 0.29906004667282104,
"learning_rate": 3.6764705882352946e-06,
"loss": 0.525,
"step": 50
},
{
"epoch": 0.11258278145695365,
"grad_norm": 0.28197285532951355,
"learning_rate": 3.7500000000000005e-06,
"loss": 0.5136,
"step": 51
},
{
"epoch": 0.11479028697571744,
"grad_norm": 0.2733807861804962,
"learning_rate": 3.8235294117647055e-06,
"loss": 0.5126,
"step": 52
},
{
"epoch": 0.11699779249448124,
"grad_norm": 0.2581369876861572,
"learning_rate": 3.897058823529412e-06,
"loss": 0.5102,
"step": 53
},
{
"epoch": 0.11920529801324503,
"grad_norm": 0.24509315192699432,
"learning_rate": 3.970588235294118e-06,
"loss": 0.4959,
"step": 54
},
{
"epoch": 0.12141280353200883,
"grad_norm": 0.23736661672592163,
"learning_rate": 4.044117647058824e-06,
"loss": 0.5024,
"step": 55
},
{
"epoch": 0.12362030905077263,
"grad_norm": 0.2297072857618332,
"learning_rate": 4.11764705882353e-06,
"loss": 0.516,
"step": 56
},
{
"epoch": 0.12582781456953643,
"grad_norm": 0.21641261875629425,
"learning_rate": 4.191176470588236e-06,
"loss": 0.5045,
"step": 57
},
{
"epoch": 0.1280353200883002,
"grad_norm": 0.2174406796693802,
"learning_rate": 4.264705882352942e-06,
"loss": 0.4954,
"step": 58
},
{
"epoch": 0.13024282560706402,
"grad_norm": 0.2182844579219818,
"learning_rate": 4.3382352941176475e-06,
"loss": 0.5073,
"step": 59
},
{
"epoch": 0.13245033112582782,
"grad_norm": 0.18668220937252045,
"learning_rate": 4.411764705882353e-06,
"loss": 0.4927,
"step": 60
},
{
"epoch": 0.1346578366445916,
"grad_norm": 0.20196162164211273,
"learning_rate": 4.485294117647059e-06,
"loss": 0.4962,
"step": 61
},
{
"epoch": 0.1368653421633554,
"grad_norm": 0.20648355782032013,
"learning_rate": 4.558823529411765e-06,
"loss": 0.4965,
"step": 62
},
{
"epoch": 0.1390728476821192,
"grad_norm": 0.1998893916606903,
"learning_rate": 4.632352941176471e-06,
"loss": 0.4857,
"step": 63
},
{
"epoch": 0.141280353200883,
"grad_norm": 0.20052312314510345,
"learning_rate": 4.705882352941177e-06,
"loss": 0.4919,
"step": 64
},
{
"epoch": 0.1434878587196468,
"grad_norm": 0.18653374910354614,
"learning_rate": 4.779411764705883e-06,
"loss": 0.4895,
"step": 65
},
{
"epoch": 0.1456953642384106,
"grad_norm": 0.17638395726680756,
"learning_rate": 4.852941176470589e-06,
"loss": 0.4831,
"step": 66
},
{
"epoch": 0.1479028697571744,
"grad_norm": 0.17102564871311188,
"learning_rate": 4.9264705882352945e-06,
"loss": 0.4817,
"step": 67
},
{
"epoch": 0.15011037527593818,
"grad_norm": 0.15633539855480194,
"learning_rate": 5e-06,
"loss": 0.4882,
"step": 68
},
{
"epoch": 0.152317880794702,
"grad_norm": 0.17420779168605804,
"learning_rate": 5.073529411764706e-06,
"loss": 0.4853,
"step": 69
},
{
"epoch": 0.1545253863134658,
"grad_norm": 0.1659373641014099,
"learning_rate": 5.147058823529411e-06,
"loss": 0.492,
"step": 70
},
{
"epoch": 0.15673289183222958,
"grad_norm": 0.17147059738636017,
"learning_rate": 5.220588235294118e-06,
"loss": 0.4744,
"step": 71
},
{
"epoch": 0.15894039735099338,
"grad_norm": 0.15868496894836426,
"learning_rate": 5.294117647058824e-06,
"loss": 0.4845,
"step": 72
},
{
"epoch": 0.16114790286975716,
"grad_norm": 0.16405610740184784,
"learning_rate": 5.36764705882353e-06,
"loss": 0.4806,
"step": 73
},
{
"epoch": 0.16335540838852097,
"grad_norm": 0.16762660443782806,
"learning_rate": 5.441176470588236e-06,
"loss": 0.465,
"step": 74
},
{
"epoch": 0.16556291390728478,
"grad_norm": 0.209846630692482,
"learning_rate": 5.514705882352942e-06,
"loss": 0.486,
"step": 75
},
{
"epoch": 0.16777041942604856,
"grad_norm": 0.17751334607601166,
"learning_rate": 5.588235294117647e-06,
"loss": 0.4761,
"step": 76
},
{
"epoch": 0.16997792494481237,
"grad_norm": 0.16121278703212738,
"learning_rate": 5.661764705882353e-06,
"loss": 0.4774,
"step": 77
},
{
"epoch": 0.17218543046357615,
"grad_norm": 0.2625029385089874,
"learning_rate": 5.735294117647059e-06,
"loss": 0.4663,
"step": 78
},
{
"epoch": 0.17439293598233996,
"grad_norm": 0.15058760344982147,
"learning_rate": 5.808823529411766e-06,
"loss": 0.4644,
"step": 79
},
{
"epoch": 0.17660044150110377,
"grad_norm": 0.16306105256080627,
"learning_rate": 5.882352941176471e-06,
"loss": 0.4745,
"step": 80
},
{
"epoch": 0.17880794701986755,
"grad_norm": 0.1497834473848343,
"learning_rate": 5.955882352941177e-06,
"loss": 0.4614,
"step": 81
},
{
"epoch": 0.18101545253863136,
"grad_norm": 0.15127182006835938,
"learning_rate": 6.029411764705883e-06,
"loss": 0.468,
"step": 82
},
{
"epoch": 0.18322295805739514,
"grad_norm": 0.14110144972801208,
"learning_rate": 6.102941176470589e-06,
"loss": 0.4709,
"step": 83
},
{
"epoch": 0.18543046357615894,
"grad_norm": 0.150424063205719,
"learning_rate": 6.176470588235295e-06,
"loss": 0.4614,
"step": 84
},
{
"epoch": 0.18763796909492272,
"grad_norm": 0.14802858233451843,
"learning_rate": 6.25e-06,
"loss": 0.4549,
"step": 85
},
{
"epoch": 0.18984547461368653,
"grad_norm": 0.1567091941833496,
"learning_rate": 6.323529411764706e-06,
"loss": 0.4487,
"step": 86
},
{
"epoch": 0.19205298013245034,
"grad_norm": 0.14341039955615997,
"learning_rate": 6.397058823529412e-06,
"loss": 0.4453,
"step": 87
},
{
"epoch": 0.19426048565121412,
"grad_norm": 0.20370961725711823,
"learning_rate": 6.470588235294119e-06,
"loss": 0.4626,
"step": 88
},
{
"epoch": 0.19646799116997793,
"grad_norm": 0.14623787999153137,
"learning_rate": 6.544117647058824e-06,
"loss": 0.4559,
"step": 89
},
{
"epoch": 0.1986754966887417,
"grad_norm": 0.1428503692150116,
"learning_rate": 6.61764705882353e-06,
"loss": 0.4464,
"step": 90
},
{
"epoch": 0.20088300220750552,
"grad_norm": 0.14742067456245422,
"learning_rate": 6.6911764705882356e-06,
"loss": 0.4663,
"step": 91
},
{
"epoch": 0.20309050772626933,
"grad_norm": 0.16871479153633118,
"learning_rate": 6.764705882352942e-06,
"loss": 0.4549,
"step": 92
},
{
"epoch": 0.2052980132450331,
"grad_norm": 0.14448532462120056,
"learning_rate": 6.838235294117648e-06,
"loss": 0.4544,
"step": 93
},
{
"epoch": 0.20750551876379691,
"grad_norm": 0.17373333871364594,
"learning_rate": 6.911764705882353e-06,
"loss": 0.4557,
"step": 94
},
{
"epoch": 0.2097130242825607,
"grad_norm": 0.15801453590393066,
"learning_rate": 6.985294117647059e-06,
"loss": 0.4507,
"step": 95
},
{
"epoch": 0.2119205298013245,
"grad_norm": 0.15145470201969147,
"learning_rate": 7.058823529411766e-06,
"loss": 0.4499,
"step": 96
},
{
"epoch": 0.2141280353200883,
"grad_norm": 0.14600904285907745,
"learning_rate": 7.132352941176472e-06,
"loss": 0.4548,
"step": 97
},
{
"epoch": 0.2163355408388521,
"grad_norm": 0.15834525227546692,
"learning_rate": 7.205882352941177e-06,
"loss": 0.4505,
"step": 98
},
{
"epoch": 0.2185430463576159,
"grad_norm": 0.15612734854221344,
"learning_rate": 7.2794117647058826e-06,
"loss": 0.4598,
"step": 99
},
{
"epoch": 0.22075055187637968,
"grad_norm": 0.18165510892868042,
"learning_rate": 7.352941176470589e-06,
"loss": 0.4479,
"step": 100
},
{
"epoch": 0.2229580573951435,
"grad_norm": 0.16820134222507477,
"learning_rate": 7.426470588235295e-06,
"loss": 0.4497,
"step": 101
},
{
"epoch": 0.2251655629139073,
"grad_norm": 0.16453172266483307,
"learning_rate": 7.500000000000001e-06,
"loss": 0.4532,
"step": 102
},
{
"epoch": 0.22737306843267108,
"grad_norm": 0.15813149511814117,
"learning_rate": 7.573529411764706e-06,
"loss": 0.4513,
"step": 103
},
{
"epoch": 0.22958057395143489,
"grad_norm": 0.1652165949344635,
"learning_rate": 7.647058823529411e-06,
"loss": 0.445,
"step": 104
},
{
"epoch": 0.23178807947019867,
"grad_norm": 0.14849768579006195,
"learning_rate": 7.720588235294119e-06,
"loss": 0.435,
"step": 105
},
{
"epoch": 0.23399558498896247,
"grad_norm": 0.17478714883327484,
"learning_rate": 7.794117647058825e-06,
"loss": 0.4343,
"step": 106
},
{
"epoch": 0.23620309050772628,
"grad_norm": 0.16727301478385925,
"learning_rate": 7.86764705882353e-06,
"loss": 0.4483,
"step": 107
},
{
"epoch": 0.23841059602649006,
"grad_norm": 0.1674540638923645,
"learning_rate": 7.941176470588236e-06,
"loss": 0.4452,
"step": 108
},
{
"epoch": 0.24061810154525387,
"grad_norm": 0.16588866710662842,
"learning_rate": 8.014705882352942e-06,
"loss": 0.4517,
"step": 109
},
{
"epoch": 0.24282560706401765,
"grad_norm": 0.176283061504364,
"learning_rate": 8.088235294117648e-06,
"loss": 0.4434,
"step": 110
},
{
"epoch": 0.24503311258278146,
"grad_norm": 0.17021676898002625,
"learning_rate": 8.161764705882354e-06,
"loss": 0.442,
"step": 111
},
{
"epoch": 0.24724061810154527,
"grad_norm": 0.15938061475753784,
"learning_rate": 8.23529411764706e-06,
"loss": 0.4436,
"step": 112
},
{
"epoch": 0.24944812362030905,
"grad_norm": 0.18990886211395264,
"learning_rate": 8.308823529411766e-06,
"loss": 0.439,
"step": 113
},
{
"epoch": 0.25165562913907286,
"grad_norm": 0.16168883442878723,
"learning_rate": 8.382352941176472e-06,
"loss": 0.4392,
"step": 114
},
{
"epoch": 0.25386313465783666,
"grad_norm": 0.2176610231399536,
"learning_rate": 8.455882352941177e-06,
"loss": 0.4488,
"step": 115
},
{
"epoch": 0.2560706401766004,
"grad_norm": 0.19106176495552063,
"learning_rate": 8.529411764705883e-06,
"loss": 0.436,
"step": 116
},
{
"epoch": 0.2582781456953642,
"grad_norm": 0.1480225920677185,
"learning_rate": 8.60294117647059e-06,
"loss": 0.4386,
"step": 117
},
{
"epoch": 0.26048565121412803,
"grad_norm": 0.20528610050678253,
"learning_rate": 8.676470588235295e-06,
"loss": 0.4441,
"step": 118
},
{
"epoch": 0.26269315673289184,
"grad_norm": 0.18629467487335205,
"learning_rate": 8.750000000000001e-06,
"loss": 0.4322,
"step": 119
},
{
"epoch": 0.26490066225165565,
"grad_norm": 0.1764117330312729,
"learning_rate": 8.823529411764707e-06,
"loss": 0.4261,
"step": 120
},
{
"epoch": 0.2671081677704194,
"grad_norm": 0.1975659281015396,
"learning_rate": 8.897058823529413e-06,
"loss": 0.4401,
"step": 121
},
{
"epoch": 0.2693156732891832,
"grad_norm": 0.2136935442686081,
"learning_rate": 8.970588235294119e-06,
"loss": 0.4416,
"step": 122
},
{
"epoch": 0.271523178807947,
"grad_norm": 0.18505676090717316,
"learning_rate": 9.044117647058824e-06,
"loss": 0.4423,
"step": 123
},
{
"epoch": 0.2737306843267108,
"grad_norm": 0.176743283867836,
"learning_rate": 9.11764705882353e-06,
"loss": 0.4435,
"step": 124
},
{
"epoch": 0.27593818984547464,
"grad_norm": 0.17542196810245514,
"learning_rate": 9.191176470588236e-06,
"loss": 0.4255,
"step": 125
},
{
"epoch": 0.2781456953642384,
"grad_norm": 0.15908770263195038,
"learning_rate": 9.264705882352942e-06,
"loss": 0.438,
"step": 126
},
{
"epoch": 0.2803532008830022,
"grad_norm": 0.16695120930671692,
"learning_rate": 9.338235294117648e-06,
"loss": 0.4418,
"step": 127
},
{
"epoch": 0.282560706401766,
"grad_norm": 0.1714697927236557,
"learning_rate": 9.411764705882354e-06,
"loss": 0.4293,
"step": 128
},
{
"epoch": 0.2847682119205298,
"grad_norm": 0.1601938009262085,
"learning_rate": 9.48529411764706e-06,
"loss": 0.4402,
"step": 129
},
{
"epoch": 0.2869757174392936,
"grad_norm": 0.16425947844982147,
"learning_rate": 9.558823529411766e-06,
"loss": 0.4328,
"step": 130
},
{
"epoch": 0.2891832229580574,
"grad_norm": 0.1931643933057785,
"learning_rate": 9.632352941176471e-06,
"loss": 0.4307,
"step": 131
},
{
"epoch": 0.2913907284768212,
"grad_norm": 0.18675902485847473,
"learning_rate": 9.705882352941177e-06,
"loss": 0.4554,
"step": 132
},
{
"epoch": 0.293598233995585,
"grad_norm": 0.16391406953334808,
"learning_rate": 9.779411764705883e-06,
"loss": 0.4272,
"step": 133
},
{
"epoch": 0.2958057395143488,
"grad_norm": 0.20189572870731354,
"learning_rate": 9.852941176470589e-06,
"loss": 0.4258,
"step": 134
},
{
"epoch": 0.2980132450331126,
"grad_norm": 0.16828037798404694,
"learning_rate": 9.926470588235295e-06,
"loss": 0.4287,
"step": 135
},
{
"epoch": 0.30022075055187636,
"grad_norm": 0.1638776957988739,
"learning_rate": 1e-05,
"loss": 0.4326,
"step": 136
},
{
"epoch": 0.30242825607064017,
"grad_norm": 0.20775483548641205,
"learning_rate": 9.999983503697906e-06,
"loss": 0.4274,
"step": 137
},
{
"epoch": 0.304635761589404,
"grad_norm": 0.17668417096138,
"learning_rate": 9.999934014900475e-06,
"loss": 0.4277,
"step": 138
},
{
"epoch": 0.3068432671081678,
"grad_norm": 0.20949822664260864,
"learning_rate": 9.999851533934259e-06,
"loss": 0.4277,
"step": 139
},
{
"epoch": 0.3090507726269316,
"grad_norm": 0.13797180354595184,
"learning_rate": 9.999736061343512e-06,
"loss": 0.4072,
"step": 140
},
{
"epoch": 0.31125827814569534,
"grad_norm": 0.20374132692813873,
"learning_rate": 9.99958759789018e-06,
"loss": 0.4302,
"step": 141
},
{
"epoch": 0.31346578366445915,
"grad_norm": 0.1946762204170227,
"learning_rate": 9.999406144553905e-06,
"loss": 0.4213,
"step": 142
},
{
"epoch": 0.31567328918322296,
"grad_norm": 0.21811267733573914,
"learning_rate": 9.999191702532008e-06,
"loss": 0.4285,
"step": 143
},
{
"epoch": 0.31788079470198677,
"grad_norm": 0.20533326268196106,
"learning_rate": 9.99894427323949e-06,
"loss": 0.4251,
"step": 144
},
{
"epoch": 0.3200883002207506,
"grad_norm": 0.20235486328601837,
"learning_rate": 9.99866385830902e-06,
"loss": 0.4237,
"step": 145
},
{
"epoch": 0.32229580573951433,
"grad_norm": 0.20962080359458923,
"learning_rate": 9.99835045959092e-06,
"loss": 0.4266,
"step": 146
},
{
"epoch": 0.32450331125827814,
"grad_norm": 0.22475510835647583,
"learning_rate": 9.998004079153156e-06,
"loss": 0.4263,
"step": 147
},
{
"epoch": 0.32671081677704195,
"grad_norm": 0.20724737644195557,
"learning_rate": 9.997624719281332e-06,
"loss": 0.416,
"step": 148
},
{
"epoch": 0.32891832229580575,
"grad_norm": 0.1801760494709015,
"learning_rate": 9.997212382478658e-06,
"loss": 0.4233,
"step": 149
},
{
"epoch": 0.33112582781456956,
"grad_norm": 0.23205707967281342,
"learning_rate": 9.996767071465947e-06,
"loss": 0.4277,
"step": 150
},
{
"epoch": 0.3333333333333333,
"grad_norm": 0.18537338078022003,
"learning_rate": 9.996288789181595e-06,
"loss": 0.4317,
"step": 151
},
{
"epoch": 0.3355408388520971,
"grad_norm": 0.17498917877674103,
"learning_rate": 9.995777538781556e-06,
"loss": 0.4288,
"step": 152
},
{
"epoch": 0.33774834437086093,
"grad_norm": 0.18248897790908813,
"learning_rate": 9.995233323639326e-06,
"loss": 0.4261,
"step": 153
},
{
"epoch": 0.33995584988962474,
"grad_norm": 0.1849048137664795,
"learning_rate": 9.994656147345922e-06,
"loss": 0.4216,
"step": 154
},
{
"epoch": 0.34216335540838855,
"grad_norm": 0.19169744849205017,
"learning_rate": 9.994046013709852e-06,
"loss": 0.423,
"step": 155
},
{
"epoch": 0.3443708609271523,
"grad_norm": 0.18524128198623657,
"learning_rate": 9.993402926757098e-06,
"loss": 0.4213,
"step": 156
},
{
"epoch": 0.3465783664459161,
"grad_norm": 0.16218866407871246,
"learning_rate": 9.99272689073108e-06,
"loss": 0.4252,
"step": 157
},
{
"epoch": 0.3487858719646799,
"grad_norm": 0.17451681196689606,
"learning_rate": 9.992017910092636e-06,
"loss": 0.4251,
"step": 158
},
{
"epoch": 0.3509933774834437,
"grad_norm": 0.15944437682628632,
"learning_rate": 9.991275989519991e-06,
"loss": 0.4123,
"step": 159
},
{
"epoch": 0.35320088300220753,
"grad_norm": 0.17371642589569092,
"learning_rate": 9.990501133908722e-06,
"loss": 0.4234,
"step": 160
},
{
"epoch": 0.3554083885209713,
"grad_norm": 0.1783660650253296,
"learning_rate": 9.98969334837173e-06,
"loss": 0.4234,
"step": 161
},
{
"epoch": 0.3576158940397351,
"grad_norm": 0.1762082278728485,
"learning_rate": 9.988852638239206e-06,
"loss": 0.418,
"step": 162
},
{
"epoch": 0.3598233995584989,
"grad_norm": 0.16491912305355072,
"learning_rate": 9.987979009058593e-06,
"loss": 0.4248,
"step": 163
},
{
"epoch": 0.3620309050772627,
"grad_norm": 0.2407284379005432,
"learning_rate": 9.98707246659455e-06,
"loss": 0.4307,
"step": 164
},
{
"epoch": 0.36423841059602646,
"grad_norm": 0.19042451679706573,
"learning_rate": 9.986133016828916e-06,
"loss": 0.4231,
"step": 165
},
{
"epoch": 0.36644591611479027,
"grad_norm": 0.18169504404067993,
"learning_rate": 9.985160665960672e-06,
"loss": 0.4266,
"step": 166
},
{
"epoch": 0.3686534216335541,
"grad_norm": 0.18646620213985443,
"learning_rate": 9.984155420405895e-06,
"loss": 0.4231,
"step": 167
},
{
"epoch": 0.3708609271523179,
"grad_norm": 0.19154079258441925,
"learning_rate": 9.983117286797718e-06,
"loss": 0.4308,
"step": 168
},
{
"epoch": 0.3730684326710817,
"grad_norm": 0.17594484984874725,
"learning_rate": 9.982046271986287e-06,
"loss": 0.4115,
"step": 169
},
{
"epoch": 0.37527593818984545,
"grad_norm": 0.18167531490325928,
"learning_rate": 9.980942383038717e-06,
"loss": 0.424,
"step": 170
},
{
"epoch": 0.37748344370860926,
"grad_norm": 0.1535561978816986,
"learning_rate": 9.97980562723904e-06,
"loss": 0.4296,
"step": 171
},
{
"epoch": 0.37969094922737306,
"grad_norm": 0.15756377577781677,
"learning_rate": 9.978636012088165e-06,
"loss": 0.4169,
"step": 172
},
{
"epoch": 0.3818984547461369,
"grad_norm": 0.15670788288116455,
"learning_rate": 9.97743354530382e-06,
"loss": 0.4394,
"step": 173
},
{
"epoch": 0.3841059602649007,
"grad_norm": 0.16224409639835358,
"learning_rate": 9.976198234820509e-06,
"loss": 0.4228,
"step": 174
},
{
"epoch": 0.38631346578366443,
"grad_norm": 0.14743737876415253,
"learning_rate": 9.974930088789452e-06,
"loss": 0.4144,
"step": 175
},
{
"epoch": 0.38852097130242824,
"grad_norm": 0.1594422310590744,
"learning_rate": 9.97362911557854e-06,
"loss": 0.4168,
"step": 176
},
{
"epoch": 0.39072847682119205,
"grad_norm": 0.16173714399337769,
"learning_rate": 9.972295323772268e-06,
"loss": 0.4166,
"step": 177
},
{
"epoch": 0.39293598233995586,
"grad_norm": 0.1668204814195633,
"learning_rate": 9.970928722171691e-06,
"loss": 0.4252,
"step": 178
},
{
"epoch": 0.39514348785871967,
"grad_norm": 0.15836164355278015,
"learning_rate": 9.96952931979436e-06,
"loss": 0.4209,
"step": 179
},
{
"epoch": 0.3973509933774834,
"grad_norm": 0.1634080708026886,
"learning_rate": 9.968097125874258e-06,
"loss": 0.4076,
"step": 180
},
{
"epoch": 0.3995584988962472,
"grad_norm": 0.16134855151176453,
"learning_rate": 9.966632149861748e-06,
"loss": 0.4276,
"step": 181
},
{
"epoch": 0.40176600441501104,
"grad_norm": 0.15993578732013702,
"learning_rate": 9.965134401423503e-06,
"loss": 0.4308,
"step": 182
},
{
"epoch": 0.40397350993377484,
"grad_norm": 0.19045297801494598,
"learning_rate": 9.963603890442448e-06,
"loss": 0.4185,
"step": 183
},
{
"epoch": 0.40618101545253865,
"grad_norm": 0.16455209255218506,
"learning_rate": 9.962040627017693e-06,
"loss": 0.4232,
"step": 184
},
{
"epoch": 0.4083885209713024,
"grad_norm": 0.16577620804309845,
"learning_rate": 9.960444621464462e-06,
"loss": 0.4149,
"step": 185
},
{
"epoch": 0.4105960264900662,
"grad_norm": 0.25177431106567383,
"learning_rate": 9.958815884314033e-06,
"loss": 0.4172,
"step": 186
},
{
"epoch": 0.41280353200883,
"grad_norm": 0.18712477385997772,
"learning_rate": 9.957154426313662e-06,
"loss": 0.4205,
"step": 187
},
{
"epoch": 0.41501103752759383,
"grad_norm": 0.1667563021183014,
"learning_rate": 9.955460258426512e-06,
"loss": 0.4207,
"step": 188
},
{
"epoch": 0.41721854304635764,
"grad_norm": 0.22865413129329681,
"learning_rate": 9.953733391831586e-06,
"loss": 0.4109,
"step": 189
},
{
"epoch": 0.4194260485651214,
"grad_norm": 0.18536990880966187,
"learning_rate": 9.951973837923652e-06,
"loss": 0.4187,
"step": 190
},
{
"epoch": 0.4216335540838852,
"grad_norm": 0.19504587352275848,
"learning_rate": 9.950181608313158e-06,
"loss": 0.4142,
"step": 191
},
{
"epoch": 0.423841059602649,
"grad_norm": 0.1892482340335846,
"learning_rate": 9.948356714826172e-06,
"loss": 0.4142,
"step": 192
},
{
"epoch": 0.4260485651214128,
"grad_norm": 0.1839127093553543,
"learning_rate": 9.946499169504294e-06,
"loss": 0.4161,
"step": 193
},
{
"epoch": 0.4282560706401766,
"grad_norm": 0.20385828614234924,
"learning_rate": 9.944608984604569e-06,
"loss": 0.4124,
"step": 194
},
{
"epoch": 0.4304635761589404,
"grad_norm": 0.1948205530643463,
"learning_rate": 9.942686172599425e-06,
"loss": 0.4251,
"step": 195
},
{
"epoch": 0.4326710816777042,
"grad_norm": 0.19438982009887695,
"learning_rate": 9.940730746176578e-06,
"loss": 0.4158,
"step": 196
},
{
"epoch": 0.434878587196468,
"grad_norm": 0.17213338613510132,
"learning_rate": 9.93874271823895e-06,
"loss": 0.4175,
"step": 197
},
{
"epoch": 0.4370860927152318,
"grad_norm": 0.22870118916034698,
"learning_rate": 9.936722101904582e-06,
"loss": 0.4267,
"step": 198
},
{
"epoch": 0.4392935982339956,
"grad_norm": 0.20383016765117645,
"learning_rate": 9.934668910506555e-06,
"loss": 0.422,
"step": 199
},
{
"epoch": 0.44150110375275936,
"grad_norm": 0.16936808824539185,
"learning_rate": 9.932583157592896e-06,
"loss": 0.4144,
"step": 200
},
{
"epoch": 0.44370860927152317,
"grad_norm": 0.19149592518806458,
"learning_rate": 9.930464856926488e-06,
"loss": 0.4077,
"step": 201
},
{
"epoch": 0.445916114790287,
"grad_norm": 0.15204112231731415,
"learning_rate": 9.928314022484982e-06,
"loss": 0.4207,
"step": 202
},
{
"epoch": 0.4481236203090508,
"grad_norm": 0.19697798788547516,
"learning_rate": 9.926130668460702e-06,
"loss": 0.4159,
"step": 203
},
{
"epoch": 0.4503311258278146,
"grad_norm": 0.1751161813735962,
"learning_rate": 9.92391480926056e-06,
"loss": 0.4179,
"step": 204
},
{
"epoch": 0.45253863134657835,
"grad_norm": 0.17122775316238403,
"learning_rate": 9.921666459505944e-06,
"loss": 0.4104,
"step": 205
},
{
"epoch": 0.45474613686534215,
"grad_norm": 0.18898116052150726,
"learning_rate": 9.91938563403264e-06,
"loss": 0.4183,
"step": 206
},
{
"epoch": 0.45695364238410596,
"grad_norm": 0.17628274857997894,
"learning_rate": 9.917072347890721e-06,
"loss": 0.411,
"step": 207
},
{
"epoch": 0.45916114790286977,
"grad_norm": 0.19427300989627838,
"learning_rate": 9.914726616344454e-06,
"loss": 0.4144,
"step": 208
},
{
"epoch": 0.4613686534216336,
"grad_norm": 0.19164274632930756,
"learning_rate": 9.912348454872196e-06,
"loss": 0.4067,
"step": 209
},
{
"epoch": 0.46357615894039733,
"grad_norm": 0.17385149002075195,
"learning_rate": 9.909937879166298e-06,
"loss": 0.408,
"step": 210
},
{
"epoch": 0.46578366445916114,
"grad_norm": 0.18648236989974976,
"learning_rate": 9.907494905132994e-06,
"loss": 0.4273,
"step": 211
},
{
"epoch": 0.46799116997792495,
"grad_norm": 0.1911754459142685,
"learning_rate": 9.905019548892296e-06,
"loss": 0.4127,
"step": 212
},
{
"epoch": 0.47019867549668876,
"grad_norm": 0.15749873220920563,
"learning_rate": 9.902511826777895e-06,
"loss": 0.4174,
"step": 213
},
{
"epoch": 0.47240618101545256,
"grad_norm": 0.19831134378910065,
"learning_rate": 9.899971755337049e-06,
"loss": 0.4045,
"step": 214
},
{
"epoch": 0.4746136865342163,
"grad_norm": 0.18536338210105896,
"learning_rate": 9.897399351330471e-06,
"loss": 0.4134,
"step": 215
},
{
"epoch": 0.4768211920529801,
"grad_norm": 0.17010532319545746,
"learning_rate": 9.894794631732223e-06,
"loss": 0.4076,
"step": 216
},
{
"epoch": 0.47902869757174393,
"grad_norm": 0.17602422833442688,
"learning_rate": 9.8921576137296e-06,
"loss": 0.4079,
"step": 217
},
{
"epoch": 0.48123620309050774,
"grad_norm": 0.1983213573694229,
"learning_rate": 9.889488314723024e-06,
"loss": 0.413,
"step": 218
},
{
"epoch": 0.48344370860927155,
"grad_norm": 0.17187613248825073,
"learning_rate": 9.886786752325917e-06,
"loss": 0.4229,
"step": 219
},
{
"epoch": 0.4856512141280353,
"grad_norm": 0.2029353827238083,
"learning_rate": 9.884052944364595e-06,
"loss": 0.4242,
"step": 220
},
{
"epoch": 0.4878587196467991,
"grad_norm": 0.18107262253761292,
"learning_rate": 9.881286908878148e-06,
"loss": 0.4157,
"step": 221
},
{
"epoch": 0.4900662251655629,
"grad_norm": 0.18853497505187988,
"learning_rate": 9.878488664118316e-06,
"loss": 0.4153,
"step": 222
},
{
"epoch": 0.4922737306843267,
"grad_norm": 0.1775682419538498,
"learning_rate": 9.875658228549379e-06,
"loss": 0.4109,
"step": 223
},
{
"epoch": 0.49448123620309054,
"grad_norm": 0.18237102031707764,
"learning_rate": 9.872795620848024e-06,
"loss": 0.407,
"step": 224
},
{
"epoch": 0.4966887417218543,
"grad_norm": 0.20489446818828583,
"learning_rate": 9.869900859903225e-06,
"loss": 0.4198,
"step": 225
},
{
"epoch": 0.4988962472406181,
"grad_norm": 0.1481681913137436,
"learning_rate": 9.866973964816126e-06,
"loss": 0.4085,
"step": 226
},
{
"epoch": 0.5011037527593819,
"grad_norm": 0.18315470218658447,
"learning_rate": 9.864014954899905e-06,
"loss": 0.4079,
"step": 227
},
{
"epoch": 0.5033112582781457,
"grad_norm": 0.18866921961307526,
"learning_rate": 9.861023849679648e-06,
"loss": 0.4249,
"step": 228
},
{
"epoch": 0.5055187637969095,
"grad_norm": 0.15121376514434814,
"learning_rate": 9.858000668892226e-06,
"loss": 0.4179,
"step": 229
},
{
"epoch": 0.5077262693156733,
"grad_norm": 0.19818085432052612,
"learning_rate": 9.85494543248616e-06,
"loss": 0.4218,
"step": 230
},
{
"epoch": 0.5099337748344371,
"grad_norm": 0.15964657068252563,
"learning_rate": 9.851858160621496e-06,
"loss": 0.4074,
"step": 231
},
{
"epoch": 0.5121412803532008,
"grad_norm": 0.18371723592281342,
"learning_rate": 9.848738873669653e-06,
"loss": 0.4103,
"step": 232
},
{
"epoch": 0.5143487858719646,
"grad_norm": 0.1854197084903717,
"learning_rate": 9.845587592213318e-06,
"loss": 0.4092,
"step": 233
},
{
"epoch": 0.5165562913907285,
"grad_norm": 0.15952259302139282,
"learning_rate": 9.842404337046284e-06,
"loss": 0.4139,
"step": 234
},
{
"epoch": 0.5187637969094923,
"grad_norm": 0.1830589473247528,
"learning_rate": 9.839189129173328e-06,
"loss": 0.4143,
"step": 235
},
{
"epoch": 0.5209713024282561,
"grad_norm": 0.1684870719909668,
"learning_rate": 9.835941989810065e-06,
"loss": 0.4088,
"step": 236
},
{
"epoch": 0.5231788079470199,
"grad_norm": 0.17676231265068054,
"learning_rate": 9.832662940382813e-06,
"loss": 0.4123,
"step": 237
},
{
"epoch": 0.5253863134657837,
"grad_norm": 0.17995555698871613,
"learning_rate": 9.829352002528449e-06,
"loss": 0.4086,
"step": 238
},
{
"epoch": 0.5275938189845475,
"grad_norm": 0.1694101095199585,
"learning_rate": 9.826009198094262e-06,
"loss": 0.4137,
"step": 239
},
{
"epoch": 0.5298013245033113,
"grad_norm": 0.16819199919700623,
"learning_rate": 9.822634549137819e-06,
"loss": 0.4072,
"step": 240
},
{
"epoch": 0.5320088300220751,
"grad_norm": 0.17408284544944763,
"learning_rate": 9.81922807792681e-06,
"loss": 0.4052,
"step": 241
},
{
"epoch": 0.5342163355408388,
"grad_norm": 0.1843319982290268,
"learning_rate": 9.815789806938909e-06,
"loss": 0.3966,
"step": 242
},
{
"epoch": 0.5364238410596026,
"grad_norm": 0.18125468492507935,
"learning_rate": 9.812319758861616e-06,
"loss": 0.4044,
"step": 243
},
{
"epoch": 0.5386313465783664,
"grad_norm": 0.16861899197101593,
"learning_rate": 9.808817956592115e-06,
"loss": 0.4092,
"step": 244
},
{
"epoch": 0.5408388520971302,
"grad_norm": 0.22598163783550262,
"learning_rate": 9.805284423237126e-06,
"loss": 0.4137,
"step": 245
},
{
"epoch": 0.543046357615894,
"grad_norm": 0.22511421144008636,
"learning_rate": 9.801719182112738e-06,
"loss": 0.4073,
"step": 246
},
{
"epoch": 0.5452538631346578,
"grad_norm": 0.1690152883529663,
"learning_rate": 9.798122256744269e-06,
"loss": 0.4097,
"step": 247
},
{
"epoch": 0.5474613686534217,
"grad_norm": 0.22893239557743073,
"learning_rate": 9.794493670866108e-06,
"loss": 0.4054,
"step": 248
},
{
"epoch": 0.5496688741721855,
"grad_norm": 0.17257164418697357,
"learning_rate": 9.790833448421554e-06,
"loss": 0.4109,
"step": 249
},
{
"epoch": 0.5518763796909493,
"grad_norm": 0.1636303812265396,
"learning_rate": 9.787141613562661e-06,
"loss": 0.3995,
"step": 250
},
{
"epoch": 0.5540838852097131,
"grad_norm": 0.17622527480125427,
"learning_rate": 9.783418190650079e-06,
"loss": 0.409,
"step": 251
},
{
"epoch": 0.5562913907284768,
"grad_norm": 0.20090465247631073,
"learning_rate": 9.779663204252887e-06,
"loss": 0.4015,
"step": 252
},
{
"epoch": 0.5584988962472406,
"grad_norm": 0.15368807315826416,
"learning_rate": 9.775876679148449e-06,
"loss": 0.4009,
"step": 253
},
{
"epoch": 0.5607064017660044,
"grad_norm": 0.2135976403951645,
"learning_rate": 9.772058640322221e-06,
"loss": 0.407,
"step": 254
},
{
"epoch": 0.5629139072847682,
"grad_norm": 0.17839573323726654,
"learning_rate": 9.768209112967619e-06,
"loss": 0.404,
"step": 255
},
{
"epoch": 0.565121412803532,
"grad_norm": 0.1953095942735672,
"learning_rate": 9.764328122485827e-06,
"loss": 0.402,
"step": 256
},
{
"epoch": 0.5673289183222958,
"grad_norm": 0.16529282927513123,
"learning_rate": 9.76041569448564e-06,
"loss": 0.4117,
"step": 257
},
{
"epoch": 0.5695364238410596,
"grad_norm": 0.17458973824977875,
"learning_rate": 9.756471854783297e-06,
"loss": 0.4144,
"step": 258
},
{
"epoch": 0.5717439293598234,
"grad_norm": 0.1776653230190277,
"learning_rate": 9.752496629402307e-06,
"loss": 0.4123,
"step": 259
},
{
"epoch": 0.5739514348785872,
"grad_norm": 0.17428044974803925,
"learning_rate": 9.748490044573275e-06,
"loss": 0.3975,
"step": 260
},
{
"epoch": 0.5761589403973509,
"grad_norm": 0.17150405049324036,
"learning_rate": 9.744452126733739e-06,
"loss": 0.4023,
"step": 261
},
{
"epoch": 0.5783664459161147,
"grad_norm": 0.1668493002653122,
"learning_rate": 9.740382902527981e-06,
"loss": 0.409,
"step": 262
},
{
"epoch": 0.5805739514348786,
"grad_norm": 0.17727597057819366,
"learning_rate": 9.736282398806862e-06,
"loss": 0.4061,
"step": 263
},
{
"epoch": 0.5827814569536424,
"grad_norm": 0.15944141149520874,
"learning_rate": 9.73215064262764e-06,
"loss": 0.402,
"step": 264
},
{
"epoch": 0.5849889624724062,
"grad_norm": 0.17232947051525116,
"learning_rate": 9.727987661253796e-06,
"loss": 0.4102,
"step": 265
},
{
"epoch": 0.58719646799117,
"grad_norm": 0.2054961770772934,
"learning_rate": 9.72379348215485e-06,
"loss": 0.3994,
"step": 266
},
{
"epoch": 0.5894039735099338,
"grad_norm": 0.1817290335893631,
"learning_rate": 9.719568133006177e-06,
"loss": 0.4033,
"step": 267
},
{
"epoch": 0.5916114790286976,
"grad_norm": 0.16904965043067932,
"learning_rate": 9.715311641688835e-06,
"loss": 0.4082,
"step": 268
},
{
"epoch": 0.5938189845474614,
"grad_norm": 0.17628441751003265,
"learning_rate": 9.71102403628937e-06,
"loss": 0.3969,
"step": 269
},
{
"epoch": 0.5960264900662252,
"grad_norm": 0.20128114521503448,
"learning_rate": 9.706705345099632e-06,
"loss": 0.4081,
"step": 270
},
{
"epoch": 0.5982339955849889,
"grad_norm": 0.17412817478179932,
"learning_rate": 9.7023555966166e-06,
"loss": 0.4076,
"step": 271
},
{
"epoch": 0.6004415011037527,
"grad_norm": 0.18813695013523102,
"learning_rate": 9.697974819542178e-06,
"loss": 0.4042,
"step": 272
},
{
"epoch": 0.6026490066225165,
"grad_norm": 0.17176660895347595,
"learning_rate": 9.693563042783011e-06,
"loss": 0.4099,
"step": 273
},
{
"epoch": 0.6048565121412803,
"grad_norm": 0.17056933045387268,
"learning_rate": 9.689120295450308e-06,
"loss": 0.4045,
"step": 274
},
{
"epoch": 0.6070640176600441,
"grad_norm": 0.20970946550369263,
"learning_rate": 9.684646606859621e-06,
"loss": 0.3944,
"step": 275
},
{
"epoch": 0.609271523178808,
"grad_norm": 0.19319604337215424,
"learning_rate": 9.680142006530684e-06,
"loss": 0.4062,
"step": 276
},
{
"epoch": 0.6114790286975718,
"grad_norm": 0.17683015763759613,
"learning_rate": 9.675606524187192e-06,
"loss": 0.4077,
"step": 277
},
{
"epoch": 0.6136865342163356,
"grad_norm": 0.16389036178588867,
"learning_rate": 9.671040189756623e-06,
"loss": 0.412,
"step": 278
},
{
"epoch": 0.6158940397350994,
"grad_norm": 0.15970537066459656,
"learning_rate": 9.666443033370026e-06,
"loss": 0.4068,
"step": 279
},
{
"epoch": 0.6181015452538632,
"grad_norm": 0.17420579493045807,
"learning_rate": 9.661815085361836e-06,
"loss": 0.4046,
"step": 280
},
{
"epoch": 0.6203090507726269,
"grad_norm": 0.16548947989940643,
"learning_rate": 9.657156376269665e-06,
"loss": 0.4122,
"step": 281
},
{
"epoch": 0.6225165562913907,
"grad_norm": 0.147056445479393,
"learning_rate": 9.652466936834101e-06,
"loss": 0.4085,
"step": 282
},
{
"epoch": 0.6247240618101545,
"grad_norm": 0.17280390858650208,
"learning_rate": 9.647746797998508e-06,
"loss": 0.4142,
"step": 283
},
{
"epoch": 0.6269315673289183,
"grad_norm": 0.16386403143405914,
"learning_rate": 9.642995990908817e-06,
"loss": 0.4103,
"step": 284
},
{
"epoch": 0.6291390728476821,
"grad_norm": 0.16252438724040985,
"learning_rate": 9.638214546913333e-06,
"loss": 0.4023,
"step": 285
},
{
"epoch": 0.6313465783664459,
"grad_norm": 0.16047868132591248,
"learning_rate": 9.633402497562512e-06,
"loss": 0.4032,
"step": 286
},
{
"epoch": 0.6335540838852097,
"grad_norm": 0.18797720968723297,
"learning_rate": 9.628559874608761e-06,
"loss": 0.4014,
"step": 287
},
{
"epoch": 0.6357615894039735,
"grad_norm": 0.15865157544612885,
"learning_rate": 9.62368671000623e-06,
"loss": 0.4022,
"step": 288
},
{
"epoch": 0.6379690949227373,
"grad_norm": 0.15757879614830017,
"learning_rate": 9.618783035910596e-06,
"loss": 0.4019,
"step": 289
},
{
"epoch": 0.6401766004415012,
"grad_norm": 0.15986581146717072,
"learning_rate": 9.613848884678851e-06,
"loss": 0.3981,
"step": 290
},
{
"epoch": 0.6423841059602649,
"grad_norm": 0.1824173629283905,
"learning_rate": 9.608884288869103e-06,
"loss": 0.407,
"step": 291
},
{
"epoch": 0.6445916114790287,
"grad_norm": 0.185842826962471,
"learning_rate": 9.603889281240334e-06,
"loss": 0.4016,
"step": 292
},
{
"epoch": 0.6467991169977925,
"grad_norm": 0.17280805110931396,
"learning_rate": 9.59886389475221e-06,
"loss": 0.4081,
"step": 293
},
{
"epoch": 0.6490066225165563,
"grad_norm": 0.1815565824508667,
"learning_rate": 9.593808162564845e-06,
"loss": 0.4141,
"step": 294
},
{
"epoch": 0.6512141280353201,
"grad_norm": 0.19602055847644806,
"learning_rate": 9.588722118038595e-06,
"loss": 0.4048,
"step": 295
},
{
"epoch": 0.6534216335540839,
"grad_norm": 0.1916995495557785,
"learning_rate": 9.583605794733833e-06,
"loss": 0.3953,
"step": 296
},
{
"epoch": 0.6556291390728477,
"grad_norm": 0.17578016221523285,
"learning_rate": 9.578459226410722e-06,
"loss": 0.4133,
"step": 297
},
{
"epoch": 0.6578366445916115,
"grad_norm": 0.18694248795509338,
"learning_rate": 9.573282447029e-06,
"loss": 0.4023,
"step": 298
},
{
"epoch": 0.6600441501103753,
"grad_norm": 0.20029006898403168,
"learning_rate": 9.568075490747756e-06,
"loss": 0.3967,
"step": 299
},
{
"epoch": 0.6622516556291391,
"grad_norm": 0.15852802991867065,
"learning_rate": 9.562838391925197e-06,
"loss": 0.4053,
"step": 300
},
{
"epoch": 0.6644591611479028,
"grad_norm": 0.18394002318382263,
"learning_rate": 9.557571185118431e-06,
"loss": 0.4001,
"step": 301
},
{
"epoch": 0.6666666666666666,
"grad_norm": 0.1864984780550003,
"learning_rate": 9.55227390508323e-06,
"loss": 0.3964,
"step": 302
},
{
"epoch": 0.6688741721854304,
"grad_norm": 0.15841950476169586,
"learning_rate": 9.546946586773808e-06,
"loss": 0.4045,
"step": 303
},
{
"epoch": 0.6710816777041942,
"grad_norm": 0.19293592870235443,
"learning_rate": 9.541589265342585e-06,
"loss": 0.405,
"step": 304
},
{
"epoch": 0.673289183222958,
"grad_norm": 0.16064338386058807,
"learning_rate": 9.536201976139958e-06,
"loss": 0.4098,
"step": 305
},
{
"epoch": 0.6754966887417219,
"grad_norm": 0.16002054512500763,
"learning_rate": 9.530784754714069e-06,
"loss": 0.3837,
"step": 306
},
{
"epoch": 0.6777041942604857,
"grad_norm": 0.16032235324382782,
"learning_rate": 9.525337636810564e-06,
"loss": 0.4095,
"step": 307
},
{
"epoch": 0.6799116997792495,
"grad_norm": 0.19734519720077515,
"learning_rate": 9.519860658372364e-06,
"loss": 0.4049,
"step": 308
},
{
"epoch": 0.6821192052980133,
"grad_norm": 0.1894775629043579,
"learning_rate": 9.514353855539428e-06,
"loss": 0.3926,
"step": 309
},
{
"epoch": 0.6843267108167771,
"grad_norm": 0.2021895796060562,
"learning_rate": 9.508817264648506e-06,
"loss": 0.3983,
"step": 310
},
{
"epoch": 0.6865342163355408,
"grad_norm": 0.20270881056785583,
"learning_rate": 9.503250922232911e-06,
"loss": 0.4105,
"step": 311
},
{
"epoch": 0.6887417218543046,
"grad_norm": 0.17310801148414612,
"learning_rate": 9.497654865022268e-06,
"loss": 0.4026,
"step": 312
},
{
"epoch": 0.6909492273730684,
"grad_norm": 0.22883597016334534,
"learning_rate": 9.492029129942277e-06,
"loss": 0.4062,
"step": 313
},
{
"epoch": 0.6931567328918322,
"grad_norm": 0.1955188363790512,
"learning_rate": 9.48637375411447e-06,
"loss": 0.4044,
"step": 314
},
{
"epoch": 0.695364238410596,
"grad_norm": 0.18373191356658936,
"learning_rate": 9.48068877485596e-06,
"loss": 0.4032,
"step": 315
},
{
"epoch": 0.6975717439293598,
"grad_norm": 0.21621759235858917,
"learning_rate": 9.474974229679201e-06,
"loss": 0.3904,
"step": 316
},
{
"epoch": 0.6997792494481236,
"grad_norm": 0.17569488286972046,
"learning_rate": 9.469230156291742e-06,
"loss": 0.4087,
"step": 317
},
{
"epoch": 0.7019867549668874,
"grad_norm": 0.2064937800168991,
"learning_rate": 9.463456592595966e-06,
"loss": 0.396,
"step": 318
},
{
"epoch": 0.7041942604856513,
"grad_norm": 0.24051177501678467,
"learning_rate": 9.457653576688857e-06,
"loss": 0.4062,
"step": 319
},
{
"epoch": 0.7064017660044151,
"grad_norm": 0.16446927189826965,
"learning_rate": 9.451821146861734e-06,
"loss": 0.3987,
"step": 320
},
{
"epoch": 0.7086092715231788,
"grad_norm": 0.20139139890670776,
"learning_rate": 9.445959341600009e-06,
"loss": 0.4036,
"step": 321
},
{
"epoch": 0.7108167770419426,
"grad_norm": 0.18157215416431427,
"learning_rate": 9.440068199582923e-06,
"loss": 0.4124,
"step": 322
},
{
"epoch": 0.7130242825607064,
"grad_norm": 0.15641391277313232,
"learning_rate": 9.434147759683303e-06,
"loss": 0.405,
"step": 323
},
{
"epoch": 0.7152317880794702,
"grad_norm": 0.16391552984714508,
"learning_rate": 9.428198060967294e-06,
"loss": 0.4005,
"step": 324
},
{
"epoch": 0.717439293598234,
"grad_norm": 0.179102823138237,
"learning_rate": 9.422219142694104e-06,
"loss": 0.3968,
"step": 325
},
{
"epoch": 0.7196467991169978,
"grad_norm": 0.17109854519367218,
"learning_rate": 9.416211044315754e-06,
"loss": 0.4049,
"step": 326
},
{
"epoch": 0.7218543046357616,
"grad_norm": 0.20250354707241058,
"learning_rate": 9.410173805476804e-06,
"loss": 0.4186,
"step": 327
},
{
"epoch": 0.7240618101545254,
"grad_norm": 0.18576852977275848,
"learning_rate": 9.404107466014101e-06,
"loss": 0.4022,
"step": 328
},
{
"epoch": 0.7262693156732892,
"grad_norm": 0.16930828988552094,
"learning_rate": 9.398012065956512e-06,
"loss": 0.3949,
"step": 329
},
{
"epoch": 0.7284768211920529,
"grad_norm": 0.1966543048620224,
"learning_rate": 9.39188764552466e-06,
"loss": 0.4103,
"step": 330
},
{
"epoch": 0.7306843267108167,
"grad_norm": 0.18573778867721558,
"learning_rate": 9.385734245130664e-06,
"loss": 0.4069,
"step": 331
},
{
"epoch": 0.7328918322295805,
"grad_norm": 0.16225308179855347,
"learning_rate": 9.379551905377863e-06,
"loss": 0.4049,
"step": 332
},
{
"epoch": 0.7350993377483444,
"grad_norm": 0.1815934032201767,
"learning_rate": 9.373340667060553e-06,
"loss": 0.3927,
"step": 333
},
{
"epoch": 0.7373068432671082,
"grad_norm": 0.1688205897808075,
"learning_rate": 9.367100571163722e-06,
"loss": 0.4019,
"step": 334
},
{
"epoch": 0.739514348785872,
"grad_norm": 0.18315370380878448,
"learning_rate": 9.360831658862774e-06,
"loss": 0.3989,
"step": 335
},
{
"epoch": 0.7417218543046358,
"grad_norm": 0.17102378606796265,
"learning_rate": 9.354533971523253e-06,
"loss": 0.4122,
"step": 336
},
{
"epoch": 0.7439293598233996,
"grad_norm": 0.17149822413921356,
"learning_rate": 9.348207550700584e-06,
"loss": 0.4087,
"step": 337
},
{
"epoch": 0.7461368653421634,
"grad_norm": 0.19260001182556152,
"learning_rate": 9.341852438139784e-06,
"loss": 0.4064,
"step": 338
},
{
"epoch": 0.7483443708609272,
"grad_norm": 0.18973691761493683,
"learning_rate": 9.335468675775196e-06,
"loss": 0.399,
"step": 339
},
{
"epoch": 0.7505518763796909,
"grad_norm": 0.19365577399730682,
"learning_rate": 9.329056305730211e-06,
"loss": 0.3959,
"step": 340
},
{
"epoch": 0.7527593818984547,
"grad_norm": 0.21033021807670593,
"learning_rate": 9.322615370316986e-06,
"loss": 0.4027,
"step": 341
},
{
"epoch": 0.7549668874172185,
"grad_norm": 0.1825421005487442,
"learning_rate": 9.316145912036165e-06,
"loss": 0.3971,
"step": 342
},
{
"epoch": 0.7571743929359823,
"grad_norm": 0.1825910061597824,
"learning_rate": 9.309647973576605e-06,
"loss": 0.3959,
"step": 343
},
{
"epoch": 0.7593818984547461,
"grad_norm": 0.1806352436542511,
"learning_rate": 9.30312159781509e-06,
"loss": 0.4047,
"step": 344
},
{
"epoch": 0.7615894039735099,
"grad_norm": 0.15455348789691925,
"learning_rate": 9.296566827816044e-06,
"loss": 0.3958,
"step": 345
},
{
"epoch": 0.7637969094922737,
"grad_norm": 0.19140276312828064,
"learning_rate": 9.289983706831254e-06,
"loss": 0.3955,
"step": 346
},
{
"epoch": 0.7660044150110376,
"grad_norm": 0.1777879148721695,
"learning_rate": 9.28337227829958e-06,
"loss": 0.3971,
"step": 347
},
{
"epoch": 0.7682119205298014,
"grad_norm": 0.16399559378623962,
"learning_rate": 9.276732585846673e-06,
"loss": 0.3998,
"step": 348
},
{
"epoch": 0.7704194260485652,
"grad_norm": 0.1594182848930359,
"learning_rate": 9.270064673284681e-06,
"loss": 0.3898,
"step": 349
},
{
"epoch": 0.7726269315673289,
"grad_norm": 0.16792874038219452,
"learning_rate": 9.263368584611965e-06,
"loss": 0.4021,
"step": 350
},
{
"epoch": 0.7748344370860927,
"grad_norm": 0.18502795696258545,
"learning_rate": 9.256644364012803e-06,
"loss": 0.3987,
"step": 351
},
{
"epoch": 0.7770419426048565,
"grad_norm": 0.18240559101104736,
"learning_rate": 9.249892055857107e-06,
"loss": 0.4074,
"step": 352
},
{
"epoch": 0.7792494481236203,
"grad_norm": 0.15263855457305908,
"learning_rate": 9.243111704700126e-06,
"loss": 0.3928,
"step": 353
},
{
"epoch": 0.7814569536423841,
"grad_norm": 0.18149858713150024,
"learning_rate": 9.236303355282142e-06,
"loss": 0.4015,
"step": 354
},
{
"epoch": 0.7836644591611479,
"grad_norm": 0.1617615818977356,
"learning_rate": 9.229467052528191e-06,
"loss": 0.4024,
"step": 355
},
{
"epoch": 0.7858719646799117,
"grad_norm": 0.19588352739810944,
"learning_rate": 9.222602841547766e-06,
"loss": 0.3946,
"step": 356
},
{
"epoch": 0.7880794701986755,
"grad_norm": 0.1633402556180954,
"learning_rate": 9.2157107676345e-06,
"loss": 0.4027,
"step": 357
},
{
"epoch": 0.7902869757174393,
"grad_norm": 0.16059032082557678,
"learning_rate": 9.208790876265887e-06,
"loss": 0.3981,
"step": 358
},
{
"epoch": 0.7924944812362031,
"grad_norm": 0.1558162271976471,
"learning_rate": 9.201843213102976e-06,
"loss": 0.3969,
"step": 359
},
{
"epoch": 0.7947019867549668,
"grad_norm": 0.1607562005519867,
"learning_rate": 9.194867823990069e-06,
"loss": 0.3972,
"step": 360
},
{
"epoch": 0.7969094922737306,
"grad_norm": 0.1428382843732834,
"learning_rate": 9.187864754954412e-06,
"loss": 0.3928,
"step": 361
},
{
"epoch": 0.7991169977924945,
"grad_norm": 0.15147483348846436,
"learning_rate": 9.180834052205903e-06,
"loss": 0.392,
"step": 362
},
{
"epoch": 0.8013245033112583,
"grad_norm": 0.14845693111419678,
"learning_rate": 9.173775762136783e-06,
"loss": 0.3989,
"step": 363
},
{
"epoch": 0.8035320088300221,
"grad_norm": 0.1665990948677063,
"learning_rate": 9.166689931321326e-06,
"loss": 0.3928,
"step": 364
},
{
"epoch": 0.8057395143487859,
"grad_norm": 0.14729240536689758,
"learning_rate": 9.159576606515532e-06,
"loss": 0.3953,
"step": 365
},
{
"epoch": 0.8079470198675497,
"grad_norm": 0.14690490067005157,
"learning_rate": 9.152435834656823e-06,
"loss": 0.4023,
"step": 366
},
{
"epoch": 0.8101545253863135,
"grad_norm": 0.15939104557037354,
"learning_rate": 9.145267662863732e-06,
"loss": 0.4013,
"step": 367
},
{
"epoch": 0.8123620309050773,
"grad_norm": 0.1510556936264038,
"learning_rate": 9.13807213843559e-06,
"loss": 0.3992,
"step": 368
},
{
"epoch": 0.8145695364238411,
"grad_norm": 0.1646500676870346,
"learning_rate": 9.130849308852217e-06,
"loss": 0.3925,
"step": 369
},
{
"epoch": 0.8167770419426048,
"grad_norm": 0.17254091799259186,
"learning_rate": 9.123599221773601e-06,
"loss": 0.4071,
"step": 370
},
{
"epoch": 0.8189845474613686,
"grad_norm": 0.1497507095336914,
"learning_rate": 9.116321925039591e-06,
"loss": 0.3883,
"step": 371
},
{
"epoch": 0.8211920529801324,
"grad_norm": 0.16002339124679565,
"learning_rate": 9.109017466669587e-06,
"loss": 0.3953,
"step": 372
},
{
"epoch": 0.8233995584988962,
"grad_norm": 0.14812716841697693,
"learning_rate": 9.101685894862206e-06,
"loss": 0.4021,
"step": 373
},
{
"epoch": 0.82560706401766,
"grad_norm": 0.15898163616657257,
"learning_rate": 9.094327257994978e-06,
"loss": 0.4102,
"step": 374
},
{
"epoch": 0.8278145695364238,
"grad_norm": 0.15488837659358978,
"learning_rate": 9.086941604624022e-06,
"loss": 0.3912,
"step": 375
},
{
"epoch": 0.8300220750551877,
"grad_norm": 0.1613994836807251,
"learning_rate": 9.079528983483726e-06,
"loss": 0.4029,
"step": 376
},
{
"epoch": 0.8322295805739515,
"grad_norm": 0.1536788046360016,
"learning_rate": 9.072089443486425e-06,
"loss": 0.3956,
"step": 377
},
{
"epoch": 0.8344370860927153,
"grad_norm": 0.1693611443042755,
"learning_rate": 9.064623033722077e-06,
"loss": 0.3984,
"step": 378
},
{
"epoch": 0.8366445916114791,
"grad_norm": 0.18005254864692688,
"learning_rate": 9.057129803457943e-06,
"loss": 0.4022,
"step": 379
},
{
"epoch": 0.8388520971302428,
"grad_norm": 0.15852075815200806,
"learning_rate": 9.049609802138262e-06,
"loss": 0.3816,
"step": 380
},
{
"epoch": 0.8410596026490066,
"grad_norm": 0.1933150738477707,
"learning_rate": 9.042063079383916e-06,
"loss": 0.4028,
"step": 381
},
{
"epoch": 0.8432671081677704,
"grad_norm": 0.18579082190990448,
"learning_rate": 9.034489684992112e-06,
"loss": 0.4057,
"step": 382
},
{
"epoch": 0.8454746136865342,
"grad_norm": 0.1803402453660965,
"learning_rate": 9.026889668936054e-06,
"loss": 0.3976,
"step": 383
},
{
"epoch": 0.847682119205298,
"grad_norm": 0.17872066795825958,
"learning_rate": 9.019263081364605e-06,
"loss": 0.3908,
"step": 384
},
{
"epoch": 0.8498896247240618,
"grad_norm": 0.18206505477428436,
"learning_rate": 9.01160997260196e-06,
"loss": 0.4072,
"step": 385
},
{
"epoch": 0.8520971302428256,
"grad_norm": 0.1718129813671112,
"learning_rate": 9.00393039314732e-06,
"loss": 0.4036,
"step": 386
},
{
"epoch": 0.8543046357615894,
"grad_norm": 0.21254006028175354,
"learning_rate": 8.996224393674545e-06,
"loss": 0.4097,
"step": 387
},
{
"epoch": 0.8565121412803532,
"grad_norm": 0.15457090735435486,
"learning_rate": 8.988492025031838e-06,
"loss": 0.4001,
"step": 388
},
{
"epoch": 0.8587196467991169,
"grad_norm": 0.18473166227340698,
"learning_rate": 8.980733338241395e-06,
"loss": 0.3965,
"step": 389
},
{
"epoch": 0.8609271523178808,
"grad_norm": 0.19067919254302979,
"learning_rate": 8.972948384499068e-06,
"loss": 0.3996,
"step": 390
},
{
"epoch": 0.8631346578366446,
"grad_norm": 0.199691504240036,
"learning_rate": 8.965137215174037e-06,
"loss": 0.4003,
"step": 391
},
{
"epoch": 0.8653421633554084,
"grad_norm": 0.22586044669151306,
"learning_rate": 8.957299881808471e-06,
"loss": 0.393,
"step": 392
},
{
"epoch": 0.8675496688741722,
"grad_norm": 0.1807902604341507,
"learning_rate": 8.949436436117172e-06,
"loss": 0.4039,
"step": 393
},
{
"epoch": 0.869757174392936,
"grad_norm": 0.22098712623119354,
"learning_rate": 8.941546929987253e-06,
"loss": 0.39,
"step": 394
},
{
"epoch": 0.8719646799116998,
"grad_norm": 0.19912059605121613,
"learning_rate": 8.933631415477785e-06,
"loss": 0.396,
"step": 395
},
{
"epoch": 0.8741721854304636,
"grad_norm": 0.16147476434707642,
"learning_rate": 8.925689944819452e-06,
"loss": 0.4011,
"step": 396
},
{
"epoch": 0.8763796909492274,
"grad_norm": 0.19136802852153778,
"learning_rate": 8.917722570414217e-06,
"loss": 0.3896,
"step": 397
},
{
"epoch": 0.8785871964679912,
"grad_norm": 0.20598876476287842,
"learning_rate": 8.909729344834965e-06,
"loss": 0.3995,
"step": 398
},
{
"epoch": 0.8807947019867549,
"grad_norm": 0.1776047945022583,
"learning_rate": 8.901710320825161e-06,
"loss": 0.4001,
"step": 399
},
{
"epoch": 0.8830022075055187,
"grad_norm": 0.18616576492786407,
"learning_rate": 8.893665551298502e-06,
"loss": 0.3995,
"step": 400
},
{
"epoch": 0.8852097130242825,
"grad_norm": 0.18072794377803802,
"learning_rate": 8.885595089338567e-06,
"loss": 0.3978,
"step": 401
},
{
"epoch": 0.8874172185430463,
"grad_norm": 0.181128591299057,
"learning_rate": 8.877498988198471e-06,
"loss": 0.4112,
"step": 402
},
{
"epoch": 0.8896247240618101,
"grad_norm": 0.17399437725543976,
"learning_rate": 8.869377301300501e-06,
"loss": 0.3937,
"step": 403
},
{
"epoch": 0.891832229580574,
"grad_norm": 0.17823876440525055,
"learning_rate": 8.86123008223578e-06,
"loss": 0.4055,
"step": 404
},
{
"epoch": 0.8940397350993378,
"grad_norm": 0.176737442612648,
"learning_rate": 8.853057384763904e-06,
"loss": 0.3991,
"step": 405
},
{
"epoch": 0.8962472406181016,
"grad_norm": 0.1630028337240219,
"learning_rate": 8.844859262812584e-06,
"loss": 0.3903,
"step": 406
},
{
"epoch": 0.8984547461368654,
"grad_norm": 0.18285702168941498,
"learning_rate": 8.8366357704773e-06,
"loss": 0.3904,
"step": 407
},
{
"epoch": 0.9006622516556292,
"grad_norm": 0.1716790795326233,
"learning_rate": 8.82838696202094e-06,
"loss": 0.4033,
"step": 408
},
{
"epoch": 0.9028697571743929,
"grad_norm": 0.15402917563915253,
"learning_rate": 8.820112891873433e-06,
"loss": 0.39,
"step": 409
},
{
"epoch": 0.9050772626931567,
"grad_norm": 0.16390980780124664,
"learning_rate": 8.811813614631411e-06,
"loss": 0.3993,
"step": 410
},
{
"epoch": 0.9072847682119205,
"grad_norm": 0.15821807086467743,
"learning_rate": 8.803489185057822e-06,
"loss": 0.3953,
"step": 411
},
{
"epoch": 0.9094922737306843,
"grad_norm": 0.17032112181186676,
"learning_rate": 8.795139658081586e-06,
"loss": 0.4035,
"step": 412
},
{
"epoch": 0.9116997792494481,
"grad_norm": 0.15845684707164764,
"learning_rate": 8.786765088797238e-06,
"loss": 0.4013,
"step": 413
},
{
"epoch": 0.9139072847682119,
"grad_norm": 0.17293021082878113,
"learning_rate": 8.778365532464543e-06,
"loss": 0.3965,
"step": 414
},
{
"epoch": 0.9161147902869757,
"grad_norm": 0.14891566336154938,
"learning_rate": 8.76994104450815e-06,
"loss": 0.392,
"step": 415
},
{
"epoch": 0.9183222958057395,
"grad_norm": 0.15123558044433594,
"learning_rate": 8.761491680517218e-06,
"loss": 0.396,
"step": 416
},
{
"epoch": 0.9205298013245033,
"grad_norm": 0.1682642549276352,
"learning_rate": 8.75301749624505e-06,
"loss": 0.3838,
"step": 417
},
{
"epoch": 0.9227373068432672,
"grad_norm": 0.1535079926252365,
"learning_rate": 8.744518547608732e-06,
"loss": 0.3921,
"step": 418
},
{
"epoch": 0.9249448123620309,
"grad_norm": 0.16250640153884888,
"learning_rate": 8.735994890688749e-06,
"loss": 0.3896,
"step": 419
},
{
"epoch": 0.9271523178807947,
"grad_norm": 0.16567584872245789,
"learning_rate": 8.72744658172863e-06,
"loss": 0.3928,
"step": 420
},
{
"epoch": 0.9293598233995585,
"grad_norm": 0.16492141783237457,
"learning_rate": 8.718873677134569e-06,
"loss": 0.405,
"step": 421
},
{
"epoch": 0.9315673289183223,
"grad_norm": 0.14573417603969574,
"learning_rate": 8.710276233475058e-06,
"loss": 0.3955,
"step": 422
},
{
"epoch": 0.9337748344370861,
"grad_norm": 0.15780314803123474,
"learning_rate": 8.701654307480508e-06,
"loss": 0.3868,
"step": 423
},
{
"epoch": 0.9359823399558499,
"grad_norm": 0.14735905826091766,
"learning_rate": 8.693007956042874e-06,
"loss": 0.3924,
"step": 424
},
{
"epoch": 0.9381898454746137,
"grad_norm": 0.16305825114250183,
"learning_rate": 8.684337236215289e-06,
"loss": 0.3866,
"step": 425
},
{
"epoch": 0.9403973509933775,
"grad_norm": 0.1656455248594284,
"learning_rate": 8.675642205211679e-06,
"loss": 0.3965,
"step": 426
},
{
"epoch": 0.9426048565121413,
"grad_norm": 0.15034431219100952,
"learning_rate": 8.666922920406384e-06,
"loss": 0.3982,
"step": 427
},
{
"epoch": 0.9448123620309051,
"grad_norm": 0.161673903465271,
"learning_rate": 8.65817943933379e-06,
"loss": 0.3933,
"step": 428
},
{
"epoch": 0.9470198675496688,
"grad_norm": 0.1589784324169159,
"learning_rate": 8.649411819687936e-06,
"loss": 0.3976,
"step": 429
},
{
"epoch": 0.9492273730684326,
"grad_norm": 0.166986882686615,
"learning_rate": 8.640620119322146e-06,
"loss": 0.4003,
"step": 430
},
{
"epoch": 0.9514348785871964,
"grad_norm": 0.15880529582500458,
"learning_rate": 8.631804396248637e-06,
"loss": 0.3926,
"step": 431
},
{
"epoch": 0.9536423841059603,
"grad_norm": 0.1741640269756317,
"learning_rate": 8.62296470863814e-06,
"loss": 0.3978,
"step": 432
},
{
"epoch": 0.9558498896247241,
"grad_norm": 0.15339982509613037,
"learning_rate": 8.61410111481952e-06,
"loss": 0.4059,
"step": 433
},
{
"epoch": 0.9580573951434879,
"grad_norm": 0.1605585813522339,
"learning_rate": 8.605213673279382e-06,
"loss": 0.3897,
"step": 434
},
{
"epoch": 0.9602649006622517,
"grad_norm": 0.154588520526886,
"learning_rate": 8.5963024426617e-06,
"loss": 0.3853,
"step": 435
},
{
"epoch": 0.9624724061810155,
"grad_norm": 0.16428066790103912,
"learning_rate": 8.587367481767409e-06,
"loss": 0.3944,
"step": 436
},
{
"epoch": 0.9646799116997793,
"grad_norm": 0.15117953717708588,
"learning_rate": 8.578408849554037e-06,
"loss": 0.3924,
"step": 437
},
{
"epoch": 0.9668874172185431,
"grad_norm": 0.15928852558135986,
"learning_rate": 8.569426605135307e-06,
"loss": 0.399,
"step": 438
},
{
"epoch": 0.9690949227373068,
"grad_norm": 0.16322654485702515,
"learning_rate": 8.560420807780742e-06,
"loss": 0.3863,
"step": 439
},
{
"epoch": 0.9713024282560706,
"grad_norm": 0.1749914139509201,
"learning_rate": 8.551391516915288e-06,
"loss": 0.4018,
"step": 440
},
{
"epoch": 0.9735099337748344,
"grad_norm": 0.15128456056118011,
"learning_rate": 8.542338792118907e-06,
"loss": 0.3784,
"step": 441
},
{
"epoch": 0.9757174392935982,
"grad_norm": 0.19126002490520477,
"learning_rate": 8.533262693126191e-06,
"loss": 0.3923,
"step": 442
},
{
"epoch": 0.977924944812362,
"grad_norm": 0.16031384468078613,
"learning_rate": 8.52416327982597e-06,
"loss": 0.4036,
"step": 443
},
{
"epoch": 0.9801324503311258,
"grad_norm": 0.16994954645633698,
"learning_rate": 8.515040612260912e-06,
"loss": 0.3865,
"step": 444
},
{
"epoch": 0.9823399558498896,
"grad_norm": 0.17667335271835327,
"learning_rate": 8.505894750627128e-06,
"loss": 0.3884,
"step": 445
},
{
"epoch": 0.9845474613686535,
"grad_norm": 0.1745615154504776,
"learning_rate": 8.496725755273778e-06,
"loss": 0.3926,
"step": 446
},
{
"epoch": 0.9867549668874173,
"grad_norm": 0.16568712890148163,
"learning_rate": 8.487533686702668e-06,
"loss": 0.393,
"step": 447
},
{
"epoch": 0.9889624724061811,
"grad_norm": 0.1762251853942871,
"learning_rate": 8.478318605567853e-06,
"loss": 0.3868,
"step": 448
},
{
"epoch": 0.9911699779249448,
"grad_norm": 0.18370142579078674,
"learning_rate": 8.46908057267524e-06,
"loss": 0.3939,
"step": 449
},
{
"epoch": 0.9933774834437086,
"grad_norm": 0.1815333366394043,
"learning_rate": 8.459819648982182e-06,
"loss": 0.3849,
"step": 450
},
{
"epoch": 0.9955849889624724,
"grad_norm": 0.19721707701683044,
"learning_rate": 8.450535895597074e-06,
"loss": 0.3953,
"step": 451
},
{
"epoch": 0.9977924944812362,
"grad_norm": 0.16212257742881775,
"learning_rate": 8.441229373778957e-06,
"loss": 0.3933,
"step": 452
},
{
"epoch": 1.0,
"grad_norm": 0.2004193216562271,
"learning_rate": 8.43190014493711e-06,
"loss": 0.3914,
"step": 453
},
{
"epoch": 1.0022075055187638,
"grad_norm": 0.15856173634529114,
"learning_rate": 8.422548270630646e-06,
"loss": 0.3806,
"step": 454
},
{
"epoch": 1.0044150110375276,
"grad_norm": 0.18225938081741333,
"learning_rate": 8.413173812568099e-06,
"loss": 0.3714,
"step": 455
},
{
"epoch": 1.0066225165562914,
"grad_norm": 0.17915207147598267,
"learning_rate": 8.403776832607028e-06,
"loss": 0.3834,
"step": 456
},
{
"epoch": 1.0088300220750552,
"grad_norm": 0.1916593611240387,
"learning_rate": 8.394357392753599e-06,
"loss": 0.3666,
"step": 457
},
{
"epoch": 1.011037527593819,
"grad_norm": 0.166362464427948,
"learning_rate": 8.384915555162183e-06,
"loss": 0.3839,
"step": 458
},
{
"epoch": 1.0132450331125828,
"grad_norm": 0.20503659546375275,
"learning_rate": 8.375451382134942e-06,
"loss": 0.381,
"step": 459
},
{
"epoch": 1.0154525386313467,
"grad_norm": 0.18665875494480133,
"learning_rate": 8.365964936121422e-06,
"loss": 0.3893,
"step": 460
},
{
"epoch": 1.0176600441501105,
"grad_norm": 0.15370848774909973,
"learning_rate": 8.35645627971813e-06,
"loss": 0.3796,
"step": 461
},
{
"epoch": 1.0198675496688743,
"grad_norm": 0.17442859709262848,
"learning_rate": 8.346925475668138e-06,
"loss": 0.3633,
"step": 462
},
{
"epoch": 1.022075055187638,
"grad_norm": 0.18989215791225433,
"learning_rate": 8.337372586860651e-06,
"loss": 0.3743,
"step": 463
},
{
"epoch": 1.0242825607064017,
"grad_norm": 0.16941452026367188,
"learning_rate": 8.327797676330604e-06,
"loss": 0.369,
"step": 464
},
{
"epoch": 1.0264900662251655,
"grad_norm": 0.14053016901016235,
"learning_rate": 8.31820080725825e-06,
"loss": 0.3724,
"step": 465
},
{
"epoch": 1.0286975717439293,
"grad_norm": 0.17336276173591614,
"learning_rate": 8.308582042968726e-06,
"loss": 0.3903,
"step": 466
},
{
"epoch": 1.030905077262693,
"grad_norm": 0.15840588510036469,
"learning_rate": 8.298941446931646e-06,
"loss": 0.3847,
"step": 467
},
{
"epoch": 1.033112582781457,
"grad_norm": 0.16392916440963745,
"learning_rate": 8.289279082760685e-06,
"loss": 0.3893,
"step": 468
},
{
"epoch": 1.0353200883002207,
"grad_norm": 0.17757326364517212,
"learning_rate": 8.279595014213158e-06,
"loss": 0.3876,
"step": 469
},
{
"epoch": 1.0375275938189845,
"grad_norm": 0.1557520627975464,
"learning_rate": 8.26988930518959e-06,
"loss": 0.3782,
"step": 470
},
{
"epoch": 1.0397350993377483,
"grad_norm": 0.16154958307743073,
"learning_rate": 8.260162019733305e-06,
"loss": 0.3933,
"step": 471
},
{
"epoch": 1.0419426048565121,
"grad_norm": 0.18913333117961884,
"learning_rate": 8.250413222029997e-06,
"loss": 0.3888,
"step": 472
},
{
"epoch": 1.044150110375276,
"grad_norm": 0.1763067990541458,
"learning_rate": 8.240642976407313e-06,
"loss": 0.3875,
"step": 473
},
{
"epoch": 1.0463576158940397,
"grad_norm": 0.16598297655582428,
"learning_rate": 8.230851347334424e-06,
"loss": 0.3894,
"step": 474
},
{
"epoch": 1.0485651214128036,
"grad_norm": 0.17947614192962646,
"learning_rate": 8.221038399421592e-06,
"loss": 0.3853,
"step": 475
},
{
"epoch": 1.0507726269315674,
"grad_norm": 0.16084261238574982,
"learning_rate": 8.211204197419766e-06,
"loss": 0.3762,
"step": 476
},
{
"epoch": 1.0529801324503312,
"grad_norm": 0.21188747882843018,
"learning_rate": 8.201348806220127e-06,
"loss": 0.3791,
"step": 477
},
{
"epoch": 1.055187637969095,
"grad_norm": 0.15903352200984955,
"learning_rate": 8.191472290853683e-06,
"loss": 0.3799,
"step": 478
},
{
"epoch": 1.0573951434878588,
"grad_norm": 0.2036534547805786,
"learning_rate": 8.181574716490823e-06,
"loss": 0.3909,
"step": 479
},
{
"epoch": 1.0596026490066226,
"grad_norm": 0.16643071174621582,
"learning_rate": 8.171656148440902e-06,
"loss": 0.3842,
"step": 480
},
{
"epoch": 1.0618101545253864,
"grad_norm": 0.16434621810913086,
"learning_rate": 8.161716652151795e-06,
"loss": 0.3748,
"step": 481
},
{
"epoch": 1.0640176600441502,
"grad_norm": 0.18316681683063507,
"learning_rate": 8.151756293209476e-06,
"loss": 0.3795,
"step": 482
},
{
"epoch": 1.0662251655629138,
"grad_norm": 0.14608311653137207,
"learning_rate": 8.14177513733758e-06,
"loss": 0.3749,
"step": 483
},
{
"epoch": 1.0684326710816776,
"grad_norm": 0.1656610518693924,
"learning_rate": 8.131773250396973e-06,
"loss": 0.3794,
"step": 484
},
{
"epoch": 1.0706401766004414,
"grad_norm": 0.14669524133205414,
"learning_rate": 8.121750698385315e-06,
"loss": 0.3825,
"step": 485
},
{
"epoch": 1.0728476821192052,
"grad_norm": 0.17891530692577362,
"learning_rate": 8.111707547436623e-06,
"loss": 0.373,
"step": 486
},
{
"epoch": 1.075055187637969,
"grad_norm": 0.15400467813014984,
"learning_rate": 8.10164386382084e-06,
"loss": 0.3792,
"step": 487
},
{
"epoch": 1.0772626931567328,
"grad_norm": 0.17551693320274353,
"learning_rate": 8.091559713943388e-06,
"loss": 0.3908,
"step": 488
},
{
"epoch": 1.0794701986754967,
"grad_norm": 0.15460826456546783,
"learning_rate": 8.081455164344745e-06,
"loss": 0.3838,
"step": 489
},
{
"epoch": 1.0816777041942605,
"grad_norm": 0.15716975927352905,
"learning_rate": 8.071330281699989e-06,
"loss": 0.3934,
"step": 490
},
{
"epoch": 1.0838852097130243,
"grad_norm": 0.15411889553070068,
"learning_rate": 8.06118513281837e-06,
"loss": 0.3792,
"step": 491
},
{
"epoch": 1.086092715231788,
"grad_norm": 0.18962320685386658,
"learning_rate": 8.051019784642864e-06,
"loss": 0.3888,
"step": 492
},
{
"epoch": 1.0883002207505519,
"grad_norm": 0.15896014869213104,
"learning_rate": 8.040834304249733e-06,
"loss": 0.3861,
"step": 493
},
{
"epoch": 1.0905077262693157,
"grad_norm": 0.1523059904575348,
"learning_rate": 8.03062875884808e-06,
"loss": 0.3727,
"step": 494
},
{
"epoch": 1.0927152317880795,
"grad_norm": 0.13958971202373505,
"learning_rate": 8.02040321577941e-06,
"loss": 0.3718,
"step": 495
},
{
"epoch": 1.0949227373068433,
"grad_norm": 0.1423843950033188,
"learning_rate": 8.010157742517185e-06,
"loss": 0.3746,
"step": 496
},
{
"epoch": 1.0971302428256071,
"grad_norm": 0.13685709238052368,
"learning_rate": 7.99989240666637e-06,
"loss": 0.3727,
"step": 497
},
{
"epoch": 1.099337748344371,
"grad_norm": 0.15423151850700378,
"learning_rate": 7.989607275963e-06,
"loss": 0.3778,
"step": 498
},
{
"epoch": 1.1015452538631347,
"grad_norm": 0.15014959871768951,
"learning_rate": 7.979302418273723e-06,
"loss": 0.3747,
"step": 499
},
{
"epoch": 1.1037527593818985,
"grad_norm": 0.16919955611228943,
"learning_rate": 7.968977901595355e-06,
"loss": 0.3782,
"step": 500
},
{
"epoch": 1.1059602649006623,
"grad_norm": 0.1481235921382904,
"learning_rate": 7.958633794054439e-06,
"loss": 0.3808,
"step": 501
},
{
"epoch": 1.108167770419426,
"grad_norm": 0.17575468122959137,
"learning_rate": 7.94827016390678e-06,
"loss": 0.3708,
"step": 502
},
{
"epoch": 1.1103752759381897,
"grad_norm": 0.139452263712883,
"learning_rate": 7.93788707953701e-06,
"loss": 0.376,
"step": 503
},
{
"epoch": 1.1125827814569536,
"grad_norm": 0.14127641916275024,
"learning_rate": 7.927484609458128e-06,
"loss": 0.3847,
"step": 504
},
{
"epoch": 1.1147902869757174,
"grad_norm": 0.15063825249671936,
"learning_rate": 7.917062822311047e-06,
"loss": 0.3842,
"step": 505
},
{
"epoch": 1.1169977924944812,
"grad_norm": 0.15780918300151825,
"learning_rate": 7.90662178686415e-06,
"loss": 0.3737,
"step": 506
},
{
"epoch": 1.119205298013245,
"grad_norm": 0.14620929956436157,
"learning_rate": 7.896161572012824e-06,
"loss": 0.3862,
"step": 507
},
{
"epoch": 1.1214128035320088,
"grad_norm": 0.1493159383535385,
"learning_rate": 7.885682246779016e-06,
"loss": 0.3809,
"step": 508
},
{
"epoch": 1.1236203090507726,
"grad_norm": 0.1471249759197235,
"learning_rate": 7.875183880310772e-06,
"loss": 0.3885,
"step": 509
},
{
"epoch": 1.1258278145695364,
"grad_norm": 0.1509741246700287,
"learning_rate": 7.86466654188178e-06,
"loss": 0.3778,
"step": 510
},
{
"epoch": 1.1280353200883002,
"grad_norm": 0.14812204241752625,
"learning_rate": 7.854130300890921e-06,
"loss": 0.3731,
"step": 511
},
{
"epoch": 1.130242825607064,
"grad_norm": 0.1415746510028839,
"learning_rate": 7.843575226861795e-06,
"loss": 0.3798,
"step": 512
},
{
"epoch": 1.1324503311258278,
"grad_norm": 0.14002346992492676,
"learning_rate": 7.833001389442283e-06,
"loss": 0.3833,
"step": 513
},
{
"epoch": 1.1346578366445916,
"grad_norm": 0.15462878346443176,
"learning_rate": 7.82240885840407e-06,
"loss": 0.3785,
"step": 514
},
{
"epoch": 1.1368653421633554,
"grad_norm": 0.15146006643772125,
"learning_rate": 7.811797703642193e-06,
"loss": 0.3767,
"step": 515
},
{
"epoch": 1.1390728476821192,
"grad_norm": 0.1545468419790268,
"learning_rate": 7.801167995174575e-06,
"loss": 0.3739,
"step": 516
},
{
"epoch": 1.141280353200883,
"grad_norm": 0.1515691876411438,
"learning_rate": 7.790519803141572e-06,
"loss": 0.394,
"step": 517
},
{
"epoch": 1.1434878587196469,
"grad_norm": 0.1403704136610031,
"learning_rate": 7.7798531978055e-06,
"loss": 0.3844,
"step": 518
},
{
"epoch": 1.1456953642384107,
"grad_norm": 0.12919290363788605,
"learning_rate": 7.769168249550176e-06,
"loss": 0.3746,
"step": 519
},
{
"epoch": 1.1479028697571745,
"grad_norm": 0.137840136885643,
"learning_rate": 7.758465028880455e-06,
"loss": 0.3748,
"step": 520
},
{
"epoch": 1.150110375275938,
"grad_norm": 0.13867942988872528,
"learning_rate": 7.747743606421761e-06,
"loss": 0.3813,
"step": 521
},
{
"epoch": 1.152317880794702,
"grad_norm": 0.1563754677772522,
"learning_rate": 7.737004052919623e-06,
"loss": 0.3673,
"step": 522
},
{
"epoch": 1.1545253863134657,
"grad_norm": 0.15125791728496552,
"learning_rate": 7.726246439239209e-06,
"loss": 0.3814,
"step": 523
},
{
"epoch": 1.1567328918322295,
"grad_norm": 0.15860167145729065,
"learning_rate": 7.715470836364857e-06,
"loss": 0.3962,
"step": 524
},
{
"epoch": 1.1589403973509933,
"grad_norm": 0.14636121690273285,
"learning_rate": 7.704677315399607e-06,
"loss": 0.3786,
"step": 525
},
{
"epoch": 1.161147902869757,
"grad_norm": 0.1441173404455185,
"learning_rate": 7.693865947564733e-06,
"loss": 0.3737,
"step": 526
},
{
"epoch": 1.163355408388521,
"grad_norm": 0.1591232568025589,
"learning_rate": 7.68303680419927e-06,
"loss": 0.3721,
"step": 527
},
{
"epoch": 1.1655629139072847,
"grad_norm": 0.1598597913980484,
"learning_rate": 7.672189956759546e-06,
"loss": 0.3793,
"step": 528
},
{
"epoch": 1.1677704194260485,
"grad_norm": 0.16569367051124573,
"learning_rate": 7.661325476818708e-06,
"loss": 0.3799,
"step": 529
},
{
"epoch": 1.1699779249448123,
"grad_norm": 0.1879713088274002,
"learning_rate": 7.65044343606626e-06,
"loss": 0.3831,
"step": 530
},
{
"epoch": 1.1721854304635762,
"grad_norm": 0.14313288033008575,
"learning_rate": 7.639543906307565e-06,
"loss": 0.3774,
"step": 531
},
{
"epoch": 1.17439293598234,
"grad_norm": 0.20557354390621185,
"learning_rate": 7.628626959463405e-06,
"loss": 0.3766,
"step": 532
},
{
"epoch": 1.1766004415011038,
"grad_norm": 0.14320634305477142,
"learning_rate": 7.6176926675694786e-06,
"loss": 0.3754,
"step": 533
},
{
"epoch": 1.1788079470198676,
"grad_norm": 0.18957629799842834,
"learning_rate": 7.606741102775936e-06,
"loss": 0.3799,
"step": 534
},
{
"epoch": 1.1810154525386314,
"grad_norm": 0.15119552612304688,
"learning_rate": 7.595772337346912e-06,
"loss": 0.3863,
"step": 535
},
{
"epoch": 1.1832229580573952,
"grad_norm": 0.16690769791603088,
"learning_rate": 7.584786443660028e-06,
"loss": 0.371,
"step": 536
},
{
"epoch": 1.185430463576159,
"grad_norm": 0.1685992032289505,
"learning_rate": 7.573783494205936e-06,
"loss": 0.3741,
"step": 537
},
{
"epoch": 1.1876379690949228,
"grad_norm": 0.14501997828483582,
"learning_rate": 7.562763561587824e-06,
"loss": 0.3701,
"step": 538
},
{
"epoch": 1.1898454746136866,
"grad_norm": 0.16112159192562103,
"learning_rate": 7.55172671852095e-06,
"loss": 0.3781,
"step": 539
},
{
"epoch": 1.1920529801324504,
"grad_norm": 0.13081905245780945,
"learning_rate": 7.5406730378321506e-06,
"loss": 0.39,
"step": 540
},
{
"epoch": 1.1942604856512142,
"grad_norm": 0.15677790343761444,
"learning_rate": 7.5296025924593705e-06,
"loss": 0.3814,
"step": 541
},
{
"epoch": 1.1964679911699778,
"grad_norm": 0.1494779884815216,
"learning_rate": 7.518515455451172e-06,
"loss": 0.3757,
"step": 542
},
{
"epoch": 1.1986754966887416,
"grad_norm": 0.1597052365541458,
"learning_rate": 7.50741169996626e-06,
"loss": 0.3772,
"step": 543
},
{
"epoch": 1.2008830022075054,
"grad_norm": 0.1445734202861786,
"learning_rate": 7.496291399273e-06,
"loss": 0.3779,
"step": 544
},
{
"epoch": 1.2030905077262692,
"grad_norm": 0.14188243448734283,
"learning_rate": 7.485154626748924e-06,
"loss": 0.3801,
"step": 545
},
{
"epoch": 1.205298013245033,
"grad_norm": 0.15295171737670898,
"learning_rate": 7.474001455880258e-06,
"loss": 0.3786,
"step": 546
},
{
"epoch": 1.2075055187637969,
"grad_norm": 0.12851738929748535,
"learning_rate": 7.4628319602614315e-06,
"loss": 0.3759,
"step": 547
},
{
"epoch": 1.2097130242825607,
"grad_norm": 0.12803953886032104,
"learning_rate": 7.451646213594597e-06,
"loss": 0.3807,
"step": 548
},
{
"epoch": 1.2119205298013245,
"grad_norm": 0.15474575757980347,
"learning_rate": 7.440444289689135e-06,
"loss": 0.3761,
"step": 549
},
{
"epoch": 1.2141280353200883,
"grad_norm": 0.14079326391220093,
"learning_rate": 7.429226262461175e-06,
"loss": 0.3744,
"step": 550
},
{
"epoch": 1.216335540838852,
"grad_norm": 0.13553716242313385,
"learning_rate": 7.417992205933104e-06,
"loss": 0.3813,
"step": 551
},
{
"epoch": 1.218543046357616,
"grad_norm": 0.14058908820152283,
"learning_rate": 7.406742194233074e-06,
"loss": 0.3722,
"step": 552
},
{
"epoch": 1.2207505518763797,
"grad_norm": 0.15452724695205688,
"learning_rate": 7.3954763015945266e-06,
"loss": 0.3628,
"step": 553
},
{
"epoch": 1.2229580573951435,
"grad_norm": 0.14715242385864258,
"learning_rate": 7.384194602355685e-06,
"loss": 0.3779,
"step": 554
},
{
"epoch": 1.2251655629139073,
"grad_norm": 0.15119391679763794,
"learning_rate": 7.37289717095908e-06,
"loss": 0.3787,
"step": 555
},
{
"epoch": 1.2273730684326711,
"grad_norm": 0.1800910234451294,
"learning_rate": 7.361584081951046e-06,
"loss": 0.3781,
"step": 556
},
{
"epoch": 1.229580573951435,
"grad_norm": 0.1489747166633606,
"learning_rate": 7.350255409981237e-06,
"loss": 0.3843,
"step": 557
},
{
"epoch": 1.2317880794701987,
"grad_norm": 0.14379991590976715,
"learning_rate": 7.338911229802133e-06,
"loss": 0.3832,
"step": 558
},
{
"epoch": 1.2339955849889626,
"grad_norm": 0.13160887360572815,
"learning_rate": 7.327551616268541e-06,
"loss": 0.3858,
"step": 559
},
{
"epoch": 1.2362030905077264,
"grad_norm": 0.1353141814470291,
"learning_rate": 7.316176644337107e-06,
"loss": 0.3688,
"step": 560
},
{
"epoch": 1.23841059602649,
"grad_norm": 0.16255126893520355,
"learning_rate": 7.304786389065823e-06,
"loss": 0.377,
"step": 561
},
{
"epoch": 1.240618101545254,
"grad_norm": 0.13924263417720795,
"learning_rate": 7.293380925613524e-06,
"loss": 0.376,
"step": 562
},
{
"epoch": 1.2428256070640176,
"grad_norm": 0.1347806751728058,
"learning_rate": 7.281960329239398e-06,
"loss": 0.3881,
"step": 563
},
{
"epoch": 1.2450331125827814,
"grad_norm": 0.1494341343641281,
"learning_rate": 7.270524675302491e-06,
"loss": 0.3843,
"step": 564
},
{
"epoch": 1.2472406181015452,
"grad_norm": 0.16562990844249725,
"learning_rate": 7.259074039261199e-06,
"loss": 0.3843,
"step": 565
},
{
"epoch": 1.249448123620309,
"grad_norm": 0.1605864018201828,
"learning_rate": 7.247608496672786e-06,
"loss": 0.3741,
"step": 566
},
{
"epoch": 1.2516556291390728,
"grad_norm": 0.1619981825351715,
"learning_rate": 7.2361281231928725e-06,
"loss": 0.392,
"step": 567
},
{
"epoch": 1.2538631346578366,
"grad_norm": 0.13447479903697968,
"learning_rate": 7.2246329945749425e-06,
"loss": 0.3755,
"step": 568
},
{
"epoch": 1.2560706401766004,
"grad_norm": 0.1463773399591446,
"learning_rate": 7.213123186669842e-06,
"loss": 0.3913,
"step": 569
},
{
"epoch": 1.2582781456953642,
"grad_norm": 0.14680610597133636,
"learning_rate": 7.201598775425278e-06,
"loss": 0.3776,
"step": 570
},
{
"epoch": 1.260485651214128,
"grad_norm": 0.14192719757556915,
"learning_rate": 7.190059836885318e-06,
"loss": 0.3829,
"step": 571
},
{
"epoch": 1.2626931567328918,
"grad_norm": 0.1412820965051651,
"learning_rate": 7.178506447189887e-06,
"loss": 0.3782,
"step": 572
},
{
"epoch": 1.2649006622516556,
"grad_norm": 0.1515144258737564,
"learning_rate": 7.166938682574272e-06,
"loss": 0.3779,
"step": 573
},
{
"epoch": 1.2671081677704195,
"grad_norm": 0.148550346493721,
"learning_rate": 7.155356619368604e-06,
"loss": 0.376,
"step": 574
},
{
"epoch": 1.2693156732891833,
"grad_norm": 0.14767727255821228,
"learning_rate": 7.14376033399737e-06,
"loss": 0.3768,
"step": 575
},
{
"epoch": 1.271523178807947,
"grad_norm": 0.1577431708574295,
"learning_rate": 7.132149902978902e-06,
"loss": 0.373,
"step": 576
},
{
"epoch": 1.2737306843267109,
"grad_norm": 0.15225905179977417,
"learning_rate": 7.120525402924871e-06,
"loss": 0.365,
"step": 577
},
{
"epoch": 1.2759381898454747,
"grad_norm": 0.16508492827415466,
"learning_rate": 7.108886910539781e-06,
"loss": 0.3866,
"step": 578
},
{
"epoch": 1.2781456953642385,
"grad_norm": 0.16345573961734772,
"learning_rate": 7.097234502620468e-06,
"loss": 0.3819,
"step": 579
},
{
"epoch": 1.280353200883002,
"grad_norm": 0.15904562175273895,
"learning_rate": 7.085568256055589e-06,
"loss": 0.3877,
"step": 580
},
{
"epoch": 1.2825607064017661,
"grad_norm": 0.17447535693645477,
"learning_rate": 7.073888247825111e-06,
"loss": 0.3902,
"step": 581
},
{
"epoch": 1.2847682119205297,
"grad_norm": 0.1613752692937851,
"learning_rate": 7.062194554999817e-06,
"loss": 0.3765,
"step": 582
},
{
"epoch": 1.2869757174392937,
"grad_norm": 0.1705675572156906,
"learning_rate": 7.05048725474078e-06,
"loss": 0.3699,
"step": 583
},
{
"epoch": 1.2891832229580573,
"grad_norm": 0.20963701605796814,
"learning_rate": 7.038766424298865e-06,
"loss": 0.3909,
"step": 584
},
{
"epoch": 1.2913907284768211,
"grad_norm": 0.16403751075267792,
"learning_rate": 7.027032141014216e-06,
"loss": 0.3753,
"step": 585
},
{
"epoch": 1.293598233995585,
"grad_norm": 0.18310756981372833,
"learning_rate": 7.0152844823157474e-06,
"loss": 0.3807,
"step": 586
},
{
"epoch": 1.2958057395143487,
"grad_norm": 0.16576433181762695,
"learning_rate": 7.003523525720626e-06,
"loss": 0.3659,
"step": 587
},
{
"epoch": 1.2980132450331126,
"grad_norm": 0.1746741682291031,
"learning_rate": 6.991749348833773e-06,
"loss": 0.3783,
"step": 588
},
{
"epoch": 1.3002207505518764,
"grad_norm": 0.16022977232933044,
"learning_rate": 6.979962029347338e-06,
"loss": 0.3745,
"step": 589
},
{
"epoch": 1.3024282560706402,
"grad_norm": 0.1607898473739624,
"learning_rate": 6.9681616450401936e-06,
"loss": 0.3732,
"step": 590
},
{
"epoch": 1.304635761589404,
"grad_norm": 0.1624252200126648,
"learning_rate": 6.956348273777424e-06,
"loss": 0.3815,
"step": 591
},
{
"epoch": 1.3068432671081678,
"grad_norm": 0.14729204773902893,
"learning_rate": 6.944521993509803e-06,
"loss": 0.3734,
"step": 592
},
{
"epoch": 1.3090507726269316,
"grad_norm": 0.18559885025024414,
"learning_rate": 6.9326828822732885e-06,
"loss": 0.386,
"step": 593
},
{
"epoch": 1.3112582781456954,
"grad_norm": 0.15120132267475128,
"learning_rate": 6.920831018188502e-06,
"loss": 0.3702,
"step": 594
},
{
"epoch": 1.3134657836644592,
"grad_norm": 0.18332970142364502,
"learning_rate": 6.908966479460219e-06,
"loss": 0.3807,
"step": 595
},
{
"epoch": 1.315673289183223,
"grad_norm": 0.17234960198402405,
"learning_rate": 6.89708934437684e-06,
"loss": 0.3644,
"step": 596
},
{
"epoch": 1.3178807947019868,
"grad_norm": 0.18331332504749298,
"learning_rate": 6.885199691309892e-06,
"loss": 0.3841,
"step": 597
},
{
"epoch": 1.3200883002207506,
"grad_norm": 0.15819165110588074,
"learning_rate": 6.873297598713497e-06,
"loss": 0.3725,
"step": 598
},
{
"epoch": 1.3222958057395142,
"grad_norm": 0.1651681512594223,
"learning_rate": 6.8613831451238636e-06,
"loss": 0.3762,
"step": 599
},
{
"epoch": 1.3245033112582782,
"grad_norm": 0.1659359186887741,
"learning_rate": 6.84945640915876e-06,
"loss": 0.3757,
"step": 600
},
{
"epoch": 1.3267108167770418,
"grad_norm": 0.15379931032657623,
"learning_rate": 6.837517469517001e-06,
"loss": 0.3735,
"step": 601
},
{
"epoch": 1.3289183222958059,
"grad_norm": 0.1771930605173111,
"learning_rate": 6.82556640497793e-06,
"loss": 0.3798,
"step": 602
},
{
"epoch": 1.3311258278145695,
"grad_norm": 0.16866551339626312,
"learning_rate": 6.813603294400895e-06,
"loss": 0.3724,
"step": 603
},
{
"epoch": 1.3333333333333333,
"grad_norm": 0.16047048568725586,
"learning_rate": 6.8016282167247325e-06,
"loss": 0.3744,
"step": 604
},
{
"epoch": 1.335540838852097,
"grad_norm": 0.13133041560649872,
"learning_rate": 6.7896412509672385e-06,
"loss": 0.3775,
"step": 605
},
{
"epoch": 1.3377483443708609,
"grad_norm": 0.16929350793361664,
"learning_rate": 6.777642476224658e-06,
"loss": 0.3724,
"step": 606
},
{
"epoch": 1.3399558498896247,
"grad_norm": 0.14370004832744598,
"learning_rate": 6.765631971671156e-06,
"loss": 0.3795,
"step": 607
},
{
"epoch": 1.3421633554083885,
"grad_norm": 0.18519333004951477,
"learning_rate": 6.753609816558297e-06,
"loss": 0.3789,
"step": 608
},
{
"epoch": 1.3443708609271523,
"grad_norm": 0.15061011910438538,
"learning_rate": 6.741576090214526e-06,
"loss": 0.3779,
"step": 609
},
{
"epoch": 1.346578366445916,
"grad_norm": 0.16702063381671906,
"learning_rate": 6.729530872044629e-06,
"loss": 0.3791,
"step": 610
},
{
"epoch": 1.34878587196468,
"grad_norm": 0.1492423117160797,
"learning_rate": 6.717474241529235e-06,
"loss": 0.3821,
"step": 611
},
{
"epoch": 1.3509933774834437,
"grad_norm": 0.15660499036312103,
"learning_rate": 6.705406278224269e-06,
"loss": 0.3764,
"step": 612
},
{
"epoch": 1.3532008830022075,
"grad_norm": 0.13613760471343994,
"learning_rate": 6.69332706176044e-06,
"loss": 0.3842,
"step": 613
},
{
"epoch": 1.3554083885209713,
"grad_norm": 0.1655375063419342,
"learning_rate": 6.681236671842709e-06,
"loss": 0.3861,
"step": 614
},
{
"epoch": 1.3576158940397351,
"grad_norm": 0.15369166433811188,
"learning_rate": 6.669135188249767e-06,
"loss": 0.3753,
"step": 615
},
{
"epoch": 1.359823399558499,
"grad_norm": 0.1611892580986023,
"learning_rate": 6.657022690833503e-06,
"loss": 0.3732,
"step": 616
},
{
"epoch": 1.3620309050772628,
"grad_norm": 0.15144729614257812,
"learning_rate": 6.644899259518485e-06,
"loss": 0.3756,
"step": 617
},
{
"epoch": 1.3642384105960264,
"grad_norm": 0.1591017246246338,
"learning_rate": 6.632764974301429e-06,
"loss": 0.3643,
"step": 618
},
{
"epoch": 1.3664459161147904,
"grad_norm": 0.15775880217552185,
"learning_rate": 6.620619915250666e-06,
"loss": 0.3773,
"step": 619
},
{
"epoch": 1.368653421633554,
"grad_norm": 0.14554527401924133,
"learning_rate": 6.608464162505621e-06,
"loss": 0.3788,
"step": 620
},
{
"epoch": 1.370860927152318,
"grad_norm": 0.15284626185894012,
"learning_rate": 6.596297796276284e-06,
"loss": 0.3743,
"step": 621
},
{
"epoch": 1.3730684326710816,
"grad_norm": 0.14736317098140717,
"learning_rate": 6.584120896842675e-06,
"loss": 0.3763,
"step": 622
},
{
"epoch": 1.3752759381898454,
"grad_norm": 0.15955647826194763,
"learning_rate": 6.571933544554319e-06,
"loss": 0.3664,
"step": 623
},
{
"epoch": 1.3774834437086092,
"grad_norm": 0.16919533908367157,
"learning_rate": 6.559735819829713e-06,
"loss": 0.3755,
"step": 624
},
{
"epoch": 1.379690949227373,
"grad_norm": 0.15306392312049866,
"learning_rate": 6.5475278031558e-06,
"loss": 0.3772,
"step": 625
},
{
"epoch": 1.3818984547461368,
"grad_norm": 0.1504804790019989,
"learning_rate": 6.5353095750874295e-06,
"loss": 0.3709,
"step": 626
},
{
"epoch": 1.3841059602649006,
"grad_norm": 0.14429201185703278,
"learning_rate": 6.523081216246838e-06,
"loss": 0.3709,
"step": 627
},
{
"epoch": 1.3863134657836644,
"grad_norm": 0.14882300794124603,
"learning_rate": 6.5108428073231e-06,
"loss": 0.3768,
"step": 628
},
{
"epoch": 1.3885209713024282,
"grad_norm": 0.1360517293214798,
"learning_rate": 6.498594429071618e-06,
"loss": 0.3768,
"step": 629
},
{
"epoch": 1.390728476821192,
"grad_norm": 0.13942170143127441,
"learning_rate": 6.486336162313568e-06,
"loss": 0.3796,
"step": 630
},
{
"epoch": 1.3929359823399559,
"grad_norm": 0.14323705434799194,
"learning_rate": 6.474068087935379e-06,
"loss": 0.379,
"step": 631
},
{
"epoch": 1.3951434878587197,
"grad_norm": 0.15375126898288727,
"learning_rate": 6.461790286888196e-06,
"loss": 0.3721,
"step": 632
},
{
"epoch": 1.3973509933774835,
"grad_norm": 0.15478521585464478,
"learning_rate": 6.449502840187344e-06,
"loss": 0.3726,
"step": 633
},
{
"epoch": 1.3995584988962473,
"grad_norm": 0.16403432190418243,
"learning_rate": 6.437205828911797e-06,
"loss": 0.3765,
"step": 634
},
{
"epoch": 1.401766004415011,
"grad_norm": 0.1496008038520813,
"learning_rate": 6.424899334203641e-06,
"loss": 0.3785,
"step": 635
},
{
"epoch": 1.403973509933775,
"grad_norm": 0.152191162109375,
"learning_rate": 6.4125834372675355e-06,
"loss": 0.3752,
"step": 636
},
{
"epoch": 1.4061810154525387,
"grad_norm": 0.15051637589931488,
"learning_rate": 6.400258219370183e-06,
"loss": 0.3743,
"step": 637
},
{
"epoch": 1.4083885209713025,
"grad_norm": 0.15709619224071503,
"learning_rate": 6.387923761839794e-06,
"loss": 0.3697,
"step": 638
},
{
"epoch": 1.410596026490066,
"grad_norm": 0.14248456060886383,
"learning_rate": 6.375580146065539e-06,
"loss": 0.373,
"step": 639
},
{
"epoch": 1.4128035320088301,
"grad_norm": 0.1371423304080963,
"learning_rate": 6.363227453497029e-06,
"loss": 0.3819,
"step": 640
},
{
"epoch": 1.4150110375275937,
"grad_norm": 0.16379839181900024,
"learning_rate": 6.350865765643759e-06,
"loss": 0.3793,
"step": 641
},
{
"epoch": 1.4172185430463577,
"grad_norm": 0.14373619854450226,
"learning_rate": 6.3384951640745865e-06,
"loss": 0.3779,
"step": 642
},
{
"epoch": 1.4194260485651213,
"grad_norm": 0.14477743208408356,
"learning_rate": 6.32611573041718e-06,
"loss": 0.3786,
"step": 643
},
{
"epoch": 1.4216335540838851,
"grad_norm": 0.17920951545238495,
"learning_rate": 6.313727546357498e-06,
"loss": 0.3901,
"step": 644
},
{
"epoch": 1.423841059602649,
"grad_norm": 0.14486798644065857,
"learning_rate": 6.301330693639224e-06,
"loss": 0.3786,
"step": 645
},
{
"epoch": 1.4260485651214128,
"grad_norm": 0.13817182183265686,
"learning_rate": 6.288925254063249e-06,
"loss": 0.3725,
"step": 646
},
{
"epoch": 1.4282560706401766,
"grad_norm": 0.14681169390678406,
"learning_rate": 6.27651130948713e-06,
"loss": 0.3851,
"step": 647
},
{
"epoch": 1.4304635761589404,
"grad_norm": 0.14171209931373596,
"learning_rate": 6.264088941824534e-06,
"loss": 0.3802,
"step": 648
},
{
"epoch": 1.4326710816777042,
"grad_norm": 0.14188838005065918,
"learning_rate": 6.251658233044715e-06,
"loss": 0.3764,
"step": 649
},
{
"epoch": 1.434878587196468,
"grad_norm": 0.15146797895431519,
"learning_rate": 6.239219265171964e-06,
"loss": 0.3791,
"step": 650
},
{
"epoch": 1.4370860927152318,
"grad_norm": 0.1602582484483719,
"learning_rate": 6.22677212028507e-06,
"loss": 0.3773,
"step": 651
},
{
"epoch": 1.4392935982339956,
"grad_norm": 0.14558373391628265,
"learning_rate": 6.214316880516775e-06,
"loss": 0.3631,
"step": 652
},
{
"epoch": 1.4415011037527594,
"grad_norm": 0.17893047630786896,
"learning_rate": 6.201853628053243e-06,
"loss": 0.374,
"step": 653
},
{
"epoch": 1.4437086092715232,
"grad_norm": 0.14639593660831451,
"learning_rate": 6.189382445133502e-06,
"loss": 0.3753,
"step": 654
},
{
"epoch": 1.445916114790287,
"grad_norm": 0.14472749829292297,
"learning_rate": 6.17690341404891e-06,
"loss": 0.3796,
"step": 655
},
{
"epoch": 1.4481236203090508,
"grad_norm": 0.1566278040409088,
"learning_rate": 6.16441661714262e-06,
"loss": 0.3741,
"step": 656
},
{
"epoch": 1.4503311258278146,
"grad_norm": 0.14471283555030823,
"learning_rate": 6.151922136809017e-06,
"loss": 0.3808,
"step": 657
},
{
"epoch": 1.4525386313465782,
"grad_norm": 0.15734338760375977,
"learning_rate": 6.13942005549319e-06,
"loss": 0.3781,
"step": 658
},
{
"epoch": 1.4547461368653423,
"grad_norm": 0.13064594566822052,
"learning_rate": 6.126910455690384e-06,
"loss": 0.377,
"step": 659
},
{
"epoch": 1.4569536423841059,
"grad_norm": 0.15152592957019806,
"learning_rate": 6.114393419945455e-06,
"loss": 0.3784,
"step": 660
},
{
"epoch": 1.4591611479028699,
"grad_norm": 0.14837685227394104,
"learning_rate": 6.101869030852321e-06,
"loss": 0.3889,
"step": 661
},
{
"epoch": 1.4613686534216335,
"grad_norm": 0.14178897440433502,
"learning_rate": 6.089337371053429e-06,
"loss": 0.3714,
"step": 662
},
{
"epoch": 1.4635761589403973,
"grad_norm": 0.15140818059444427,
"learning_rate": 6.076798523239194e-06,
"loss": 0.3816,
"step": 663
},
{
"epoch": 1.465783664459161,
"grad_norm": 0.14195586740970612,
"learning_rate": 6.064252570147464e-06,
"loss": 0.3738,
"step": 664
},
{
"epoch": 1.467991169977925,
"grad_norm": 0.13242661952972412,
"learning_rate": 6.051699594562976e-06,
"loss": 0.3669,
"step": 665
},
{
"epoch": 1.4701986754966887,
"grad_norm": 0.1379902958869934,
"learning_rate": 6.039139679316797e-06,
"loss": 0.362,
"step": 666
},
{
"epoch": 1.4724061810154525,
"grad_norm": 0.1443612426519394,
"learning_rate": 6.026572907285791e-06,
"loss": 0.3711,
"step": 667
},
{
"epoch": 1.4746136865342163,
"grad_norm": 0.14281077682971954,
"learning_rate": 6.013999361392064e-06,
"loss": 0.3753,
"step": 668
},
{
"epoch": 1.4768211920529801,
"grad_norm": 0.14133213460445404,
"learning_rate": 6.001419124602422e-06,
"loss": 0.377,
"step": 669
},
{
"epoch": 1.479028697571744,
"grad_norm": 0.14672201871871948,
"learning_rate": 5.988832279927818e-06,
"loss": 0.3763,
"step": 670
},
{
"epoch": 1.4812362030905077,
"grad_norm": 0.13457442820072174,
"learning_rate": 5.976238910422811e-06,
"loss": 0.381,
"step": 671
},
{
"epoch": 1.4834437086092715,
"grad_norm": 0.15843527019023895,
"learning_rate": 5.963639099185011e-06,
"loss": 0.3773,
"step": 672
},
{
"epoch": 1.4856512141280354,
"grad_norm": 0.15779507160186768,
"learning_rate": 5.9510329293545356e-06,
"loss": 0.3789,
"step": 673
},
{
"epoch": 1.4878587196467992,
"grad_norm": 0.14860793948173523,
"learning_rate": 5.938420484113458e-06,
"loss": 0.3765,
"step": 674
},
{
"epoch": 1.490066225165563,
"grad_norm": 0.13150885701179504,
"learning_rate": 5.925801846685262e-06,
"loss": 0.37,
"step": 675
},
{
"epoch": 1.4922737306843268,
"grad_norm": 0.13564042747020721,
"learning_rate": 5.913177100334287e-06,
"loss": 0.3788,
"step": 676
},
{
"epoch": 1.4944812362030906,
"grad_norm": 0.14382179081439972,
"learning_rate": 5.900546328365189e-06,
"loss": 0.381,
"step": 677
},
{
"epoch": 1.4966887417218544,
"grad_norm": 0.14150989055633545,
"learning_rate": 5.887909614122378e-06,
"loss": 0.3683,
"step": 678
},
{
"epoch": 1.498896247240618,
"grad_norm": 0.13479964435100555,
"learning_rate": 5.875267040989475e-06,
"loss": 0.3651,
"step": 679
},
{
"epoch": 1.501103752759382,
"grad_norm": 0.14743578433990479,
"learning_rate": 5.862618692388765e-06,
"loss": 0.3676,
"step": 680
},
{
"epoch": 1.5033112582781456,
"grad_norm": 0.1419716477394104,
"learning_rate": 5.849964651780637e-06,
"loss": 0.367,
"step": 681
},
{
"epoch": 1.5055187637969096,
"grad_norm": 0.15260837972164154,
"learning_rate": 5.837305002663043e-06,
"loss": 0.3814,
"step": 682
},
{
"epoch": 1.5077262693156732,
"grad_norm": 0.12932439148426056,
"learning_rate": 5.824639828570943e-06,
"loss": 0.3868,
"step": 683
},
{
"epoch": 1.5099337748344372,
"grad_norm": 0.15234258770942688,
"learning_rate": 5.811969213075749e-06,
"loss": 0.3695,
"step": 684
},
{
"epoch": 1.5121412803532008,
"grad_norm": 0.1671096384525299,
"learning_rate": 5.799293239784783e-06,
"loss": 0.3821,
"step": 685
},
{
"epoch": 1.5143487858719646,
"grad_norm": 0.14161601662635803,
"learning_rate": 5.786611992340719e-06,
"loss": 0.3657,
"step": 686
},
{
"epoch": 1.5165562913907285,
"grad_norm": 0.16359050571918488,
"learning_rate": 5.773925554421033e-06,
"loss": 0.3788,
"step": 687
},
{
"epoch": 1.5187637969094923,
"grad_norm": 0.14034205675125122,
"learning_rate": 5.761234009737449e-06,
"loss": 0.3779,
"step": 688
},
{
"epoch": 1.520971302428256,
"grad_norm": 0.14073529839515686,
"learning_rate": 5.74853744203539e-06,
"loss": 0.3756,
"step": 689
},
{
"epoch": 1.5231788079470199,
"grad_norm": 0.14516787230968475,
"learning_rate": 5.735835935093421e-06,
"loss": 0.3886,
"step": 690
},
{
"epoch": 1.5253863134657837,
"grad_norm": 0.1415787935256958,
"learning_rate": 5.723129572722701e-06,
"loss": 0.375,
"step": 691
},
{
"epoch": 1.5275938189845475,
"grad_norm": 0.13726189732551575,
"learning_rate": 5.710418438766428e-06,
"loss": 0.3735,
"step": 692
},
{
"epoch": 1.5298013245033113,
"grad_norm": 0.14881715178489685,
"learning_rate": 5.6977026170992834e-06,
"loss": 0.3735,
"step": 693
},
{
"epoch": 1.532008830022075,
"grad_norm": 0.1551184505224228,
"learning_rate": 5.68498219162688e-06,
"loss": 0.3815,
"step": 694
},
{
"epoch": 1.534216335540839,
"grad_norm": 0.15148243308067322,
"learning_rate": 5.672257246285212e-06,
"loss": 0.3736,
"step": 695
},
{
"epoch": 1.5364238410596025,
"grad_norm": 0.15206126868724823,
"learning_rate": 5.659527865040097e-06,
"loss": 0.3771,
"step": 696
},
{
"epoch": 1.5386313465783665,
"grad_norm": 0.14559738337993622,
"learning_rate": 5.6467941318866214e-06,
"loss": 0.3742,
"step": 697
},
{
"epoch": 1.5408388520971301,
"grad_norm": 0.13176165521144867,
"learning_rate": 5.634056130848589e-06,
"loss": 0.3715,
"step": 698
},
{
"epoch": 1.5430463576158941,
"grad_norm": 0.14761999249458313,
"learning_rate": 5.621313945977964e-06,
"loss": 0.377,
"step": 699
},
{
"epoch": 1.5452538631346577,
"grad_norm": 0.12837889790534973,
"learning_rate": 5.60856766135432e-06,
"loss": 0.371,
"step": 700
},
{
"epoch": 1.5474613686534218,
"grad_norm": 0.1346467137336731,
"learning_rate": 5.595817361084283e-06,
"loss": 0.3757,
"step": 701
},
{
"epoch": 1.5496688741721854,
"grad_norm": 0.129667267203331,
"learning_rate": 5.583063129300971e-06,
"loss": 0.3821,
"step": 702
},
{
"epoch": 1.5518763796909494,
"grad_norm": 0.1456080973148346,
"learning_rate": 5.570305050163453e-06,
"loss": 0.3659,
"step": 703
},
{
"epoch": 1.554083885209713,
"grad_norm": 0.13144616782665253,
"learning_rate": 5.557543207856175e-06,
"loss": 0.3755,
"step": 704
},
{
"epoch": 1.5562913907284768,
"grad_norm": 0.13695232570171356,
"learning_rate": 5.544777686588424e-06,
"loss": 0.3704,
"step": 705
},
{
"epoch": 1.5584988962472406,
"grad_norm": 0.1521306186914444,
"learning_rate": 5.532008570593756e-06,
"loss": 0.3829,
"step": 706
},
{
"epoch": 1.5607064017660044,
"grad_norm": 0.1419302225112915,
"learning_rate": 5.519235944129448e-06,
"loss": 0.371,
"step": 707
},
{
"epoch": 1.5629139072847682,
"grad_norm": 0.14487622678279877,
"learning_rate": 5.506459891475944e-06,
"loss": 0.3785,
"step": 708
},
{
"epoch": 1.565121412803532,
"grad_norm": 0.15528716146945953,
"learning_rate": 5.4936804969362924e-06,
"loss": 0.3853,
"step": 709
},
{
"epoch": 1.5673289183222958,
"grad_norm": 0.14845651388168335,
"learning_rate": 5.480897844835595e-06,
"loss": 0.3853,
"step": 710
},
{
"epoch": 1.5695364238410596,
"grad_norm": 0.13595354557037354,
"learning_rate": 5.4681120195204466e-06,
"loss": 0.3718,
"step": 711
},
{
"epoch": 1.5717439293598234,
"grad_norm": 0.15339161455631256,
"learning_rate": 5.455323105358384e-06,
"loss": 0.3791,
"step": 712
},
{
"epoch": 1.5739514348785872,
"grad_norm": 0.1527385264635086,
"learning_rate": 5.442531186737322e-06,
"loss": 0.3703,
"step": 713
},
{
"epoch": 1.576158940397351,
"grad_norm": 0.13763274252414703,
"learning_rate": 5.429736348065003e-06,
"loss": 0.3638,
"step": 714
},
{
"epoch": 1.5783664459161146,
"grad_norm": 0.16397210955619812,
"learning_rate": 5.416938673768437e-06,
"loss": 0.3761,
"step": 715
},
{
"epoch": 1.5805739514348787,
"grad_norm": 0.14585243165493011,
"learning_rate": 5.404138248293346e-06,
"loss": 0.3762,
"step": 716
},
{
"epoch": 1.5827814569536423,
"grad_norm": 0.13906413316726685,
"learning_rate": 5.3913351561036e-06,
"loss": 0.3669,
"step": 717
},
{
"epoch": 1.5849889624724063,
"grad_norm": 0.13377633690834045,
"learning_rate": 5.378529481680672e-06,
"loss": 0.3721,
"step": 718
},
{
"epoch": 1.5871964679911699,
"grad_norm": 0.14944018423557281,
"learning_rate": 5.365721309523072e-06,
"loss": 0.3739,
"step": 719
},
{
"epoch": 1.589403973509934,
"grad_norm": 0.14180952310562134,
"learning_rate": 5.352910724145789e-06,
"loss": 0.373,
"step": 720
},
{
"epoch": 1.5916114790286975,
"grad_norm": 0.1462966352701187,
"learning_rate": 5.340097810079741e-06,
"loss": 0.3768,
"step": 721
},
{
"epoch": 1.5938189845474615,
"grad_norm": 0.12976975739002228,
"learning_rate": 5.327282651871205e-06,
"loss": 0.3638,
"step": 722
},
{
"epoch": 1.596026490066225,
"grad_norm": 0.15462516248226166,
"learning_rate": 5.3144653340812715e-06,
"loss": 0.3854,
"step": 723
},
{
"epoch": 1.598233995584989,
"grad_norm": 0.13364708423614502,
"learning_rate": 5.301645941285278e-06,
"loss": 0.3862,
"step": 724
},
{
"epoch": 1.6004415011037527,
"grad_norm": 0.14875876903533936,
"learning_rate": 5.288824558072257e-06,
"loss": 0.3847,
"step": 725
},
{
"epoch": 1.6026490066225165,
"grad_norm": 0.13750705122947693,
"learning_rate": 5.276001269044369e-06,
"loss": 0.3792,
"step": 726
},
{
"epoch": 1.6048565121412803,
"grad_norm": 0.13552795350551605,
"learning_rate": 5.263176158816355e-06,
"loss": 0.3824,
"step": 727
},
{
"epoch": 1.6070640176600441,
"grad_norm": 0.13398821651935577,
"learning_rate": 5.250349312014976e-06,
"loss": 0.3728,
"step": 728
},
{
"epoch": 1.609271523178808,
"grad_norm": 0.14047983288764954,
"learning_rate": 5.237520813278443e-06,
"loss": 0.3833,
"step": 729
},
{
"epoch": 1.6114790286975718,
"grad_norm": 0.1419335901737213,
"learning_rate": 5.224690747255875e-06,
"loss": 0.3738,
"step": 730
},
{
"epoch": 1.6136865342163356,
"grad_norm": 0.1423855870962143,
"learning_rate": 5.211859198606729e-06,
"loss": 0.3702,
"step": 731
},
{
"epoch": 1.6158940397350994,
"grad_norm": 0.14920298755168915,
"learning_rate": 5.199026252000245e-06,
"loss": 0.3719,
"step": 732
},
{
"epoch": 1.6181015452538632,
"grad_norm": 0.13715742528438568,
"learning_rate": 5.186191992114892e-06,
"loss": 0.3766,
"step": 733
},
{
"epoch": 1.6203090507726268,
"grad_norm": 0.142774298787117,
"learning_rate": 5.173356503637799e-06,
"loss": 0.3693,
"step": 734
},
{
"epoch": 1.6225165562913908,
"grad_norm": 0.12615923583507538,
"learning_rate": 5.160519871264204e-06,
"loss": 0.375,
"step": 735
},
{
"epoch": 1.6247240618101544,
"grad_norm": 0.14882232248783112,
"learning_rate": 5.147682179696893e-06,
"loss": 0.3746,
"step": 736
},
{
"epoch": 1.6269315673289184,
"grad_norm": 0.12539036571979523,
"learning_rate": 5.134843513645642e-06,
"loss": 0.3786,
"step": 737
},
{
"epoch": 1.629139072847682,
"grad_norm": 0.14418825507164001,
"learning_rate": 5.122003957826657e-06,
"loss": 0.3693,
"step": 738
},
{
"epoch": 1.631346578366446,
"grad_norm": 0.14610478281974792,
"learning_rate": 5.109163596962013e-06,
"loss": 0.3759,
"step": 739
},
{
"epoch": 1.6335540838852096,
"grad_norm": 0.13532301783561707,
"learning_rate": 5.096322515779101e-06,
"loss": 0.373,
"step": 740
},
{
"epoch": 1.6357615894039736,
"grad_norm": 0.14194665849208832,
"learning_rate": 5.083480799010061e-06,
"loss": 0.3762,
"step": 741
},
{
"epoch": 1.6379690949227372,
"grad_norm": 0.12951509654521942,
"learning_rate": 5.070638531391229e-06,
"loss": 0.3723,
"step": 742
},
{
"epoch": 1.6401766004415013,
"grad_norm": 0.14456267654895782,
"learning_rate": 5.057795797662577e-06,
"loss": 0.3666,
"step": 743
},
{
"epoch": 1.6423841059602649,
"grad_norm": 0.13711507618427277,
"learning_rate": 5.044952682567148e-06,
"loss": 0.3749,
"step": 744
},
{
"epoch": 1.6445916114790287,
"grad_norm": 0.12987838685512543,
"learning_rate": 5.032109270850507e-06,
"loss": 0.3779,
"step": 745
},
{
"epoch": 1.6467991169977925,
"grad_norm": 0.1495376080274582,
"learning_rate": 5.019265647260173e-06,
"loss": 0.3704,
"step": 746
},
{
"epoch": 1.6490066225165563,
"grad_norm": 0.13476036489009857,
"learning_rate": 5.006421896545064e-06,
"loss": 0.3663,
"step": 747
},
{
"epoch": 1.65121412803532,
"grad_norm": 0.1335316300392151,
"learning_rate": 4.993578103454938e-06,
"loss": 0.3843,
"step": 748
},
{
"epoch": 1.653421633554084,
"grad_norm": 0.13593193888664246,
"learning_rate": 4.980734352739829e-06,
"loss": 0.3707,
"step": 749
},
{
"epoch": 1.6556291390728477,
"grad_norm": 0.13324475288391113,
"learning_rate": 4.967890729149494e-06,
"loss": 0.3746,
"step": 750
},
{
"epoch": 1.6578366445916115,
"grad_norm": 0.12398661673069,
"learning_rate": 4.955047317432854e-06,
"loss": 0.3783,
"step": 751
},
{
"epoch": 1.6600441501103753,
"grad_norm": 0.12614096701145172,
"learning_rate": 4.942204202337425e-06,
"loss": 0.3806,
"step": 752
},
{
"epoch": 1.6622516556291391,
"grad_norm": 0.14331743121147156,
"learning_rate": 4.929361468608773e-06,
"loss": 0.3833,
"step": 753
},
{
"epoch": 1.664459161147903,
"grad_norm": 0.1342383772134781,
"learning_rate": 4.9165192009899414e-06,
"loss": 0.3799,
"step": 754
},
{
"epoch": 1.6666666666666665,
"grad_norm": 0.13297194242477417,
"learning_rate": 4.9036774842209e-06,
"loss": 0.3702,
"step": 755
},
{
"epoch": 1.6688741721854305,
"grad_norm": 0.12575410306453705,
"learning_rate": 4.890836403037988e-06,
"loss": 0.364,
"step": 756
},
{
"epoch": 1.6710816777041941,
"grad_norm": 0.1317722052335739,
"learning_rate": 4.877996042173345e-06,
"loss": 0.3782,
"step": 757
},
{
"epoch": 1.6732891832229582,
"grad_norm": 0.13363225758075714,
"learning_rate": 4.865156486354359e-06,
"loss": 0.3804,
"step": 758
},
{
"epoch": 1.6754966887417218,
"grad_norm": 0.14083701372146606,
"learning_rate": 4.852317820303107e-06,
"loss": 0.38,
"step": 759
},
{
"epoch": 1.6777041942604858,
"grad_norm": 0.1284651756286621,
"learning_rate": 4.839480128735798e-06,
"loss": 0.3655,
"step": 760
},
{
"epoch": 1.6799116997792494,
"grad_norm": 0.1379299908876419,
"learning_rate": 4.826643496362202e-06,
"loss": 0.3779,
"step": 761
},
{
"epoch": 1.6821192052980134,
"grad_norm": 0.13508039712905884,
"learning_rate": 4.81380800788511e-06,
"loss": 0.3785,
"step": 762
},
{
"epoch": 1.684326710816777,
"grad_norm": 0.14211615920066833,
"learning_rate": 4.800973747999757e-06,
"loss": 0.3648,
"step": 763
},
{
"epoch": 1.6865342163355408,
"grad_norm": 0.1293489634990692,
"learning_rate": 4.788140801393273e-06,
"loss": 0.3663,
"step": 764
},
{
"epoch": 1.6887417218543046,
"grad_norm": 0.13812755048274994,
"learning_rate": 4.775309252744126e-06,
"loss": 0.3643,
"step": 765
},
{
"epoch": 1.6909492273730684,
"grad_norm": 0.13112276792526245,
"learning_rate": 4.762479186721559e-06,
"loss": 0.3703,
"step": 766
},
{
"epoch": 1.6931567328918322,
"grad_norm": 0.12749050557613373,
"learning_rate": 4.7496506879850264e-06,
"loss": 0.3844,
"step": 767
},
{
"epoch": 1.695364238410596,
"grad_norm": 0.1374199390411377,
"learning_rate": 4.736823841183645e-06,
"loss": 0.3689,
"step": 768
},
{
"epoch": 1.6975717439293598,
"grad_norm": 0.12659025192260742,
"learning_rate": 4.723998730955633e-06,
"loss": 0.3606,
"step": 769
},
{
"epoch": 1.6997792494481236,
"grad_norm": 0.11276005953550339,
"learning_rate": 4.711175441927746e-06,
"loss": 0.3722,
"step": 770
},
{
"epoch": 1.7019867549668874,
"grad_norm": 0.12801696360111237,
"learning_rate": 4.698354058714724e-06,
"loss": 0.3789,
"step": 771
},
{
"epoch": 1.7041942604856513,
"grad_norm": 0.12910126149654388,
"learning_rate": 4.685534665918731e-06,
"loss": 0.3669,
"step": 772
},
{
"epoch": 1.706401766004415,
"grad_norm": 0.12605293095111847,
"learning_rate": 4.672717348128796e-06,
"loss": 0.377,
"step": 773
},
{
"epoch": 1.7086092715231787,
"grad_norm": 0.12655854225158691,
"learning_rate": 4.65990218992026e-06,
"loss": 0.3738,
"step": 774
},
{
"epoch": 1.7108167770419427,
"grad_norm": 0.1326448917388916,
"learning_rate": 4.647089275854212e-06,
"loss": 0.3847,
"step": 775
},
{
"epoch": 1.7130242825607063,
"grad_norm": 0.12570975720882416,
"learning_rate": 4.63427869047693e-06,
"loss": 0.3652,
"step": 776
},
{
"epoch": 1.7152317880794703,
"grad_norm": 0.13106000423431396,
"learning_rate": 4.621470518319329e-06,
"loss": 0.3781,
"step": 777
},
{
"epoch": 1.7174392935982339,
"grad_norm": 0.12707674503326416,
"learning_rate": 4.608664843896402e-06,
"loss": 0.367,
"step": 778
},
{
"epoch": 1.719646799116998,
"grad_norm": 0.1342884600162506,
"learning_rate": 4.595861751706656e-06,
"loss": 0.3827,
"step": 779
},
{
"epoch": 1.7218543046357615,
"grad_norm": 0.15550127625465393,
"learning_rate": 4.583061326231564e-06,
"loss": 0.384,
"step": 780
},
{
"epoch": 1.7240618101545255,
"grad_norm": 0.1267244517803192,
"learning_rate": 4.570263651934998e-06,
"loss": 0.38,
"step": 781
},
{
"epoch": 1.7262693156732891,
"grad_norm": 0.13444717228412628,
"learning_rate": 4.557468813262678e-06,
"loss": 0.3856,
"step": 782
},
{
"epoch": 1.728476821192053,
"grad_norm": 0.12585113942623138,
"learning_rate": 4.544676894641617e-06,
"loss": 0.3603,
"step": 783
},
{
"epoch": 1.7306843267108167,
"grad_norm": 0.14303074777126312,
"learning_rate": 4.531887980479555e-06,
"loss": 0.3769,
"step": 784
},
{
"epoch": 1.7328918322295805,
"grad_norm": 0.14289291203022003,
"learning_rate": 4.519102155164407e-06,
"loss": 0.3679,
"step": 785
},
{
"epoch": 1.7350993377483444,
"grad_norm": 0.13349799811840057,
"learning_rate": 4.506319503063708e-06,
"loss": 0.3702,
"step": 786
},
{
"epoch": 1.7373068432671082,
"grad_norm": 0.15247680246829987,
"learning_rate": 4.493540108524057e-06,
"loss": 0.3713,
"step": 787
},
{
"epoch": 1.739514348785872,
"grad_norm": 0.14266639947891235,
"learning_rate": 4.480764055870553e-06,
"loss": 0.3742,
"step": 788
},
{
"epoch": 1.7417218543046358,
"grad_norm": 0.13838379085063934,
"learning_rate": 4.467991429406247e-06,
"loss": 0.3857,
"step": 789
},
{
"epoch": 1.7439293598233996,
"grad_norm": 0.13494673371315002,
"learning_rate": 4.455222313411578e-06,
"loss": 0.3783,
"step": 790
},
{
"epoch": 1.7461368653421634,
"grad_norm": 0.12723763287067413,
"learning_rate": 4.442456792143825e-06,
"loss": 0.3657,
"step": 791
},
{
"epoch": 1.7483443708609272,
"grad_norm": 0.1302330642938614,
"learning_rate": 4.4296949498365496e-06,
"loss": 0.3671,
"step": 792
},
{
"epoch": 1.7505518763796908,
"grad_norm": 0.12404655665159225,
"learning_rate": 4.416936870699031e-06,
"loss": 0.3755,
"step": 793
},
{
"epoch": 1.7527593818984548,
"grad_norm": 0.13801135122776031,
"learning_rate": 4.40418263891572e-06,
"loss": 0.377,
"step": 794
},
{
"epoch": 1.7549668874172184,
"grad_norm": 0.1327562779188156,
"learning_rate": 4.391432338645681e-06,
"loss": 0.3649,
"step": 795
},
{
"epoch": 1.7571743929359824,
"grad_norm": 0.13369016349315643,
"learning_rate": 4.378686054022037e-06,
"loss": 0.3882,
"step": 796
},
{
"epoch": 1.759381898454746,
"grad_norm": 0.14249202609062195,
"learning_rate": 4.365943869151412e-06,
"loss": 0.3763,
"step": 797
},
{
"epoch": 1.76158940397351,
"grad_norm": 0.1298934519290924,
"learning_rate": 4.35320586811338e-06,
"loss": 0.3742,
"step": 798
},
{
"epoch": 1.7637969094922736,
"grad_norm": 0.14034999907016754,
"learning_rate": 4.3404721349599044e-06,
"loss": 0.3685,
"step": 799
},
{
"epoch": 1.7660044150110377,
"grad_norm": 0.12729597091674805,
"learning_rate": 4.327742753714788e-06,
"loss": 0.3742,
"step": 800
},
{
"epoch": 1.7682119205298013,
"grad_norm": 0.13787223398685455,
"learning_rate": 4.315017808373121e-06,
"loss": 0.3718,
"step": 801
},
{
"epoch": 1.7704194260485653,
"grad_norm": 0.13648386299610138,
"learning_rate": 4.302297382900718e-06,
"loss": 0.3722,
"step": 802
},
{
"epoch": 1.7726269315673289,
"grad_norm": 0.12655648589134216,
"learning_rate": 4.289581561233574e-06,
"loss": 0.3667,
"step": 803
},
{
"epoch": 1.7748344370860927,
"grad_norm": 0.12654156982898712,
"learning_rate": 4.276870427277299e-06,
"loss": 0.3666,
"step": 804
},
{
"epoch": 1.7770419426048565,
"grad_norm": 0.15781165659427643,
"learning_rate": 4.264164064906581e-06,
"loss": 0.3787,
"step": 805
},
{
"epoch": 1.7792494481236203,
"grad_norm": 0.12070343643426895,
"learning_rate": 4.251462557964612e-06,
"loss": 0.3897,
"step": 806
},
{
"epoch": 1.781456953642384,
"grad_norm": 0.1305495798587799,
"learning_rate": 4.238765990262554e-06,
"loss": 0.3622,
"step": 807
},
{
"epoch": 1.783664459161148,
"grad_norm": 0.12424585223197937,
"learning_rate": 4.226074445578969e-06,
"loss": 0.3732,
"step": 808
},
{
"epoch": 1.7858719646799117,
"grad_norm": 0.12352015823125839,
"learning_rate": 4.213388007659281e-06,
"loss": 0.367,
"step": 809
},
{
"epoch": 1.7880794701986755,
"grad_norm": 0.1388520896434784,
"learning_rate": 4.200706760215219e-06,
"loss": 0.3723,
"step": 810
},
{
"epoch": 1.7902869757174393,
"grad_norm": 0.13200843334197998,
"learning_rate": 4.188030786924252e-06,
"loss": 0.3765,
"step": 811
},
{
"epoch": 1.7924944812362031,
"grad_norm": 0.11911409348249435,
"learning_rate": 4.17536017142906e-06,
"loss": 0.3694,
"step": 812
},
{
"epoch": 1.794701986754967,
"grad_norm": 0.1263428032398224,
"learning_rate": 4.162694997336957e-06,
"loss": 0.3656,
"step": 813
},
{
"epoch": 1.7969094922737305,
"grad_norm": 0.125799298286438,
"learning_rate": 4.1500353482193646e-06,
"loss": 0.3686,
"step": 814
},
{
"epoch": 1.7991169977924946,
"grad_norm": 0.13606040179729462,
"learning_rate": 4.137381307611236e-06,
"loss": 0.3579,
"step": 815
},
{
"epoch": 1.8013245033112582,
"grad_norm": 0.13091742992401123,
"learning_rate": 4.124732959010526e-06,
"loss": 0.369,
"step": 816
},
{
"epoch": 1.8035320088300222,
"grad_norm": 0.13721388578414917,
"learning_rate": 4.112090385877624e-06,
"loss": 0.3673,
"step": 817
},
{
"epoch": 1.8057395143487858,
"grad_norm": 0.12669645249843597,
"learning_rate": 4.099453671634811e-06,
"loss": 0.3778,
"step": 818
},
{
"epoch": 1.8079470198675498,
"grad_norm": 0.12563163042068481,
"learning_rate": 4.086822899665713e-06,
"loss": 0.3687,
"step": 819
},
{
"epoch": 1.8101545253863134,
"grad_norm": 0.1383393555879593,
"learning_rate": 4.07419815331474e-06,
"loss": 0.3787,
"step": 820
},
{
"epoch": 1.8123620309050774,
"grad_norm": 0.1277354508638382,
"learning_rate": 4.061579515886544e-06,
"loss": 0.3752,
"step": 821
},
{
"epoch": 1.814569536423841,
"grad_norm": 0.12579599022865295,
"learning_rate": 4.048967070645465e-06,
"loss": 0.3767,
"step": 822
},
{
"epoch": 1.8167770419426048,
"grad_norm": 0.1388178914785385,
"learning_rate": 4.03636090081499e-06,
"loss": 0.3795,
"step": 823
},
{
"epoch": 1.8189845474613686,
"grad_norm": 0.13451789319515228,
"learning_rate": 4.0237610895771895e-06,
"loss": 0.3636,
"step": 824
},
{
"epoch": 1.8211920529801324,
"grad_norm": 0.1298052817583084,
"learning_rate": 4.011167720072183e-06,
"loss": 0.3831,
"step": 825
},
{
"epoch": 1.8233995584988962,
"grad_norm": 0.13578468561172485,
"learning_rate": 3.99858087539758e-06,
"loss": 0.3787,
"step": 826
},
{
"epoch": 1.82560706401766,
"grad_norm": 0.1320340931415558,
"learning_rate": 3.986000638607937e-06,
"loss": 0.3697,
"step": 827
},
{
"epoch": 1.8278145695364238,
"grad_norm": 0.13334733247756958,
"learning_rate": 3.973427092714212e-06,
"loss": 0.361,
"step": 828
},
{
"epoch": 1.8300220750551877,
"grad_norm": 0.12365536391735077,
"learning_rate": 3.960860320683205e-06,
"loss": 0.3716,
"step": 829
},
{
"epoch": 1.8322295805739515,
"grad_norm": 0.1276884227991104,
"learning_rate": 3.948300405437026e-06,
"loss": 0.3717,
"step": 830
},
{
"epoch": 1.8344370860927153,
"grad_norm": 0.13205379247665405,
"learning_rate": 3.9357474298525356e-06,
"loss": 0.3632,
"step": 831
},
{
"epoch": 1.836644591611479,
"grad_norm": 0.12293750792741776,
"learning_rate": 3.9232014767608065e-06,
"loss": 0.3809,
"step": 832
},
{
"epoch": 1.8388520971302427,
"grad_norm": 0.1293838620185852,
"learning_rate": 3.910662628946573e-06,
"loss": 0.3782,
"step": 833
},
{
"epoch": 1.8410596026490067,
"grad_norm": 0.1325875222682953,
"learning_rate": 3.89813096914768e-06,
"loss": 0.3734,
"step": 834
},
{
"epoch": 1.8432671081677703,
"grad_norm": 0.12951777875423431,
"learning_rate": 3.8856065800545475e-06,
"loss": 0.3828,
"step": 835
},
{
"epoch": 1.8454746136865343,
"grad_norm": 0.12658190727233887,
"learning_rate": 3.873089544309616e-06,
"loss": 0.3753,
"step": 836
},
{
"epoch": 1.847682119205298,
"grad_norm": 0.12527629733085632,
"learning_rate": 3.860579944506811e-06,
"loss": 0.3716,
"step": 837
},
{
"epoch": 1.849889624724062,
"grad_norm": 0.12750259041786194,
"learning_rate": 3.848077863190985e-06,
"loss": 0.3727,
"step": 838
},
{
"epoch": 1.8520971302428255,
"grad_norm": 0.13881564140319824,
"learning_rate": 3.835583382857382e-06,
"loss": 0.3743,
"step": 839
},
{
"epoch": 1.8543046357615895,
"grad_norm": 0.12845860421657562,
"learning_rate": 3.82309658595109e-06,
"loss": 0.3692,
"step": 840
},
{
"epoch": 1.8565121412803531,
"grad_norm": 0.12954264879226685,
"learning_rate": 3.8106175548664994e-06,
"loss": 0.3778,
"step": 841
},
{
"epoch": 1.858719646799117,
"grad_norm": 0.13028046488761902,
"learning_rate": 3.7981463719467587e-06,
"loss": 0.3825,
"step": 842
},
{
"epoch": 1.8609271523178808,
"grad_norm": 0.1258542686700821,
"learning_rate": 3.7856831194832262e-06,
"loss": 0.3817,
"step": 843
},
{
"epoch": 1.8631346578366446,
"grad_norm": 0.1192869246006012,
"learning_rate": 3.7732278797149324e-06,
"loss": 0.3706,
"step": 844
},
{
"epoch": 1.8653421633554084,
"grad_norm": 0.13153821229934692,
"learning_rate": 3.7607807348280364e-06,
"loss": 0.3626,
"step": 845
},
{
"epoch": 1.8675496688741722,
"grad_norm": 0.12072479724884033,
"learning_rate": 3.7483417669552857e-06,
"loss": 0.3557,
"step": 846
},
{
"epoch": 1.869757174392936,
"grad_norm": 0.12467250972986221,
"learning_rate": 3.735911058175467e-06,
"loss": 0.3621,
"step": 847
},
{
"epoch": 1.8719646799116998,
"grad_norm": 0.12176632136106491,
"learning_rate": 3.7234886905128724e-06,
"loss": 0.3726,
"step": 848
},
{
"epoch": 1.8741721854304636,
"grad_norm": 0.1291275918483734,
"learning_rate": 3.711074745936751e-06,
"loss": 0.3663,
"step": 849
},
{
"epoch": 1.8763796909492274,
"grad_norm": 0.1401110738515854,
"learning_rate": 3.698669306360778e-06,
"loss": 0.3763,
"step": 850
},
{
"epoch": 1.8785871964679912,
"grad_norm": 0.12075633555650711,
"learning_rate": 3.6862724536425042e-06,
"loss": 0.3761,
"step": 851
},
{
"epoch": 1.8807947019867548,
"grad_norm": 0.12913060188293457,
"learning_rate": 3.67388426958282e-06,
"loss": 0.3735,
"step": 852
},
{
"epoch": 1.8830022075055188,
"grad_norm": 0.13538530468940735,
"learning_rate": 3.6615048359254155e-06,
"loss": 0.3737,
"step": 853
},
{
"epoch": 1.8852097130242824,
"grad_norm": 0.1396217793226242,
"learning_rate": 3.649134234356242e-06,
"loss": 0.3765,
"step": 854
},
{
"epoch": 1.8874172185430464,
"grad_norm": 0.12726326286792755,
"learning_rate": 3.636772546502973e-06,
"loss": 0.3706,
"step": 855
},
{
"epoch": 1.88962472406181,
"grad_norm": 0.12331200391054153,
"learning_rate": 3.624419853934461e-06,
"loss": 0.3698,
"step": 856
},
{
"epoch": 1.891832229580574,
"grad_norm": 0.12955856323242188,
"learning_rate": 3.612076238160209e-06,
"loss": 0.3779,
"step": 857
},
{
"epoch": 1.8940397350993377,
"grad_norm": 0.1281704604625702,
"learning_rate": 3.5997417806298172e-06,
"loss": 0.3711,
"step": 858
},
{
"epoch": 1.8962472406181017,
"grad_norm": 0.13094061613082886,
"learning_rate": 3.5874165627324665e-06,
"loss": 0.3706,
"step": 859
},
{
"epoch": 1.8984547461368653,
"grad_norm": 0.1244734600186348,
"learning_rate": 3.575100665796362e-06,
"loss": 0.3701,
"step": 860
},
{
"epoch": 1.9006622516556293,
"grad_norm": 0.13025392591953278,
"learning_rate": 3.5627941710882042e-06,
"loss": 0.3774,
"step": 861
},
{
"epoch": 1.9028697571743929,
"grad_norm": 0.13566212356090546,
"learning_rate": 3.550497159812658e-06,
"loss": 0.376,
"step": 862
},
{
"epoch": 1.9050772626931567,
"grad_norm": 0.13212575018405914,
"learning_rate": 3.5382097131118052e-06,
"loss": 0.3759,
"step": 863
},
{
"epoch": 1.9072847682119205,
"grad_norm": 0.13214178383350372,
"learning_rate": 3.525931912064623e-06,
"loss": 0.3706,
"step": 864
},
{
"epoch": 1.9094922737306843,
"grad_norm": 0.1290368288755417,
"learning_rate": 3.5136638376864333e-06,
"loss": 0.3767,
"step": 865
},
{
"epoch": 1.9116997792494481,
"grad_norm": 0.1320725828409195,
"learning_rate": 3.5014055709283847e-06,
"loss": 0.3704,
"step": 866
},
{
"epoch": 1.913907284768212,
"grad_norm": 0.13395248353481293,
"learning_rate": 3.4891571926769e-06,
"loss": 0.3774,
"step": 867
},
{
"epoch": 1.9161147902869757,
"grad_norm": 0.12245456129312515,
"learning_rate": 3.4769187837531637e-06,
"loss": 0.3872,
"step": 868
},
{
"epoch": 1.9183222958057395,
"grad_norm": 0.1327778398990631,
"learning_rate": 3.4646904249125713e-06,
"loss": 0.3666,
"step": 869
},
{
"epoch": 1.9205298013245033,
"grad_norm": 0.13339614868164062,
"learning_rate": 3.4524721968442017e-06,
"loss": 0.3782,
"step": 870
},
{
"epoch": 1.9227373068432672,
"grad_norm": 0.12259134650230408,
"learning_rate": 3.4402641801702883e-06,
"loss": 0.38,
"step": 871
},
{
"epoch": 1.924944812362031,
"grad_norm": 0.12124523520469666,
"learning_rate": 3.4280664554456823e-06,
"loss": 0.3746,
"step": 872
},
{
"epoch": 1.9271523178807946,
"grad_norm": 0.12291016429662704,
"learning_rate": 3.415879103157327e-06,
"loss": 0.3697,
"step": 873
},
{
"epoch": 1.9293598233995586,
"grad_norm": 0.12792356312274933,
"learning_rate": 3.403702203723718e-06,
"loss": 0.3755,
"step": 874
},
{
"epoch": 1.9315673289183222,
"grad_norm": 0.1225101426243782,
"learning_rate": 3.3915358374943813e-06,
"loss": 0.3821,
"step": 875
},
{
"epoch": 1.9337748344370862,
"grad_norm": 0.1265992522239685,
"learning_rate": 3.379380084749336e-06,
"loss": 0.3672,
"step": 876
},
{
"epoch": 1.9359823399558498,
"grad_norm": 0.12633873522281647,
"learning_rate": 3.3672350256985727e-06,
"loss": 0.3766,
"step": 877
},
{
"epoch": 1.9381898454746138,
"grad_norm": 0.1319660246372223,
"learning_rate": 3.3551007404815162e-06,
"loss": 0.374,
"step": 878
},
{
"epoch": 1.9403973509933774,
"grad_norm": 0.12871594727039337,
"learning_rate": 3.3429773091664985e-06,
"loss": 0.3673,
"step": 879
},
{
"epoch": 1.9426048565121414,
"grad_norm": 0.13455092906951904,
"learning_rate": 3.330864811750235e-06,
"loss": 0.3805,
"step": 880
},
{
"epoch": 1.944812362030905,
"grad_norm": 0.13253872096538544,
"learning_rate": 3.3187633281572913e-06,
"loss": 0.3695,
"step": 881
},
{
"epoch": 1.9470198675496688,
"grad_norm": 0.11975626647472382,
"learning_rate": 3.306672938239561e-06,
"loss": 0.3767,
"step": 882
},
{
"epoch": 1.9492273730684326,
"grad_norm": 0.12552706897258759,
"learning_rate": 3.2945937217757324e-06,
"loss": 0.363,
"step": 883
},
{
"epoch": 1.9514348785871964,
"grad_norm": 0.13104894757270813,
"learning_rate": 3.282525758470768e-06,
"loss": 0.3699,
"step": 884
},
{
"epoch": 1.9536423841059603,
"grad_norm": 0.12783929705619812,
"learning_rate": 3.2704691279553725e-06,
"loss": 0.3681,
"step": 885
},
{
"epoch": 1.955849889624724,
"grad_norm": 0.12647448480129242,
"learning_rate": 3.2584239097854763e-06,
"loss": 0.3873,
"step": 886
},
{
"epoch": 1.9580573951434879,
"grad_norm": 0.11794891208410263,
"learning_rate": 3.246390183441703e-06,
"loss": 0.3689,
"step": 887
},
{
"epoch": 1.9602649006622517,
"grad_norm": 0.1255064755678177,
"learning_rate": 3.234368028328845e-06,
"loss": 0.3874,
"step": 888
},
{
"epoch": 1.9624724061810155,
"grad_norm": 0.12461333721876144,
"learning_rate": 3.2223575237753448e-06,
"loss": 0.3719,
"step": 889
},
{
"epoch": 1.9646799116997793,
"grad_norm": 0.13075792789459229,
"learning_rate": 3.2103587490327624e-06,
"loss": 0.3746,
"step": 890
},
{
"epoch": 1.966887417218543,
"grad_norm": 0.13413941860198975,
"learning_rate": 3.198371783275269e-06,
"loss": 0.3723,
"step": 891
},
{
"epoch": 1.9690949227373067,
"grad_norm": 0.13534380495548248,
"learning_rate": 3.186396705599106e-06,
"loss": 0.3762,
"step": 892
},
{
"epoch": 1.9713024282560707,
"grad_norm": 0.12702573835849762,
"learning_rate": 3.1744335950220716e-06,
"loss": 0.3763,
"step": 893
},
{
"epoch": 1.9735099337748343,
"grad_norm": 0.131902277469635,
"learning_rate": 3.162482530483e-06,
"loss": 0.3647,
"step": 894
},
{
"epoch": 1.9757174392935983,
"grad_norm": 0.12826719880104065,
"learning_rate": 3.1505435908412415e-06,
"loss": 0.3705,
"step": 895
},
{
"epoch": 1.977924944812362,
"grad_norm": 0.1235160231590271,
"learning_rate": 3.1386168548761377e-06,
"loss": 0.3806,
"step": 896
},
{
"epoch": 1.980132450331126,
"grad_norm": 0.13026514649391174,
"learning_rate": 3.126702401286503e-06,
"loss": 0.3714,
"step": 897
},
{
"epoch": 1.9823399558498895,
"grad_norm": 0.12384995073080063,
"learning_rate": 3.11480030869011e-06,
"loss": 0.3775,
"step": 898
},
{
"epoch": 1.9845474613686536,
"grad_norm": 0.11622393131256104,
"learning_rate": 3.1029106556231615e-06,
"loss": 0.3685,
"step": 899
},
{
"epoch": 1.9867549668874172,
"grad_norm": 0.12467402219772339,
"learning_rate": 3.0910335205397834e-06,
"loss": 0.3753,
"step": 900
},
{
"epoch": 1.9889624724061812,
"grad_norm": 0.12134755402803421,
"learning_rate": 3.0791689818114988e-06,
"loss": 0.3763,
"step": 901
},
{
"epoch": 1.9911699779249448,
"grad_norm": 0.118636354804039,
"learning_rate": 3.067317117726715e-06,
"loss": 0.3779,
"step": 902
},
{
"epoch": 1.9933774834437086,
"grad_norm": 0.12621286511421204,
"learning_rate": 3.055478006490199e-06,
"loss": 0.3755,
"step": 903
},
{
"epoch": 1.9955849889624724,
"grad_norm": 0.13427864015102386,
"learning_rate": 3.043651726222578e-06,
"loss": 0.3677,
"step": 904
},
{
"epoch": 1.9977924944812362,
"grad_norm": 0.1349465250968933,
"learning_rate": 3.0318383549598085e-06,
"loss": 0.3821,
"step": 905
},
{
"epoch": 2.0,
"grad_norm": 0.1383267641067505,
"learning_rate": 3.020037970652664e-06,
"loss": 0.3605,
"step": 906
},
{
"epoch": 2.0022075055187636,
"grad_norm": 0.1400349736213684,
"learning_rate": 3.00825065116623e-06,
"loss": 0.3587,
"step": 907
},
{
"epoch": 2.0044150110375276,
"grad_norm": 0.12129946798086166,
"learning_rate": 2.996476474279375e-06,
"loss": 0.3564,
"step": 908
},
{
"epoch": 2.006622516556291,
"grad_norm": 0.13566596806049347,
"learning_rate": 2.9847155176842547e-06,
"loss": 0.3628,
"step": 909
},
{
"epoch": 2.0088300220750552,
"grad_norm": 0.12823748588562012,
"learning_rate": 2.9729678589857852e-06,
"loss": 0.3589,
"step": 910
},
{
"epoch": 2.011037527593819,
"grad_norm": 0.1251477748155594,
"learning_rate": 2.9612335757011362e-06,
"loss": 0.3583,
"step": 911
},
{
"epoch": 2.013245033112583,
"grad_norm": 0.1308368593454361,
"learning_rate": 2.94951274525922e-06,
"loss": 0.354,
"step": 912
},
{
"epoch": 2.0154525386313464,
"grad_norm": 0.1357547789812088,
"learning_rate": 2.9378054450001836e-06,
"loss": 0.3664,
"step": 913
},
{
"epoch": 2.0176600441501105,
"grad_norm": 0.13018201291561127,
"learning_rate": 2.9261117521748904e-06,
"loss": 0.3627,
"step": 914
},
{
"epoch": 2.019867549668874,
"grad_norm": 0.12550939619541168,
"learning_rate": 2.914431743944414e-06,
"loss": 0.3518,
"step": 915
},
{
"epoch": 2.022075055187638,
"grad_norm": 0.12784212827682495,
"learning_rate": 2.902765497379534e-06,
"loss": 0.3547,
"step": 916
},
{
"epoch": 2.0242825607064017,
"grad_norm": 0.13502328097820282,
"learning_rate": 2.8911130894602198e-06,
"loss": 0.3648,
"step": 917
},
{
"epoch": 2.0264900662251657,
"grad_norm": 0.12056614458560944,
"learning_rate": 2.8794745970751308e-06,
"loss": 0.3612,
"step": 918
},
{
"epoch": 2.0286975717439293,
"grad_norm": 0.12976831197738647,
"learning_rate": 2.8678500970210977e-06,
"loss": 0.3619,
"step": 919
},
{
"epoch": 2.0309050772626933,
"grad_norm": 0.13974756002426147,
"learning_rate": 2.8562396660026304e-06,
"loss": 0.3585,
"step": 920
},
{
"epoch": 2.033112582781457,
"grad_norm": 0.1298479437828064,
"learning_rate": 2.8446433806313966e-06,
"loss": 0.3533,
"step": 921
},
{
"epoch": 2.035320088300221,
"grad_norm": 0.13592906296253204,
"learning_rate": 2.83306131742573e-06,
"loss": 0.36,
"step": 922
},
{
"epoch": 2.0375275938189845,
"grad_norm": 0.13170795142650604,
"learning_rate": 2.821493552810114e-06,
"loss": 0.3535,
"step": 923
},
{
"epoch": 2.0397350993377485,
"grad_norm": 0.14562806487083435,
"learning_rate": 2.809940163114685e-06,
"loss": 0.3737,
"step": 924
},
{
"epoch": 2.041942604856512,
"grad_norm": 0.12969909608364105,
"learning_rate": 2.7984012245747237e-06,
"loss": 0.3597,
"step": 925
},
{
"epoch": 2.044150110375276,
"grad_norm": 0.13986288011074066,
"learning_rate": 2.786876813330158e-06,
"loss": 0.3608,
"step": 926
},
{
"epoch": 2.0463576158940397,
"grad_norm": 0.13949252665042877,
"learning_rate": 2.7753670054250583e-06,
"loss": 0.3507,
"step": 927
},
{
"epoch": 2.0485651214128033,
"grad_norm": 0.12819348275661469,
"learning_rate": 2.763871876807129e-06,
"loss": 0.356,
"step": 928
},
{
"epoch": 2.0507726269315674,
"grad_norm": 0.13626167178153992,
"learning_rate": 2.7523915033272163e-06,
"loss": 0.3534,
"step": 929
},
{
"epoch": 2.052980132450331,
"grad_norm": 0.14214330911636353,
"learning_rate": 2.740925960738802e-06,
"loss": 0.3627,
"step": 930
},
{
"epoch": 2.055187637969095,
"grad_norm": 0.1371086835861206,
"learning_rate": 2.72947532469751e-06,
"loss": 0.3594,
"step": 931
},
{
"epoch": 2.0573951434878586,
"grad_norm": 0.15132765471935272,
"learning_rate": 2.7180396707606023e-06,
"loss": 0.3487,
"step": 932
},
{
"epoch": 2.0596026490066226,
"grad_norm": 0.1387225240468979,
"learning_rate": 2.7066190743864774e-06,
"loss": 0.3625,
"step": 933
},
{
"epoch": 2.061810154525386,
"grad_norm": 0.13822317123413086,
"learning_rate": 2.695213610934179e-06,
"loss": 0.3542,
"step": 934
},
{
"epoch": 2.06401766004415,
"grad_norm": 0.1481810212135315,
"learning_rate": 2.6838233556628932e-06,
"loss": 0.3543,
"step": 935
},
{
"epoch": 2.066225165562914,
"grad_norm": 0.1475333869457245,
"learning_rate": 2.672448383731461e-06,
"loss": 0.3587,
"step": 936
},
{
"epoch": 2.068432671081678,
"grad_norm": 0.13259293138980865,
"learning_rate": 2.661088770197868e-06,
"loss": 0.3531,
"step": 937
},
{
"epoch": 2.0706401766004414,
"grad_norm": 0.129195436835289,
"learning_rate": 2.6497445900187635e-06,
"loss": 0.3557,
"step": 938
},
{
"epoch": 2.0728476821192054,
"grad_norm": 0.12847961485385895,
"learning_rate": 2.6384159180489542e-06,
"loss": 0.3477,
"step": 939
},
{
"epoch": 2.075055187637969,
"grad_norm": 0.13196007907390594,
"learning_rate": 2.6271028290409216e-06,
"loss": 0.3584,
"step": 940
},
{
"epoch": 2.077262693156733,
"grad_norm": 0.1364007443189621,
"learning_rate": 2.6158053976443164e-06,
"loss": 0.3601,
"step": 941
},
{
"epoch": 2.0794701986754967,
"grad_norm": 0.1247749775648117,
"learning_rate": 2.604523698405477e-06,
"loss": 0.3647,
"step": 942
},
{
"epoch": 2.0816777041942607,
"grad_norm": 0.12793247401714325,
"learning_rate": 2.5932578057669273e-06,
"loss": 0.3656,
"step": 943
},
{
"epoch": 2.0838852097130243,
"grad_norm": 0.12500843405723572,
"learning_rate": 2.5820077940668975e-06,
"loss": 0.3511,
"step": 944
},
{
"epoch": 2.0860927152317883,
"grad_norm": 0.13672660291194916,
"learning_rate": 2.570773737538825e-06,
"loss": 0.3709,
"step": 945
},
{
"epoch": 2.088300220750552,
"grad_norm": 0.1374640017747879,
"learning_rate": 2.559555710310866e-06,
"loss": 0.3528,
"step": 946
},
{
"epoch": 2.0905077262693155,
"grad_norm": 0.13424260914325714,
"learning_rate": 2.5483537864054055e-06,
"loss": 0.354,
"step": 947
},
{
"epoch": 2.0927152317880795,
"grad_norm": 0.13996686041355133,
"learning_rate": 2.537168039738569e-06,
"loss": 0.3655,
"step": 948
},
{
"epoch": 2.094922737306843,
"grad_norm": 0.12471663951873779,
"learning_rate": 2.525998544119743e-06,
"loss": 0.3606,
"step": 949
},
{
"epoch": 2.097130242825607,
"grad_norm": 0.13321642577648163,
"learning_rate": 2.514845373251078e-06,
"loss": 0.3477,
"step": 950
},
{
"epoch": 2.0993377483443707,
"grad_norm": 0.1407744139432907,
"learning_rate": 2.5037086007270017e-06,
"loss": 0.3566,
"step": 951
},
{
"epoch": 2.1015452538631347,
"grad_norm": 0.1319989413022995,
"learning_rate": 2.4925883000337407e-06,
"loss": 0.3511,
"step": 952
},
{
"epoch": 2.1037527593818983,
"grad_norm": 0.12347118556499481,
"learning_rate": 2.4814845445488288e-06,
"loss": 0.3609,
"step": 953
},
{
"epoch": 2.1059602649006623,
"grad_norm": 0.15142813324928284,
"learning_rate": 2.4703974075406316e-06,
"loss": 0.3581,
"step": 954
},
{
"epoch": 2.108167770419426,
"grad_norm": 0.12996384501457214,
"learning_rate": 2.4593269621678502e-06,
"loss": 0.3586,
"step": 955
},
{
"epoch": 2.11037527593819,
"grad_norm": 0.12407524138689041,
"learning_rate": 2.448273281479052e-06,
"loss": 0.3561,
"step": 956
},
{
"epoch": 2.1125827814569536,
"grad_norm": 0.12946733832359314,
"learning_rate": 2.437236438412177e-06,
"loss": 0.3637,
"step": 957
},
{
"epoch": 2.1147902869757176,
"grad_norm": 0.14091919362545013,
"learning_rate": 2.426216505794066e-06,
"loss": 0.3572,
"step": 958
},
{
"epoch": 2.116997792494481,
"grad_norm": 0.13770008087158203,
"learning_rate": 2.4152135563399743e-06,
"loss": 0.3617,
"step": 959
},
{
"epoch": 2.119205298013245,
"grad_norm": 0.1215134859085083,
"learning_rate": 2.404227662653089e-06,
"loss": 0.3613,
"step": 960
},
{
"epoch": 2.121412803532009,
"grad_norm": 0.13981877267360687,
"learning_rate": 2.3932588972240642e-06,
"loss": 0.3601,
"step": 961
},
{
"epoch": 2.123620309050773,
"grad_norm": 0.13256755471229553,
"learning_rate": 2.3823073324305222e-06,
"loss": 0.3551,
"step": 962
},
{
"epoch": 2.1258278145695364,
"grad_norm": 0.14095091819763184,
"learning_rate": 2.3713730405365965e-06,
"loss": 0.3671,
"step": 963
},
{
"epoch": 2.1280353200883004,
"grad_norm": 0.11968285590410233,
"learning_rate": 2.3604560936924363e-06,
"loss": 0.3558,
"step": 964
},
{
"epoch": 2.130242825607064,
"grad_norm": 0.13734926283359528,
"learning_rate": 2.3495565639337446e-06,
"loss": 0.3653,
"step": 965
},
{
"epoch": 2.1324503311258276,
"grad_norm": 0.13910116255283356,
"learning_rate": 2.3386745231812923e-06,
"loss": 0.3559,
"step": 966
},
{
"epoch": 2.1346578366445916,
"grad_norm": 0.13142208755016327,
"learning_rate": 2.3278100432404554e-06,
"loss": 0.3571,
"step": 967
},
{
"epoch": 2.136865342163355,
"grad_norm": 0.12986791133880615,
"learning_rate": 2.3169631958007316e-06,
"loss": 0.3613,
"step": 968
},
{
"epoch": 2.1390728476821192,
"grad_norm": 0.12400130182504654,
"learning_rate": 2.3061340524352687e-06,
"loss": 0.3596,
"step": 969
},
{
"epoch": 2.141280353200883,
"grad_norm": 0.12488456070423126,
"learning_rate": 2.295322684600395e-06,
"loss": 0.3571,
"step": 970
},
{
"epoch": 2.143487858719647,
"grad_norm": 0.13525961339473724,
"learning_rate": 2.284529163635144e-06,
"loss": 0.352,
"step": 971
},
{
"epoch": 2.1456953642384105,
"grad_norm": 0.14086809754371643,
"learning_rate": 2.273753560760793e-06,
"loss": 0.3529,
"step": 972
},
{
"epoch": 2.1479028697571745,
"grad_norm": 0.13002382218837738,
"learning_rate": 2.262995947080378e-06,
"loss": 0.3628,
"step": 973
},
{
"epoch": 2.150110375275938,
"grad_norm": 0.13714134693145752,
"learning_rate": 2.252256393578241e-06,
"loss": 0.3602,
"step": 974
},
{
"epoch": 2.152317880794702,
"grad_norm": 0.1353500783443451,
"learning_rate": 2.2415349711195455e-06,
"loss": 0.3562,
"step": 975
},
{
"epoch": 2.1545253863134657,
"grad_norm": 0.14100465178489685,
"learning_rate": 2.230831750449825e-06,
"loss": 0.3562,
"step": 976
},
{
"epoch": 2.1567328918322297,
"grad_norm": 0.11746090650558472,
"learning_rate": 2.2201468021945024e-06,
"loss": 0.3679,
"step": 977
},
{
"epoch": 2.1589403973509933,
"grad_norm": 0.13000161945819855,
"learning_rate": 2.209480196858429e-06,
"loss": 0.3579,
"step": 978
},
{
"epoch": 2.1611479028697573,
"grad_norm": 0.14291274547576904,
"learning_rate": 2.198832004825427e-06,
"loss": 0.366,
"step": 979
},
{
"epoch": 2.163355408388521,
"grad_norm": 0.1327233463525772,
"learning_rate": 2.1882022963578087e-06,
"loss": 0.3576,
"step": 980
},
{
"epoch": 2.165562913907285,
"grad_norm": 0.12592221796512604,
"learning_rate": 2.177591141595931e-06,
"loss": 0.3554,
"step": 981
},
{
"epoch": 2.1677704194260485,
"grad_norm": 0.13044218719005585,
"learning_rate": 2.166998610557718e-06,
"loss": 0.3489,
"step": 982
},
{
"epoch": 2.1699779249448126,
"grad_norm": 0.14413036406040192,
"learning_rate": 2.1564247731382063e-06,
"loss": 0.3653,
"step": 983
},
{
"epoch": 2.172185430463576,
"grad_norm": 0.12830859422683716,
"learning_rate": 2.145869699109081e-06,
"loss": 0.3566,
"step": 984
},
{
"epoch": 2.1743929359823397,
"grad_norm": 0.12629762291908264,
"learning_rate": 2.1353334581182193e-06,
"loss": 0.3581,
"step": 985
},
{
"epoch": 2.1766004415011038,
"grad_norm": 0.12835480272769928,
"learning_rate": 2.1248161196892295e-06,
"loss": 0.3655,
"step": 986
},
{
"epoch": 2.1788079470198674,
"grad_norm": 0.12383104115724564,
"learning_rate": 2.1143177532209855e-06,
"loss": 0.3564,
"step": 987
},
{
"epoch": 2.1810154525386314,
"grad_norm": 0.1490948498249054,
"learning_rate": 2.1038384279871786e-06,
"loss": 0.37,
"step": 988
},
{
"epoch": 2.183222958057395,
"grad_norm": 0.12230035662651062,
"learning_rate": 2.0933782131358516e-06,
"loss": 0.3549,
"step": 989
},
{
"epoch": 2.185430463576159,
"grad_norm": 0.13511480391025543,
"learning_rate": 2.082937177688952e-06,
"loss": 0.3576,
"step": 990
},
{
"epoch": 2.1876379690949226,
"grad_norm": 0.135064497590065,
"learning_rate": 2.0725153905418726e-06,
"loss": 0.3544,
"step": 991
},
{
"epoch": 2.1898454746136866,
"grad_norm": 0.12062691897153854,
"learning_rate": 2.0621129204629907e-06,
"loss": 0.3523,
"step": 992
},
{
"epoch": 2.19205298013245,
"grad_norm": 0.14147672057151794,
"learning_rate": 2.0517298360932202e-06,
"loss": 0.3583,
"step": 993
},
{
"epoch": 2.1942604856512142,
"grad_norm": 0.12477768212556839,
"learning_rate": 2.041366205945563e-06,
"loss": 0.3604,
"step": 994
},
{
"epoch": 2.196467991169978,
"grad_norm": 0.14072567224502563,
"learning_rate": 2.0310220984046467e-06,
"loss": 0.3693,
"step": 995
},
{
"epoch": 2.198675496688742,
"grad_norm": 0.13882459700107574,
"learning_rate": 2.020697581726279e-06,
"loss": 0.3604,
"step": 996
},
{
"epoch": 2.2008830022075054,
"grad_norm": 0.12994110584259033,
"learning_rate": 2.010392724037002e-06,
"loss": 0.3589,
"step": 997
},
{
"epoch": 2.2030905077262695,
"grad_norm": 0.13540343940258026,
"learning_rate": 2.0001075933336302e-06,
"loss": 0.3664,
"step": 998
},
{
"epoch": 2.205298013245033,
"grad_norm": 0.12710383534431458,
"learning_rate": 1.9898422574828163e-06,
"loss": 0.3606,
"step": 999
},
{
"epoch": 2.207505518763797,
"grad_norm": 0.13436554372310638,
"learning_rate": 1.979596784220591e-06,
"loss": 0.3583,
"step": 1000
},
{
"epoch": 2.2097130242825607,
"grad_norm": 0.1312997192144394,
"learning_rate": 1.969371241151923e-06,
"loss": 0.3485,
"step": 1001
},
{
"epoch": 2.2119205298013247,
"grad_norm": 0.12719684839248657,
"learning_rate": 1.9591656957502696e-06,
"loss": 0.3646,
"step": 1002
},
{
"epoch": 2.2141280353200883,
"grad_norm": 0.11936060339212418,
"learning_rate": 1.9489802153571373e-06,
"loss": 0.3578,
"step": 1003
},
{
"epoch": 2.216335540838852,
"grad_norm": 0.11757774651050568,
"learning_rate": 1.938814867181632e-06,
"loss": 0.3633,
"step": 1004
},
{
"epoch": 2.218543046357616,
"grad_norm": 0.12061762064695358,
"learning_rate": 1.928669718300013e-06,
"loss": 0.3539,
"step": 1005
},
{
"epoch": 2.2207505518763795,
"grad_norm": 0.11802493780851364,
"learning_rate": 1.9185448356552575e-06,
"loss": 0.3519,
"step": 1006
},
{
"epoch": 2.2229580573951435,
"grad_norm": 0.12381359189748764,
"learning_rate": 1.9084402860566128e-06,
"loss": 0.3528,
"step": 1007
},
{
"epoch": 2.225165562913907,
"grad_norm": 0.12773284316062927,
"learning_rate": 1.8983561361791608e-06,
"loss": 0.3618,
"step": 1008
},
{
"epoch": 2.227373068432671,
"grad_norm": 0.1377752125263214,
"learning_rate": 1.8882924525633778e-06,
"loss": 0.36,
"step": 1009
},
{
"epoch": 2.2295805739514347,
"grad_norm": 0.13043536245822906,
"learning_rate": 1.8782493016146868e-06,
"loss": 0.3588,
"step": 1010
},
{
"epoch": 2.2317880794701987,
"grad_norm": 0.12282190471887589,
"learning_rate": 1.8682267496030276e-06,
"loss": 0.3661,
"step": 1011
},
{
"epoch": 2.2339955849889623,
"grad_norm": 0.11887135356664658,
"learning_rate": 1.8582248626624217e-06,
"loss": 0.3585,
"step": 1012
},
{
"epoch": 2.2362030905077264,
"grad_norm": 0.13011059165000916,
"learning_rate": 1.8482437067905268e-06,
"loss": 0.3648,
"step": 1013
},
{
"epoch": 2.23841059602649,
"grad_norm": 0.12632228434085846,
"learning_rate": 1.8382833478482066e-06,
"loss": 0.3711,
"step": 1014
},
{
"epoch": 2.240618101545254,
"grad_norm": 0.13252626359462738,
"learning_rate": 1.8283438515590996e-06,
"loss": 0.3691,
"step": 1015
},
{
"epoch": 2.2428256070640176,
"grad_norm": 0.12210851162672043,
"learning_rate": 1.8184252835091764e-06,
"loss": 0.36,
"step": 1016
},
{
"epoch": 2.2450331125827816,
"grad_norm": 0.1259019374847412,
"learning_rate": 1.8085277091463188e-06,
"loss": 0.3507,
"step": 1017
},
{
"epoch": 2.247240618101545,
"grad_norm": 0.11658161133527756,
"learning_rate": 1.798651193779875e-06,
"loss": 0.3512,
"step": 1018
},
{
"epoch": 2.249448123620309,
"grad_norm": 0.12686073780059814,
"learning_rate": 1.788795802580236e-06,
"loss": 0.3554,
"step": 1019
},
{
"epoch": 2.251655629139073,
"grad_norm": 0.13531404733657837,
"learning_rate": 1.7789616005784077e-06,
"loss": 0.3569,
"step": 1020
},
{
"epoch": 2.253863134657837,
"grad_norm": 0.12338366359472275,
"learning_rate": 1.7691486526655782e-06,
"loss": 0.361,
"step": 1021
},
{
"epoch": 2.2560706401766004,
"grad_norm": 0.1204574853181839,
"learning_rate": 1.7593570235926883e-06,
"loss": 0.3665,
"step": 1022
},
{
"epoch": 2.258278145695364,
"grad_norm": 0.12831008434295654,
"learning_rate": 1.7495867779700053e-06,
"loss": 0.3617,
"step": 1023
},
{
"epoch": 2.260485651214128,
"grad_norm": 0.1243155300617218,
"learning_rate": 1.7398379802666993e-06,
"loss": 0.3485,
"step": 1024
},
{
"epoch": 2.262693156732892,
"grad_norm": 0.12370171397924423,
"learning_rate": 1.7301106948104123e-06,
"loss": 0.363,
"step": 1025
},
{
"epoch": 2.2649006622516556,
"grad_norm": 0.12362192571163177,
"learning_rate": 1.7204049857868433e-06,
"loss": 0.3553,
"step": 1026
},
{
"epoch": 2.2671081677704192,
"grad_norm": 0.1303797960281372,
"learning_rate": 1.7107209172393158e-06,
"loss": 0.3582,
"step": 1027
},
{
"epoch": 2.2693156732891833,
"grad_norm": 0.12913931906223297,
"learning_rate": 1.701058553068357e-06,
"loss": 0.3621,
"step": 1028
},
{
"epoch": 2.271523178807947,
"grad_norm": 0.13231627643108368,
"learning_rate": 1.6914179570312767e-06,
"loss": 0.3529,
"step": 1029
},
{
"epoch": 2.273730684326711,
"grad_norm": 0.12977519631385803,
"learning_rate": 1.6817991927417516e-06,
"loss": 0.3619,
"step": 1030
},
{
"epoch": 2.2759381898454745,
"grad_norm": 0.12845906615257263,
"learning_rate": 1.672202323669397e-06,
"loss": 0.3613,
"step": 1031
},
{
"epoch": 2.2781456953642385,
"grad_norm": 0.1263485550880432,
"learning_rate": 1.662627413139351e-06,
"loss": 0.3608,
"step": 1032
},
{
"epoch": 2.280353200883002,
"grad_norm": 0.12105315178632736,
"learning_rate": 1.6530745243318646e-06,
"loss": 0.3694,
"step": 1033
},
{
"epoch": 2.282560706401766,
"grad_norm": 0.12585538625717163,
"learning_rate": 1.64354372028187e-06,
"loss": 0.3674,
"step": 1034
},
{
"epoch": 2.2847682119205297,
"grad_norm": 0.13143764436244965,
"learning_rate": 1.634035063878579e-06,
"loss": 0.3654,
"step": 1035
},
{
"epoch": 2.2869757174392937,
"grad_norm": 0.1330743134021759,
"learning_rate": 1.6245486178650582e-06,
"loss": 0.3579,
"step": 1036
},
{
"epoch": 2.2891832229580573,
"grad_norm": 0.12009730935096741,
"learning_rate": 1.6150844448378178e-06,
"loss": 0.3531,
"step": 1037
},
{
"epoch": 2.2913907284768213,
"grad_norm": 0.13864392042160034,
"learning_rate": 1.6056426072464015e-06,
"loss": 0.3628,
"step": 1038
},
{
"epoch": 2.293598233995585,
"grad_norm": 0.13580797612667084,
"learning_rate": 1.5962231673929735e-06,
"loss": 0.3722,
"step": 1039
},
{
"epoch": 2.295805739514349,
"grad_norm": 0.12375160306692123,
"learning_rate": 1.586826187431902e-06,
"loss": 0.364,
"step": 1040
},
{
"epoch": 2.2980132450331126,
"grad_norm": 0.12487831711769104,
"learning_rate": 1.5774517293693558e-06,
"loss": 0.3592,
"step": 1041
},
{
"epoch": 2.300220750551876,
"grad_norm": 0.14687076210975647,
"learning_rate": 1.5680998550628912e-06,
"loss": 0.3506,
"step": 1042
},
{
"epoch": 2.30242825607064,
"grad_norm": 0.12587517499923706,
"learning_rate": 1.558770626221044e-06,
"loss": 0.3577,
"step": 1043
},
{
"epoch": 2.304635761589404,
"grad_norm": 0.13661305606365204,
"learning_rate": 1.5494641044029268e-06,
"loss": 0.3516,
"step": 1044
},
{
"epoch": 2.306843267108168,
"grad_norm": 0.1377173811197281,
"learning_rate": 1.5401803510178197e-06,
"loss": 0.3603,
"step": 1045
},
{
"epoch": 2.3090507726269314,
"grad_norm": 0.13497798144817352,
"learning_rate": 1.5309194273247612e-06,
"loss": 0.3545,
"step": 1046
},
{
"epoch": 2.3112582781456954,
"grad_norm": 0.12721391022205353,
"learning_rate": 1.5216813944321473e-06,
"loss": 0.3447,
"step": 1047
},
{
"epoch": 2.313465783664459,
"grad_norm": 0.12541592121124268,
"learning_rate": 1.5124663132973338e-06,
"loss": 0.3637,
"step": 1048
},
{
"epoch": 2.315673289183223,
"grad_norm": 0.13365043699741364,
"learning_rate": 1.5032742447262228e-06,
"loss": 0.3535,
"step": 1049
},
{
"epoch": 2.3178807947019866,
"grad_norm": 0.13111183047294617,
"learning_rate": 1.4941052493728731e-06,
"loss": 0.3528,
"step": 1050
},
{
"epoch": 2.3200883002207506,
"grad_norm": 0.1259627640247345,
"learning_rate": 1.48495938773909e-06,
"loss": 0.3599,
"step": 1051
},
{
"epoch": 2.322295805739514,
"grad_norm": 0.12294916808605194,
"learning_rate": 1.4758367201740303e-06,
"loss": 0.3604,
"step": 1052
},
{
"epoch": 2.3245033112582782,
"grad_norm": 0.12867586314678192,
"learning_rate": 1.46673730687381e-06,
"loss": 0.3649,
"step": 1053
},
{
"epoch": 2.326710816777042,
"grad_norm": 0.12218500673770905,
"learning_rate": 1.4576612078810953e-06,
"loss": 0.3637,
"step": 1054
},
{
"epoch": 2.328918322295806,
"grad_norm": 0.12343169003725052,
"learning_rate": 1.448608483084713e-06,
"loss": 0.3552,
"step": 1055
},
{
"epoch": 2.3311258278145695,
"grad_norm": 0.12273656576871872,
"learning_rate": 1.4395791922192575e-06,
"loss": 0.3613,
"step": 1056
},
{
"epoch": 2.3333333333333335,
"grad_norm": 0.1312197744846344,
"learning_rate": 1.430573394864695e-06,
"loss": 0.3603,
"step": 1057
},
{
"epoch": 2.335540838852097,
"grad_norm": 0.13838644325733185,
"learning_rate": 1.4215911504459645e-06,
"loss": 0.3642,
"step": 1058
},
{
"epoch": 2.337748344370861,
"grad_norm": 0.11605527251958847,
"learning_rate": 1.412632518232594e-06,
"loss": 0.3626,
"step": 1059
},
{
"epoch": 2.3399558498896247,
"grad_norm": 0.13165433704853058,
"learning_rate": 1.4036975573383028e-06,
"loss": 0.3636,
"step": 1060
},
{
"epoch": 2.3421633554083887,
"grad_norm": 0.1277674436569214,
"learning_rate": 1.3947863267206174e-06,
"loss": 0.3517,
"step": 1061
},
{
"epoch": 2.3443708609271523,
"grad_norm": 0.13928262889385223,
"learning_rate": 1.3858988851804816e-06,
"loss": 0.3649,
"step": 1062
},
{
"epoch": 2.3465783664459163,
"grad_norm": 0.13180677592754364,
"learning_rate": 1.3770352913618613e-06,
"loss": 0.3506,
"step": 1063
},
{
"epoch": 2.34878587196468,
"grad_norm": 0.13746923208236694,
"learning_rate": 1.3681956037513656e-06,
"loss": 0.3583,
"step": 1064
},
{
"epoch": 2.3509933774834435,
"grad_norm": 0.13763363659381866,
"learning_rate": 1.3593798806778546e-06,
"loss": 0.354,
"step": 1065
},
{
"epoch": 2.3532008830022075,
"grad_norm": 0.12797822058200836,
"learning_rate": 1.3505881803120647e-06,
"loss": 0.3719,
"step": 1066
},
{
"epoch": 2.355408388520971,
"grad_norm": 0.12430882453918457,
"learning_rate": 1.341820560666211e-06,
"loss": 0.3622,
"step": 1067
},
{
"epoch": 2.357615894039735,
"grad_norm": 0.13003040850162506,
"learning_rate": 1.3330770795936172e-06,
"loss": 0.3644,
"step": 1068
},
{
"epoch": 2.3598233995584987,
"grad_norm": 0.13834795355796814,
"learning_rate": 1.3243577947883223e-06,
"loss": 0.3539,
"step": 1069
},
{
"epoch": 2.3620309050772628,
"grad_norm": 0.12837594747543335,
"learning_rate": 1.315662763784712e-06,
"loss": 0.366,
"step": 1070
},
{
"epoch": 2.3642384105960264,
"grad_norm": 0.12611278891563416,
"learning_rate": 1.3069920439571277e-06,
"loss": 0.3567,
"step": 1071
},
{
"epoch": 2.3664459161147904,
"grad_norm": 0.1504361927509308,
"learning_rate": 1.2983456925194953e-06,
"loss": 0.3539,
"step": 1072
},
{
"epoch": 2.368653421633554,
"grad_norm": 0.13299378752708435,
"learning_rate": 1.2897237665249429e-06,
"loss": 0.3612,
"step": 1073
},
{
"epoch": 2.370860927152318,
"grad_norm": 0.13157807290554047,
"learning_rate": 1.2811263228654308e-06,
"loss": 0.3652,
"step": 1074
},
{
"epoch": 2.3730684326710816,
"grad_norm": 0.11843698471784592,
"learning_rate": 1.2725534182713717e-06,
"loss": 0.3664,
"step": 1075
},
{
"epoch": 2.3752759381898456,
"grad_norm": 0.1275041550397873,
"learning_rate": 1.2640051093112532e-06,
"loss": 0.3595,
"step": 1076
},
{
"epoch": 2.377483443708609,
"grad_norm": 0.12319236993789673,
"learning_rate": 1.25548145239127e-06,
"loss": 0.3458,
"step": 1077
},
{
"epoch": 2.3796909492273732,
"grad_norm": 0.1297508329153061,
"learning_rate": 1.2469825037549493e-06,
"loss": 0.3599,
"step": 1078
},
{
"epoch": 2.381898454746137,
"grad_norm": 0.13835854828357697,
"learning_rate": 1.2385083194827818e-06,
"loss": 0.3631,
"step": 1079
},
{
"epoch": 2.384105960264901,
"grad_norm": 0.14326325058937073,
"learning_rate": 1.2300589554918502e-06,
"loss": 0.3612,
"step": 1080
},
{
"epoch": 2.3863134657836644,
"grad_norm": 0.13561151921749115,
"learning_rate": 1.221634467535458e-06,
"loss": 0.3623,
"step": 1081
},
{
"epoch": 2.3885209713024285,
"grad_norm": 0.12151821702718735,
"learning_rate": 1.2132349112027636e-06,
"loss": 0.347,
"step": 1082
},
{
"epoch": 2.390728476821192,
"grad_norm": 0.12261461466550827,
"learning_rate": 1.204860341918414e-06,
"loss": 0.3615,
"step": 1083
},
{
"epoch": 2.3929359823399556,
"grad_norm": 0.1293371468782425,
"learning_rate": 1.1965108149421812e-06,
"loss": 0.3611,
"step": 1084
},
{
"epoch": 2.3951434878587197,
"grad_norm": 0.12331968545913696,
"learning_rate": 1.1881863853685904e-06,
"loss": 0.3635,
"step": 1085
},
{
"epoch": 2.3973509933774833,
"grad_norm": 0.13170567154884338,
"learning_rate": 1.1798871081265672e-06,
"loss": 0.3698,
"step": 1086
},
{
"epoch": 2.3995584988962473,
"grad_norm": 0.1309044063091278,
"learning_rate": 1.1716130379790613e-06,
"loss": 0.3474,
"step": 1087
},
{
"epoch": 2.401766004415011,
"grad_norm": 0.12268619239330292,
"learning_rate": 1.1633642295227005e-06,
"loss": 0.3634,
"step": 1088
},
{
"epoch": 2.403973509933775,
"grad_norm": 0.12474343925714493,
"learning_rate": 1.155140737187418e-06,
"loss": 0.3651,
"step": 1089
},
{
"epoch": 2.4061810154525385,
"grad_norm": 0.13622671365737915,
"learning_rate": 1.1469426152360974e-06,
"loss": 0.3626,
"step": 1090
},
{
"epoch": 2.4083885209713025,
"grad_norm": 0.1363033503293991,
"learning_rate": 1.138769917764221e-06,
"loss": 0.3608,
"step": 1091
},
{
"epoch": 2.410596026490066,
"grad_norm": 0.12666958570480347,
"learning_rate": 1.1306226986994989e-06,
"loss": 0.3571,
"step": 1092
},
{
"epoch": 2.41280353200883,
"grad_norm": 0.13340440392494202,
"learning_rate": 1.1225010118015306e-06,
"loss": 0.3602,
"step": 1093
},
{
"epoch": 2.4150110375275937,
"grad_norm": 0.14224812388420105,
"learning_rate": 1.1144049106614335e-06,
"loss": 0.3525,
"step": 1094
},
{
"epoch": 2.4172185430463577,
"grad_norm": 0.13309240341186523,
"learning_rate": 1.1063344487015e-06,
"loss": 0.3595,
"step": 1095
},
{
"epoch": 2.4194260485651213,
"grad_norm": 0.1310422122478485,
"learning_rate": 1.098289679174841e-06,
"loss": 0.3674,
"step": 1096
},
{
"epoch": 2.4216335540838854,
"grad_norm": 0.12578153610229492,
"learning_rate": 1.090270655165036e-06,
"loss": 0.3543,
"step": 1097
},
{
"epoch": 2.423841059602649,
"grad_norm": 0.13076968491077423,
"learning_rate": 1.082277429585784e-06,
"loss": 0.3724,
"step": 1098
},
{
"epoch": 2.426048565121413,
"grad_norm": 0.14538316428661346,
"learning_rate": 1.074310055180549e-06,
"loss": 0.3539,
"step": 1099
},
{
"epoch": 2.4282560706401766,
"grad_norm": 0.13270893692970276,
"learning_rate": 1.0663685845222177e-06,
"loss": 0.3637,
"step": 1100
},
{
"epoch": 2.4304635761589406,
"grad_norm": 0.13810580968856812,
"learning_rate": 1.0584530700127478e-06,
"loss": 0.3526,
"step": 1101
},
{
"epoch": 2.432671081677704,
"grad_norm": 0.1298057585954666,
"learning_rate": 1.0505635638828288e-06,
"loss": 0.3551,
"step": 1102
},
{
"epoch": 2.4348785871964678,
"grad_norm": 0.13741736114025116,
"learning_rate": 1.0427001181915298e-06,
"loss": 0.3512,
"step": 1103
},
{
"epoch": 2.437086092715232,
"grad_norm": 0.1301170289516449,
"learning_rate": 1.034862784825963e-06,
"loss": 0.3568,
"step": 1104
},
{
"epoch": 2.439293598233996,
"grad_norm": 0.12320411205291748,
"learning_rate": 1.0270516155009336e-06,
"loss": 0.3701,
"step": 1105
},
{
"epoch": 2.4415011037527594,
"grad_norm": 0.1281418800354004,
"learning_rate": 1.0192666617586072e-06,
"loss": 0.3657,
"step": 1106
},
{
"epoch": 2.443708609271523,
"grad_norm": 0.14360150694847107,
"learning_rate": 1.0115079749681628e-06,
"loss": 0.3598,
"step": 1107
},
{
"epoch": 2.445916114790287,
"grad_norm": 0.13790085911750793,
"learning_rate": 1.0037756063254555e-06,
"loss": 0.3612,
"step": 1108
},
{
"epoch": 2.4481236203090506,
"grad_norm": 0.1285860389471054,
"learning_rate": 9.960696068526826e-07,
"loss": 0.3615,
"step": 1109
},
{
"epoch": 2.4503311258278146,
"grad_norm": 0.1286444067955017,
"learning_rate": 9.883900273980408e-07,
"loss": 0.3587,
"step": 1110
},
{
"epoch": 2.4525386313465782,
"grad_norm": 0.12481298297643661,
"learning_rate": 9.807369186353965e-07,
"loss": 0.356,
"step": 1111
},
{
"epoch": 2.4547461368653423,
"grad_norm": 0.1254754513502121,
"learning_rate": 9.731103310639473e-07,
"loss": 0.3575,
"step": 1112
},
{
"epoch": 2.456953642384106,
"grad_norm": 0.12286875396966934,
"learning_rate": 9.655103150078892e-07,
"loss": 0.3608,
"step": 1113
},
{
"epoch": 2.45916114790287,
"grad_norm": 0.11990831047296524,
"learning_rate": 9.57936920616086e-07,
"loss": 0.3586,
"step": 1114
},
{
"epoch": 2.4613686534216335,
"grad_norm": 0.11828765273094177,
"learning_rate": 9.503901978617392e-07,
"loss": 0.3621,
"step": 1115
},
{
"epoch": 2.4635761589403975,
"grad_norm": 0.13093435764312744,
"learning_rate": 9.428701965420572e-07,
"loss": 0.361,
"step": 1116
},
{
"epoch": 2.465783664459161,
"grad_norm": 0.12574456632137299,
"learning_rate": 9.353769662779249e-07,
"loss": 0.3556,
"step": 1117
},
{
"epoch": 2.467991169977925,
"grad_norm": 0.136729896068573,
"learning_rate": 9.279105565135777e-07,
"loss": 0.3661,
"step": 1118
},
{
"epoch": 2.4701986754966887,
"grad_norm": 0.12174253910779953,
"learning_rate": 9.204710165162751e-07,
"loss": 0.3487,
"step": 1119
},
{
"epoch": 2.4724061810154527,
"grad_norm": 0.13417866826057434,
"learning_rate": 9.130583953759781e-07,
"loss": 0.358,
"step": 1120
},
{
"epoch": 2.4746136865342163,
"grad_norm": 0.14360179007053375,
"learning_rate": 9.056727420050227e-07,
"loss": 0.3628,
"step": 1121
},
{
"epoch": 2.47682119205298,
"grad_norm": 0.12731443345546722,
"learning_rate": 8.983141051377953e-07,
"loss": 0.3603,
"step": 1122
},
{
"epoch": 2.479028697571744,
"grad_norm": 0.13204246759414673,
"learning_rate": 8.909825333304134e-07,
"loss": 0.355,
"step": 1123
},
{
"epoch": 2.481236203090508,
"grad_norm": 0.1328221708536148,
"learning_rate": 8.836780749604096e-07,
"loss": 0.3541,
"step": 1124
},
{
"epoch": 2.4834437086092715,
"grad_norm": 0.13670921325683594,
"learning_rate": 8.764007782264022e-07,
"loss": 0.3649,
"step": 1125
},
{
"epoch": 2.485651214128035,
"grad_norm": 0.124913290143013,
"learning_rate": 8.691506911477848e-07,
"loss": 0.3474,
"step": 1126
},
{
"epoch": 2.487858719646799,
"grad_norm": 0.12466172128915787,
"learning_rate": 8.619278615644106e-07,
"loss": 0.3559,
"step": 1127
},
{
"epoch": 2.4900662251655628,
"grad_norm": 0.13907819986343384,
"learning_rate": 8.547323371362682e-07,
"loss": 0.3544,
"step": 1128
},
{
"epoch": 2.492273730684327,
"grad_norm": 0.13403339684009552,
"learning_rate": 8.475641653431782e-07,
"loss": 0.3621,
"step": 1129
},
{
"epoch": 2.4944812362030904,
"grad_norm": 0.13107538223266602,
"learning_rate": 8.404233934844707e-07,
"loss": 0.3629,
"step": 1130
},
{
"epoch": 2.4966887417218544,
"grad_norm": 0.12382876127958298,
"learning_rate": 8.333100686786766e-07,
"loss": 0.3425,
"step": 1131
},
{
"epoch": 2.498896247240618,
"grad_norm": 0.1340240240097046,
"learning_rate": 8.262242378632179e-07,
"loss": 0.3637,
"step": 1132
},
{
"epoch": 2.501103752759382,
"grad_norm": 0.12560687959194183,
"learning_rate": 8.191659477940972e-07,
"loss": 0.3694,
"step": 1133
},
{
"epoch": 2.5033112582781456,
"grad_norm": 0.13509927690029144,
"learning_rate": 8.121352450455899e-07,
"loss": 0.3588,
"step": 1134
},
{
"epoch": 2.5055187637969096,
"grad_norm": 0.13506671786308289,
"learning_rate": 8.051321760099334e-07,
"loss": 0.3657,
"step": 1135
},
{
"epoch": 2.507726269315673,
"grad_norm": 0.12824004888534546,
"learning_rate": 7.981567868970252e-07,
"loss": 0.3541,
"step": 1136
},
{
"epoch": 2.5099337748344372,
"grad_norm": 0.12654347717761993,
"learning_rate": 7.91209123734114e-07,
"loss": 0.3628,
"step": 1137
},
{
"epoch": 2.512141280353201,
"grad_norm": 0.13879314064979553,
"learning_rate": 7.84289232365501e-07,
"loss": 0.3631,
"step": 1138
},
{
"epoch": 2.514348785871965,
"grad_norm": 0.11704026162624359,
"learning_rate": 7.773971584522355e-07,
"loss": 0.3589,
"step": 1139
},
{
"epoch": 2.5165562913907285,
"grad_norm": 0.12439953535795212,
"learning_rate": 7.705329474718093e-07,
"loss": 0.3537,
"step": 1140
},
{
"epoch": 2.518763796909492,
"grad_norm": 0.13436546921730042,
"learning_rate": 7.636966447178601e-07,
"loss": 0.356,
"step": 1141
},
{
"epoch": 2.520971302428256,
"grad_norm": 0.12255270779132843,
"learning_rate": 7.568882952998762e-07,
"loss": 0.3568,
"step": 1142
},
{
"epoch": 2.52317880794702,
"grad_norm": 0.12233200669288635,
"learning_rate": 7.501079441428927e-07,
"loss": 0.3538,
"step": 1143
},
{
"epoch": 2.5253863134657837,
"grad_norm": 0.12473509460687637,
"learning_rate": 7.433556359871968e-07,
"loss": 0.3711,
"step": 1144
},
{
"epoch": 2.5275938189845473,
"grad_norm": 0.13673733174800873,
"learning_rate": 7.366314153880361e-07,
"loss": 0.3613,
"step": 1145
},
{
"epoch": 2.5298013245033113,
"grad_norm": 0.1345418095588684,
"learning_rate": 7.299353267153192e-07,
"loss": 0.3632,
"step": 1146
},
{
"epoch": 2.5320088300220753,
"grad_norm": 0.1333194524049759,
"learning_rate": 7.232674141533274e-07,
"loss": 0.3653,
"step": 1147
},
{
"epoch": 2.534216335540839,
"grad_norm": 0.13697415590286255,
"learning_rate": 7.166277217004214e-07,
"loss": 0.3571,
"step": 1148
},
{
"epoch": 2.5364238410596025,
"grad_norm": 0.13772854208946228,
"learning_rate": 7.100162931687476e-07,
"loss": 0.3737,
"step": 1149
},
{
"epoch": 2.5386313465783665,
"grad_norm": 0.1227809488773346,
"learning_rate": 7.034331721839566e-07,
"loss": 0.3622,
"step": 1150
},
{
"epoch": 2.54083885209713,
"grad_norm": 0.12854185700416565,
"learning_rate": 6.968784021849106e-07,
"loss": 0.3612,
"step": 1151
},
{
"epoch": 2.543046357615894,
"grad_norm": 0.11819867044687271,
"learning_rate": 6.903520264233954e-07,
"loss": 0.3625,
"step": 1152
},
{
"epoch": 2.5452538631346577,
"grad_norm": 0.1528642773628235,
"learning_rate": 6.838540879638367e-07,
"loss": 0.3644,
"step": 1153
},
{
"epoch": 2.5474613686534218,
"grad_norm": 0.1466141641139984,
"learning_rate": 6.773846296830167e-07,
"loss": 0.3681,
"step": 1154
},
{
"epoch": 2.5496688741721854,
"grad_norm": 0.13698391616344452,
"learning_rate": 6.7094369426979e-07,
"loss": 0.3569,
"step": 1155
},
{
"epoch": 2.5518763796909494,
"grad_norm": 0.1320303976535797,
"learning_rate": 6.645313242248042e-07,
"loss": 0.3531,
"step": 1156
},
{
"epoch": 2.554083885209713,
"grad_norm": 0.14429907500743866,
"learning_rate": 6.581475618602174e-07,
"loss": 0.3587,
"step": 1157
},
{
"epoch": 2.556291390728477,
"grad_norm": 0.12878787517547607,
"learning_rate": 6.517924492994182e-07,
"loss": 0.3648,
"step": 1158
},
{
"epoch": 2.5584988962472406,
"grad_norm": 0.13368941843509674,
"learning_rate": 6.454660284767477e-07,
"loss": 0.3569,
"step": 1159
},
{
"epoch": 2.560706401766004,
"grad_norm": 0.13522258400917053,
"learning_rate": 6.391683411372279e-07,
"loss": 0.3624,
"step": 1160
},
{
"epoch": 2.562913907284768,
"grad_norm": 0.1288643628358841,
"learning_rate": 6.328994288362783e-07,
"loss": 0.3571,
"step": 1161
},
{
"epoch": 2.5651214128035322,
"grad_norm": 0.13528789579868317,
"learning_rate": 6.266593329394471e-07,
"loss": 0.3614,
"step": 1162
},
{
"epoch": 2.567328918322296,
"grad_norm": 0.11771933734416962,
"learning_rate": 6.204480946221386e-07,
"loss": 0.3503,
"step": 1163
},
{
"epoch": 2.5695364238410594,
"grad_norm": 0.13160669803619385,
"learning_rate": 6.142657548693364e-07,
"loss": 0.3662,
"step": 1164
},
{
"epoch": 2.5717439293598234,
"grad_norm": 0.13713331520557404,
"learning_rate": 6.081123544753404e-07,
"loss": 0.3474,
"step": 1165
},
{
"epoch": 2.5739514348785875,
"grad_norm": 0.12216666340827942,
"learning_rate": 6.019879340434904e-07,
"loss": 0.3414,
"step": 1166
},
{
"epoch": 2.576158940397351,
"grad_norm": 0.12497063726186752,
"learning_rate": 5.958925339859001e-07,
"loss": 0.3501,
"step": 1167
},
{
"epoch": 2.5783664459161146,
"grad_norm": 0.1288682073354721,
"learning_rate": 5.898261945231965e-07,
"loss": 0.3518,
"step": 1168
},
{
"epoch": 2.5805739514348787,
"grad_norm": 0.13969705998897552,
"learning_rate": 5.837889556842469e-07,
"loss": 0.3595,
"step": 1169
},
{
"epoch": 2.5827814569536423,
"grad_norm": 0.13971775770187378,
"learning_rate": 5.777808573058969e-07,
"loss": 0.3635,
"step": 1170
},
{
"epoch": 2.5849889624724063,
"grad_norm": 0.12668545544147491,
"learning_rate": 5.718019390327084e-07,
"loss": 0.3519,
"step": 1171
},
{
"epoch": 2.58719646799117,
"grad_norm": 0.13675172626972198,
"learning_rate": 5.658522403166989e-07,
"loss": 0.363,
"step": 1172
},
{
"epoch": 2.589403973509934,
"grad_norm": 0.145385280251503,
"learning_rate": 5.599318004170778e-07,
"loss": 0.3561,
"step": 1173
},
{
"epoch": 2.5916114790286975,
"grad_norm": 0.14911803603172302,
"learning_rate": 5.540406583999925e-07,
"loss": 0.3578,
"step": 1174
},
{
"epoch": 2.5938189845474615,
"grad_norm": 0.12694287300109863,
"learning_rate": 5.481788531382671e-07,
"loss": 0.364,
"step": 1175
},
{
"epoch": 2.596026490066225,
"grad_norm": 0.13262110948562622,
"learning_rate": 5.423464233111448e-07,
"loss": 0.3718,
"step": 1176
},
{
"epoch": 2.598233995584989,
"grad_norm": 0.13582631945610046,
"learning_rate": 5.365434074040343e-07,
"loss": 0.3654,
"step": 1177
},
{
"epoch": 2.6004415011037527,
"grad_norm": 0.13755717873573303,
"learning_rate": 5.307698437082598e-07,
"loss": 0.3522,
"step": 1178
},
{
"epoch": 2.6026490066225163,
"grad_norm": 0.1361360102891922,
"learning_rate": 5.250257703207984e-07,
"loss": 0.3683,
"step": 1179
},
{
"epoch": 2.6048565121412803,
"grad_norm": 0.13162481784820557,
"learning_rate": 5.193112251440407e-07,
"loss": 0.3559,
"step": 1180
},
{
"epoch": 2.6070640176600444,
"grad_norm": 0.1336871236562729,
"learning_rate": 5.136262458855312e-07,
"loss": 0.3539,
"step": 1181
},
{
"epoch": 2.609271523178808,
"grad_norm": 0.12142772227525711,
"learning_rate": 5.079708700577229e-07,
"loss": 0.3633,
"step": 1182
},
{
"epoch": 2.6114790286975715,
"grad_norm": 0.12398800998926163,
"learning_rate": 5.023451349777331e-07,
"loss": 0.3666,
"step": 1183
},
{
"epoch": 2.6136865342163356,
"grad_norm": 0.1366734653711319,
"learning_rate": 4.967490777670903e-07,
"loss": 0.3552,
"step": 1184
},
{
"epoch": 2.6158940397350996,
"grad_norm": 0.13601034879684448,
"learning_rate": 4.911827353514947e-07,
"loss": 0.3606,
"step": 1185
},
{
"epoch": 2.618101545253863,
"grad_norm": 0.13659435510635376,
"learning_rate": 4.856461444605732e-07,
"loss": 0.356,
"step": 1186
},
{
"epoch": 2.6203090507726268,
"grad_norm": 0.13636133074760437,
"learning_rate": 4.801393416276368e-07,
"loss": 0.3591,
"step": 1187
},
{
"epoch": 2.622516556291391,
"grad_norm": 0.1251705437898636,
"learning_rate": 4.7466236318943816e-07,
"loss": 0.3557,
"step": 1188
},
{
"epoch": 2.6247240618101544,
"grad_norm": 0.14416959881782532,
"learning_rate": 4.692152452859333e-07,
"loss": 0.3587,
"step": 1189
},
{
"epoch": 2.6269315673289184,
"grad_norm": 0.15531152486801147,
"learning_rate": 4.637980238600437e-07,
"loss": 0.3494,
"step": 1190
},
{
"epoch": 2.629139072847682,
"grad_norm": 0.1295311003923416,
"learning_rate": 4.584107346574168e-07,
"loss": 0.3595,
"step": 1191
},
{
"epoch": 2.631346578366446,
"grad_norm": 0.13844533264636993,
"learning_rate": 4.530534132261932e-07,
"loss": 0.3589,
"step": 1192
},
{
"epoch": 2.6335540838852096,
"grad_norm": 0.13652655482292175,
"learning_rate": 4.477260949167711e-07,
"loss": 0.3492,
"step": 1193
},
{
"epoch": 2.6357615894039736,
"grad_norm": 0.14151211082935333,
"learning_rate": 4.4242881488157083e-07,
"loss": 0.3565,
"step": 1194
},
{
"epoch": 2.6379690949227372,
"grad_norm": 0.13526977598667145,
"learning_rate": 4.371616080748037e-07,
"loss": 0.355,
"step": 1195
},
{
"epoch": 2.6401766004415013,
"grad_norm": 0.13757188618183136,
"learning_rate": 4.319245092522456e-07,
"loss": 0.3547,
"step": 1196
},
{
"epoch": 2.642384105960265,
"grad_norm": 0.1314619481563568,
"learning_rate": 4.2671755297100047e-07,
"loss": 0.36,
"step": 1197
},
{
"epoch": 2.6445916114790284,
"grad_norm": 0.13309867680072784,
"learning_rate": 4.215407735892796e-07,
"loss": 0.3578,
"step": 1198
},
{
"epoch": 2.6467991169977925,
"grad_norm": 0.14237742125988007,
"learning_rate": 4.1639420526616845e-07,
"loss": 0.3567,
"step": 1199
},
{
"epoch": 2.6490066225165565,
"grad_norm": 0.13557758927345276,
"learning_rate": 4.1127788196140437e-07,
"loss": 0.3627,
"step": 1200
},
{
"epoch": 2.65121412803532,
"grad_norm": 0.14701923727989197,
"learning_rate": 4.061918374351559e-07,
"loss": 0.3518,
"step": 1201
},
{
"epoch": 2.6534216335540837,
"grad_norm": 0.1429191380739212,
"learning_rate": 4.0113610524779246e-07,
"loss": 0.3579,
"step": 1202
},
{
"epoch": 2.6556291390728477,
"grad_norm": 0.13261815905570984,
"learning_rate": 3.961107187596669e-07,
"loss": 0.3558,
"step": 1203
},
{
"epoch": 2.6578366445916117,
"grad_norm": 0.15215714275836945,
"learning_rate": 3.911157111308983e-07,
"loss": 0.3582,
"step": 1204
},
{
"epoch": 2.6600441501103753,
"grad_norm": 0.134977787733078,
"learning_rate": 3.861511153211489e-07,
"loss": 0.3484,
"step": 1205
},
{
"epoch": 2.662251655629139,
"grad_norm": 0.13735589385032654,
"learning_rate": 3.81216964089407e-07,
"loss": 0.3633,
"step": 1206
},
{
"epoch": 2.664459161147903,
"grad_norm": 0.13418880105018616,
"learning_rate": 3.763132899937721e-07,
"loss": 0.3646,
"step": 1207
},
{
"epoch": 2.6666666666666665,
"grad_norm": 0.13318879902362823,
"learning_rate": 3.7144012539123973e-07,
"loss": 0.3449,
"step": 1208
},
{
"epoch": 2.6688741721854305,
"grad_norm": 0.14231936633586884,
"learning_rate": 3.665975024374879e-07,
"loss": 0.3661,
"step": 1209
},
{
"epoch": 2.671081677704194,
"grad_norm": 0.13167575001716614,
"learning_rate": 3.617854530866671e-07,
"loss": 0.3627,
"step": 1210
},
{
"epoch": 2.673289183222958,
"grad_norm": 0.13379094004631042,
"learning_rate": 3.5700400909118306e-07,
"loss": 0.3585,
"step": 1211
},
{
"epoch": 2.6754966887417218,
"grad_norm": 0.1270161271095276,
"learning_rate": 3.522532020014946e-07,
"loss": 0.3571,
"step": 1212
},
{
"epoch": 2.677704194260486,
"grad_norm": 0.1238294392824173,
"learning_rate": 3.475330631659002e-07,
"loss": 0.3653,
"step": 1213
},
{
"epoch": 2.6799116997792494,
"grad_norm": 0.12466707080602646,
"learning_rate": 3.4284362373033566e-07,
"loss": 0.3535,
"step": 1214
},
{
"epoch": 2.6821192052980134,
"grad_norm": 0.12770843505859375,
"learning_rate": 3.3818491463816385e-07,
"loss": 0.3573,
"step": 1215
},
{
"epoch": 2.684326710816777,
"grad_norm": 0.1300242394208908,
"learning_rate": 3.335569666299748e-07,
"loss": 0.3566,
"step": 1216
},
{
"epoch": 2.6865342163355406,
"grad_norm": 0.1325581669807434,
"learning_rate": 3.289598102433794e-07,
"loss": 0.3575,
"step": 1217
},
{
"epoch": 2.6887417218543046,
"grad_norm": 0.1356075257062912,
"learning_rate": 3.2439347581280865e-07,
"loss": 0.3614,
"step": 1218
},
{
"epoch": 2.6909492273730686,
"grad_norm": 0.12728764116764069,
"learning_rate": 3.19857993469318e-07,
"loss": 0.3586,
"step": 1219
},
{
"epoch": 2.693156732891832,
"grad_norm": 0.1246216669678688,
"learning_rate": 3.1535339314038015e-07,
"loss": 0.3581,
"step": 1220
},
{
"epoch": 2.695364238410596,
"grad_norm": 0.1401343196630478,
"learning_rate": 3.108797045496942e-07,
"loss": 0.3611,
"step": 1221
},
{
"epoch": 2.69757174392936,
"grad_norm": 0.1308797299861908,
"learning_rate": 3.0643695721698783e-07,
"loss": 0.3664,
"step": 1222
},
{
"epoch": 2.699779249448124,
"grad_norm": 0.12849809229373932,
"learning_rate": 3.0202518045782337e-07,
"loss": 0.3605,
"step": 1223
},
{
"epoch": 2.7019867549668874,
"grad_norm": 0.128408744931221,
"learning_rate": 2.9764440338340083e-07,
"loss": 0.3583,
"step": 1224
},
{
"epoch": 2.704194260485651,
"grad_norm": 0.1296570599079132,
"learning_rate": 2.9329465490036844e-07,
"loss": 0.3528,
"step": 1225
},
{
"epoch": 2.706401766004415,
"grad_norm": 0.14551834762096405,
"learning_rate": 2.8897596371063153e-07,
"loss": 0.3497,
"step": 1226
},
{
"epoch": 2.7086092715231787,
"grad_norm": 0.13115178048610687,
"learning_rate": 2.846883583111648e-07,
"loss": 0.3594,
"step": 1227
},
{
"epoch": 2.7108167770419427,
"grad_norm": 0.12846940755844116,
"learning_rate": 2.804318669938233e-07,
"loss": 0.3583,
"step": 1228
},
{
"epoch": 2.7130242825607063,
"grad_norm": 0.1380581110715866,
"learning_rate": 2.762065178451517e-07,
"loss": 0.3585,
"step": 1229
},
{
"epoch": 2.7152317880794703,
"grad_norm": 0.12574802339076996,
"learning_rate": 2.7201233874620534e-07,
"loss": 0.3598,
"step": 1230
},
{
"epoch": 2.717439293598234,
"grad_norm": 0.13319121301174164,
"learning_rate": 2.678493573723612e-07,
"loss": 0.3669,
"step": 1231
},
{
"epoch": 2.719646799116998,
"grad_norm": 0.12967029213905334,
"learning_rate": 2.6371760119314026e-07,
"loss": 0.3637,
"step": 1232
},
{
"epoch": 2.7218543046357615,
"grad_norm": 0.13314829766750336,
"learning_rate": 2.596170974720202e-07,
"loss": 0.3505,
"step": 1233
},
{
"epoch": 2.7240618101545255,
"grad_norm": 0.13859589397907257,
"learning_rate": 2.5554787326626194e-07,
"loss": 0.3508,
"step": 1234
},
{
"epoch": 2.726269315673289,
"grad_norm": 0.13054342567920685,
"learning_rate": 2.515099554267247e-07,
"loss": 0.3459,
"step": 1235
},
{
"epoch": 2.7284768211920527,
"grad_norm": 0.13782578706741333,
"learning_rate": 2.4750337059769425e-07,
"loss": 0.3653,
"step": 1236
},
{
"epoch": 2.7306843267108167,
"grad_norm": 0.1294957846403122,
"learning_rate": 2.4352814521670375e-07,
"loss": 0.3614,
"step": 1237
},
{
"epoch": 2.7328918322295808,
"grad_norm": 0.1325952708721161,
"learning_rate": 2.3958430551436095e-07,
"loss": 0.3638,
"step": 1238
},
{
"epoch": 2.7350993377483444,
"grad_norm": 0.12079501897096634,
"learning_rate": 2.3567187751417475e-07,
"loss": 0.3605,
"step": 1239
},
{
"epoch": 2.737306843267108,
"grad_norm": 0.13967706263065338,
"learning_rate": 2.3179088703238096e-07,
"loss": 0.3721,
"step": 1240
},
{
"epoch": 2.739514348785872,
"grad_norm": 0.133636474609375,
"learning_rate": 2.2794135967777908e-07,
"loss": 0.3655,
"step": 1241
},
{
"epoch": 2.741721854304636,
"grad_norm": 0.13323010504245758,
"learning_rate": 2.2412332085155364e-07,
"loss": 0.3531,
"step": 1242
},
{
"epoch": 2.7439293598233996,
"grad_norm": 0.13812537491321564,
"learning_rate": 2.2033679574711365e-07,
"loss": 0.3453,
"step": 1243
},
{
"epoch": 2.746136865342163,
"grad_norm": 0.12806718051433563,
"learning_rate": 2.1658180934992333e-07,
"loss": 0.3632,
"step": 1244
},
{
"epoch": 2.748344370860927,
"grad_norm": 0.13933822512626648,
"learning_rate": 2.1285838643733958e-07,
"loss": 0.3595,
"step": 1245
},
{
"epoch": 2.750551876379691,
"grad_norm": 0.12600216269493103,
"learning_rate": 2.0916655157844634e-07,
"loss": 0.352,
"step": 1246
},
{
"epoch": 2.752759381898455,
"grad_norm": 0.12791283428668976,
"learning_rate": 2.0550632913389213e-07,
"loss": 0.3556,
"step": 1247
},
{
"epoch": 2.7549668874172184,
"grad_norm": 0.13090188801288605,
"learning_rate": 2.0187774325573174e-07,
"loss": 0.3632,
"step": 1248
},
{
"epoch": 2.7571743929359824,
"grad_norm": 0.1253674477338791,
"learning_rate": 1.9828081788726307e-07,
"loss": 0.3499,
"step": 1249
},
{
"epoch": 2.759381898454746,
"grad_norm": 0.13384896516799927,
"learning_rate": 1.9471557676287501e-07,
"loss": 0.3556,
"step": 1250
},
{
"epoch": 2.76158940397351,
"grad_norm": 0.13008341193199158,
"learning_rate": 1.9118204340788426e-07,
"loss": 0.3521,
"step": 1251
},
{
"epoch": 2.7637969094922736,
"grad_norm": 0.14128050208091736,
"learning_rate": 1.8768024113838546e-07,
"loss": 0.3629,
"step": 1252
},
{
"epoch": 2.7660044150110377,
"grad_norm": 0.13070915639400482,
"learning_rate": 1.8421019306109288e-07,
"loss": 0.3576,
"step": 1253
},
{
"epoch": 2.7682119205298013,
"grad_norm": 0.13155661523342133,
"learning_rate": 1.8077192207319072e-07,
"loss": 0.3614,
"step": 1254
},
{
"epoch": 2.7704194260485653,
"grad_norm": 0.14038477838039398,
"learning_rate": 1.773654508621825e-07,
"loss": 0.354,
"step": 1255
},
{
"epoch": 2.772626931567329,
"grad_norm": 0.13292455673217773,
"learning_rate": 1.7399080190573903e-07,
"loss": 0.3635,
"step": 1256
},
{
"epoch": 2.774834437086093,
"grad_norm": 0.1393938958644867,
"learning_rate": 1.7064799747155248e-07,
"loss": 0.3502,
"step": 1257
},
{
"epoch": 2.7770419426048565,
"grad_norm": 0.13566245138645172,
"learning_rate": 1.6733705961718694e-07,
"loss": 0.353,
"step": 1258
},
{
"epoch": 2.77924944812362,
"grad_norm": 0.1313687562942505,
"learning_rate": 1.640580101899353e-07,
"loss": 0.3519,
"step": 1259
},
{
"epoch": 2.781456953642384,
"grad_norm": 0.13328541815280914,
"learning_rate": 1.6081087082667314e-07,
"loss": 0.3513,
"step": 1260
},
{
"epoch": 2.783664459161148,
"grad_norm": 0.13640496134757996,
"learning_rate": 1.575956629537173e-07,
"loss": 0.3559,
"step": 1261
},
{
"epoch": 2.7858719646799117,
"grad_norm": 0.13169561326503754,
"learning_rate": 1.5441240778668321e-07,
"loss": 0.3589,
"step": 1262
},
{
"epoch": 2.7880794701986753,
"grad_norm": 0.1369987428188324,
"learning_rate": 1.5126112633034761e-07,
"loss": 0.368,
"step": 1263
},
{
"epoch": 2.7902869757174393,
"grad_norm": 0.12467605620622635,
"learning_rate": 1.4814183937850668e-07,
"loss": 0.3563,
"step": 1264
},
{
"epoch": 2.7924944812362034,
"grad_norm": 0.1307917833328247,
"learning_rate": 1.4505456751383985e-07,
"loss": 0.3516,
"step": 1265
},
{
"epoch": 2.794701986754967,
"grad_norm": 0.14058281481266022,
"learning_rate": 1.4199933110777553e-07,
"loss": 0.3546,
"step": 1266
},
{
"epoch": 2.7969094922737305,
"grad_norm": 0.12925602495670319,
"learning_rate": 1.389761503203535e-07,
"loss": 0.3615,
"step": 1267
},
{
"epoch": 2.7991169977924946,
"grad_norm": 0.1409684419631958,
"learning_rate": 1.3598504510009602e-07,
"loss": 0.3611,
"step": 1268
},
{
"epoch": 2.801324503311258,
"grad_norm": 0.13529111444950104,
"learning_rate": 1.3302603518387358e-07,
"loss": 0.3556,
"step": 1269
},
{
"epoch": 2.803532008830022,
"grad_norm": 0.12611360847949982,
"learning_rate": 1.3009914009677493e-07,
"loss": 0.3695,
"step": 1270
},
{
"epoch": 2.8057395143487858,
"grad_norm": 0.1310908943414688,
"learning_rate": 1.272043791519778e-07,
"loss": 0.3571,
"step": 1271
},
{
"epoch": 2.80794701986755,
"grad_norm": 0.13993841409683228,
"learning_rate": 1.2434177145062177e-07,
"loss": 0.3572,
"step": 1272
},
{
"epoch": 2.8101545253863134,
"grad_norm": 0.1199953481554985,
"learning_rate": 1.215113358816844e-07,
"loss": 0.3523,
"step": 1273
},
{
"epoch": 2.8123620309050774,
"grad_norm": 0.12030813843011856,
"learning_rate": 1.187130911218537e-07,
"loss": 0.3533,
"step": 1274
},
{
"epoch": 2.814569536423841,
"grad_norm": 0.1288137286901474,
"learning_rate": 1.1594705563540642e-07,
"loss": 0.3587,
"step": 1275
},
{
"epoch": 2.816777041942605,
"grad_norm": 0.1391681730747223,
"learning_rate": 1.1321324767408382e-07,
"loss": 0.3592,
"step": 1276
},
{
"epoch": 2.8189845474613686,
"grad_norm": 0.12392991036176682,
"learning_rate": 1.1051168527697665e-07,
"loss": 0.3605,
"step": 1277
},
{
"epoch": 2.821192052980132,
"grad_norm": 0.12354668974876404,
"learning_rate": 1.0784238627039977e-07,
"loss": 0.3593,
"step": 1278
},
{
"epoch": 2.8233995584988962,
"grad_norm": 0.13385632634162903,
"learning_rate": 1.0520536826777783e-07,
"loss": 0.3573,
"step": 1279
},
{
"epoch": 2.8256070640176603,
"grad_norm": 0.12551288306713104,
"learning_rate": 1.0260064866952968e-07,
"loss": 0.3563,
"step": 1280
},
{
"epoch": 2.827814569536424,
"grad_norm": 0.119255430996418,
"learning_rate": 1.0002824466295191e-07,
"loss": 0.3553,
"step": 1281
},
{
"epoch": 2.8300220750551874,
"grad_norm": 0.12346068769693375,
"learning_rate": 9.748817322210558e-08,
"loss": 0.3547,
"step": 1282
},
{
"epoch": 2.8322295805739515,
"grad_norm": 0.12130031734704971,
"learning_rate": 9.498045110770571e-08,
"loss": 0.3476,
"step": 1283
},
{
"epoch": 2.8344370860927155,
"grad_norm": 0.13486534357070923,
"learning_rate": 9.250509486700809e-08,
"loss": 0.3652,
"step": 1284
},
{
"epoch": 2.836644591611479,
"grad_norm": 0.12220917642116547,
"learning_rate": 9.006212083370213e-08,
"loss": 0.3548,
"step": 1285
},
{
"epoch": 2.8388520971302427,
"grad_norm": 0.12486063688993454,
"learning_rate": 8.765154512780428e-08,
"loss": 0.3502,
"step": 1286
},
{
"epoch": 2.8410596026490067,
"grad_norm": 0.11943277716636658,
"learning_rate": 8.527338365554749e-08,
"loss": 0.3462,
"step": 1287
},
{
"epoch": 2.8432671081677703,
"grad_norm": 0.13910432159900665,
"learning_rate": 8.292765210928089e-08,
"loss": 0.3585,
"step": 1288
},
{
"epoch": 2.8454746136865343,
"grad_norm": 0.12948773801326752,
"learning_rate": 8.061436596736139e-08,
"loss": 0.3589,
"step": 1289
},
{
"epoch": 2.847682119205298,
"grad_norm": 0.13374410569667816,
"learning_rate": 7.833354049405717e-08,
"loss": 0.3505,
"step": 1290
},
{
"epoch": 2.849889624724062,
"grad_norm": 0.1347481608390808,
"learning_rate": 7.608519073944165e-08,
"loss": 0.3687,
"step": 1291
},
{
"epoch": 2.8520971302428255,
"grad_norm": 0.13521930575370789,
"learning_rate": 7.386933153929798e-08,
"loss": 0.366,
"step": 1292
},
{
"epoch": 2.8543046357615895,
"grad_norm": 0.13502496480941772,
"learning_rate": 7.168597751501972e-08,
"loss": 0.366,
"step": 1293
},
{
"epoch": 2.856512141280353,
"grad_norm": 0.1404918134212494,
"learning_rate": 6.953514307351306e-08,
"loss": 0.357,
"step": 1294
},
{
"epoch": 2.858719646799117,
"grad_norm": 0.12767677009105682,
"learning_rate": 6.741684240710477e-08,
"loss": 0.3517,
"step": 1295
},
{
"epoch": 2.8609271523178808,
"grad_norm": 0.13126912713050842,
"learning_rate": 6.533108949344558e-08,
"loss": 0.3564,
"step": 1296
},
{
"epoch": 2.8631346578366443,
"grad_norm": 0.13056641817092896,
"learning_rate": 6.32778980954185e-08,
"loss": 0.3587,
"step": 1297
},
{
"epoch": 2.8653421633554084,
"grad_norm": 0.13267594575881958,
"learning_rate": 6.125728176105129e-08,
"loss": 0.3686,
"step": 1298
},
{
"epoch": 2.8675496688741724,
"grad_norm": 0.12520645558834076,
"learning_rate": 5.9269253823421855e-08,
"loss": 0.3703,
"step": 1299
},
{
"epoch": 2.869757174392936,
"grad_norm": 0.1308923214673996,
"learning_rate": 5.731382740057523e-08,
"loss": 0.3549,
"step": 1300
},
{
"epoch": 2.8719646799116996,
"grad_norm": 0.13203322887420654,
"learning_rate": 5.5391015395432346e-08,
"loss": 0.353,
"step": 1301
},
{
"epoch": 2.8741721854304636,
"grad_norm": 0.13061153888702393,
"learning_rate": 5.350083049570853e-08,
"loss": 0.3556,
"step": 1302
},
{
"epoch": 2.8763796909492276,
"grad_norm": 0.13314051926136017,
"learning_rate": 5.164328517382744e-08,
"loss": 0.3578,
"step": 1303
},
{
"epoch": 2.878587196467991,
"grad_norm": 0.13481754064559937,
"learning_rate": 4.9818391686842214e-08,
"loss": 0.362,
"step": 1304
},
{
"epoch": 2.880794701986755,
"grad_norm": 0.12934917211532593,
"learning_rate": 4.802616207634947e-08,
"loss": 0.3489,
"step": 1305
},
{
"epoch": 2.883002207505519,
"grad_norm": 0.1276032030582428,
"learning_rate": 4.626660816841433e-08,
"loss": 0.3554,
"step": 1306
},
{
"epoch": 2.8852097130242824,
"grad_norm": 0.12852588295936584,
"learning_rate": 4.4539741573489395e-08,
"loss": 0.3519,
"step": 1307
},
{
"epoch": 2.8874172185430464,
"grad_norm": 0.1251831203699112,
"learning_rate": 4.2845573686339235e-08,
"loss": 0.3586,
"step": 1308
},
{
"epoch": 2.88962472406181,
"grad_norm": 0.12635353207588196,
"learning_rate": 4.118411568596714e-08,
"loss": 0.3521,
"step": 1309
},
{
"epoch": 2.891832229580574,
"grad_norm": 0.13549165427684784,
"learning_rate": 3.9555378535537925e-08,
"loss": 0.3573,
"step": 1310
},
{
"epoch": 2.8940397350993377,
"grad_norm": 0.12129613012075424,
"learning_rate": 3.795937298230801e-08,
"loss": 0.3644,
"step": 1311
},
{
"epoch": 2.8962472406181017,
"grad_norm": 0.12583400309085846,
"learning_rate": 3.639610955755213e-08,
"loss": 0.3628,
"step": 1312
},
{
"epoch": 2.8984547461368653,
"grad_norm": 0.11590792238712311,
"learning_rate": 3.486559857649785e-08,
"loss": 0.3541,
"step": 1313
},
{
"epoch": 2.9006622516556293,
"grad_norm": 0.1254376322031021,
"learning_rate": 3.336785013825339e-08,
"loss": 0.3591,
"step": 1314
},
{
"epoch": 2.902869757174393,
"grad_norm": 0.12448057532310486,
"learning_rate": 3.190287412574267e-08,
"loss": 0.3685,
"step": 1315
},
{
"epoch": 2.9050772626931565,
"grad_norm": 0.12795038521289825,
"learning_rate": 3.047068020564037e-08,
"loss": 0.3661,
"step": 1316
},
{
"epoch": 2.9072847682119205,
"grad_norm": 0.1275874376296997,
"learning_rate": 2.9071277828308654e-08,
"loss": 0.3547,
"step": 1317
},
{
"epoch": 2.9094922737306845,
"grad_norm": 0.13140250742435455,
"learning_rate": 2.7704676227732764e-08,
"loss": 0.3476,
"step": 1318
},
{
"epoch": 2.911699779249448,
"grad_norm": 0.13054263591766357,
"learning_rate": 2.637088442146163e-08,
"loss": 0.3555,
"step": 1319
},
{
"epoch": 2.9139072847682117,
"grad_norm": 0.14265048503875732,
"learning_rate": 2.506991121054847e-08,
"loss": 0.3552,
"step": 1320
},
{
"epoch": 2.9161147902869757,
"grad_norm": 0.12008768320083618,
"learning_rate": 2.380176517949251e-08,
"loss": 0.3555,
"step": 1321
},
{
"epoch": 2.9183222958057398,
"grad_norm": 0.13723327219486237,
"learning_rate": 2.256645469618124e-08,
"loss": 0.3578,
"step": 1322
},
{
"epoch": 2.9205298013245033,
"grad_norm": 0.13760052621364594,
"learning_rate": 2.136398791183658e-08,
"loss": 0.3543,
"step": 1323
},
{
"epoch": 2.922737306843267,
"grad_norm": 0.129390150308609,
"learning_rate": 2.0194372760961034e-08,
"loss": 0.362,
"step": 1324
},
{
"epoch": 2.924944812362031,
"grad_norm": 0.15059252083301544,
"learning_rate": 1.905761696128494e-08,
"loss": 0.3611,
"step": 1325
},
{
"epoch": 2.9271523178807946,
"grad_norm": 0.13671807944774628,
"learning_rate": 1.795372801371431e-08,
"loss": 0.3601,
"step": 1326
},
{
"epoch": 2.9293598233995586,
"grad_norm": 0.138703852891922,
"learning_rate": 1.6882713202283076e-08,
"loss": 0.3525,
"step": 1327
},
{
"epoch": 2.931567328918322,
"grad_norm": 0.12975488603115082,
"learning_rate": 1.5844579594105904e-08,
"loss": 0.3636,
"step": 1328
},
{
"epoch": 2.933774834437086,
"grad_norm": 0.15756423771381378,
"learning_rate": 1.48393340393288e-08,
"loss": 0.3586,
"step": 1329
},
{
"epoch": 2.93598233995585,
"grad_norm": 0.13145272433757782,
"learning_rate": 1.3866983171084703e-08,
"loss": 0.3563,
"step": 1330
},
{
"epoch": 2.938189845474614,
"grad_norm": 0.15646247565746307,
"learning_rate": 1.292753340545183e-08,
"loss": 0.3567,
"step": 1331
},
{
"epoch": 2.9403973509933774,
"grad_norm": 0.1439492553472519,
"learning_rate": 1.2020990941408739e-08,
"loss": 0.3678,
"step": 1332
},
{
"epoch": 2.9426048565121414,
"grad_norm": 0.1377602368593216,
"learning_rate": 1.1147361760794895e-08,
"loss": 0.3547,
"step": 1333
},
{
"epoch": 2.944812362030905,
"grad_norm": 0.14668287336826324,
"learning_rate": 1.0306651628270715e-08,
"loss": 0.3672,
"step": 1334
},
{
"epoch": 2.9470198675496686,
"grad_norm": 0.12802091240882874,
"learning_rate": 9.498866091278702e-09,
"loss": 0.3596,
"step": 1335
},
{
"epoch": 2.9492273730684326,
"grad_norm": 0.15803688764572144,
"learning_rate": 8.724010480010147e-09,
"loss": 0.3586,
"step": 1336
},
{
"epoch": 2.9514348785871967,
"grad_norm": 0.13727080821990967,
"learning_rate": 7.982089907364598e-09,
"loss": 0.3515,
"step": 1337
},
{
"epoch": 2.9536423841059603,
"grad_norm": 0.14836713671684265,
"learning_rate": 7.273109268920997e-09,
"loss": 0.3572,
"step": 1338
},
{
"epoch": 2.955849889624724,
"grad_norm": 0.15154053270816803,
"learning_rate": 6.597073242902707e-09,
"loss": 0.3632,
"step": 1339
},
{
"epoch": 2.958057395143488,
"grad_norm": 0.14268819987773895,
"learning_rate": 5.95398629014754e-09,
"loss": 0.3737,
"step": 1340
},
{
"epoch": 2.960264900662252,
"grad_norm": 0.13592885434627533,
"learning_rate": 5.3438526540777745e-09,
"loss": 0.3647,
"step": 1341
},
{
"epoch": 2.9624724061810155,
"grad_norm": 0.15282437205314636,
"learning_rate": 4.766676360674072e-09,
"loss": 0.3666,
"step": 1342
},
{
"epoch": 2.964679911699779,
"grad_norm": 0.14543361961841583,
"learning_rate": 4.22246121844494e-09,
"loss": 0.3576,
"step": 1343
},
{
"epoch": 2.966887417218543,
"grad_norm": 0.1356423795223236,
"learning_rate": 3.7112108184061966e-09,
"loss": 0.3646,
"step": 1344
},
{
"epoch": 2.9690949227373067,
"grad_norm": 0.13516850769519806,
"learning_rate": 3.2329285340537696e-09,
"loss": 0.3577,
"step": 1345
},
{
"epoch": 2.9713024282560707,
"grad_norm": 0.12816190719604492,
"learning_rate": 2.7876175213431557e-09,
"loss": 0.3568,
"step": 1346
},
{
"epoch": 2.9735099337748343,
"grad_norm": 0.13490551710128784,
"learning_rate": 2.375280718668882e-09,
"loss": 0.3548,
"step": 1347
},
{
"epoch": 2.9757174392935983,
"grad_norm": 0.14126001298427582,
"learning_rate": 1.995920846843968e-09,
"loss": 0.3666,
"step": 1348
},
{
"epoch": 2.977924944812362,
"grad_norm": 0.12279074639081955,
"learning_rate": 1.649540409081607e-09,
"loss": 0.3562,
"step": 1349
},
{
"epoch": 2.980132450331126,
"grad_norm": 0.12752176821231842,
"learning_rate": 1.3361416909812852e-09,
"loss": 0.365,
"step": 1350
},
{
"epoch": 2.9823399558498895,
"grad_norm": 0.13065557181835175,
"learning_rate": 1.055726760510467e-09,
"loss": 0.3615,
"step": 1351
},
{
"epoch": 2.9845474613686536,
"grad_norm": 0.11860226094722748,
"learning_rate": 8.082974679929357e-10,
"loss": 0.3564,
"step": 1352
},
{
"epoch": 2.986754966887417,
"grad_norm": 0.1291607767343521,
"learning_rate": 5.938554460965807e-10,
"loss": 0.359,
"step": 1353
},
{
"epoch": 2.988962472406181,
"grad_norm": 0.146859273314476,
"learning_rate": 4.12402109820631e-10,
"loss": 0.3614,
"step": 1354
},
{
"epoch": 2.9911699779249448,
"grad_norm": 0.12649129331111908,
"learning_rate": 2.639386564889934e-10,
"loss": 0.3566,
"step": 1355
},
{
"epoch": 2.993377483443709,
"grad_norm": 0.14126034080982208,
"learning_rate": 1.4846606574137058e-10,
"loss": 0.3608,
"step": 1356
},
{
"epoch": 2.9955849889624724,
"grad_norm": 0.13480675220489502,
"learning_rate": 6.59850995254896e-11,
"loss": 0.3555,
"step": 1357
},
{
"epoch": 2.997792494481236,
"grad_norm": 0.14647357165813446,
"learning_rate": 1.649630209432651e-11,
"loss": 0.3643,
"step": 1358
},
{
"epoch": 3.0,
"grad_norm": 0.13931505382061005,
"learning_rate": 0.0,
"loss": 0.3486,
"step": 1359
},
{
"epoch": 3.0,
"step": 1359,
"total_flos": 2.7571726585954304e+16,
"train_loss": 0.13500037013037053,
"train_runtime": 42654.1727,
"train_samples_per_second": 12.208,
"train_steps_per_second": 0.032
}
],
"logging_steps": 1,
"max_steps": 1359,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 50,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 2.7571726585954304e+16,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}