EtashGuha's picture
Upload folder using huggingface_hub
61f3858 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.993684210526316,
"eval_steps": 500,
"global_step": 948,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.003157894736842105,
"grad_norm": 6.864107608795166,
"learning_rate": 1.0526315789473685e-07,
"loss": 1.2166,
"step": 1
},
{
"epoch": 0.00631578947368421,
"grad_norm": 7.149033546447754,
"learning_rate": 2.105263157894737e-07,
"loss": 1.2292,
"step": 2
},
{
"epoch": 0.009473684210526316,
"grad_norm": 6.731090545654297,
"learning_rate": 3.1578947368421055e-07,
"loss": 1.1835,
"step": 3
},
{
"epoch": 0.01263157894736842,
"grad_norm": 7.016407012939453,
"learning_rate": 4.210526315789474e-07,
"loss": 1.2222,
"step": 4
},
{
"epoch": 0.015789473684210527,
"grad_norm": 6.585211753845215,
"learning_rate": 5.263157894736843e-07,
"loss": 1.2031,
"step": 5
},
{
"epoch": 0.018947368421052633,
"grad_norm": 6.720523834228516,
"learning_rate": 6.315789473684211e-07,
"loss": 1.1972,
"step": 6
},
{
"epoch": 0.022105263157894735,
"grad_norm": 6.992764472961426,
"learning_rate": 7.368421052631579e-07,
"loss": 1.1867,
"step": 7
},
{
"epoch": 0.02526315789473684,
"grad_norm": 6.5448174476623535,
"learning_rate": 8.421052631578948e-07,
"loss": 1.2113,
"step": 8
},
{
"epoch": 0.028421052631578948,
"grad_norm": 6.110010147094727,
"learning_rate": 9.473684210526317e-07,
"loss": 1.1666,
"step": 9
},
{
"epoch": 0.031578947368421054,
"grad_norm": 6.425546169281006,
"learning_rate": 1.0526315789473685e-06,
"loss": 1.1722,
"step": 10
},
{
"epoch": 0.034736842105263156,
"grad_norm": 4.949359893798828,
"learning_rate": 1.1578947368421053e-06,
"loss": 1.1464,
"step": 11
},
{
"epoch": 0.037894736842105266,
"grad_norm": 4.857133865356445,
"learning_rate": 1.2631578947368422e-06,
"loss": 1.1495,
"step": 12
},
{
"epoch": 0.04105263157894737,
"grad_norm": 4.663426876068115,
"learning_rate": 1.3684210526315791e-06,
"loss": 1.1108,
"step": 13
},
{
"epoch": 0.04421052631578947,
"grad_norm": 2.899270534515381,
"learning_rate": 1.4736842105263159e-06,
"loss": 1.0779,
"step": 14
},
{
"epoch": 0.04736842105263158,
"grad_norm": 2.799199104309082,
"learning_rate": 1.5789473684210526e-06,
"loss": 1.0596,
"step": 15
},
{
"epoch": 0.05052631578947368,
"grad_norm": 2.7462809085845947,
"learning_rate": 1.6842105263157895e-06,
"loss": 1.0846,
"step": 16
},
{
"epoch": 0.05368421052631579,
"grad_norm": 2.6440353393554688,
"learning_rate": 1.7894736842105265e-06,
"loss": 1.0408,
"step": 17
},
{
"epoch": 0.056842105263157895,
"grad_norm": 2.496100664138794,
"learning_rate": 1.8947368421052634e-06,
"loss": 1.0447,
"step": 18
},
{
"epoch": 0.06,
"grad_norm": 2.5369954109191895,
"learning_rate": 2.0000000000000003e-06,
"loss": 1.0338,
"step": 19
},
{
"epoch": 0.06315789473684211,
"grad_norm": 3.36142635345459,
"learning_rate": 2.105263157894737e-06,
"loss": 1.0461,
"step": 20
},
{
"epoch": 0.06631578947368422,
"grad_norm": 3.2484495639801025,
"learning_rate": 2.2105263157894738e-06,
"loss": 1.0202,
"step": 21
},
{
"epoch": 0.06947368421052631,
"grad_norm": 3.0597786903381348,
"learning_rate": 2.3157894736842105e-06,
"loss": 1.0088,
"step": 22
},
{
"epoch": 0.07263157894736842,
"grad_norm": 2.594919204711914,
"learning_rate": 2.4210526315789477e-06,
"loss": 0.9878,
"step": 23
},
{
"epoch": 0.07578947368421053,
"grad_norm": 2.3219525814056396,
"learning_rate": 2.5263157894736844e-06,
"loss": 1.0149,
"step": 24
},
{
"epoch": 0.07894736842105263,
"grad_norm": 1.812867283821106,
"learning_rate": 2.631578947368421e-06,
"loss": 0.9818,
"step": 25
},
{
"epoch": 0.08210526315789474,
"grad_norm": 1.581230640411377,
"learning_rate": 2.7368421052631583e-06,
"loss": 0.9757,
"step": 26
},
{
"epoch": 0.08526315789473685,
"grad_norm": 1.8027225732803345,
"learning_rate": 2.842105263157895e-06,
"loss": 0.9676,
"step": 27
},
{
"epoch": 0.08842105263157894,
"grad_norm": 1.7598600387573242,
"learning_rate": 2.9473684210526317e-06,
"loss": 0.9181,
"step": 28
},
{
"epoch": 0.09157894736842105,
"grad_norm": 1.7967835664749146,
"learning_rate": 3.052631578947369e-06,
"loss": 0.9336,
"step": 29
},
{
"epoch": 0.09473684210526316,
"grad_norm": 1.3898029327392578,
"learning_rate": 3.157894736842105e-06,
"loss": 0.9156,
"step": 30
},
{
"epoch": 0.09789473684210526,
"grad_norm": 1.2304682731628418,
"learning_rate": 3.2631578947368423e-06,
"loss": 0.936,
"step": 31
},
{
"epoch": 0.10105263157894737,
"grad_norm": 1.1240628957748413,
"learning_rate": 3.368421052631579e-06,
"loss": 0.8984,
"step": 32
},
{
"epoch": 0.10421052631578948,
"grad_norm": 1.1114623546600342,
"learning_rate": 3.473684210526316e-06,
"loss": 0.9032,
"step": 33
},
{
"epoch": 0.10736842105263159,
"grad_norm": 1.2093194723129272,
"learning_rate": 3.578947368421053e-06,
"loss": 0.9466,
"step": 34
},
{
"epoch": 0.11052631578947368,
"grad_norm": 1.1614065170288086,
"learning_rate": 3.6842105263157896e-06,
"loss": 0.8888,
"step": 35
},
{
"epoch": 0.11368421052631579,
"grad_norm": 0.9620938897132874,
"learning_rate": 3.789473684210527e-06,
"loss": 0.8985,
"step": 36
},
{
"epoch": 0.1168421052631579,
"grad_norm": 0.8165697455406189,
"learning_rate": 3.894736842105263e-06,
"loss": 0.8922,
"step": 37
},
{
"epoch": 0.12,
"grad_norm": 0.9624667763710022,
"learning_rate": 4.000000000000001e-06,
"loss": 0.9059,
"step": 38
},
{
"epoch": 0.1231578947368421,
"grad_norm": 0.9526475667953491,
"learning_rate": 4.105263157894737e-06,
"loss": 0.8697,
"step": 39
},
{
"epoch": 0.12631578947368421,
"grad_norm": 0.9052897691726685,
"learning_rate": 4.210526315789474e-06,
"loss": 0.8752,
"step": 40
},
{
"epoch": 0.12947368421052632,
"grad_norm": 0.8724302053451538,
"learning_rate": 4.315789473684211e-06,
"loss": 0.8745,
"step": 41
},
{
"epoch": 0.13263157894736843,
"grad_norm": 0.8744531273841858,
"learning_rate": 4.4210526315789476e-06,
"loss": 0.833,
"step": 42
},
{
"epoch": 0.13578947368421052,
"grad_norm": 0.8191508650779724,
"learning_rate": 4.526315789473685e-06,
"loss": 0.8681,
"step": 43
},
{
"epoch": 0.13894736842105262,
"grad_norm": 0.8925843238830566,
"learning_rate": 4.631578947368421e-06,
"loss": 0.8522,
"step": 44
},
{
"epoch": 0.14210526315789473,
"grad_norm": 0.89957195520401,
"learning_rate": 4.736842105263158e-06,
"loss": 0.8732,
"step": 45
},
{
"epoch": 0.14526315789473684,
"grad_norm": 0.8851773738861084,
"learning_rate": 4.842105263157895e-06,
"loss": 0.8991,
"step": 46
},
{
"epoch": 0.14842105263157895,
"grad_norm": 0.7817597985267639,
"learning_rate": 4.947368421052632e-06,
"loss": 0.8711,
"step": 47
},
{
"epoch": 0.15157894736842106,
"grad_norm": 0.7551989555358887,
"learning_rate": 5.052631578947369e-06,
"loss": 0.8306,
"step": 48
},
{
"epoch": 0.15473684210526314,
"grad_norm": 0.8163503408432007,
"learning_rate": 5.157894736842106e-06,
"loss": 0.8403,
"step": 49
},
{
"epoch": 0.15789473684210525,
"grad_norm": 0.7936963438987732,
"learning_rate": 5.263157894736842e-06,
"loss": 0.837,
"step": 50
},
{
"epoch": 0.16105263157894736,
"grad_norm": 0.7440241575241089,
"learning_rate": 5.36842105263158e-06,
"loss": 0.8538,
"step": 51
},
{
"epoch": 0.16421052631578947,
"grad_norm": 0.7333534359931946,
"learning_rate": 5.4736842105263165e-06,
"loss": 0.819,
"step": 52
},
{
"epoch": 0.16736842105263158,
"grad_norm": 0.6583252549171448,
"learning_rate": 5.578947368421052e-06,
"loss": 0.7879,
"step": 53
},
{
"epoch": 0.1705263157894737,
"grad_norm": 0.684806764125824,
"learning_rate": 5.68421052631579e-06,
"loss": 0.8325,
"step": 54
},
{
"epoch": 0.1736842105263158,
"grad_norm": 0.7598180174827576,
"learning_rate": 5.789473684210527e-06,
"loss": 0.8341,
"step": 55
},
{
"epoch": 0.17684210526315788,
"grad_norm": 0.8507996201515198,
"learning_rate": 5.8947368421052634e-06,
"loss": 0.8428,
"step": 56
},
{
"epoch": 0.18,
"grad_norm": 0.7600204348564148,
"learning_rate": 6e-06,
"loss": 0.818,
"step": 57
},
{
"epoch": 0.1831578947368421,
"grad_norm": 0.6492787599563599,
"learning_rate": 6.105263157894738e-06,
"loss": 0.8346,
"step": 58
},
{
"epoch": 0.1863157894736842,
"grad_norm": 0.6897916793823242,
"learning_rate": 6.2105263157894745e-06,
"loss": 0.8409,
"step": 59
},
{
"epoch": 0.18947368421052632,
"grad_norm": 0.7259018421173096,
"learning_rate": 6.31578947368421e-06,
"loss": 0.8285,
"step": 60
},
{
"epoch": 0.19263157894736843,
"grad_norm": 0.8188095688819885,
"learning_rate": 6.421052631578948e-06,
"loss": 0.8816,
"step": 61
},
{
"epoch": 0.1957894736842105,
"grad_norm": 0.6784875988960266,
"learning_rate": 6.526315789473685e-06,
"loss": 0.8153,
"step": 62
},
{
"epoch": 0.19894736842105262,
"grad_norm": 0.740426242351532,
"learning_rate": 6.631578947368421e-06,
"loss": 0.8243,
"step": 63
},
{
"epoch": 0.20210526315789473,
"grad_norm": 0.7158690094947815,
"learning_rate": 6.736842105263158e-06,
"loss": 0.828,
"step": 64
},
{
"epoch": 0.20526315789473684,
"grad_norm": 0.7803067564964294,
"learning_rate": 6.842105263157896e-06,
"loss": 0.7897,
"step": 65
},
{
"epoch": 0.20842105263157895,
"grad_norm": 0.6472960710525513,
"learning_rate": 6.947368421052632e-06,
"loss": 0.7908,
"step": 66
},
{
"epoch": 0.21157894736842106,
"grad_norm": 0.7025761008262634,
"learning_rate": 7.052631578947369e-06,
"loss": 0.8408,
"step": 67
},
{
"epoch": 0.21473684210526317,
"grad_norm": 0.7511818408966064,
"learning_rate": 7.157894736842106e-06,
"loss": 0.7913,
"step": 68
},
{
"epoch": 0.21789473684210525,
"grad_norm": 0.6920287609100342,
"learning_rate": 7.263157894736843e-06,
"loss": 0.8173,
"step": 69
},
{
"epoch": 0.22105263157894736,
"grad_norm": 0.6897710561752319,
"learning_rate": 7.368421052631579e-06,
"loss": 0.8213,
"step": 70
},
{
"epoch": 0.22421052631578947,
"grad_norm": 0.7388243675231934,
"learning_rate": 7.473684210526316e-06,
"loss": 0.8039,
"step": 71
},
{
"epoch": 0.22736842105263158,
"grad_norm": 0.8480104804039001,
"learning_rate": 7.578947368421054e-06,
"loss": 0.8235,
"step": 72
},
{
"epoch": 0.2305263157894737,
"grad_norm": 0.7924239635467529,
"learning_rate": 7.68421052631579e-06,
"loss": 0.8256,
"step": 73
},
{
"epoch": 0.2336842105263158,
"grad_norm": 0.7806857228279114,
"learning_rate": 7.789473684210526e-06,
"loss": 0.8284,
"step": 74
},
{
"epoch": 0.23684210526315788,
"grad_norm": 0.7032557129859924,
"learning_rate": 7.894736842105265e-06,
"loss": 0.7895,
"step": 75
},
{
"epoch": 0.24,
"grad_norm": 0.8388243913650513,
"learning_rate": 8.000000000000001e-06,
"loss": 0.8032,
"step": 76
},
{
"epoch": 0.2431578947368421,
"grad_norm": 0.8523675203323364,
"learning_rate": 8.105263157894736e-06,
"loss": 0.8118,
"step": 77
},
{
"epoch": 0.2463157894736842,
"grad_norm": 0.7454504370689392,
"learning_rate": 8.210526315789475e-06,
"loss": 0.7955,
"step": 78
},
{
"epoch": 0.24947368421052632,
"grad_norm": 0.737902820110321,
"learning_rate": 8.315789473684212e-06,
"loss": 0.8337,
"step": 79
},
{
"epoch": 0.25263157894736843,
"grad_norm": 0.7588112950325012,
"learning_rate": 8.421052631578948e-06,
"loss": 0.7851,
"step": 80
},
{
"epoch": 0.2557894736842105,
"grad_norm": 0.812930166721344,
"learning_rate": 8.526315789473685e-06,
"loss": 0.8024,
"step": 81
},
{
"epoch": 0.25894736842105265,
"grad_norm": 0.9560610055923462,
"learning_rate": 8.631578947368422e-06,
"loss": 0.8093,
"step": 82
},
{
"epoch": 0.26210526315789473,
"grad_norm": 0.9282992482185364,
"learning_rate": 8.736842105263158e-06,
"loss": 0.8169,
"step": 83
},
{
"epoch": 0.26526315789473687,
"grad_norm": 0.7830009460449219,
"learning_rate": 8.842105263157895e-06,
"loss": 0.8062,
"step": 84
},
{
"epoch": 0.26842105263157895,
"grad_norm": 0.8576416969299316,
"learning_rate": 8.947368421052632e-06,
"loss": 0.8023,
"step": 85
},
{
"epoch": 0.27157894736842103,
"grad_norm": 0.8227952122688293,
"learning_rate": 9.05263157894737e-06,
"loss": 0.805,
"step": 86
},
{
"epoch": 0.27473684210526317,
"grad_norm": 0.8914945721626282,
"learning_rate": 9.157894736842105e-06,
"loss": 0.8269,
"step": 87
},
{
"epoch": 0.27789473684210525,
"grad_norm": 0.81199711561203,
"learning_rate": 9.263157894736842e-06,
"loss": 0.8056,
"step": 88
},
{
"epoch": 0.2810526315789474,
"grad_norm": 1.0146549940109253,
"learning_rate": 9.36842105263158e-06,
"loss": 0.7905,
"step": 89
},
{
"epoch": 0.28421052631578947,
"grad_norm": 0.8971161246299744,
"learning_rate": 9.473684210526315e-06,
"loss": 0.8079,
"step": 90
},
{
"epoch": 0.2873684210526316,
"grad_norm": 0.9439178705215454,
"learning_rate": 9.578947368421054e-06,
"loss": 0.8126,
"step": 91
},
{
"epoch": 0.2905263157894737,
"grad_norm": 0.9185954332351685,
"learning_rate": 9.68421052631579e-06,
"loss": 0.7945,
"step": 92
},
{
"epoch": 0.29368421052631577,
"grad_norm": 0.7818227410316467,
"learning_rate": 9.789473684210527e-06,
"loss": 0.7825,
"step": 93
},
{
"epoch": 0.2968421052631579,
"grad_norm": 0.7514786124229431,
"learning_rate": 9.894736842105264e-06,
"loss": 0.8049,
"step": 94
},
{
"epoch": 0.3,
"grad_norm": 1.0437406301498413,
"learning_rate": 1e-05,
"loss": 0.8204,
"step": 95
},
{
"epoch": 0.3031578947368421,
"grad_norm": 0.921838104724884,
"learning_rate": 9.999966088952842e-06,
"loss": 0.8107,
"step": 96
},
{
"epoch": 0.3063157894736842,
"grad_norm": 0.7265084981918335,
"learning_rate": 9.99986435627135e-06,
"loss": 0.8067,
"step": 97
},
{
"epoch": 0.3094736842105263,
"grad_norm": 1.0714375972747803,
"learning_rate": 9.999694803335468e-06,
"loss": 0.8052,
"step": 98
},
{
"epoch": 0.3126315789473684,
"grad_norm": 0.9621460437774658,
"learning_rate": 9.999457432445087e-06,
"loss": 0.8122,
"step": 99
},
{
"epoch": 0.3157894736842105,
"grad_norm": 0.7225667834281921,
"learning_rate": 9.999152246820001e-06,
"loss": 0.8104,
"step": 100
},
{
"epoch": 0.31894736842105265,
"grad_norm": 0.8660763502120972,
"learning_rate": 9.998779250599877e-06,
"loss": 0.8005,
"step": 101
},
{
"epoch": 0.32210526315789473,
"grad_norm": 0.951799213886261,
"learning_rate": 9.998338448844193e-06,
"loss": 0.8022,
"step": 102
},
{
"epoch": 0.32526315789473687,
"grad_norm": 0.8598304986953735,
"learning_rate": 9.997829847532165e-06,
"loss": 0.7862,
"step": 103
},
{
"epoch": 0.32842105263157895,
"grad_norm": 0.7481747269630432,
"learning_rate": 9.99725345356268e-06,
"loss": 0.7933,
"step": 104
},
{
"epoch": 0.33157894736842103,
"grad_norm": 0.9335681200027466,
"learning_rate": 9.996609274754183e-06,
"loss": 0.7822,
"step": 105
},
{
"epoch": 0.33473684210526317,
"grad_norm": 0.8390291929244995,
"learning_rate": 9.995897319844588e-06,
"loss": 0.7838,
"step": 106
},
{
"epoch": 0.33789473684210525,
"grad_norm": 0.7386859655380249,
"learning_rate": 9.995117598491146e-06,
"loss": 0.7872,
"step": 107
},
{
"epoch": 0.3410526315789474,
"grad_norm": 0.8380564451217651,
"learning_rate": 9.994270121270327e-06,
"loss": 0.7568,
"step": 108
},
{
"epoch": 0.34421052631578947,
"grad_norm": 0.7734573483467102,
"learning_rate": 9.993354899677665e-06,
"loss": 0.8004,
"step": 109
},
{
"epoch": 0.3473684210526316,
"grad_norm": 0.8333052396774292,
"learning_rate": 9.99237194612761e-06,
"loss": 0.7686,
"step": 110
},
{
"epoch": 0.3505263157894737,
"grad_norm": 0.9416193962097168,
"learning_rate": 9.991321273953357e-06,
"loss": 0.7973,
"step": 111
},
{
"epoch": 0.35368421052631577,
"grad_norm": 0.7098072171211243,
"learning_rate": 9.99020289740666e-06,
"loss": 0.7547,
"step": 112
},
{
"epoch": 0.3568421052631579,
"grad_norm": 0.9404763579368591,
"learning_rate": 9.989016831657652e-06,
"loss": 0.8093,
"step": 113
},
{
"epoch": 0.36,
"grad_norm": 1.0210902690887451,
"learning_rate": 9.987763092794621e-06,
"loss": 0.7939,
"step": 114
},
{
"epoch": 0.3631578947368421,
"grad_norm": 0.7902900576591492,
"learning_rate": 9.986441697823808e-06,
"loss": 0.7889,
"step": 115
},
{
"epoch": 0.3663157894736842,
"grad_norm": 0.9824138283729553,
"learning_rate": 9.985052664669168e-06,
"loss": 0.7888,
"step": 116
},
{
"epoch": 0.36947368421052634,
"grad_norm": 0.9166287183761597,
"learning_rate": 9.983596012172127e-06,
"loss": 0.7845,
"step": 117
},
{
"epoch": 0.3726315789473684,
"grad_norm": 0.7522174715995789,
"learning_rate": 9.982071760091334e-06,
"loss": 0.7619,
"step": 118
},
{
"epoch": 0.3757894736842105,
"grad_norm": 0.9811515212059021,
"learning_rate": 9.980479929102377e-06,
"loss": 0.7943,
"step": 119
},
{
"epoch": 0.37894736842105264,
"grad_norm": 0.7924212217330933,
"learning_rate": 9.978820540797521e-06,
"loss": 0.7928,
"step": 120
},
{
"epoch": 0.3821052631578947,
"grad_norm": 0.8440699577331543,
"learning_rate": 9.977093617685404e-06,
"loss": 0.8005,
"step": 121
},
{
"epoch": 0.38526315789473686,
"grad_norm": 0.7170959115028381,
"learning_rate": 9.975299183190734e-06,
"loss": 0.7844,
"step": 122
},
{
"epoch": 0.38842105263157894,
"grad_norm": 0.9932653307914734,
"learning_rate": 9.973437261653973e-06,
"loss": 0.7932,
"step": 123
},
{
"epoch": 0.391578947368421,
"grad_norm": 0.7866117358207703,
"learning_rate": 9.971507878331005e-06,
"loss": 0.7672,
"step": 124
},
{
"epoch": 0.39473684210526316,
"grad_norm": 0.791227400302887,
"learning_rate": 9.96951105939279e-06,
"loss": 0.7678,
"step": 125
},
{
"epoch": 0.39789473684210525,
"grad_norm": 1.0364530086517334,
"learning_rate": 9.96744683192502e-06,
"loss": 0.7814,
"step": 126
},
{
"epoch": 0.4010526315789474,
"grad_norm": 0.6801534295082092,
"learning_rate": 9.96531522392774e-06,
"loss": 0.751,
"step": 127
},
{
"epoch": 0.40421052631578946,
"grad_norm": 1.0359455347061157,
"learning_rate": 9.963116264314974e-06,
"loss": 0.7701,
"step": 128
},
{
"epoch": 0.4073684210526316,
"grad_norm": 0.8113856315612793,
"learning_rate": 9.960849982914332e-06,
"loss": 0.7924,
"step": 129
},
{
"epoch": 0.4105263157894737,
"grad_norm": 0.7722939848899841,
"learning_rate": 9.958516410466601e-06,
"loss": 0.7891,
"step": 130
},
{
"epoch": 0.41368421052631577,
"grad_norm": 0.8695185780525208,
"learning_rate": 9.956115578625339e-06,
"loss": 0.7795,
"step": 131
},
{
"epoch": 0.4168421052631579,
"grad_norm": 0.6542952060699463,
"learning_rate": 9.953647519956432e-06,
"loss": 0.7518,
"step": 132
},
{
"epoch": 0.42,
"grad_norm": 0.7686191201210022,
"learning_rate": 9.951112267937663e-06,
"loss": 0.7851,
"step": 133
},
{
"epoch": 0.4231578947368421,
"grad_norm": 0.7790025472640991,
"learning_rate": 9.948509856958253e-06,
"loss": 0.7453,
"step": 134
},
{
"epoch": 0.4263157894736842,
"grad_norm": 0.7110816240310669,
"learning_rate": 9.945840322318391e-06,
"loss": 0.8121,
"step": 135
},
{
"epoch": 0.42947368421052634,
"grad_norm": 0.7633756995201111,
"learning_rate": 9.943103700228768e-06,
"loss": 0.7982,
"step": 136
},
{
"epoch": 0.4326315789473684,
"grad_norm": 0.7634323239326477,
"learning_rate": 9.940300027810067e-06,
"loss": 0.8097,
"step": 137
},
{
"epoch": 0.4357894736842105,
"grad_norm": 0.7589544057846069,
"learning_rate": 9.93742934309248e-06,
"loss": 0.7697,
"step": 138
},
{
"epoch": 0.43894736842105264,
"grad_norm": 0.820608377456665,
"learning_rate": 9.934491685015173e-06,
"loss": 0.7917,
"step": 139
},
{
"epoch": 0.4421052631578947,
"grad_norm": 0.801396906375885,
"learning_rate": 9.931487093425775e-06,
"loss": 0.7908,
"step": 140
},
{
"epoch": 0.44526315789473686,
"grad_norm": 0.9706858992576599,
"learning_rate": 9.928415609079821e-06,
"loss": 0.7753,
"step": 141
},
{
"epoch": 0.44842105263157894,
"grad_norm": 0.818867027759552,
"learning_rate": 9.925277273640211e-06,
"loss": 0.7884,
"step": 142
},
{
"epoch": 0.4515789473684211,
"grad_norm": 0.9297605156898499,
"learning_rate": 9.922072129676644e-06,
"loss": 0.8113,
"step": 143
},
{
"epoch": 0.45473684210526316,
"grad_norm": 0.9068427681922913,
"learning_rate": 9.918800220665035e-06,
"loss": 0.808,
"step": 144
},
{
"epoch": 0.45789473684210524,
"grad_norm": 0.7996309399604797,
"learning_rate": 9.915461590986926e-06,
"loss": 0.7967,
"step": 145
},
{
"epoch": 0.4610526315789474,
"grad_norm": 0.777631938457489,
"learning_rate": 9.912056285928891e-06,
"loss": 0.8123,
"step": 146
},
{
"epoch": 0.46421052631578946,
"grad_norm": 0.7640287280082703,
"learning_rate": 9.908584351681911e-06,
"loss": 0.7942,
"step": 147
},
{
"epoch": 0.4673684210526316,
"grad_norm": 0.900375247001648,
"learning_rate": 9.90504583534076e-06,
"loss": 0.7916,
"step": 148
},
{
"epoch": 0.4705263157894737,
"grad_norm": 0.8545629382133484,
"learning_rate": 9.901440784903354e-06,
"loss": 0.7946,
"step": 149
},
{
"epoch": 0.47368421052631576,
"grad_norm": 0.739754319190979,
"learning_rate": 9.897769249270106e-06,
"loss": 0.797,
"step": 150
},
{
"epoch": 0.4768421052631579,
"grad_norm": 0.814207136631012,
"learning_rate": 9.894031278243266e-06,
"loss": 0.7636,
"step": 151
},
{
"epoch": 0.48,
"grad_norm": 0.8900481462478638,
"learning_rate": 9.890226922526238e-06,
"loss": 0.7918,
"step": 152
},
{
"epoch": 0.4831578947368421,
"grad_norm": 0.7793235182762146,
"learning_rate": 9.886356233722894e-06,
"loss": 0.7956,
"step": 153
},
{
"epoch": 0.4863157894736842,
"grad_norm": 0.890493631362915,
"learning_rate": 9.88241926433688e-06,
"loss": 0.7643,
"step": 154
},
{
"epoch": 0.48947368421052634,
"grad_norm": 0.8794576525688171,
"learning_rate": 9.878416067770898e-06,
"loss": 0.7929,
"step": 155
},
{
"epoch": 0.4926315789473684,
"grad_norm": 0.7571336627006531,
"learning_rate": 9.874346698325983e-06,
"loss": 0.7852,
"step": 156
},
{
"epoch": 0.4957894736842105,
"grad_norm": 0.8208326101303101,
"learning_rate": 9.870211211200766e-06,
"loss": 0.8021,
"step": 157
},
{
"epoch": 0.49894736842105264,
"grad_norm": 0.904834508895874,
"learning_rate": 9.866009662490727e-06,
"loss": 0.7809,
"step": 158
},
{
"epoch": 0.5021052631578947,
"grad_norm": 0.8771081566810608,
"learning_rate": 9.861742109187433e-06,
"loss": 0.8044,
"step": 159
},
{
"epoch": 0.5052631578947369,
"grad_norm": 0.7162814140319824,
"learning_rate": 9.857408609177763e-06,
"loss": 0.7935,
"step": 160
},
{
"epoch": 0.508421052631579,
"grad_norm": 0.8736246824264526,
"learning_rate": 9.853009221243129e-06,
"loss": 0.781,
"step": 161
},
{
"epoch": 0.511578947368421,
"grad_norm": 0.7425425052642822,
"learning_rate": 9.848544005058668e-06,
"loss": 0.7619,
"step": 162
},
{
"epoch": 0.5147368421052632,
"grad_norm": 0.7853212356567383,
"learning_rate": 9.844013021192447e-06,
"loss": 0.77,
"step": 163
},
{
"epoch": 0.5178947368421053,
"grad_norm": 0.7985186576843262,
"learning_rate": 9.839416331104625e-06,
"loss": 0.7823,
"step": 164
},
{
"epoch": 0.5210526315789473,
"grad_norm": 0.8637293577194214,
"learning_rate": 9.834753997146633e-06,
"loss": 0.7598,
"step": 165
},
{
"epoch": 0.5242105263157895,
"grad_norm": 0.7811316847801208,
"learning_rate": 9.830026082560324e-06,
"loss": 0.7787,
"step": 166
},
{
"epoch": 0.5273684210526316,
"grad_norm": 0.795226514339447,
"learning_rate": 9.825232651477109e-06,
"loss": 0.7843,
"step": 167
},
{
"epoch": 0.5305263157894737,
"grad_norm": 0.9816672205924988,
"learning_rate": 9.820373768917095e-06,
"loss": 0.778,
"step": 168
},
{
"epoch": 0.5336842105263158,
"grad_norm": 0.7347379922866821,
"learning_rate": 9.815449500788203e-06,
"loss": 0.8002,
"step": 169
},
{
"epoch": 0.5368421052631579,
"grad_norm": 0.8775848746299744,
"learning_rate": 9.810459913885265e-06,
"loss": 0.7484,
"step": 170
},
{
"epoch": 0.54,
"grad_norm": 0.8044248819351196,
"learning_rate": 9.805405075889129e-06,
"loss": 0.7607,
"step": 171
},
{
"epoch": 0.5431578947368421,
"grad_norm": 0.8173661231994629,
"learning_rate": 9.800285055365737e-06,
"loss": 0.7654,
"step": 172
},
{
"epoch": 0.5463157894736842,
"grad_norm": 0.7608327269554138,
"learning_rate": 9.79509992176519e-06,
"loss": 0.7716,
"step": 173
},
{
"epoch": 0.5494736842105263,
"grad_norm": 0.7798534631729126,
"learning_rate": 9.789849745420811e-06,
"loss": 0.7906,
"step": 174
},
{
"epoch": 0.5526315789473685,
"grad_norm": 0.7842231392860413,
"learning_rate": 9.784534597548194e-06,
"loss": 0.7752,
"step": 175
},
{
"epoch": 0.5557894736842105,
"grad_norm": 0.7228990793228149,
"learning_rate": 9.779154550244228e-06,
"loss": 0.7626,
"step": 176
},
{
"epoch": 0.5589473684210526,
"grad_norm": 0.8362434506416321,
"learning_rate": 9.77370967648613e-06,
"loss": 0.7547,
"step": 177
},
{
"epoch": 0.5621052631578948,
"grad_norm": 0.7536957859992981,
"learning_rate": 9.768200050130446e-06,
"loss": 0.7658,
"step": 178
},
{
"epoch": 0.5652631578947368,
"grad_norm": 0.7978958487510681,
"learning_rate": 9.76262574591206e-06,
"loss": 0.7753,
"step": 179
},
{
"epoch": 0.5684210526315789,
"grad_norm": 0.7004671692848206,
"learning_rate": 9.756986839443166e-06,
"loss": 0.8162,
"step": 180
},
{
"epoch": 0.5715789473684211,
"grad_norm": 0.7977942228317261,
"learning_rate": 9.751283407212253e-06,
"loss": 0.7756,
"step": 181
},
{
"epoch": 0.5747368421052632,
"grad_norm": 0.7080273628234863,
"learning_rate": 9.745515526583066e-06,
"loss": 0.7631,
"step": 182
},
{
"epoch": 0.5778947368421052,
"grad_norm": 0.8024187088012695,
"learning_rate": 9.739683275793554e-06,
"loss": 0.7908,
"step": 183
},
{
"epoch": 0.5810526315789474,
"grad_norm": 0.7169841527938843,
"learning_rate": 9.73378673395481e-06,
"loss": 0.7602,
"step": 184
},
{
"epoch": 0.5842105263157895,
"grad_norm": 0.7052045464515686,
"learning_rate": 9.727825981049994e-06,
"loss": 0.7795,
"step": 185
},
{
"epoch": 0.5873684210526315,
"grad_norm": 0.7155394554138184,
"learning_rate": 9.72180109793326e-06,
"loss": 0.7919,
"step": 186
},
{
"epoch": 0.5905263157894737,
"grad_norm": 0.7038570046424866,
"learning_rate": 9.715712166328643e-06,
"loss": 0.799,
"step": 187
},
{
"epoch": 0.5936842105263158,
"grad_norm": 0.6392240524291992,
"learning_rate": 9.709559268828963e-06,
"loss": 0.7868,
"step": 188
},
{
"epoch": 0.5968421052631578,
"grad_norm": 0.748368501663208,
"learning_rate": 9.703342488894699e-06,
"loss": 0.7572,
"step": 189
},
{
"epoch": 0.6,
"grad_norm": 0.6144984364509583,
"learning_rate": 9.697061910852857e-06,
"loss": 0.7579,
"step": 190
},
{
"epoch": 0.6031578947368421,
"grad_norm": 0.7086066007614136,
"learning_rate": 9.690717619895828e-06,
"loss": 0.7773,
"step": 191
},
{
"epoch": 0.6063157894736843,
"grad_norm": 0.7267181277275085,
"learning_rate": 9.684309702080234e-06,
"loss": 0.8062,
"step": 192
},
{
"epoch": 0.6094736842105263,
"grad_norm": 0.6966003179550171,
"learning_rate": 9.677838244325754e-06,
"loss": 0.7542,
"step": 193
},
{
"epoch": 0.6126315789473684,
"grad_norm": 0.7743439674377441,
"learning_rate": 9.671303334413952e-06,
"loss": 0.7607,
"step": 194
},
{
"epoch": 0.6157894736842106,
"grad_norm": 0.698222279548645,
"learning_rate": 9.664705060987085e-06,
"loss": 0.7181,
"step": 195
},
{
"epoch": 0.6189473684210526,
"grad_norm": 0.738317608833313,
"learning_rate": 9.658043513546898e-06,
"loss": 0.8108,
"step": 196
},
{
"epoch": 0.6221052631578947,
"grad_norm": 0.7287928462028503,
"learning_rate": 9.651318782453407e-06,
"loss": 0.7593,
"step": 197
},
{
"epoch": 0.6252631578947369,
"grad_norm": 0.7296643853187561,
"learning_rate": 9.644530958923683e-06,
"loss": 0.7683,
"step": 198
},
{
"epoch": 0.628421052631579,
"grad_norm": 0.6896747946739197,
"learning_rate": 9.637680135030609e-06,
"loss": 0.7851,
"step": 199
},
{
"epoch": 0.631578947368421,
"grad_norm": 0.722059428691864,
"learning_rate": 9.63076640370163e-06,
"loss": 0.7837,
"step": 200
},
{
"epoch": 0.6347368421052632,
"grad_norm": 0.7174692749977112,
"learning_rate": 9.623789858717491e-06,
"loss": 0.7959,
"step": 201
},
{
"epoch": 0.6378947368421053,
"grad_norm": 0.769080400466919,
"learning_rate": 9.616750594710972e-06,
"loss": 0.7711,
"step": 202
},
{
"epoch": 0.6410526315789473,
"grad_norm": 0.857413113117218,
"learning_rate": 9.6096487071656e-06,
"loss": 0.7967,
"step": 203
},
{
"epoch": 0.6442105263157895,
"grad_norm": 0.6896303296089172,
"learning_rate": 9.602484292414348e-06,
"loss": 0.7541,
"step": 204
},
{
"epoch": 0.6473684210526316,
"grad_norm": 0.7579444050788879,
"learning_rate": 9.595257447638344e-06,
"loss": 0.7721,
"step": 205
},
{
"epoch": 0.6505263157894737,
"grad_norm": 0.8411123752593994,
"learning_rate": 9.587968270865534e-06,
"loss": 0.8087,
"step": 206
},
{
"epoch": 0.6536842105263158,
"grad_norm": 0.769237220287323,
"learning_rate": 9.580616860969365e-06,
"loss": 0.7511,
"step": 207
},
{
"epoch": 0.6568421052631579,
"grad_norm": 0.6318850517272949,
"learning_rate": 9.573203317667442e-06,
"loss": 0.7552,
"step": 208
},
{
"epoch": 0.66,
"grad_norm": 0.7551825046539307,
"learning_rate": 9.56572774152017e-06,
"loss": 0.7795,
"step": 209
},
{
"epoch": 0.6631578947368421,
"grad_norm": 0.8102651834487915,
"learning_rate": 9.558190233929396e-06,
"loss": 0.7973,
"step": 210
},
{
"epoch": 0.6663157894736842,
"grad_norm": 0.7229571342468262,
"learning_rate": 9.55059089713703e-06,
"loss": 0.7906,
"step": 211
},
{
"epoch": 0.6694736842105263,
"grad_norm": 0.6960732936859131,
"learning_rate": 9.54292983422366e-06,
"loss": 0.755,
"step": 212
},
{
"epoch": 0.6726315789473685,
"grad_norm": 0.7564675211906433,
"learning_rate": 9.53520714910715e-06,
"loss": 0.7993,
"step": 213
},
{
"epoch": 0.6757894736842105,
"grad_norm": 0.7037273645401001,
"learning_rate": 9.527422946541238e-06,
"loss": 0.7898,
"step": 214
},
{
"epoch": 0.6789473684210526,
"grad_norm": 0.735377311706543,
"learning_rate": 9.519577332114107e-06,
"loss": 0.7599,
"step": 215
},
{
"epoch": 0.6821052631578948,
"grad_norm": 0.6795523762702942,
"learning_rate": 9.511670412246956e-06,
"loss": 0.7779,
"step": 216
},
{
"epoch": 0.6852631578947368,
"grad_norm": 0.6760138273239136,
"learning_rate": 9.503702294192563e-06,
"loss": 0.7759,
"step": 217
},
{
"epoch": 0.6884210526315789,
"grad_norm": 0.6858091354370117,
"learning_rate": 9.495673086033813e-06,
"loss": 0.7508,
"step": 218
},
{
"epoch": 0.6915789473684211,
"grad_norm": 0.8286890983581543,
"learning_rate": 9.487582896682252e-06,
"loss": 0.7454,
"step": 219
},
{
"epoch": 0.6947368421052632,
"grad_norm": 0.8973384499549866,
"learning_rate": 9.479431835876596e-06,
"loss": 0.7616,
"step": 220
},
{
"epoch": 0.6978947368421052,
"grad_norm": 0.7554185390472412,
"learning_rate": 9.471220014181247e-06,
"loss": 0.8202,
"step": 221
},
{
"epoch": 0.7010526315789474,
"grad_norm": 0.7141987681388855,
"learning_rate": 9.462947542984795e-06,
"loss": 0.7863,
"step": 222
},
{
"epoch": 0.7042105263157895,
"grad_norm": 0.9621738195419312,
"learning_rate": 9.454614534498506e-06,
"loss": 0.7748,
"step": 223
},
{
"epoch": 0.7073684210526315,
"grad_norm": 0.8891428709030151,
"learning_rate": 9.446221101754795e-06,
"loss": 0.7568,
"step": 224
},
{
"epoch": 0.7105263157894737,
"grad_norm": 0.8448644876480103,
"learning_rate": 9.4377673586057e-06,
"loss": 0.7561,
"step": 225
},
{
"epoch": 0.7136842105263158,
"grad_norm": 0.744163453578949,
"learning_rate": 9.429253419721335e-06,
"loss": 0.8085,
"step": 226
},
{
"epoch": 0.716842105263158,
"grad_norm": 0.7832540273666382,
"learning_rate": 9.420679400588334e-06,
"loss": 0.7592,
"step": 227
},
{
"epoch": 0.72,
"grad_norm": 0.8211809992790222,
"learning_rate": 9.412045417508281e-06,
"loss": 0.7546,
"step": 228
},
{
"epoch": 0.7231578947368421,
"grad_norm": 0.7208123803138733,
"learning_rate": 9.40335158759614e-06,
"loss": 0.7848,
"step": 229
},
{
"epoch": 0.7263157894736842,
"grad_norm": 0.770294725894928,
"learning_rate": 9.394598028778664e-06,
"loss": 0.7693,
"step": 230
},
{
"epoch": 0.7294736842105263,
"grad_norm": 0.7772082686424255,
"learning_rate": 9.385784859792787e-06,
"loss": 0.7752,
"step": 231
},
{
"epoch": 0.7326315789473684,
"grad_norm": 0.7851942181587219,
"learning_rate": 9.376912200184029e-06,
"loss": 0.7858,
"step": 232
},
{
"epoch": 0.7357894736842105,
"grad_norm": 0.9354181885719299,
"learning_rate": 9.367980170304857e-06,
"loss": 0.7474,
"step": 233
},
{
"epoch": 0.7389473684210527,
"grad_norm": 0.7733820676803589,
"learning_rate": 9.35898889131307e-06,
"loss": 0.7727,
"step": 234
},
{
"epoch": 0.7421052631578947,
"grad_norm": 0.746245265007019,
"learning_rate": 9.349938485170139e-06,
"loss": 0.7459,
"step": 235
},
{
"epoch": 0.7452631578947368,
"grad_norm": 0.8601150512695312,
"learning_rate": 9.340829074639566e-06,
"loss": 0.768,
"step": 236
},
{
"epoch": 0.748421052631579,
"grad_norm": 0.7379437685012817,
"learning_rate": 9.331660783285208e-06,
"loss": 0.7667,
"step": 237
},
{
"epoch": 0.751578947368421,
"grad_norm": 0.7364124059677124,
"learning_rate": 9.322433735469614e-06,
"loss": 0.799,
"step": 238
},
{
"epoch": 0.7547368421052632,
"grad_norm": 0.7081237435340881,
"learning_rate": 9.313148056352321e-06,
"loss": 0.7464,
"step": 239
},
{
"epoch": 0.7578947368421053,
"grad_norm": 0.7287322878837585,
"learning_rate": 9.303803871888172e-06,
"loss": 0.7915,
"step": 240
},
{
"epoch": 0.7610526315789473,
"grad_norm": 0.7040133476257324,
"learning_rate": 9.2944013088256e-06,
"loss": 0.7631,
"step": 241
},
{
"epoch": 0.7642105263157895,
"grad_norm": 0.7509044408798218,
"learning_rate": 9.284940494704906e-06,
"loss": 0.7836,
"step": 242
},
{
"epoch": 0.7673684210526316,
"grad_norm": 0.7864567041397095,
"learning_rate": 9.275421557856536e-06,
"loss": 0.7547,
"step": 243
},
{
"epoch": 0.7705263157894737,
"grad_norm": 0.696180522441864,
"learning_rate": 9.26584462739934e-06,
"loss": 0.7908,
"step": 244
},
{
"epoch": 0.7736842105263158,
"grad_norm": 0.7591786980628967,
"learning_rate": 9.25620983323881e-06,
"loss": 0.7663,
"step": 245
},
{
"epoch": 0.7768421052631579,
"grad_norm": 0.6878987550735474,
"learning_rate": 9.246517306065332e-06,
"loss": 0.7363,
"step": 246
},
{
"epoch": 0.78,
"grad_norm": 0.7305520176887512,
"learning_rate": 9.236767177352403e-06,
"loss": 0.7649,
"step": 247
},
{
"epoch": 0.783157894736842,
"grad_norm": 0.7379259467124939,
"learning_rate": 9.226959579354855e-06,
"loss": 0.7733,
"step": 248
},
{
"epoch": 0.7863157894736842,
"grad_norm": 0.7051509618759155,
"learning_rate": 9.217094645107052e-06,
"loss": 0.771,
"step": 249
},
{
"epoch": 0.7894736842105263,
"grad_norm": 0.7959958910942078,
"learning_rate": 9.207172508421099e-06,
"loss": 0.7529,
"step": 250
},
{
"epoch": 0.7926315789473685,
"grad_norm": 0.6911975145339966,
"learning_rate": 9.197193303885008e-06,
"loss": 0.7722,
"step": 251
},
{
"epoch": 0.7957894736842105,
"grad_norm": 0.8246911764144897,
"learning_rate": 9.187157166860894e-06,
"loss": 0.768,
"step": 252
},
{
"epoch": 0.7989473684210526,
"grad_norm": 0.7509826421737671,
"learning_rate": 9.177064233483121e-06,
"loss": 0.7869,
"step": 253
},
{
"epoch": 0.8021052631578948,
"grad_norm": 0.7473223209381104,
"learning_rate": 9.166914640656467e-06,
"loss": 0.7515,
"step": 254
},
{
"epoch": 0.8052631578947368,
"grad_norm": 0.8410593867301941,
"learning_rate": 9.156708526054257e-06,
"loss": 0.8106,
"step": 255
},
{
"epoch": 0.8084210526315789,
"grad_norm": 0.7072636485099792,
"learning_rate": 9.146446028116508e-06,
"loss": 0.7475,
"step": 256
},
{
"epoch": 0.8115789473684211,
"grad_norm": 0.7030969858169556,
"learning_rate": 9.136127286048038e-06,
"loss": 0.7567,
"step": 257
},
{
"epoch": 0.8147368421052632,
"grad_norm": 0.8154286742210388,
"learning_rate": 9.125752439816588e-06,
"loss": 0.7884,
"step": 258
},
{
"epoch": 0.8178947368421052,
"grad_norm": 0.6962621212005615,
"learning_rate": 9.115321630150918e-06,
"loss": 0.7688,
"step": 259
},
{
"epoch": 0.8210526315789474,
"grad_norm": 0.7069746255874634,
"learning_rate": 9.104834998538899e-06,
"loss": 0.7313,
"step": 260
},
{
"epoch": 0.8242105263157895,
"grad_norm": 0.7636252045631409,
"learning_rate": 9.094292687225594e-06,
"loss": 0.7617,
"step": 261
},
{
"epoch": 0.8273684210526315,
"grad_norm": 0.7339772582054138,
"learning_rate": 9.08369483921133e-06,
"loss": 0.7469,
"step": 262
},
{
"epoch": 0.8305263157894737,
"grad_norm": 0.71903395652771,
"learning_rate": 9.073041598249757e-06,
"loss": 0.7662,
"step": 263
},
{
"epoch": 0.8336842105263158,
"grad_norm": 0.7993423938751221,
"learning_rate": 9.062333108845897e-06,
"loss": 0.7514,
"step": 264
},
{
"epoch": 0.8368421052631579,
"grad_norm": 0.8029423356056213,
"learning_rate": 9.051569516254186e-06,
"loss": 0.7457,
"step": 265
},
{
"epoch": 0.84,
"grad_norm": 0.6745309233665466,
"learning_rate": 9.040750966476502e-06,
"loss": 0.7591,
"step": 266
},
{
"epoch": 0.8431578947368421,
"grad_norm": 0.8080049753189087,
"learning_rate": 9.029877606260187e-06,
"loss": 0.7712,
"step": 267
},
{
"epoch": 0.8463157894736842,
"grad_norm": 0.6188151836395264,
"learning_rate": 9.018949583096051e-06,
"loss": 0.7339,
"step": 268
},
{
"epoch": 0.8494736842105263,
"grad_norm": 0.7027499079704285,
"learning_rate": 9.00796704521638e-06,
"loss": 0.7651,
"step": 269
},
{
"epoch": 0.8526315789473684,
"grad_norm": 0.6698938608169556,
"learning_rate": 8.996930141592915e-06,
"loss": 0.7648,
"step": 270
},
{
"epoch": 0.8557894736842105,
"grad_norm": 0.610205352306366,
"learning_rate": 8.985839021934843e-06,
"loss": 0.7663,
"step": 271
},
{
"epoch": 0.8589473684210527,
"grad_norm": 0.6673887372016907,
"learning_rate": 8.974693836686755e-06,
"loss": 0.7512,
"step": 272
},
{
"epoch": 0.8621052631578947,
"grad_norm": 0.6559451818466187,
"learning_rate": 8.963494737026612e-06,
"loss": 0.7553,
"step": 273
},
{
"epoch": 0.8652631578947368,
"grad_norm": 0.679578959941864,
"learning_rate": 8.952241874863695e-06,
"loss": 0.7456,
"step": 274
},
{
"epoch": 0.868421052631579,
"grad_norm": 0.6281573176383972,
"learning_rate": 8.940935402836535e-06,
"loss": 0.7251,
"step": 275
},
{
"epoch": 0.871578947368421,
"grad_norm": 0.6603413820266724,
"learning_rate": 8.92957547431086e-06,
"loss": 0.7518,
"step": 276
},
{
"epoch": 0.8747368421052631,
"grad_norm": 0.759972870349884,
"learning_rate": 8.918162243377494e-06,
"loss": 0.755,
"step": 277
},
{
"epoch": 0.8778947368421053,
"grad_norm": 0.7080827951431274,
"learning_rate": 8.906695864850284e-06,
"loss": 0.7868,
"step": 278
},
{
"epoch": 0.8810526315789474,
"grad_norm": 0.6595768928527832,
"learning_rate": 8.895176494263993e-06,
"loss": 0.7536,
"step": 279
},
{
"epoch": 0.8842105263157894,
"grad_norm": 0.6742393374443054,
"learning_rate": 8.883604287872186e-06,
"loss": 0.7771,
"step": 280
},
{
"epoch": 0.8873684210526316,
"grad_norm": 0.7637141942977905,
"learning_rate": 8.871979402645116e-06,
"loss": 0.7723,
"step": 281
},
{
"epoch": 0.8905263157894737,
"grad_norm": 0.6951268315315247,
"learning_rate": 8.860301996267601e-06,
"loss": 0.7668,
"step": 282
},
{
"epoch": 0.8936842105263157,
"grad_norm": 0.672383189201355,
"learning_rate": 8.848572227136869e-06,
"loss": 0.759,
"step": 283
},
{
"epoch": 0.8968421052631579,
"grad_norm": 0.8001788854598999,
"learning_rate": 8.83679025436042e-06,
"loss": 0.7903,
"step": 284
},
{
"epoch": 0.9,
"grad_norm": 0.7648778557777405,
"learning_rate": 8.824956237753872e-06,
"loss": 0.7746,
"step": 285
},
{
"epoch": 0.9031578947368422,
"grad_norm": 0.6406739354133606,
"learning_rate": 8.813070337838781e-06,
"loss": 0.7771,
"step": 286
},
{
"epoch": 0.9063157894736842,
"grad_norm": 0.6822999715805054,
"learning_rate": 8.80113271584047e-06,
"loss": 0.7592,
"step": 287
},
{
"epoch": 0.9094736842105263,
"grad_norm": 0.7665005326271057,
"learning_rate": 8.789143533685847e-06,
"loss": 0.747,
"step": 288
},
{
"epoch": 0.9126315789473685,
"grad_norm": 0.7119230628013611,
"learning_rate": 8.777102954001199e-06,
"loss": 0.7686,
"step": 289
},
{
"epoch": 0.9157894736842105,
"grad_norm": 0.6693398356437683,
"learning_rate": 8.765011140109993e-06,
"loss": 0.7496,
"step": 290
},
{
"epoch": 0.9189473684210526,
"grad_norm": 0.6795002818107605,
"learning_rate": 8.752868256030658e-06,
"loss": 0.7543,
"step": 291
},
{
"epoch": 0.9221052631578948,
"grad_norm": 0.773586094379425,
"learning_rate": 8.740674466474357e-06,
"loss": 0.7822,
"step": 292
},
{
"epoch": 0.9252631578947368,
"grad_norm": 0.7866982817649841,
"learning_rate": 8.728429936842762e-06,
"loss": 0.7792,
"step": 293
},
{
"epoch": 0.9284210526315789,
"grad_norm": 0.6205121874809265,
"learning_rate": 8.716134833225803e-06,
"loss": 0.7541,
"step": 294
},
{
"epoch": 0.9315789473684211,
"grad_norm": 0.819891095161438,
"learning_rate": 8.70378932239941e-06,
"loss": 0.7477,
"step": 295
},
{
"epoch": 0.9347368421052632,
"grad_norm": 0.7897220253944397,
"learning_rate": 8.691393571823266e-06,
"loss": 0.7401,
"step": 296
},
{
"epoch": 0.9378947368421052,
"grad_norm": 0.772665798664093,
"learning_rate": 8.678947749638525e-06,
"loss": 0.7888,
"step": 297
},
{
"epoch": 0.9410526315789474,
"grad_norm": 0.7343323826789856,
"learning_rate": 8.666452024665533e-06,
"loss": 0.7321,
"step": 298
},
{
"epoch": 0.9442105263157895,
"grad_norm": 0.9545820355415344,
"learning_rate": 8.653906566401533e-06,
"loss": 0.761,
"step": 299
},
{
"epoch": 0.9473684210526315,
"grad_norm": 0.8397694230079651,
"learning_rate": 8.64131154501838e-06,
"loss": 0.7766,
"step": 300
},
{
"epoch": 0.9505263157894737,
"grad_norm": 0.7189294695854187,
"learning_rate": 8.628667131360218e-06,
"loss": 0.7279,
"step": 301
},
{
"epoch": 0.9536842105263158,
"grad_norm": 0.7687074542045593,
"learning_rate": 8.61597349694117e-06,
"loss": 0.8155,
"step": 302
},
{
"epoch": 0.9568421052631579,
"grad_norm": 0.7857553362846375,
"learning_rate": 8.60323081394301e-06,
"loss": 0.7422,
"step": 303
},
{
"epoch": 0.96,
"grad_norm": 0.7681856751441956,
"learning_rate": 8.59043925521283e-06,
"loss": 0.7551,
"step": 304
},
{
"epoch": 0.9631578947368421,
"grad_norm": 0.7847979068756104,
"learning_rate": 8.577598994260687e-06,
"loss": 0.7978,
"step": 305
},
{
"epoch": 0.9663157894736842,
"grad_norm": 0.7032335996627808,
"learning_rate": 8.56471020525726e-06,
"loss": 0.7596,
"step": 306
},
{
"epoch": 0.9694736842105263,
"grad_norm": 0.8031393885612488,
"learning_rate": 8.551773063031484e-06,
"loss": 0.7615,
"step": 307
},
{
"epoch": 0.9726315789473684,
"grad_norm": 0.9480533599853516,
"learning_rate": 8.538787743068172e-06,
"loss": 0.7728,
"step": 308
},
{
"epoch": 0.9757894736842105,
"grad_norm": 0.7876819968223572,
"learning_rate": 8.525754421505646e-06,
"loss": 0.7871,
"step": 309
},
{
"epoch": 0.9789473684210527,
"grad_norm": 0.8743879199028015,
"learning_rate": 8.512673275133334e-06,
"loss": 0.7953,
"step": 310
},
{
"epoch": 0.9821052631578947,
"grad_norm": 0.8548572063446045,
"learning_rate": 8.49954448138939e-06,
"loss": 0.751,
"step": 311
},
{
"epoch": 0.9852631578947368,
"grad_norm": 0.801235020160675,
"learning_rate": 8.486368218358268e-06,
"loss": 0.7325,
"step": 312
},
{
"epoch": 0.988421052631579,
"grad_norm": 0.6868360042572021,
"learning_rate": 8.473144664768322e-06,
"loss": 0.7565,
"step": 313
},
{
"epoch": 0.991578947368421,
"grad_norm": 0.6946709156036377,
"learning_rate": 8.459873999989367e-06,
"loss": 0.7436,
"step": 314
},
{
"epoch": 0.9947368421052631,
"grad_norm": 0.7966560125350952,
"learning_rate": 8.446556404030263e-06,
"loss": 0.7818,
"step": 315
},
{
"epoch": 0.9978947368421053,
"grad_norm": 0.7011910080909729,
"learning_rate": 8.433192057536458e-06,
"loss": 0.8047,
"step": 316
},
{
"epoch": 1.0010526315789474,
"grad_norm": 0.776292085647583,
"learning_rate": 8.419781141787549e-06,
"loss": 0.7393,
"step": 317
},
{
"epoch": 1.0042105263157894,
"grad_norm": 0.7284736037254333,
"learning_rate": 8.406323838694808e-06,
"loss": 0.6665,
"step": 318
},
{
"epoch": 1.0073684210526315,
"grad_norm": 0.6233306527137756,
"learning_rate": 8.392820330798734e-06,
"loss": 0.6677,
"step": 319
},
{
"epoch": 1.0105263157894737,
"grad_norm": 0.6907612085342407,
"learning_rate": 8.379270801266569e-06,
"loss": 0.6921,
"step": 320
},
{
"epoch": 1.0136842105263157,
"grad_norm": 0.7600920796394348,
"learning_rate": 8.3656754338898e-06,
"loss": 0.7129,
"step": 321
},
{
"epoch": 1.016842105263158,
"grad_norm": 0.7222225069999695,
"learning_rate": 8.352034413081687e-06,
"loss": 0.6633,
"step": 322
},
{
"epoch": 1.02,
"grad_norm": 0.8816447854042053,
"learning_rate": 8.33834792387475e-06,
"loss": 0.7188,
"step": 323
},
{
"epoch": 1.023157894736842,
"grad_norm": 0.7416877150535583,
"learning_rate": 8.324616151918263e-06,
"loss": 0.7279,
"step": 324
},
{
"epoch": 1.0263157894736843,
"grad_norm": 0.7744479775428772,
"learning_rate": 8.31083928347573e-06,
"loss": 0.6765,
"step": 325
},
{
"epoch": 1.0294736842105263,
"grad_norm": 0.7218484282493591,
"learning_rate": 8.297017505422366e-06,
"loss": 0.6872,
"step": 326
},
{
"epoch": 1.0326315789473683,
"grad_norm": 0.6902859807014465,
"learning_rate": 8.28315100524256e-06,
"loss": 0.6696,
"step": 327
},
{
"epoch": 1.0357894736842106,
"grad_norm": 0.6810570359230042,
"learning_rate": 8.269239971027328e-06,
"loss": 0.692,
"step": 328
},
{
"epoch": 1.0389473684210526,
"grad_norm": 0.7604600191116333,
"learning_rate": 8.255284591471762e-06,
"loss": 0.7042,
"step": 329
},
{
"epoch": 1.0421052631578946,
"grad_norm": 0.6768532991409302,
"learning_rate": 8.241285055872478e-06,
"loss": 0.7111,
"step": 330
},
{
"epoch": 1.045263157894737,
"grad_norm": 0.7930629849433899,
"learning_rate": 8.227241554125041e-06,
"loss": 0.7196,
"step": 331
},
{
"epoch": 1.048421052631579,
"grad_norm": 0.7775673270225525,
"learning_rate": 8.213154276721388e-06,
"loss": 0.6706,
"step": 332
},
{
"epoch": 1.051578947368421,
"grad_norm": 0.7330057621002197,
"learning_rate": 8.199023414747257e-06,
"loss": 0.6841,
"step": 333
},
{
"epoch": 1.0547368421052632,
"grad_norm": 0.6644675731658936,
"learning_rate": 8.18484915987957e-06,
"loss": 0.6813,
"step": 334
},
{
"epoch": 1.0578947368421052,
"grad_norm": 0.7848517894744873,
"learning_rate": 8.170631704383865e-06,
"loss": 0.6907,
"step": 335
},
{
"epoch": 1.0610526315789475,
"grad_norm": 0.6873379349708557,
"learning_rate": 8.15637124111166e-06,
"loss": 0.7008,
"step": 336
},
{
"epoch": 1.0642105263157895,
"grad_norm": 0.6563950777053833,
"learning_rate": 8.14206796349785e-06,
"loss": 0.7082,
"step": 337
},
{
"epoch": 1.0673684210526315,
"grad_norm": 0.6331363916397095,
"learning_rate": 8.127722065558087e-06,
"loss": 0.6702,
"step": 338
},
{
"epoch": 1.0705263157894738,
"grad_norm": 0.7190977931022644,
"learning_rate": 8.113333741886137e-06,
"loss": 0.6687,
"step": 339
},
{
"epoch": 1.0736842105263158,
"grad_norm": 0.6932594180107117,
"learning_rate": 8.098903187651252e-06,
"loss": 0.6677,
"step": 340
},
{
"epoch": 1.0768421052631578,
"grad_norm": 0.7676414251327515,
"learning_rate": 8.084430598595514e-06,
"loss": 0.6945,
"step": 341
},
{
"epoch": 1.08,
"grad_norm": 0.7621206045150757,
"learning_rate": 8.069916171031181e-06,
"loss": 0.6908,
"step": 342
},
{
"epoch": 1.083157894736842,
"grad_norm": 0.7197576761245728,
"learning_rate": 8.055360101838026e-06,
"loss": 0.7144,
"step": 343
},
{
"epoch": 1.0863157894736841,
"grad_norm": 0.7100708484649658,
"learning_rate": 8.04076258846067e-06,
"loss": 0.6983,
"step": 344
},
{
"epoch": 1.0894736842105264,
"grad_norm": 0.6768266558647156,
"learning_rate": 8.026123828905902e-06,
"loss": 0.6769,
"step": 345
},
{
"epoch": 1.0926315789473684,
"grad_norm": 0.7214046716690063,
"learning_rate": 8.011444021739986e-06,
"loss": 0.7022,
"step": 346
},
{
"epoch": 1.0957894736842104,
"grad_norm": 0.7675743699073792,
"learning_rate": 7.996723366085978e-06,
"loss": 0.6718,
"step": 347
},
{
"epoch": 1.0989473684210527,
"grad_norm": 0.7916297912597656,
"learning_rate": 7.981962061621012e-06,
"loss": 0.6947,
"step": 348
},
{
"epoch": 1.1021052631578947,
"grad_norm": 0.7070953845977783,
"learning_rate": 7.967160308573607e-06,
"loss": 0.6223,
"step": 349
},
{
"epoch": 1.1052631578947367,
"grad_norm": 0.8087365031242371,
"learning_rate": 7.952318307720943e-06,
"loss": 0.724,
"step": 350
},
{
"epoch": 1.108421052631579,
"grad_norm": 0.6387585401535034,
"learning_rate": 7.937436260386134e-06,
"loss": 0.6732,
"step": 351
},
{
"epoch": 1.111578947368421,
"grad_norm": 0.7180246114730835,
"learning_rate": 7.922514368435506e-06,
"loss": 0.669,
"step": 352
},
{
"epoch": 1.1147368421052632,
"grad_norm": 0.7500141859054565,
"learning_rate": 7.907552834275847e-06,
"loss": 0.6946,
"step": 353
},
{
"epoch": 1.1178947368421053,
"grad_norm": 0.6470015048980713,
"learning_rate": 7.892551860851679e-06,
"loss": 0.6826,
"step": 354
},
{
"epoch": 1.1210526315789473,
"grad_norm": 0.668079674243927,
"learning_rate": 7.877511651642486e-06,
"loss": 0.6672,
"step": 355
},
{
"epoch": 1.1242105263157895,
"grad_norm": 0.7336689829826355,
"learning_rate": 7.862432410659964e-06,
"loss": 0.6932,
"step": 356
},
{
"epoch": 1.1273684210526316,
"grad_norm": 0.6308228969573975,
"learning_rate": 7.847314342445258e-06,
"loss": 0.7253,
"step": 357
},
{
"epoch": 1.1305263157894736,
"grad_norm": 0.6617457866668701,
"learning_rate": 7.832157652066173e-06,
"loss": 0.7053,
"step": 358
},
{
"epoch": 1.1336842105263158,
"grad_norm": 0.7797347903251648,
"learning_rate": 7.81696254511441e-06,
"loss": 0.6927,
"step": 359
},
{
"epoch": 1.1368421052631579,
"grad_norm": 0.6679685711860657,
"learning_rate": 7.80172922770276e-06,
"loss": 0.6917,
"step": 360
},
{
"epoch": 1.1400000000000001,
"grad_norm": 0.7145143151283264,
"learning_rate": 7.786457906462329e-06,
"loss": 0.6967,
"step": 361
},
{
"epoch": 1.1431578947368422,
"grad_norm": 0.8625593185424805,
"learning_rate": 7.771148788539704e-06,
"loss": 0.6728,
"step": 362
},
{
"epoch": 1.1463157894736842,
"grad_norm": 0.7024162411689758,
"learning_rate": 7.755802081594179e-06,
"loss": 0.7053,
"step": 363
},
{
"epoch": 1.1494736842105264,
"grad_norm": 0.7329344153404236,
"learning_rate": 7.740417993794918e-06,
"loss": 0.644,
"step": 364
},
{
"epoch": 1.1526315789473685,
"grad_norm": 0.7487715482711792,
"learning_rate": 7.724996733818124e-06,
"loss": 0.6838,
"step": 365
},
{
"epoch": 1.1557894736842105,
"grad_norm": 0.7509961128234863,
"learning_rate": 7.709538510844234e-06,
"loss": 0.6527,
"step": 366
},
{
"epoch": 1.1589473684210527,
"grad_norm": 0.7864775657653809,
"learning_rate": 7.694043534555055e-06,
"loss": 0.7205,
"step": 367
},
{
"epoch": 1.1621052631578948,
"grad_norm": 0.821261465549469,
"learning_rate": 7.678512015130936e-06,
"loss": 0.6832,
"step": 368
},
{
"epoch": 1.1652631578947368,
"grad_norm": 0.7063290476799011,
"learning_rate": 7.662944163247916e-06,
"loss": 0.6964,
"step": 369
},
{
"epoch": 1.168421052631579,
"grad_norm": 0.7851418256759644,
"learning_rate": 7.647340190074854e-06,
"loss": 0.7028,
"step": 370
},
{
"epoch": 1.171578947368421,
"grad_norm": 0.747036337852478,
"learning_rate": 7.63170030727058e-06,
"loss": 0.637,
"step": 371
},
{
"epoch": 1.174736842105263,
"grad_norm": 0.6792373657226562,
"learning_rate": 7.616024726981015e-06,
"loss": 0.6903,
"step": 372
},
{
"epoch": 1.1778947368421053,
"grad_norm": 0.805324912071228,
"learning_rate": 7.600313661836298e-06,
"loss": 0.6826,
"step": 373
},
{
"epoch": 1.1810526315789474,
"grad_norm": 0.6382725238800049,
"learning_rate": 7.584567324947893e-06,
"loss": 0.6337,
"step": 374
},
{
"epoch": 1.1842105263157894,
"grad_norm": 0.606036365032196,
"learning_rate": 7.568785929905713e-06,
"loss": 0.7219,
"step": 375
},
{
"epoch": 1.1873684210526316,
"grad_norm": 0.6815652847290039,
"learning_rate": 7.552969690775209e-06,
"loss": 0.6729,
"step": 376
},
{
"epoch": 1.1905263157894737,
"grad_norm": 0.6756043434143066,
"learning_rate": 7.537118822094474e-06,
"loss": 0.7,
"step": 377
},
{
"epoch": 1.1936842105263157,
"grad_norm": 0.6447840929031372,
"learning_rate": 7.521233538871329e-06,
"loss": 0.6982,
"step": 378
},
{
"epoch": 1.196842105263158,
"grad_norm": 0.5919721126556396,
"learning_rate": 7.505314056580411e-06,
"loss": 0.6923,
"step": 379
},
{
"epoch": 1.2,
"grad_norm": 0.6429916620254517,
"learning_rate": 7.489360591160245e-06,
"loss": 0.6523,
"step": 380
},
{
"epoch": 1.203157894736842,
"grad_norm": 0.7389147281646729,
"learning_rate": 7.4733733590103185e-06,
"loss": 0.6774,
"step": 381
},
{
"epoch": 1.2063157894736842,
"grad_norm": 0.6661043763160706,
"learning_rate": 7.457352576988144e-06,
"loss": 0.7086,
"step": 382
},
{
"epoch": 1.2094736842105263,
"grad_norm": 0.6406580209732056,
"learning_rate": 7.441298462406321e-06,
"loss": 0.7328,
"step": 383
},
{
"epoch": 1.2126315789473685,
"grad_norm": 0.613018810749054,
"learning_rate": 7.4252112330295835e-06,
"loss": 0.7078,
"step": 384
},
{
"epoch": 1.2157894736842105,
"grad_norm": 0.696648359298706,
"learning_rate": 7.409091107071849e-06,
"loss": 0.7006,
"step": 385
},
{
"epoch": 1.2189473684210526,
"grad_norm": 0.6505605578422546,
"learning_rate": 7.392938303193257e-06,
"loss": 0.6805,
"step": 386
},
{
"epoch": 1.2221052631578948,
"grad_norm": 0.6389300227165222,
"learning_rate": 7.376753040497207e-06,
"loss": 0.6447,
"step": 387
},
{
"epoch": 1.2252631578947368,
"grad_norm": 0.6349870562553406,
"learning_rate": 7.3605355385273805e-06,
"loss": 0.6565,
"step": 388
},
{
"epoch": 1.2284210526315789,
"grad_norm": 0.6041010618209839,
"learning_rate": 7.344286017264765e-06,
"loss": 0.708,
"step": 389
},
{
"epoch": 1.231578947368421,
"grad_norm": 0.6843451261520386,
"learning_rate": 7.3280046971246786e-06,
"loss": 0.7028,
"step": 390
},
{
"epoch": 1.2347368421052631,
"grad_norm": 0.6678407788276672,
"learning_rate": 7.311691798953765e-06,
"loss": 0.702,
"step": 391
},
{
"epoch": 1.2378947368421054,
"grad_norm": 0.6551707983016968,
"learning_rate": 7.295347544027006e-06,
"loss": 0.6968,
"step": 392
},
{
"epoch": 1.2410526315789474,
"grad_norm": 0.6763453483581543,
"learning_rate": 7.278972154044722e-06,
"loss": 0.7002,
"step": 393
},
{
"epoch": 1.2442105263157894,
"grad_norm": 0.6329587697982788,
"learning_rate": 7.2625658511295635e-06,
"loss": 0.6731,
"step": 394
},
{
"epoch": 1.2473684210526317,
"grad_norm": 0.6213243007659912,
"learning_rate": 7.2461288578234955e-06,
"loss": 0.6926,
"step": 395
},
{
"epoch": 1.2505263157894737,
"grad_norm": 0.6473403573036194,
"learning_rate": 7.229661397084775e-06,
"loss": 0.684,
"step": 396
},
{
"epoch": 1.2536842105263157,
"grad_norm": 0.6111834049224854,
"learning_rate": 7.213163692284943e-06,
"loss": 0.7092,
"step": 397
},
{
"epoch": 1.256842105263158,
"grad_norm": 0.693008303642273,
"learning_rate": 7.196635967205776e-06,
"loss": 0.7175,
"step": 398
},
{
"epoch": 1.26,
"grad_norm": 0.6312058568000793,
"learning_rate": 7.180078446036259e-06,
"loss": 0.6499,
"step": 399
},
{
"epoch": 1.263157894736842,
"grad_norm": 0.6444021463394165,
"learning_rate": 7.163491353369545e-06,
"loss": 0.6972,
"step": 400
},
{
"epoch": 1.2663157894736843,
"grad_norm": 0.6179962158203125,
"learning_rate": 7.146874914199906e-06,
"loss": 0.7037,
"step": 401
},
{
"epoch": 1.2694736842105263,
"grad_norm": 0.6444630026817322,
"learning_rate": 7.130229353919685e-06,
"loss": 0.6928,
"step": 402
},
{
"epoch": 1.2726315789473683,
"grad_norm": 0.666993260383606,
"learning_rate": 7.113554898316231e-06,
"loss": 0.6872,
"step": 403
},
{
"epoch": 1.2757894736842106,
"grad_norm": 0.6936898231506348,
"learning_rate": 7.0968517735688445e-06,
"loss": 0.6788,
"step": 404
},
{
"epoch": 1.2789473684210526,
"grad_norm": 0.5810768604278564,
"learning_rate": 7.080120206245709e-06,
"loss": 0.7012,
"step": 405
},
{
"epoch": 1.2821052631578946,
"grad_norm": 0.6620011925697327,
"learning_rate": 7.063360423300808e-06,
"loss": 0.6811,
"step": 406
},
{
"epoch": 1.2852631578947369,
"grad_norm": 0.6566019058227539,
"learning_rate": 7.04657265207086e-06,
"loss": 0.6951,
"step": 407
},
{
"epoch": 1.288421052631579,
"grad_norm": 0.6208463311195374,
"learning_rate": 7.029757120272222e-06,
"loss": 0.7054,
"step": 408
},
{
"epoch": 1.291578947368421,
"grad_norm": 0.6561064124107361,
"learning_rate": 7.0129140559978184e-06,
"loss": 0.6849,
"step": 409
},
{
"epoch": 1.2947368421052632,
"grad_norm": 0.6156150698661804,
"learning_rate": 6.99604368771402e-06,
"loss": 0.6664,
"step": 410
},
{
"epoch": 1.2978947368421052,
"grad_norm": 0.7128709554672241,
"learning_rate": 6.979146244257573e-06,
"loss": 0.6856,
"step": 411
},
{
"epoch": 1.3010526315789472,
"grad_norm": 0.7587366700172424,
"learning_rate": 6.962221954832476e-06,
"loss": 0.6407,
"step": 412
},
{
"epoch": 1.3042105263157895,
"grad_norm": 0.7253438234329224,
"learning_rate": 6.945271049006882e-06,
"loss": 0.6383,
"step": 413
},
{
"epoch": 1.3073684210526315,
"grad_norm": 0.7263516187667847,
"learning_rate": 6.928293756709976e-06,
"loss": 0.7091,
"step": 414
},
{
"epoch": 1.3105263157894738,
"grad_norm": 0.7259902358055115,
"learning_rate": 6.911290308228861e-06,
"loss": 0.658,
"step": 415
},
{
"epoch": 1.3136842105263158,
"grad_norm": 0.6816960573196411,
"learning_rate": 6.894260934205437e-06,
"loss": 0.6276,
"step": 416
},
{
"epoch": 1.3168421052631578,
"grad_norm": 0.6776388883590698,
"learning_rate": 6.8772058656332626e-06,
"loss": 0.7156,
"step": 417
},
{
"epoch": 1.32,
"grad_norm": 0.658682644367218,
"learning_rate": 6.860125333854437e-06,
"loss": 0.6832,
"step": 418
},
{
"epoch": 1.323157894736842,
"grad_norm": 0.6834143400192261,
"learning_rate": 6.843019570556443e-06,
"loss": 0.7103,
"step": 419
},
{
"epoch": 1.3263157894736843,
"grad_norm": 0.7089201211929321,
"learning_rate": 6.82588880776902e-06,
"loss": 0.7263,
"step": 420
},
{
"epoch": 1.3294736842105264,
"grad_norm": 0.6268842220306396,
"learning_rate": 6.8087332778610116e-06,
"loss": 0.6906,
"step": 421
},
{
"epoch": 1.3326315789473684,
"grad_norm": 0.6012834906578064,
"learning_rate": 6.791553213537209e-06,
"loss": 0.7038,
"step": 422
},
{
"epoch": 1.3357894736842106,
"grad_norm": 0.6678085923194885,
"learning_rate": 6.774348847835203e-06,
"loss": 0.7062,
"step": 423
},
{
"epoch": 1.3389473684210527,
"grad_norm": 0.7389436364173889,
"learning_rate": 6.757120414122214e-06,
"loss": 0.6959,
"step": 424
},
{
"epoch": 1.3421052631578947,
"grad_norm": 0.6289021968841553,
"learning_rate": 6.739868146091934e-06,
"loss": 0.7158,
"step": 425
},
{
"epoch": 1.345263157894737,
"grad_norm": 0.6698178052902222,
"learning_rate": 6.722592277761355e-06,
"loss": 0.6744,
"step": 426
},
{
"epoch": 1.348421052631579,
"grad_norm": 0.7639443874359131,
"learning_rate": 6.705293043467589e-06,
"loss": 0.6726,
"step": 427
},
{
"epoch": 1.351578947368421,
"grad_norm": 0.6818212270736694,
"learning_rate": 6.687970677864696e-06,
"loss": 0.7182,
"step": 428
},
{
"epoch": 1.3547368421052632,
"grad_norm": 0.6343548893928528,
"learning_rate": 6.6706254159205e-06,
"loss": 0.6823,
"step": 429
},
{
"epoch": 1.3578947368421053,
"grad_norm": 0.6729276776313782,
"learning_rate": 6.653257492913398e-06,
"loss": 0.6729,
"step": 430
},
{
"epoch": 1.3610526315789473,
"grad_norm": 0.6306588053703308,
"learning_rate": 6.6358671444291735e-06,
"loss": 0.6753,
"step": 431
},
{
"epoch": 1.3642105263157895,
"grad_norm": 0.5845309495925903,
"learning_rate": 6.618454606357796e-06,
"loss": 0.6878,
"step": 432
},
{
"epoch": 1.3673684210526316,
"grad_norm": 0.7026935815811157,
"learning_rate": 6.601020114890227e-06,
"loss": 0.691,
"step": 433
},
{
"epoch": 1.3705263157894736,
"grad_norm": 0.7399420738220215,
"learning_rate": 6.5835639065152104e-06,
"loss": 0.7315,
"step": 434
},
{
"epoch": 1.3736842105263158,
"grad_norm": 0.6496444344520569,
"learning_rate": 6.56608621801607e-06,
"loss": 0.7043,
"step": 435
},
{
"epoch": 1.3768421052631579,
"grad_norm": 0.6120364665985107,
"learning_rate": 6.548587286467491e-06,
"loss": 0.6685,
"step": 436
},
{
"epoch": 1.38,
"grad_norm": 0.6417269706726074,
"learning_rate": 6.531067349232314e-06,
"loss": 0.6831,
"step": 437
},
{
"epoch": 1.3831578947368421,
"grad_norm": 0.8120979070663452,
"learning_rate": 6.5135266439583015e-06,
"loss": 0.6841,
"step": 438
},
{
"epoch": 1.3863157894736842,
"grad_norm": 0.7373508214950562,
"learning_rate": 6.495965408574929e-06,
"loss": 0.6976,
"step": 439
},
{
"epoch": 1.3894736842105262,
"grad_norm": 0.6089206337928772,
"learning_rate": 6.478383881290152e-06,
"loss": 0.6892,
"step": 440
},
{
"epoch": 1.3926315789473684,
"grad_norm": 0.7098264098167419,
"learning_rate": 6.460782300587166e-06,
"loss": 0.7105,
"step": 441
},
{
"epoch": 1.3957894736842105,
"grad_norm": 0.7377170920372009,
"learning_rate": 6.443160905221188e-06,
"loss": 0.6848,
"step": 442
},
{
"epoch": 1.3989473684210525,
"grad_norm": 0.6914153099060059,
"learning_rate": 6.425519934216204e-06,
"loss": 0.7005,
"step": 443
},
{
"epoch": 1.4021052631578947,
"grad_norm": 0.6577550172805786,
"learning_rate": 6.407859626861734e-06,
"loss": 0.6872,
"step": 444
},
{
"epoch": 1.4052631578947368,
"grad_norm": 0.7401857972145081,
"learning_rate": 6.390180222709583e-06,
"loss": 0.6642,
"step": 445
},
{
"epoch": 1.408421052631579,
"grad_norm": 0.653255820274353,
"learning_rate": 6.372481961570597e-06,
"loss": 0.6641,
"step": 446
},
{
"epoch": 1.411578947368421,
"grad_norm": 0.5870925784111023,
"learning_rate": 6.3547650835114014e-06,
"loss": 0.6783,
"step": 447
},
{
"epoch": 1.4147368421052633,
"grad_norm": 0.7446110248565674,
"learning_rate": 6.337029828851151e-06,
"loss": 0.6965,
"step": 448
},
{
"epoch": 1.4178947368421053,
"grad_norm": 0.7110916972160339,
"learning_rate": 6.319276438158271e-06,
"loss": 0.6985,
"step": 449
},
{
"epoch": 1.4210526315789473,
"grad_norm": 0.6280571222305298,
"learning_rate": 6.301505152247185e-06,
"loss": 0.6855,
"step": 450
},
{
"epoch": 1.4242105263157896,
"grad_norm": 0.6359561681747437,
"learning_rate": 6.283716212175062e-06,
"loss": 0.6676,
"step": 451
},
{
"epoch": 1.4273684210526316,
"grad_norm": 0.670707106590271,
"learning_rate": 6.265909859238536e-06,
"loss": 0.6898,
"step": 452
},
{
"epoch": 1.4305263157894736,
"grad_norm": 0.6176190376281738,
"learning_rate": 6.248086334970435e-06,
"loss": 0.6481,
"step": 453
},
{
"epoch": 1.433684210526316,
"grad_norm": 0.6055241823196411,
"learning_rate": 6.230245881136509e-06,
"loss": 0.6382,
"step": 454
},
{
"epoch": 1.436842105263158,
"grad_norm": 0.6523144245147705,
"learning_rate": 6.2123887397321456e-06,
"loss": 0.7032,
"step": 455
},
{
"epoch": 1.44,
"grad_norm": 0.6355512142181396,
"learning_rate": 6.194515152979093e-06,
"loss": 0.6795,
"step": 456
},
{
"epoch": 1.4431578947368422,
"grad_norm": 0.5738744735717773,
"learning_rate": 6.176625363322164e-06,
"loss": 0.7088,
"step": 457
},
{
"epoch": 1.4463157894736842,
"grad_norm": 0.5846467018127441,
"learning_rate": 6.158719613425964e-06,
"loss": 0.7205,
"step": 458
},
{
"epoch": 1.4494736842105262,
"grad_norm": 0.6887890696525574,
"learning_rate": 6.140798146171581e-06,
"loss": 0.7071,
"step": 459
},
{
"epoch": 1.4526315789473685,
"grad_norm": 0.6898921132087708,
"learning_rate": 6.122861204653304e-06,
"loss": 0.6667,
"step": 460
},
{
"epoch": 1.4557894736842105,
"grad_norm": 0.6479648351669312,
"learning_rate": 6.104909032175323e-06,
"loss": 0.6884,
"step": 461
},
{
"epoch": 1.4589473684210525,
"grad_norm": 0.7151727676391602,
"learning_rate": 6.086941872248424e-06,
"loss": 0.6796,
"step": 462
},
{
"epoch": 1.4621052631578948,
"grad_norm": 0.649958074092865,
"learning_rate": 6.068959968586689e-06,
"loss": 0.7084,
"step": 463
},
{
"epoch": 1.4652631578947368,
"grad_norm": 0.590797483921051,
"learning_rate": 6.050963565104191e-06,
"loss": 0.6626,
"step": 464
},
{
"epoch": 1.4684210526315788,
"grad_norm": 0.6719523072242737,
"learning_rate": 6.032952905911686e-06,
"loss": 0.6946,
"step": 465
},
{
"epoch": 1.471578947368421,
"grad_norm": 0.6793899536132812,
"learning_rate": 6.014928235313301e-06,
"loss": 0.6863,
"step": 466
},
{
"epoch": 1.4747368421052631,
"grad_norm": 0.6513259410858154,
"learning_rate": 5.996889797803214e-06,
"loss": 0.7068,
"step": 467
},
{
"epoch": 1.4778947368421052,
"grad_norm": 0.5904105305671692,
"learning_rate": 5.978837838062348e-06,
"loss": 0.6878,
"step": 468
},
{
"epoch": 1.4810526315789474,
"grad_norm": 0.6193195581436157,
"learning_rate": 5.9607726009550494e-06,
"loss": 0.6824,
"step": 469
},
{
"epoch": 1.4842105263157894,
"grad_norm": 0.6615768074989319,
"learning_rate": 5.942694331525758e-06,
"loss": 0.6957,
"step": 470
},
{
"epoch": 1.4873684210526315,
"grad_norm": 0.5977884531021118,
"learning_rate": 5.924603274995693e-06,
"loss": 0.6732,
"step": 471
},
{
"epoch": 1.4905263157894737,
"grad_norm": 0.6094414591789246,
"learning_rate": 5.906499676759524e-06,
"loss": 0.6896,
"step": 472
},
{
"epoch": 1.4936842105263157,
"grad_norm": 0.5905715227127075,
"learning_rate": 5.88838378238204e-06,
"loss": 0.6964,
"step": 473
},
{
"epoch": 1.496842105263158,
"grad_norm": 0.6121339797973633,
"learning_rate": 5.8702558375948206e-06,
"loss": 0.6704,
"step": 474
},
{
"epoch": 1.5,
"grad_norm": 0.5825003385543823,
"learning_rate": 5.852116088292901e-06,
"loss": 0.6957,
"step": 475
},
{
"epoch": 1.5031578947368422,
"grad_norm": 0.6180065870285034,
"learning_rate": 5.8339647805314404e-06,
"loss": 0.7134,
"step": 476
},
{
"epoch": 1.506315789473684,
"grad_norm": 0.6621575355529785,
"learning_rate": 5.815802160522379e-06,
"loss": 0.6579,
"step": 477
},
{
"epoch": 1.5094736842105263,
"grad_norm": 0.6555248498916626,
"learning_rate": 5.797628474631102e-06,
"loss": 0.6424,
"step": 478
},
{
"epoch": 1.5126315789473685,
"grad_norm": 0.6568667888641357,
"learning_rate": 5.7794439693730975e-06,
"loss": 0.6965,
"step": 479
},
{
"epoch": 1.5157894736842106,
"grad_norm": 0.6549687385559082,
"learning_rate": 5.761248891410613e-06,
"loss": 0.7017,
"step": 480
},
{
"epoch": 1.5189473684210526,
"grad_norm": 0.6165953874588013,
"learning_rate": 5.743043487549306e-06,
"loss": 0.719,
"step": 481
},
{
"epoch": 1.5221052631578948,
"grad_norm": 0.6918132901191711,
"learning_rate": 5.7248280047348995e-06,
"loss": 0.6964,
"step": 482
},
{
"epoch": 1.5252631578947369,
"grad_norm": 0.6982928514480591,
"learning_rate": 5.706602690049832e-06,
"loss": 0.6933,
"step": 483
},
{
"epoch": 1.528421052631579,
"grad_norm": 0.6259163618087769,
"learning_rate": 5.688367790709909e-06,
"loss": 0.7202,
"step": 484
},
{
"epoch": 1.5315789473684212,
"grad_norm": 0.7041128873825073,
"learning_rate": 5.6701235540609405e-06,
"loss": 0.7115,
"step": 485
},
{
"epoch": 1.5347368421052632,
"grad_norm": 0.6209255456924438,
"learning_rate": 5.651870227575391e-06,
"loss": 0.6945,
"step": 486
},
{
"epoch": 1.5378947368421052,
"grad_norm": 0.6310893893241882,
"learning_rate": 5.633608058849033e-06,
"loss": 0.6985,
"step": 487
},
{
"epoch": 1.5410526315789475,
"grad_norm": 0.6265409588813782,
"learning_rate": 5.61533729559757e-06,
"loss": 0.7472,
"step": 488
},
{
"epoch": 1.5442105263157895,
"grad_norm": 0.621372640132904,
"learning_rate": 5.5970581856532864e-06,
"loss": 0.6869,
"step": 489
},
{
"epoch": 1.5473684210526315,
"grad_norm": 0.6584547758102417,
"learning_rate": 5.578770976961685e-06,
"loss": 0.6797,
"step": 490
},
{
"epoch": 1.5505263157894738,
"grad_norm": 0.6084906458854675,
"learning_rate": 5.560475917578129e-06,
"loss": 0.6897,
"step": 491
},
{
"epoch": 1.5536842105263158,
"grad_norm": 0.6554166674613953,
"learning_rate": 5.542173255664463e-06,
"loss": 0.6844,
"step": 492
},
{
"epoch": 1.5568421052631578,
"grad_norm": 0.5814923048019409,
"learning_rate": 5.523863239485661e-06,
"loss": 0.7098,
"step": 493
},
{
"epoch": 1.56,
"grad_norm": 0.6252988576889038,
"learning_rate": 5.505546117406449e-06,
"loss": 0.7227,
"step": 494
},
{
"epoch": 1.563157894736842,
"grad_norm": 0.6618260145187378,
"learning_rate": 5.487222137887949e-06,
"loss": 0.6874,
"step": 495
},
{
"epoch": 1.566315789473684,
"grad_norm": 0.6781257390975952,
"learning_rate": 5.4688915494842886e-06,
"loss": 0.6582,
"step": 496
},
{
"epoch": 1.5694736842105264,
"grad_norm": 0.7620246410369873,
"learning_rate": 5.450554600839251e-06,
"loss": 0.6982,
"step": 497
},
{
"epoch": 1.5726315789473684,
"grad_norm": 0.646678626537323,
"learning_rate": 5.432211540682887e-06,
"loss": 0.6742,
"step": 498
},
{
"epoch": 1.5757894736842104,
"grad_norm": 0.7081868648529053,
"learning_rate": 5.413862617828147e-06,
"loss": 0.656,
"step": 499
},
{
"epoch": 1.5789473684210527,
"grad_norm": 0.6851574182510376,
"learning_rate": 5.395508081167506e-06,
"loss": 0.6885,
"step": 500
},
{
"epoch": 1.582105263157895,
"grad_norm": 0.679471492767334,
"learning_rate": 5.37714817966959e-06,
"loss": 0.6661,
"step": 501
},
{
"epoch": 1.5852631578947367,
"grad_norm": 0.6246220469474792,
"learning_rate": 5.35878316237579e-06,
"loss": 0.6632,
"step": 502
},
{
"epoch": 1.588421052631579,
"grad_norm": 0.6367597579956055,
"learning_rate": 5.340413278396896e-06,
"loss": 0.6997,
"step": 503
},
{
"epoch": 1.5915789473684212,
"grad_norm": 0.7005538940429688,
"learning_rate": 5.322038776909705e-06,
"loss": 0.7041,
"step": 504
},
{
"epoch": 1.594736842105263,
"grad_norm": 0.6432095766067505,
"learning_rate": 5.303659907153654e-06,
"loss": 0.6719,
"step": 505
},
{
"epoch": 1.5978947368421053,
"grad_norm": 0.5566903352737427,
"learning_rate": 5.285276918427432e-06,
"loss": 0.6733,
"step": 506
},
{
"epoch": 1.6010526315789475,
"grad_norm": 0.5807555317878723,
"learning_rate": 5.2668900600855955e-06,
"loss": 0.7036,
"step": 507
},
{
"epoch": 1.6042105263157893,
"grad_norm": 0.5966797471046448,
"learning_rate": 5.248499581535193e-06,
"loss": 0.682,
"step": 508
},
{
"epoch": 1.6073684210526316,
"grad_norm": 0.646507740020752,
"learning_rate": 5.2301057322323786e-06,
"loss": 0.7059,
"step": 509
},
{
"epoch": 1.6105263157894738,
"grad_norm": 0.6112558841705322,
"learning_rate": 5.211708761679031e-06,
"loss": 0.6732,
"step": 510
},
{
"epoch": 1.6136842105263158,
"grad_norm": 0.6396105885505676,
"learning_rate": 5.193308919419363e-06,
"loss": 0.7091,
"step": 511
},
{
"epoch": 1.6168421052631579,
"grad_norm": 0.678657054901123,
"learning_rate": 5.1749064550365414e-06,
"loss": 0.6975,
"step": 512
},
{
"epoch": 1.62,
"grad_norm": 0.6183331608772278,
"learning_rate": 5.156501618149301e-06,
"loss": 0.6832,
"step": 513
},
{
"epoch": 1.6231578947368421,
"grad_norm": 0.6091780066490173,
"learning_rate": 5.13809465840856e-06,
"loss": 0.6911,
"step": 514
},
{
"epoch": 1.6263157894736842,
"grad_norm": 0.6304209232330322,
"learning_rate": 5.11968582549403e-06,
"loss": 0.6461,
"step": 515
},
{
"epoch": 1.6294736842105264,
"grad_norm": 0.6831485629081726,
"learning_rate": 5.10127536911083e-06,
"loss": 0.6757,
"step": 516
},
{
"epoch": 1.6326315789473684,
"grad_norm": 0.6476657390594482,
"learning_rate": 5.082863538986103e-06,
"loss": 0.6935,
"step": 517
},
{
"epoch": 1.6357894736842105,
"grad_norm": 0.6213446259498596,
"learning_rate": 5.064450584865624e-06,
"loss": 0.668,
"step": 518
},
{
"epoch": 1.6389473684210527,
"grad_norm": 0.6051395535469055,
"learning_rate": 5.046036756510417e-06,
"loss": 0.6926,
"step": 519
},
{
"epoch": 1.6421052631578947,
"grad_norm": 0.6724497079849243,
"learning_rate": 5.027622303693363e-06,
"loss": 0.6912,
"step": 520
},
{
"epoch": 1.6452631578947368,
"grad_norm": 0.6436070799827576,
"learning_rate": 5.0092074761958085e-06,
"loss": 0.6791,
"step": 521
},
{
"epoch": 1.648421052631579,
"grad_norm": 0.602382242679596,
"learning_rate": 4.990792523804192e-06,
"loss": 0.7104,
"step": 522
},
{
"epoch": 1.651578947368421,
"grad_norm": 0.6710056662559509,
"learning_rate": 4.972377696306639e-06,
"loss": 0.6412,
"step": 523
},
{
"epoch": 1.654736842105263,
"grad_norm": 0.6339631676673889,
"learning_rate": 4.953963243489583e-06,
"loss": 0.696,
"step": 524
},
{
"epoch": 1.6578947368421053,
"grad_norm": 0.6232373118400574,
"learning_rate": 4.935549415134376e-06,
"loss": 0.6824,
"step": 525
},
{
"epoch": 1.6610526315789473,
"grad_norm": 0.5875207781791687,
"learning_rate": 4.9171364610139e-06,
"loss": 0.7114,
"step": 526
},
{
"epoch": 1.6642105263157894,
"grad_norm": 0.6042991876602173,
"learning_rate": 4.898724630889172e-06,
"loss": 0.6792,
"step": 527
},
{
"epoch": 1.6673684210526316,
"grad_norm": 0.6709006428718567,
"learning_rate": 4.880314174505972e-06,
"loss": 0.6597,
"step": 528
},
{
"epoch": 1.6705263157894736,
"grad_norm": 0.6271777153015137,
"learning_rate": 4.861905341591442e-06,
"loss": 0.6763,
"step": 529
},
{
"epoch": 1.6736842105263157,
"grad_norm": 0.6083865165710449,
"learning_rate": 4.843498381850701e-06,
"loss": 0.6882,
"step": 530
},
{
"epoch": 1.676842105263158,
"grad_norm": 0.544291079044342,
"learning_rate": 4.82509354496346e-06,
"loss": 0.705,
"step": 531
},
{
"epoch": 1.6800000000000002,
"grad_norm": 0.6035980582237244,
"learning_rate": 4.8066910805806384e-06,
"loss": 0.6771,
"step": 532
},
{
"epoch": 1.683157894736842,
"grad_norm": 0.6175979971885681,
"learning_rate": 4.78829123832097e-06,
"loss": 0.7164,
"step": 533
},
{
"epoch": 1.6863157894736842,
"grad_norm": 0.6440612077713013,
"learning_rate": 4.769894267767621e-06,
"loss": 0.6812,
"step": 534
},
{
"epoch": 1.6894736842105265,
"grad_norm": 0.6156215071678162,
"learning_rate": 4.751500418464809e-06,
"loss": 0.6916,
"step": 535
},
{
"epoch": 1.6926315789473683,
"grad_norm": 0.5670990943908691,
"learning_rate": 4.733109939914407e-06,
"loss": 0.6901,
"step": 536
},
{
"epoch": 1.6957894736842105,
"grad_norm": 0.6149737238883972,
"learning_rate": 4.714723081572571e-06,
"loss": 0.6412,
"step": 537
},
{
"epoch": 1.6989473684210528,
"grad_norm": 0.6131264567375183,
"learning_rate": 4.696340092846347e-06,
"loss": 0.6881,
"step": 538
},
{
"epoch": 1.7021052631578948,
"grad_norm": 0.5654208660125732,
"learning_rate": 4.677961223090297e-06,
"loss": 0.7101,
"step": 539
},
{
"epoch": 1.7052631578947368,
"grad_norm": 0.6652705073356628,
"learning_rate": 4.659586721603107e-06,
"loss": 0.7,
"step": 540
},
{
"epoch": 1.708421052631579,
"grad_norm": 0.6081032752990723,
"learning_rate": 4.641216837624211e-06,
"loss": 0.7095,
"step": 541
},
{
"epoch": 1.711578947368421,
"grad_norm": 0.5956637263298035,
"learning_rate": 4.622851820330412e-06,
"loss": 0.6858,
"step": 542
},
{
"epoch": 1.7147368421052631,
"grad_norm": 0.6079306602478027,
"learning_rate": 4.604491918832494e-06,
"loss": 0.6929,
"step": 543
},
{
"epoch": 1.7178947368421054,
"grad_norm": 0.5816709399223328,
"learning_rate": 4.586137382171856e-06,
"loss": 0.6681,
"step": 544
},
{
"epoch": 1.7210526315789474,
"grad_norm": 0.6434882879257202,
"learning_rate": 4.567788459317116e-06,
"loss": 0.6812,
"step": 545
},
{
"epoch": 1.7242105263157894,
"grad_norm": 0.6063170433044434,
"learning_rate": 4.54944539916075e-06,
"loss": 0.6832,
"step": 546
},
{
"epoch": 1.7273684210526317,
"grad_norm": 0.6341161727905273,
"learning_rate": 4.531108450515712e-06,
"loss": 0.692,
"step": 547
},
{
"epoch": 1.7305263157894737,
"grad_norm": 0.6072157621383667,
"learning_rate": 4.512777862112053e-06,
"loss": 0.6582,
"step": 548
},
{
"epoch": 1.7336842105263157,
"grad_norm": 0.5953732132911682,
"learning_rate": 4.494453882593552e-06,
"loss": 0.663,
"step": 549
},
{
"epoch": 1.736842105263158,
"grad_norm": 0.6301018595695496,
"learning_rate": 4.476136760514341e-06,
"loss": 0.6686,
"step": 550
},
{
"epoch": 1.74,
"grad_norm": 0.628161609172821,
"learning_rate": 4.457826744335538e-06,
"loss": 0.6909,
"step": 551
},
{
"epoch": 1.743157894736842,
"grad_norm": 0.5877235531806946,
"learning_rate": 4.439524082421872e-06,
"loss": 0.6515,
"step": 552
},
{
"epoch": 1.7463157894736843,
"grad_norm": 0.6606840491294861,
"learning_rate": 4.421229023038316e-06,
"loss": 0.6738,
"step": 553
},
{
"epoch": 1.7494736842105263,
"grad_norm": 0.5608782768249512,
"learning_rate": 4.402941814346716e-06,
"loss": 0.6991,
"step": 554
},
{
"epoch": 1.7526315789473683,
"grad_norm": 0.5692189335823059,
"learning_rate": 4.384662704402433e-06,
"loss": 0.6547,
"step": 555
},
{
"epoch": 1.7557894736842106,
"grad_norm": 0.589918315410614,
"learning_rate": 4.366391941150969e-06,
"loss": 0.691,
"step": 556
},
{
"epoch": 1.7589473684210526,
"grad_norm": 0.5952433943748474,
"learning_rate": 4.34812977242461e-06,
"loss": 0.6866,
"step": 557
},
{
"epoch": 1.7621052631578946,
"grad_norm": 0.6127429008483887,
"learning_rate": 4.329876445939062e-06,
"loss": 0.7041,
"step": 558
},
{
"epoch": 1.7652631578947369,
"grad_norm": 0.6347705125808716,
"learning_rate": 4.3116322092900925e-06,
"loss": 0.6565,
"step": 559
},
{
"epoch": 1.768421052631579,
"grad_norm": 0.5670740604400635,
"learning_rate": 4.293397309950168e-06,
"loss": 0.7044,
"step": 560
},
{
"epoch": 1.771578947368421,
"grad_norm": 0.6550533175468445,
"learning_rate": 4.275171995265101e-06,
"loss": 0.6867,
"step": 561
},
{
"epoch": 1.7747368421052632,
"grad_norm": 0.5752691030502319,
"learning_rate": 4.256956512450697e-06,
"loss": 0.7015,
"step": 562
},
{
"epoch": 1.7778947368421054,
"grad_norm": 0.6235054731369019,
"learning_rate": 4.238751108589389e-06,
"loss": 0.6954,
"step": 563
},
{
"epoch": 1.7810526315789472,
"grad_norm": 0.6959113478660583,
"learning_rate": 4.220556030626904e-06,
"loss": 0.6945,
"step": 564
},
{
"epoch": 1.7842105263157895,
"grad_norm": 0.5536004304885864,
"learning_rate": 4.202371525368899e-06,
"loss": 0.6849,
"step": 565
},
{
"epoch": 1.7873684210526317,
"grad_norm": 0.5802682042121887,
"learning_rate": 4.184197839477622e-06,
"loss": 0.6898,
"step": 566
},
{
"epoch": 1.7905263157894735,
"grad_norm": 0.6382439136505127,
"learning_rate": 4.166035219468561e-06,
"loss": 0.6784,
"step": 567
},
{
"epoch": 1.7936842105263158,
"grad_norm": 0.5798212289810181,
"learning_rate": 4.1478839117071e-06,
"loss": 0.7353,
"step": 568
},
{
"epoch": 1.796842105263158,
"grad_norm": 0.6376686096191406,
"learning_rate": 4.12974416240518e-06,
"loss": 0.665,
"step": 569
},
{
"epoch": 1.8,
"grad_norm": 0.566200852394104,
"learning_rate": 4.11161621761796e-06,
"loss": 0.6689,
"step": 570
},
{
"epoch": 1.803157894736842,
"grad_norm": 0.6278012990951538,
"learning_rate": 4.093500323240479e-06,
"loss": 0.6716,
"step": 571
},
{
"epoch": 1.8063157894736843,
"grad_norm": 0.5599969625473022,
"learning_rate": 4.075396725004308e-06,
"loss": 0.6789,
"step": 572
},
{
"epoch": 1.8094736842105263,
"grad_norm": 0.6141947507858276,
"learning_rate": 4.057305668474244e-06,
"loss": 0.6696,
"step": 573
},
{
"epoch": 1.8126315789473684,
"grad_norm": 0.6866646409034729,
"learning_rate": 4.039227399044952e-06,
"loss": 0.7151,
"step": 574
},
{
"epoch": 1.8157894736842106,
"grad_norm": 0.5738588571548462,
"learning_rate": 4.021162161937653e-06,
"loss": 0.6942,
"step": 575
},
{
"epoch": 1.8189473684210526,
"grad_norm": 0.6379348039627075,
"learning_rate": 4.003110202196787e-06,
"loss": 0.6812,
"step": 576
},
{
"epoch": 1.8221052631578947,
"grad_norm": 0.6298931837081909,
"learning_rate": 3.985071764686701e-06,
"loss": 0.6711,
"step": 577
},
{
"epoch": 1.825263157894737,
"grad_norm": 0.6079388856887817,
"learning_rate": 3.9670470940883144e-06,
"loss": 0.6989,
"step": 578
},
{
"epoch": 1.828421052631579,
"grad_norm": 0.5470216870307922,
"learning_rate": 3.94903643489581e-06,
"loss": 0.7104,
"step": 579
},
{
"epoch": 1.831578947368421,
"grad_norm": 0.5332590341567993,
"learning_rate": 3.931040031413313e-06,
"loss": 0.6709,
"step": 580
},
{
"epoch": 1.8347368421052632,
"grad_norm": 0.5865632891654968,
"learning_rate": 3.913058127751578e-06,
"loss": 0.6886,
"step": 581
},
{
"epoch": 1.8378947368421052,
"grad_norm": 0.55133056640625,
"learning_rate": 3.895090967824678e-06,
"loss": 0.6543,
"step": 582
},
{
"epoch": 1.8410526315789473,
"grad_norm": 0.5989318490028381,
"learning_rate": 3.877138795346697e-06,
"loss": 0.64,
"step": 583
},
{
"epoch": 1.8442105263157895,
"grad_norm": 0.7691587805747986,
"learning_rate": 3.85920185382842e-06,
"loss": 0.6826,
"step": 584
},
{
"epoch": 1.8473684210526315,
"grad_norm": 0.6429767608642578,
"learning_rate": 3.841280386574037e-06,
"loss": 0.7013,
"step": 585
},
{
"epoch": 1.8505263157894736,
"grad_norm": 0.6208791732788086,
"learning_rate": 3.823374636677837e-06,
"loss": 0.6459,
"step": 586
},
{
"epoch": 1.8536842105263158,
"grad_norm": 0.5830709934234619,
"learning_rate": 3.8054848470209094e-06,
"loss": 0.678,
"step": 587
},
{
"epoch": 1.8568421052631578,
"grad_norm": 0.7179412245750427,
"learning_rate": 3.7876112602678544e-06,
"loss": 0.7003,
"step": 588
},
{
"epoch": 1.8599999999999999,
"grad_norm": 0.72615647315979,
"learning_rate": 3.7697541188634934e-06,
"loss": 0.6505,
"step": 589
},
{
"epoch": 1.8631578947368421,
"grad_norm": 0.6515791416168213,
"learning_rate": 3.7519136650295673e-06,
"loss": 0.7111,
"step": 590
},
{
"epoch": 1.8663157894736844,
"grad_norm": 0.6129830479621887,
"learning_rate": 3.734090140761466e-06,
"loss": 0.6769,
"step": 591
},
{
"epoch": 1.8694736842105262,
"grad_norm": 0.6099033951759338,
"learning_rate": 3.716283787824939e-06,
"loss": 0.6742,
"step": 592
},
{
"epoch": 1.8726315789473684,
"grad_norm": 0.6122466325759888,
"learning_rate": 3.698494847752816e-06,
"loss": 0.6786,
"step": 593
},
{
"epoch": 1.8757894736842107,
"grad_norm": 0.6633381843566895,
"learning_rate": 3.6807235618417314e-06,
"loss": 0.6878,
"step": 594
},
{
"epoch": 1.8789473684210525,
"grad_norm": 0.669661819934845,
"learning_rate": 3.6629701711488485e-06,
"loss": 0.6794,
"step": 595
},
{
"epoch": 1.8821052631578947,
"grad_norm": 0.5703924894332886,
"learning_rate": 3.645234916488599e-06,
"loss": 0.6814,
"step": 596
},
{
"epoch": 1.885263157894737,
"grad_norm": 0.6017104387283325,
"learning_rate": 3.6275180384294033e-06,
"loss": 0.6442,
"step": 597
},
{
"epoch": 1.888421052631579,
"grad_norm": 0.5789917707443237,
"learning_rate": 3.609819777290418e-06,
"loss": 0.6669,
"step": 598
},
{
"epoch": 1.891578947368421,
"grad_norm": 0.6235988140106201,
"learning_rate": 3.5921403731382685e-06,
"loss": 0.636,
"step": 599
},
{
"epoch": 1.8947368421052633,
"grad_norm": 0.6223317384719849,
"learning_rate": 3.5744800657837984e-06,
"loss": 0.6911,
"step": 600
},
{
"epoch": 1.8978947368421053,
"grad_norm": 0.6065788269042969,
"learning_rate": 3.556839094778814e-06,
"loss": 0.6495,
"step": 601
},
{
"epoch": 1.9010526315789473,
"grad_norm": 0.6049926280975342,
"learning_rate": 3.5392176994128357e-06,
"loss": 0.7152,
"step": 602
},
{
"epoch": 1.9042105263157896,
"grad_norm": 0.5944041609764099,
"learning_rate": 3.5216161187098497e-06,
"loss": 0.6835,
"step": 603
},
{
"epoch": 1.9073684210526316,
"grad_norm": 0.6229305863380432,
"learning_rate": 3.504034591425071e-06,
"loss": 0.674,
"step": 604
},
{
"epoch": 1.9105263157894736,
"grad_norm": 0.6037994623184204,
"learning_rate": 3.4864733560416998e-06,
"loss": 0.6526,
"step": 605
},
{
"epoch": 1.9136842105263159,
"grad_norm": 0.6101519465446472,
"learning_rate": 3.468932650767689e-06,
"loss": 0.6693,
"step": 606
},
{
"epoch": 1.916842105263158,
"grad_norm": 0.5631049871444702,
"learning_rate": 3.4514127135325105e-06,
"loss": 0.6967,
"step": 607
},
{
"epoch": 1.92,
"grad_norm": 0.5985934734344482,
"learning_rate": 3.433913781983932e-06,
"loss": 0.6955,
"step": 608
},
{
"epoch": 1.9231578947368422,
"grad_norm": 0.5818445086479187,
"learning_rate": 3.4164360934847912e-06,
"loss": 0.6642,
"step": 609
},
{
"epoch": 1.9263157894736842,
"grad_norm": 0.6571829319000244,
"learning_rate": 3.3989798851097744e-06,
"loss": 0.6848,
"step": 610
},
{
"epoch": 1.9294736842105262,
"grad_norm": 0.5725364685058594,
"learning_rate": 3.381545393642205e-06,
"loss": 0.6863,
"step": 611
},
{
"epoch": 1.9326315789473685,
"grad_norm": 0.6822765469551086,
"learning_rate": 3.3641328555708286e-06,
"loss": 0.7075,
"step": 612
},
{
"epoch": 1.9357894736842105,
"grad_norm": 0.6487634778022766,
"learning_rate": 3.3467425070866034e-06,
"loss": 0.675,
"step": 613
},
{
"epoch": 1.9389473684210525,
"grad_norm": 0.6720603108406067,
"learning_rate": 3.3293745840795004e-06,
"loss": 0.6709,
"step": 614
},
{
"epoch": 1.9421052631578948,
"grad_norm": 0.5721080303192139,
"learning_rate": 3.312029322135306e-06,
"loss": 0.6994,
"step": 615
},
{
"epoch": 1.9452631578947368,
"grad_norm": 0.6841775178909302,
"learning_rate": 3.2947069565324134e-06,
"loss": 0.6535,
"step": 616
},
{
"epoch": 1.9484210526315788,
"grad_norm": 0.6342321038246155,
"learning_rate": 3.2774077222386465e-06,
"loss": 0.6453,
"step": 617
},
{
"epoch": 1.951578947368421,
"grad_norm": 0.667797863483429,
"learning_rate": 3.260131853908066e-06,
"loss": 0.7042,
"step": 618
},
{
"epoch": 1.954736842105263,
"grad_norm": 0.6030858755111694,
"learning_rate": 3.2428795858777873e-06,
"loss": 0.6576,
"step": 619
},
{
"epoch": 1.9578947368421051,
"grad_norm": 0.6035648584365845,
"learning_rate": 3.225651152164799e-06,
"loss": 0.7307,
"step": 620
},
{
"epoch": 1.9610526315789474,
"grad_norm": 0.6750203371047974,
"learning_rate": 3.208446786462791e-06,
"loss": 0.7058,
"step": 621
},
{
"epoch": 1.9642105263157896,
"grad_norm": 0.6008321046829224,
"learning_rate": 3.1912667221389892e-06,
"loss": 0.6851,
"step": 622
},
{
"epoch": 1.9673684210526314,
"grad_norm": 0.6270667910575867,
"learning_rate": 3.1741111922309797e-06,
"loss": 0.6851,
"step": 623
},
{
"epoch": 1.9705263157894737,
"grad_norm": 0.5572888255119324,
"learning_rate": 3.156980429443559e-06,
"loss": 0.6896,
"step": 624
},
{
"epoch": 1.973684210526316,
"grad_norm": 0.6308818459510803,
"learning_rate": 3.1398746661455647e-06,
"loss": 0.7103,
"step": 625
},
{
"epoch": 1.9768421052631577,
"grad_norm": 0.6266995668411255,
"learning_rate": 3.122794134366738e-06,
"loss": 0.6679,
"step": 626
},
{
"epoch": 1.98,
"grad_norm": 0.5579388737678528,
"learning_rate": 3.105739065794565e-06,
"loss": 0.677,
"step": 627
},
{
"epoch": 1.9831578947368422,
"grad_norm": 0.6134604215621948,
"learning_rate": 3.0887096917711408e-06,
"loss": 0.6466,
"step": 628
},
{
"epoch": 1.9863157894736843,
"grad_norm": 0.6037647128105164,
"learning_rate": 3.071706243290026e-06,
"loss": 0.6906,
"step": 629
},
{
"epoch": 1.9894736842105263,
"grad_norm": 0.636805534362793,
"learning_rate": 3.0547289509931194e-06,
"loss": 0.6838,
"step": 630
},
{
"epoch": 1.9926315789473685,
"grad_norm": 0.6072667837142944,
"learning_rate": 3.0377780451675243e-06,
"loss": 0.6751,
"step": 631
},
{
"epoch": 1.9957894736842106,
"grad_norm": 0.5715926885604858,
"learning_rate": 3.020853755742428e-06,
"loss": 0.6666,
"step": 632
},
{
"epoch": 1.9989473684210526,
"grad_norm": 0.634506106376648,
"learning_rate": 3.0039563122859815e-06,
"loss": 0.706,
"step": 633
},
{
"epoch": 2.002105263157895,
"grad_norm": 0.6592942476272583,
"learning_rate": 2.9870859440021845e-06,
"loss": 0.6761,
"step": 634
},
{
"epoch": 2.0052631578947366,
"grad_norm": 0.7126978635787964,
"learning_rate": 2.970242879727778e-06,
"loss": 0.6233,
"step": 635
},
{
"epoch": 2.008421052631579,
"grad_norm": 0.5990601778030396,
"learning_rate": 2.953427347929142e-06,
"loss": 0.6303,
"step": 636
},
{
"epoch": 2.011578947368421,
"grad_norm": 0.5881582498550415,
"learning_rate": 2.936639576699194e-06,
"loss": 0.5866,
"step": 637
},
{
"epoch": 2.014736842105263,
"grad_norm": 0.6993144750595093,
"learning_rate": 2.9198797937542935e-06,
"loss": 0.6256,
"step": 638
},
{
"epoch": 2.017894736842105,
"grad_norm": 0.6200084686279297,
"learning_rate": 2.903148226431155e-06,
"loss": 0.6453,
"step": 639
},
{
"epoch": 2.0210526315789474,
"grad_norm": 0.6157467365264893,
"learning_rate": 2.8864451016837703e-06,
"loss": 0.606,
"step": 640
},
{
"epoch": 2.0242105263157897,
"grad_norm": 0.6459230780601501,
"learning_rate": 2.869770646080316e-06,
"loss": 0.6267,
"step": 641
},
{
"epoch": 2.0273684210526315,
"grad_norm": 0.7284769415855408,
"learning_rate": 2.853125085800096e-06,
"loss": 0.6036,
"step": 642
},
{
"epoch": 2.0305263157894737,
"grad_norm": 0.65320885181427,
"learning_rate": 2.836508646630457e-06,
"loss": 0.6798,
"step": 643
},
{
"epoch": 2.033684210526316,
"grad_norm": 0.6911686658859253,
"learning_rate": 2.8199215539637427e-06,
"loss": 0.6431,
"step": 644
},
{
"epoch": 2.036842105263158,
"grad_norm": 0.6208088994026184,
"learning_rate": 2.8033640327942235e-06,
"loss": 0.673,
"step": 645
},
{
"epoch": 2.04,
"grad_norm": 0.6126965284347534,
"learning_rate": 2.786836307715056e-06,
"loss": 0.6152,
"step": 646
},
{
"epoch": 2.0431578947368423,
"grad_norm": 0.7363206148147583,
"learning_rate": 2.7703386029152246e-06,
"loss": 0.6492,
"step": 647
},
{
"epoch": 2.046315789473684,
"grad_norm": 0.6911171078681946,
"learning_rate": 2.753871142176506e-06,
"loss": 0.6315,
"step": 648
},
{
"epoch": 2.0494736842105263,
"grad_norm": 0.6493732929229736,
"learning_rate": 2.737434148870437e-06,
"loss": 0.6249,
"step": 649
},
{
"epoch": 2.0526315789473686,
"grad_norm": 0.6811050772666931,
"learning_rate": 2.7210278459552786e-06,
"loss": 0.6038,
"step": 650
},
{
"epoch": 2.0557894736842104,
"grad_norm": 0.6253058910369873,
"learning_rate": 2.704652455972997e-06,
"loss": 0.6369,
"step": 651
},
{
"epoch": 2.0589473684210526,
"grad_norm": 0.5684403777122498,
"learning_rate": 2.688308201046236e-06,
"loss": 0.6551,
"step": 652
},
{
"epoch": 2.062105263157895,
"grad_norm": 0.6057170629501343,
"learning_rate": 2.6719953028753214e-06,
"loss": 0.5905,
"step": 653
},
{
"epoch": 2.0652631578947367,
"grad_norm": 0.603461742401123,
"learning_rate": 2.655713982735234e-06,
"loss": 0.6049,
"step": 654
},
{
"epoch": 2.068421052631579,
"grad_norm": 0.5477900505065918,
"learning_rate": 2.6394644614726215e-06,
"loss": 0.6091,
"step": 655
},
{
"epoch": 2.071578947368421,
"grad_norm": 0.6187466382980347,
"learning_rate": 2.623246959502795e-06,
"loss": 0.6247,
"step": 656
},
{
"epoch": 2.074736842105263,
"grad_norm": 0.5489943623542786,
"learning_rate": 2.6070616968067446e-06,
"loss": 0.6522,
"step": 657
},
{
"epoch": 2.0778947368421052,
"grad_norm": 0.5987683534622192,
"learning_rate": 2.5909088929281534e-06,
"loss": 0.613,
"step": 658
},
{
"epoch": 2.0810526315789475,
"grad_norm": 0.5765239000320435,
"learning_rate": 2.574788766970418e-06,
"loss": 0.6277,
"step": 659
},
{
"epoch": 2.0842105263157893,
"grad_norm": 0.566733717918396,
"learning_rate": 2.55870153759368e-06,
"loss": 0.5948,
"step": 660
},
{
"epoch": 2.0873684210526315,
"grad_norm": 0.5783083438873291,
"learning_rate": 2.542647423011857e-06,
"loss": 0.6302,
"step": 661
},
{
"epoch": 2.090526315789474,
"grad_norm": 0.6213259696960449,
"learning_rate": 2.526626640989683e-06,
"loss": 0.607,
"step": 662
},
{
"epoch": 2.0936842105263156,
"grad_norm": 0.5684741735458374,
"learning_rate": 2.510639408839757e-06,
"loss": 0.6122,
"step": 663
},
{
"epoch": 2.096842105263158,
"grad_norm": 0.5746645927429199,
"learning_rate": 2.4946859434195904e-06,
"loss": 0.6336,
"step": 664
},
{
"epoch": 2.1,
"grad_norm": 0.554362416267395,
"learning_rate": 2.478766461128672e-06,
"loss": 0.6155,
"step": 665
},
{
"epoch": 2.103157894736842,
"grad_norm": 0.5717495679855347,
"learning_rate": 2.4628811779055277e-06,
"loss": 0.6356,
"step": 666
},
{
"epoch": 2.106315789473684,
"grad_norm": 0.5523346662521362,
"learning_rate": 2.4470303092247926e-06,
"loss": 0.5855,
"step": 667
},
{
"epoch": 2.1094736842105264,
"grad_norm": 0.5319191813468933,
"learning_rate": 2.431214070094289e-06,
"loss": 0.5865,
"step": 668
},
{
"epoch": 2.1126315789473686,
"grad_norm": 0.52125483751297,
"learning_rate": 2.4154326750521084e-06,
"loss": 0.6155,
"step": 669
},
{
"epoch": 2.1157894736842104,
"grad_norm": 0.5644091963768005,
"learning_rate": 2.3996863381637046e-06,
"loss": 0.5797,
"step": 670
},
{
"epoch": 2.1189473684210527,
"grad_norm": 0.5224770307540894,
"learning_rate": 2.383975273018986e-06,
"loss": 0.6275,
"step": 671
},
{
"epoch": 2.122105263157895,
"grad_norm": 0.5387361645698547,
"learning_rate": 2.3682996927294216e-06,
"loss": 0.5888,
"step": 672
},
{
"epoch": 2.1252631578947367,
"grad_norm": 0.6090015172958374,
"learning_rate": 2.3526598099251473e-06,
"loss": 0.5825,
"step": 673
},
{
"epoch": 2.128421052631579,
"grad_norm": 0.5374984741210938,
"learning_rate": 2.3370558367520856e-06,
"loss": 0.6262,
"step": 674
},
{
"epoch": 2.1315789473684212,
"grad_norm": 0.581964373588562,
"learning_rate": 2.321487984869064e-06,
"loss": 0.5949,
"step": 675
},
{
"epoch": 2.134736842105263,
"grad_norm": 0.5389418601989746,
"learning_rate": 2.305956465444945e-06,
"loss": 0.6104,
"step": 676
},
{
"epoch": 2.1378947368421053,
"grad_norm": 0.5349326729774475,
"learning_rate": 2.290461489155768e-06,
"loss": 0.6828,
"step": 677
},
{
"epoch": 2.1410526315789475,
"grad_norm": 0.5772843360900879,
"learning_rate": 2.275003266181877e-06,
"loss": 0.5864,
"step": 678
},
{
"epoch": 2.1442105263157893,
"grad_norm": 0.6210830211639404,
"learning_rate": 2.2595820062050854e-06,
"loss": 0.6047,
"step": 679
},
{
"epoch": 2.1473684210526316,
"grad_norm": 0.5432112812995911,
"learning_rate": 2.2441979184058223e-06,
"loss": 0.6402,
"step": 680
},
{
"epoch": 2.150526315789474,
"grad_norm": 0.5438613891601562,
"learning_rate": 2.2288512114602986e-06,
"loss": 0.5905,
"step": 681
},
{
"epoch": 2.1536842105263156,
"grad_norm": 0.5581168532371521,
"learning_rate": 2.213542093537675e-06,
"loss": 0.5887,
"step": 682
},
{
"epoch": 2.156842105263158,
"grad_norm": 0.5446808338165283,
"learning_rate": 2.1982707722972383e-06,
"loss": 0.6573,
"step": 683
},
{
"epoch": 2.16,
"grad_norm": 0.6228784322738647,
"learning_rate": 2.1830374548855905e-06,
"loss": 0.6083,
"step": 684
},
{
"epoch": 2.163157894736842,
"grad_norm": 0.577155351638794,
"learning_rate": 2.167842347933826e-06,
"loss": 0.614,
"step": 685
},
{
"epoch": 2.166315789473684,
"grad_norm": 0.560425877571106,
"learning_rate": 2.1526856575547444e-06,
"loss": 0.6321,
"step": 686
},
{
"epoch": 2.1694736842105264,
"grad_norm": 0.547867476940155,
"learning_rate": 2.1375675893400373e-06,
"loss": 0.6192,
"step": 687
},
{
"epoch": 2.1726315789473682,
"grad_norm": 0.5950310826301575,
"learning_rate": 2.1224883483575166e-06,
"loss": 0.6292,
"step": 688
},
{
"epoch": 2.1757894736842105,
"grad_norm": 0.595939576625824,
"learning_rate": 2.1074481391483233e-06,
"loss": 0.6098,
"step": 689
},
{
"epoch": 2.1789473684210527,
"grad_norm": 0.6173756122589111,
"learning_rate": 2.0924471657241526e-06,
"loss": 0.6209,
"step": 690
},
{
"epoch": 2.1821052631578945,
"grad_norm": 0.5339811444282532,
"learning_rate": 2.0774856315644955e-06,
"loss": 0.6303,
"step": 691
},
{
"epoch": 2.185263157894737,
"grad_norm": 0.5910825133323669,
"learning_rate": 2.0625637396138666e-06,
"loss": 0.578,
"step": 692
},
{
"epoch": 2.188421052631579,
"grad_norm": 0.574038565158844,
"learning_rate": 2.0476816922790575e-06,
"loss": 0.5743,
"step": 693
},
{
"epoch": 2.191578947368421,
"grad_norm": 0.537571370601654,
"learning_rate": 2.0328396914263925e-06,
"loss": 0.6386,
"step": 694
},
{
"epoch": 2.194736842105263,
"grad_norm": 0.562436580657959,
"learning_rate": 2.0180379383789907e-06,
"loss": 0.6322,
"step": 695
},
{
"epoch": 2.1978947368421053,
"grad_norm": 0.5552737712860107,
"learning_rate": 2.0032766339140246e-06,
"loss": 0.6069,
"step": 696
},
{
"epoch": 2.201052631578947,
"grad_norm": 0.5092212557792664,
"learning_rate": 1.988555978260013e-06,
"loss": 0.6607,
"step": 697
},
{
"epoch": 2.2042105263157894,
"grad_norm": 0.558074414730072,
"learning_rate": 1.973876171094097e-06,
"loss": 0.6178,
"step": 698
},
{
"epoch": 2.2073684210526316,
"grad_norm": 0.5705953240394592,
"learning_rate": 1.9592374115393293e-06,
"loss": 0.6321,
"step": 699
},
{
"epoch": 2.2105263157894735,
"grad_norm": 0.5584784150123596,
"learning_rate": 1.9446398981619757e-06,
"loss": 0.5915,
"step": 700
},
{
"epoch": 2.2136842105263157,
"grad_norm": 0.5504412055015564,
"learning_rate": 1.9300838289688216e-06,
"loss": 0.6143,
"step": 701
},
{
"epoch": 2.216842105263158,
"grad_norm": 0.5174974203109741,
"learning_rate": 1.915569401404488e-06,
"loss": 0.5982,
"step": 702
},
{
"epoch": 2.22,
"grad_norm": 0.5328599810600281,
"learning_rate": 1.9010968123487478e-06,
"loss": 0.6022,
"step": 703
},
{
"epoch": 2.223157894736842,
"grad_norm": 0.5421804785728455,
"learning_rate": 1.8866662581138646e-06,
"loss": 0.5817,
"step": 704
},
{
"epoch": 2.2263157894736842,
"grad_norm": 0.57915198802948,
"learning_rate": 1.8722779344419139e-06,
"loss": 0.613,
"step": 705
},
{
"epoch": 2.2294736842105265,
"grad_norm": 0.5778079628944397,
"learning_rate": 1.8579320365021508e-06,
"loss": 0.6301,
"step": 706
},
{
"epoch": 2.2326315789473683,
"grad_norm": 0.5267239212989807,
"learning_rate": 1.8436287588883416e-06,
"loss": 0.6269,
"step": 707
},
{
"epoch": 2.2357894736842105,
"grad_norm": 0.5386923551559448,
"learning_rate": 1.8293682956161357e-06,
"loss": 0.5903,
"step": 708
},
{
"epoch": 2.238947368421053,
"grad_norm": 0.5258237719535828,
"learning_rate": 1.8151508401204298e-06,
"loss": 0.6086,
"step": 709
},
{
"epoch": 2.2421052631578946,
"grad_norm": 0.5495336055755615,
"learning_rate": 1.800976585252745e-06,
"loss": 0.6247,
"step": 710
},
{
"epoch": 2.245263157894737,
"grad_norm": 0.55051589012146,
"learning_rate": 1.7868457232786117e-06,
"loss": 0.5962,
"step": 711
},
{
"epoch": 2.248421052631579,
"grad_norm": 0.5342631936073303,
"learning_rate": 1.7727584458749608e-06,
"loss": 0.587,
"step": 712
},
{
"epoch": 2.251578947368421,
"grad_norm": 0.5018291473388672,
"learning_rate": 1.7587149441275236e-06,
"loss": 0.6476,
"step": 713
},
{
"epoch": 2.254736842105263,
"grad_norm": 0.5503069758415222,
"learning_rate": 1.7447154085282398e-06,
"loss": 0.613,
"step": 714
},
{
"epoch": 2.2578947368421054,
"grad_norm": 0.5181307196617126,
"learning_rate": 1.7307600289726745e-06,
"loss": 0.6221,
"step": 715
},
{
"epoch": 2.261052631578947,
"grad_norm": 0.5215415358543396,
"learning_rate": 1.7168489947574407e-06,
"loss": 0.6311,
"step": 716
},
{
"epoch": 2.2642105263157895,
"grad_norm": 0.559654712677002,
"learning_rate": 1.7029824945776346e-06,
"loss": 0.5764,
"step": 717
},
{
"epoch": 2.2673684210526317,
"grad_norm": 0.5514314770698547,
"learning_rate": 1.6891607165242718e-06,
"loss": 0.629,
"step": 718
},
{
"epoch": 2.2705263157894735,
"grad_norm": 0.4893876910209656,
"learning_rate": 1.6753838480817397e-06,
"loss": 0.6328,
"step": 719
},
{
"epoch": 2.2736842105263158,
"grad_norm": 0.5276650190353394,
"learning_rate": 1.661652076125252e-06,
"loss": 0.5912,
"step": 720
},
{
"epoch": 2.276842105263158,
"grad_norm": 0.5329832434654236,
"learning_rate": 1.6479655869183142e-06,
"loss": 0.61,
"step": 721
},
{
"epoch": 2.2800000000000002,
"grad_norm": 0.535116970539093,
"learning_rate": 1.6343245661102031e-06,
"loss": 0.6346,
"step": 722
},
{
"epoch": 2.283157894736842,
"grad_norm": 0.5258433818817139,
"learning_rate": 1.620729198733434e-06,
"loss": 0.6336,
"step": 723
},
{
"epoch": 2.2863157894736843,
"grad_norm": 0.5541167259216309,
"learning_rate": 1.6071796692012663e-06,
"loss": 0.6221,
"step": 724
},
{
"epoch": 2.2894736842105265,
"grad_norm": 0.5115761160850525,
"learning_rate": 1.5936761613051937e-06,
"loss": 0.6489,
"step": 725
},
{
"epoch": 2.2926315789473684,
"grad_norm": 0.5598406195640564,
"learning_rate": 1.580218858212454e-06,
"loss": 0.6,
"step": 726
},
{
"epoch": 2.2957894736842106,
"grad_norm": 0.5273876190185547,
"learning_rate": 1.5668079424635424e-06,
"loss": 0.6227,
"step": 727
},
{
"epoch": 2.298947368421053,
"grad_norm": 0.5451186299324036,
"learning_rate": 1.5534435959697363e-06,
"loss": 0.5881,
"step": 728
},
{
"epoch": 2.3021052631578947,
"grad_norm": 0.5124930739402771,
"learning_rate": 1.5401260000106321e-06,
"loss": 0.6739,
"step": 729
},
{
"epoch": 2.305263157894737,
"grad_norm": 0.5419645309448242,
"learning_rate": 1.526855335231679e-06,
"loss": 0.6209,
"step": 730
},
{
"epoch": 2.308421052631579,
"grad_norm": 0.5517361760139465,
"learning_rate": 1.5136317816417333e-06,
"loss": 0.6345,
"step": 731
},
{
"epoch": 2.311578947368421,
"grad_norm": 0.5226176977157593,
"learning_rate": 1.5004555186106124e-06,
"loss": 0.6008,
"step": 732
},
{
"epoch": 2.314736842105263,
"grad_norm": 0.5727024078369141,
"learning_rate": 1.487326724866668e-06,
"loss": 0.5842,
"step": 733
},
{
"epoch": 2.3178947368421055,
"grad_norm": 0.5469027161598206,
"learning_rate": 1.4742455784943576e-06,
"loss": 0.6151,
"step": 734
},
{
"epoch": 2.3210526315789473,
"grad_norm": 0.5690147280693054,
"learning_rate": 1.4612122569318282e-06,
"loss": 0.566,
"step": 735
},
{
"epoch": 2.3242105263157895,
"grad_norm": 0.5304024815559387,
"learning_rate": 1.448226936968517e-06,
"loss": 0.6047,
"step": 736
},
{
"epoch": 2.3273684210526318,
"grad_norm": 0.5012169480323792,
"learning_rate": 1.4352897947427396e-06,
"loss": 0.6352,
"step": 737
},
{
"epoch": 2.3305263157894736,
"grad_norm": 0.49926042556762695,
"learning_rate": 1.422401005739314e-06,
"loss": 0.6673,
"step": 738
},
{
"epoch": 2.333684210526316,
"grad_norm": 0.5833967924118042,
"learning_rate": 1.4095607447871711e-06,
"loss": 0.5515,
"step": 739
},
{
"epoch": 2.336842105263158,
"grad_norm": 0.6136566996574402,
"learning_rate": 1.3967691860569915e-06,
"loss": 0.5947,
"step": 740
},
{
"epoch": 2.34,
"grad_norm": 0.5746322274208069,
"learning_rate": 1.3840265030588323e-06,
"loss": 0.6308,
"step": 741
},
{
"epoch": 2.343157894736842,
"grad_norm": 0.5507466197013855,
"learning_rate": 1.3713328686397832e-06,
"loss": 0.6041,
"step": 742
},
{
"epoch": 2.3463157894736844,
"grad_norm": 0.524361252784729,
"learning_rate": 1.358688454981621e-06,
"loss": 0.5928,
"step": 743
},
{
"epoch": 2.349473684210526,
"grad_norm": 0.5451714396476746,
"learning_rate": 1.3460934335984677e-06,
"loss": 0.6086,
"step": 744
},
{
"epoch": 2.3526315789473684,
"grad_norm": 0.5549687147140503,
"learning_rate": 1.3335479753344688e-06,
"loss": 0.6241,
"step": 745
},
{
"epoch": 2.3557894736842107,
"grad_norm": 0.5424072742462158,
"learning_rate": 1.3210522503614753e-06,
"loss": 0.6078,
"step": 746
},
{
"epoch": 2.3589473684210525,
"grad_norm": 0.5681266188621521,
"learning_rate": 1.3086064281767346e-06,
"loss": 0.5994,
"step": 747
},
{
"epoch": 2.3621052631578947,
"grad_norm": 0.5240947008132935,
"learning_rate": 1.2962106776005917e-06,
"loss": 0.6148,
"step": 748
},
{
"epoch": 2.365263157894737,
"grad_norm": 0.5335183143615723,
"learning_rate": 1.2838651667742014e-06,
"loss": 0.6244,
"step": 749
},
{
"epoch": 2.3684210526315788,
"grad_norm": 0.5286924839019775,
"learning_rate": 1.2715700631572387e-06,
"loss": 0.6306,
"step": 750
},
{
"epoch": 2.371578947368421,
"grad_norm": 0.527306854724884,
"learning_rate": 1.2593255335256438e-06,
"loss": 0.64,
"step": 751
},
{
"epoch": 2.3747368421052633,
"grad_norm": 0.5724322199821472,
"learning_rate": 1.2471317439693436e-06,
"loss": 0.6075,
"step": 752
},
{
"epoch": 2.377894736842105,
"grad_norm": 0.5209251046180725,
"learning_rate": 1.2349888598900078e-06,
"loss": 0.5929,
"step": 753
},
{
"epoch": 2.3810526315789473,
"grad_norm": 0.53974848985672,
"learning_rate": 1.2228970459988015e-06,
"loss": 0.5841,
"step": 754
},
{
"epoch": 2.3842105263157896,
"grad_norm": 0.5406898260116577,
"learning_rate": 1.2108564663141541e-06,
"loss": 0.6004,
"step": 755
},
{
"epoch": 2.3873684210526314,
"grad_norm": 0.5027428269386292,
"learning_rate": 1.1988672841595312e-06,
"loss": 0.6057,
"step": 756
},
{
"epoch": 2.3905263157894736,
"grad_norm": 0.5120511054992676,
"learning_rate": 1.186929662161221e-06,
"loss": 0.6129,
"step": 757
},
{
"epoch": 2.393684210526316,
"grad_norm": 0.5876862406730652,
"learning_rate": 1.1750437622461293e-06,
"loss": 0.5984,
"step": 758
},
{
"epoch": 2.3968421052631577,
"grad_norm": 0.5318094491958618,
"learning_rate": 1.1632097456395802e-06,
"loss": 0.631,
"step": 759
},
{
"epoch": 2.4,
"grad_norm": 0.48409977555274963,
"learning_rate": 1.1514277728631323e-06,
"loss": 0.6322,
"step": 760
},
{
"epoch": 2.403157894736842,
"grad_norm": 0.5341342687606812,
"learning_rate": 1.1396980037324e-06,
"loss": 0.6,
"step": 761
},
{
"epoch": 2.406315789473684,
"grad_norm": 0.5514808893203735,
"learning_rate": 1.128020597354884e-06,
"loss": 0.6348,
"step": 762
},
{
"epoch": 2.409473684210526,
"grad_norm": 0.5363562107086182,
"learning_rate": 1.1163957121278163e-06,
"loss": 0.5762,
"step": 763
},
{
"epoch": 2.4126315789473685,
"grad_norm": 0.5589635968208313,
"learning_rate": 1.104823505736009e-06,
"loss": 0.6406,
"step": 764
},
{
"epoch": 2.4157894736842107,
"grad_norm": 0.5272628664970398,
"learning_rate": 1.093304135149717e-06,
"loss": 0.5923,
"step": 765
},
{
"epoch": 2.4189473684210525,
"grad_norm": 0.54401695728302,
"learning_rate": 1.0818377566225075e-06,
"loss": 0.5936,
"step": 766
},
{
"epoch": 2.4221052631578948,
"grad_norm": 0.5256767272949219,
"learning_rate": 1.070424525689142e-06,
"loss": 0.6173,
"step": 767
},
{
"epoch": 2.425263157894737,
"grad_norm": 0.5305107831954956,
"learning_rate": 1.0590645971634655e-06,
"loss": 0.5995,
"step": 768
},
{
"epoch": 2.428421052631579,
"grad_norm": 0.5133773684501648,
"learning_rate": 1.0477581251363066e-06,
"loss": 0.6294,
"step": 769
},
{
"epoch": 2.431578947368421,
"grad_norm": 0.5675320625305176,
"learning_rate": 1.0365052629733884e-06,
"loss": 0.5941,
"step": 770
},
{
"epoch": 2.4347368421052633,
"grad_norm": 0.5107877254486084,
"learning_rate": 1.025306163313246e-06,
"loss": 0.6195,
"step": 771
},
{
"epoch": 2.437894736842105,
"grad_norm": 0.5630270838737488,
"learning_rate": 1.0141609780651585e-06,
"loss": 0.6353,
"step": 772
},
{
"epoch": 2.4410526315789474,
"grad_norm": 0.4963136911392212,
"learning_rate": 1.0030698584070848e-06,
"loss": 0.6019,
"step": 773
},
{
"epoch": 2.4442105263157896,
"grad_norm": 0.5430152416229248,
"learning_rate": 9.92032954783621e-07,
"loss": 0.6518,
"step": 774
},
{
"epoch": 2.4473684210526314,
"grad_norm": 0.5275354981422424,
"learning_rate": 9.81050416903951e-07,
"loss": 0.6547,
"step": 775
},
{
"epoch": 2.4505263157894737,
"grad_norm": 0.5483253598213196,
"learning_rate": 9.701223937398152e-07,
"loss": 0.634,
"step": 776
},
{
"epoch": 2.453684210526316,
"grad_norm": 0.5382000803947449,
"learning_rate": 9.592490335234993e-07,
"loss": 0.648,
"step": 777
},
{
"epoch": 2.4568421052631577,
"grad_norm": 0.5335687398910522,
"learning_rate": 9.484304837458158e-07,
"loss": 0.5985,
"step": 778
},
{
"epoch": 2.46,
"grad_norm": 0.5186282992362976,
"learning_rate": 9.376668911541042e-07,
"loss": 0.6092,
"step": 779
},
{
"epoch": 2.463157894736842,
"grad_norm": 0.517254114151001,
"learning_rate": 9.269584017502431e-07,
"loss": 0.6524,
"step": 780
},
{
"epoch": 2.466315789473684,
"grad_norm": 0.5538238286972046,
"learning_rate": 9.163051607886703e-07,
"loss": 0.6008,
"step": 781
},
{
"epoch": 2.4694736842105263,
"grad_norm": 0.5268839597702026,
"learning_rate": 9.057073127744065e-07,
"loss": 0.6115,
"step": 782
},
{
"epoch": 2.4726315789473685,
"grad_norm": 0.5400952696800232,
"learning_rate": 8.951650014611019e-07,
"loss": 0.5816,
"step": 783
},
{
"epoch": 2.4757894736842108,
"grad_norm": 0.571423351764679,
"learning_rate": 8.846783698490835e-07,
"loss": 0.6134,
"step": 784
},
{
"epoch": 2.4789473684210526,
"grad_norm": 0.5169550180435181,
"learning_rate": 8.742475601834133e-07,
"loss": 0.6033,
"step": 785
},
{
"epoch": 2.482105263157895,
"grad_norm": 0.5374566912651062,
"learning_rate": 8.638727139519637e-07,
"loss": 0.597,
"step": 786
},
{
"epoch": 2.485263157894737,
"grad_norm": 0.5302397012710571,
"learning_rate": 8.535539718834929e-07,
"loss": 0.6735,
"step": 787
},
{
"epoch": 2.488421052631579,
"grad_norm": 0.5013649463653564,
"learning_rate": 8.432914739457432e-07,
"loss": 0.6714,
"step": 788
},
{
"epoch": 2.491578947368421,
"grad_norm": 0.5414134860038757,
"learning_rate": 8.330853593435345e-07,
"loss": 0.6105,
"step": 789
},
{
"epoch": 2.4947368421052634,
"grad_norm": 0.5048105120658875,
"learning_rate": 8.229357665168791e-07,
"loss": 0.6186,
"step": 790
},
{
"epoch": 2.497894736842105,
"grad_norm": 0.5406049489974976,
"learning_rate": 8.12842833139107e-07,
"loss": 0.6146,
"step": 791
},
{
"epoch": 2.5010526315789474,
"grad_norm": 0.5164625644683838,
"learning_rate": 8.028066961149921e-07,
"loss": 0.6194,
"step": 792
},
{
"epoch": 2.5042105263157897,
"grad_norm": 0.5256004929542542,
"learning_rate": 7.928274915789035e-07,
"loss": 0.6136,
"step": 793
},
{
"epoch": 2.5073684210526315,
"grad_norm": 0.5570123791694641,
"learning_rate": 7.829053548929488e-07,
"loss": 0.6325,
"step": 794
},
{
"epoch": 2.5105263157894737,
"grad_norm": 0.5029011964797974,
"learning_rate": 7.730404206451459e-07,
"loss": 0.599,
"step": 795
},
{
"epoch": 2.513684210526316,
"grad_norm": 0.5093344449996948,
"learning_rate": 7.632328226475971e-07,
"loss": 0.6007,
"step": 796
},
{
"epoch": 2.5168421052631578,
"grad_norm": 0.4984530508518219,
"learning_rate": 7.53482693934669e-07,
"loss": 0.639,
"step": 797
},
{
"epoch": 2.52,
"grad_norm": 0.5288375616073608,
"learning_rate": 7.437901667611908e-07,
"loss": 0.6021,
"step": 798
},
{
"epoch": 2.5231578947368423,
"grad_norm": 0.5130994915962219,
"learning_rate": 7.341553726006611e-07,
"loss": 0.6125,
"step": 799
},
{
"epoch": 2.526315789473684,
"grad_norm": 0.5277830958366394,
"learning_rate": 7.245784421434643e-07,
"loss": 0.6388,
"step": 800
},
{
"epoch": 2.5294736842105263,
"grad_norm": 0.5155095458030701,
"learning_rate": 7.150595052950954e-07,
"loss": 0.6319,
"step": 801
},
{
"epoch": 2.5326315789473686,
"grad_norm": 0.5809716582298279,
"learning_rate": 7.055986911744017e-07,
"loss": 0.578,
"step": 802
},
{
"epoch": 2.5357894736842104,
"grad_norm": 0.5170502066612244,
"learning_rate": 6.961961281118285e-07,
"loss": 0.6074,
"step": 803
},
{
"epoch": 2.5389473684210526,
"grad_norm": 0.5173423290252686,
"learning_rate": 6.868519436476795e-07,
"loss": 0.631,
"step": 804
},
{
"epoch": 2.542105263157895,
"grad_norm": 0.49472615122795105,
"learning_rate": 6.775662645303871e-07,
"loss": 0.6439,
"step": 805
},
{
"epoch": 2.5452631578947367,
"grad_norm": 0.5428845882415771,
"learning_rate": 6.683392167147917e-07,
"loss": 0.6203,
"step": 806
},
{
"epoch": 2.548421052631579,
"grad_norm": 0.5209921002388,
"learning_rate": 6.591709253604356e-07,
"loss": 0.6588,
"step": 807
},
{
"epoch": 2.551578947368421,
"grad_norm": 0.5301423668861389,
"learning_rate": 6.500615148298617e-07,
"loss": 0.6003,
"step": 808
},
{
"epoch": 2.554736842105263,
"grad_norm": 0.5289193391799927,
"learning_rate": 6.410111086869314e-07,
"loss": 0.6389,
"step": 809
},
{
"epoch": 2.557894736842105,
"grad_norm": 0.5304815769195557,
"learning_rate": 6.320198296951435e-07,
"loss": 0.5732,
"step": 810
},
{
"epoch": 2.5610526315789475,
"grad_norm": 0.5228027701377869,
"learning_rate": 6.230877998159724e-07,
"loss": 0.6278,
"step": 811
},
{
"epoch": 2.5642105263157893,
"grad_norm": 0.5397827625274658,
"learning_rate": 6.142151402072133e-07,
"loss": 0.5979,
"step": 812
},
{
"epoch": 2.5673684210526315,
"grad_norm": 0.5186311602592468,
"learning_rate": 6.054019712213377e-07,
"loss": 0.5872,
"step": 813
},
{
"epoch": 2.5705263157894738,
"grad_norm": 0.5260378122329712,
"learning_rate": 5.966484124038602e-07,
"loss": 0.6337,
"step": 814
},
{
"epoch": 2.5736842105263156,
"grad_norm": 0.556025505065918,
"learning_rate": 5.879545824917199e-07,
"loss": 0.5965,
"step": 815
},
{
"epoch": 2.576842105263158,
"grad_norm": 0.5116238594055176,
"learning_rate": 5.793205994116674e-07,
"loss": 0.6147,
"step": 816
},
{
"epoch": 2.58,
"grad_norm": 0.5269439816474915,
"learning_rate": 5.707465802786655e-07,
"loss": 0.5967,
"step": 817
},
{
"epoch": 2.583157894736842,
"grad_norm": 0.5351024270057678,
"learning_rate": 5.622326413942997e-07,
"loss": 0.6241,
"step": 818
},
{
"epoch": 2.586315789473684,
"grad_norm": 0.536216676235199,
"learning_rate": 5.537788982452052e-07,
"loss": 0.631,
"step": 819
},
{
"epoch": 2.5894736842105264,
"grad_norm": 0.5194189548492432,
"learning_rate": 5.453854655014956e-07,
"loss": 0.6394,
"step": 820
},
{
"epoch": 2.592631578947368,
"grad_norm": 0.5215575098991394,
"learning_rate": 5.370524570152059e-07,
"loss": 0.5948,
"step": 821
},
{
"epoch": 2.5957894736842104,
"grad_norm": 0.5416327118873596,
"learning_rate": 5.287799858187548e-07,
"loss": 0.6352,
"step": 822
},
{
"epoch": 2.5989473684210527,
"grad_norm": 0.5264502167701721,
"learning_rate": 5.205681641234062e-07,
"loss": 0.61,
"step": 823
},
{
"epoch": 2.6021052631578945,
"grad_norm": 0.5675118565559387,
"learning_rate": 5.1241710331775e-07,
"loss": 0.5791,
"step": 824
},
{
"epoch": 2.6052631578947367,
"grad_norm": 0.49685797095298767,
"learning_rate": 5.043269139661872e-07,
"loss": 0.6316,
"step": 825
},
{
"epoch": 2.608421052631579,
"grad_norm": 0.5199131965637207,
"learning_rate": 4.962977058074381e-07,
"loss": 0.6217,
"step": 826
},
{
"epoch": 2.6115789473684208,
"grad_norm": 0.5157626271247864,
"learning_rate": 4.883295877530431e-07,
"loss": 0.6394,
"step": 827
},
{
"epoch": 2.614736842105263,
"grad_norm": 0.5056947469711304,
"learning_rate": 4.804226678858936e-07,
"loss": 0.6168,
"step": 828
},
{
"epoch": 2.6178947368421053,
"grad_norm": 0.5152086019515991,
"learning_rate": 4.725770534587637e-07,
"loss": 0.5963,
"step": 829
},
{
"epoch": 2.6210526315789475,
"grad_norm": 0.4946594536304474,
"learning_rate": 4.647928508928512e-07,
"loss": 0.6344,
"step": 830
},
{
"epoch": 2.6242105263157893,
"grad_norm": 0.5005388855934143,
"learning_rate": 4.5707016577634156e-07,
"loss": 0.6797,
"step": 831
},
{
"epoch": 2.6273684210526316,
"grad_norm": 0.49225690960884094,
"learning_rate": 4.494091028629699e-07,
"loss": 0.6552,
"step": 832
},
{
"epoch": 2.630526315789474,
"grad_norm": 0.5395132303237915,
"learning_rate": 4.418097660706039e-07,
"loss": 0.6197,
"step": 833
},
{
"epoch": 2.6336842105263156,
"grad_norm": 0.5258274674415588,
"learning_rate": 4.342722584798298e-07,
"loss": 0.6151,
"step": 834
},
{
"epoch": 2.636842105263158,
"grad_norm": 0.507228672504425,
"learning_rate": 4.267966823325581e-07,
"loss": 0.6213,
"step": 835
},
{
"epoch": 2.64,
"grad_norm": 0.5385991334915161,
"learning_rate": 4.193831390306352e-07,
"loss": 0.6423,
"step": 836
},
{
"epoch": 2.6431578947368424,
"grad_norm": 0.5073803067207336,
"learning_rate": 4.1203172913446774e-07,
"loss": 0.5812,
"step": 837
},
{
"epoch": 2.646315789473684,
"grad_norm": 0.5264935493469238,
"learning_rate": 4.047425523616577e-07,
"loss": 0.6305,
"step": 838
},
{
"epoch": 2.6494736842105264,
"grad_norm": 0.4937891364097595,
"learning_rate": 3.9751570758565284e-07,
"loss": 0.6197,
"step": 839
},
{
"epoch": 2.6526315789473687,
"grad_norm": 0.5468772053718567,
"learning_rate": 3.9035129283440165e-07,
"loss": 0.592,
"step": 840
},
{
"epoch": 2.6557894736842105,
"grad_norm": 0.5518105626106262,
"learning_rate": 3.8324940528902845e-07,
"loss": 0.6332,
"step": 841
},
{
"epoch": 2.6589473684210527,
"grad_norm": 0.5344856381416321,
"learning_rate": 3.762101412825098e-07,
"loss": 0.6221,
"step": 842
},
{
"epoch": 2.662105263157895,
"grad_norm": 0.527391791343689,
"learning_rate": 3.6923359629837117e-07,
"loss": 0.5941,
"step": 843
},
{
"epoch": 2.665263157894737,
"grad_norm": 0.5052513480186462,
"learning_rate": 3.6231986496939153e-07,
"loss": 0.6352,
"step": 844
},
{
"epoch": 2.668421052631579,
"grad_norm": 0.5419393181800842,
"learning_rate": 3.554690410763173e-07,
"loss": 0.6058,
"step": 845
},
{
"epoch": 2.6715789473684213,
"grad_norm": 0.5119031667709351,
"learning_rate": 3.4868121754659533e-07,
"loss": 0.5942,
"step": 846
},
{
"epoch": 2.674736842105263,
"grad_norm": 0.4917730689048767,
"learning_rate": 3.4195648645310443e-07,
"loss": 0.6145,
"step": 847
},
{
"epoch": 2.6778947368421053,
"grad_norm": 0.5071961283683777,
"learning_rate": 3.3529493901291567e-07,
"loss": 0.6481,
"step": 848
},
{
"epoch": 2.6810526315789476,
"grad_norm": 0.5443476438522339,
"learning_rate": 3.286966655860485e-07,
"loss": 0.5775,
"step": 849
},
{
"epoch": 2.6842105263157894,
"grad_norm": 0.5274863839149475,
"learning_rate": 3.2216175567424737e-07,
"loss": 0.5835,
"step": 850
},
{
"epoch": 2.6873684210526316,
"grad_norm": 0.5420557856559753,
"learning_rate": 3.156902979197679e-07,
"loss": 0.5913,
"step": 851
},
{
"epoch": 2.690526315789474,
"grad_norm": 0.5218223929405212,
"learning_rate": 3.0928238010417275e-07,
"loss": 0.6144,
"step": 852
},
{
"epoch": 2.6936842105263157,
"grad_norm": 0.5233944654464722,
"learning_rate": 3.029380891471445e-07,
"loss": 0.632,
"step": 853
},
{
"epoch": 2.696842105263158,
"grad_norm": 0.5044023394584656,
"learning_rate": 2.966575111053027e-07,
"loss": 0.5874,
"step": 854
},
{
"epoch": 2.7,
"grad_norm": 0.5117988586425781,
"learning_rate": 2.9044073117103777e-07,
"loss": 0.6533,
"step": 855
},
{
"epoch": 2.703157894736842,
"grad_norm": 0.511157751083374,
"learning_rate": 2.842878336713578e-07,
"loss": 0.6282,
"step": 856
},
{
"epoch": 2.7063157894736842,
"grad_norm": 0.5114948749542236,
"learning_rate": 2.7819890206674083e-07,
"loss": 0.6355,
"step": 857
},
{
"epoch": 2.7094736842105265,
"grad_norm": 0.549742579460144,
"learning_rate": 2.7217401895000664e-07,
"loss": 0.6086,
"step": 858
},
{
"epoch": 2.7126315789473683,
"grad_norm": 0.5112407803535461,
"learning_rate": 2.6621326604519216e-07,
"loss": 0.6062,
"step": 859
},
{
"epoch": 2.7157894736842105,
"grad_norm": 0.5380560755729675,
"learning_rate": 2.6031672420644694e-07,
"loss": 0.5904,
"step": 860
},
{
"epoch": 2.718947368421053,
"grad_norm": 0.49735158681869507,
"learning_rate": 2.5448447341693493e-07,
"loss": 0.593,
"step": 861
},
{
"epoch": 2.7221052631578946,
"grad_norm": 0.4785706698894501,
"learning_rate": 2.4871659278774884e-07,
"loss": 0.6211,
"step": 862
},
{
"epoch": 2.725263157894737,
"grad_norm": 0.5243927240371704,
"learning_rate": 2.430131605568353e-07,
"loss": 0.6143,
"step": 863
},
{
"epoch": 2.728421052631579,
"grad_norm": 0.5355093479156494,
"learning_rate": 2.3737425408794202e-07,
"loss": 0.6162,
"step": 864
},
{
"epoch": 2.731578947368421,
"grad_norm": 0.5540766716003418,
"learning_rate": 2.31799949869555e-07,
"loss": 0.5777,
"step": 865
},
{
"epoch": 2.734736842105263,
"grad_norm": 0.5112879276275635,
"learning_rate": 2.2629032351387247e-07,
"loss": 0.5949,
"step": 866
},
{
"epoch": 2.7378947368421054,
"grad_norm": 0.5168077945709229,
"learning_rate": 2.2084544975577383e-07,
"loss": 0.6391,
"step": 867
},
{
"epoch": 2.741052631578947,
"grad_norm": 0.5143442749977112,
"learning_rate": 2.1546540245180825e-07,
"loss": 0.6142,
"step": 868
},
{
"epoch": 2.7442105263157894,
"grad_norm": 0.49836885929107666,
"learning_rate": 2.1015025457919002e-07,
"loss": 0.5926,
"step": 869
},
{
"epoch": 2.7473684210526317,
"grad_norm": 0.5337677597999573,
"learning_rate": 2.0490007823481096e-07,
"loss": 0.5758,
"step": 870
},
{
"epoch": 2.7505263157894735,
"grad_norm": 0.5162203311920166,
"learning_rate": 1.9971494463426332e-07,
"loss": 0.6044,
"step": 871
},
{
"epoch": 2.7536842105263157,
"grad_norm": 0.5083566308021545,
"learning_rate": 1.9459492411087078e-07,
"loss": 0.6377,
"step": 872
},
{
"epoch": 2.756842105263158,
"grad_norm": 0.5102013349533081,
"learning_rate": 1.8954008611473618e-07,
"loss": 0.6287,
"step": 873
},
{
"epoch": 2.76,
"grad_norm": 0.5197393894195557,
"learning_rate": 1.8455049921179858e-07,
"loss": 0.6224,
"step": 874
},
{
"epoch": 2.763157894736842,
"grad_norm": 0.4867171347141266,
"learning_rate": 1.7962623108290556e-07,
"loss": 0.6475,
"step": 875
},
{
"epoch": 2.7663157894736843,
"grad_norm": 0.5163273215293884,
"learning_rate": 1.7476734852289235e-07,
"loss": 0.6408,
"step": 876
},
{
"epoch": 2.769473684210526,
"grad_norm": 0.5355550646781921,
"learning_rate": 1.6997391743967696e-07,
"loss": 0.6129,
"step": 877
},
{
"epoch": 2.7726315789473683,
"grad_norm": 0.5214855670928955,
"learning_rate": 1.65246002853367e-07,
"loss": 0.5987,
"step": 878
},
{
"epoch": 2.7757894736842106,
"grad_norm": 0.5373660922050476,
"learning_rate": 1.6058366889537546e-07,
"loss": 0.5472,
"step": 879
},
{
"epoch": 2.7789473684210524,
"grad_norm": 0.49646681547164917,
"learning_rate": 1.559869788075541e-07,
"loss": 0.612,
"step": 880
},
{
"epoch": 2.7821052631578946,
"grad_norm": 0.5255357623100281,
"learning_rate": 1.514559949413319e-07,
"loss": 0.6083,
"step": 881
},
{
"epoch": 2.785263157894737,
"grad_norm": 0.5195873379707336,
"learning_rate": 1.4699077875687252e-07,
"loss": 0.6032,
"step": 882
},
{
"epoch": 2.7884210526315787,
"grad_norm": 0.5284191370010376,
"learning_rate": 1.4259139082223761e-07,
"loss": 0.6466,
"step": 883
},
{
"epoch": 2.791578947368421,
"grad_norm": 0.5327157378196716,
"learning_rate": 1.3825789081256812e-07,
"loss": 0.5883,
"step": 884
},
{
"epoch": 2.794736842105263,
"grad_norm": 0.5047375559806824,
"learning_rate": 1.3399033750927327e-07,
"loss": 0.6168,
"step": 885
},
{
"epoch": 2.797894736842105,
"grad_norm": 0.5161851644515991,
"learning_rate": 1.297887887992344e-07,
"loss": 0.615,
"step": 886
},
{
"epoch": 2.8010526315789472,
"grad_norm": 0.5051410794258118,
"learning_rate": 1.2565330167401747e-07,
"loss": 0.6444,
"step": 887
},
{
"epoch": 2.8042105263157895,
"grad_norm": 0.5022399425506592,
"learning_rate": 1.2158393222910235e-07,
"loss": 0.6318,
"step": 888
},
{
"epoch": 2.8073684210526317,
"grad_norm": 0.4974055588245392,
"learning_rate": 1.175807356631209e-07,
"loss": 0.6291,
"step": 889
},
{
"epoch": 2.8105263157894735,
"grad_norm": 0.4854937493801117,
"learning_rate": 1.1364376627710727e-07,
"loss": 0.6456,
"step": 890
},
{
"epoch": 2.813684210526316,
"grad_norm": 0.5069525837898254,
"learning_rate": 1.0977307747376431e-07,
"loss": 0.6226,
"step": 891
},
{
"epoch": 2.816842105263158,
"grad_norm": 0.5431550145149231,
"learning_rate": 1.0596872175673456e-07,
"loss": 0.6005,
"step": 892
},
{
"epoch": 2.82,
"grad_norm": 0.5111134052276611,
"learning_rate": 1.0223075072989418e-07,
"loss": 0.6243,
"step": 893
},
{
"epoch": 2.823157894736842,
"grad_norm": 0.5054842829704285,
"learning_rate": 9.855921509664745e-08,
"loss": 0.6145,
"step": 894
},
{
"epoch": 2.8263157894736843,
"grad_norm": 0.5049132108688354,
"learning_rate": 9.495416465924113e-08,
"loss": 0.6283,
"step": 895
},
{
"epoch": 2.8294736842105266,
"grad_norm": 0.5091620087623596,
"learning_rate": 9.141564831808947e-08,
"loss": 0.6245,
"step": 896
},
{
"epoch": 2.8326315789473684,
"grad_norm": 0.5132347941398621,
"learning_rate": 8.794371407111091e-08,
"loss": 0.6149,
"step": 897
},
{
"epoch": 2.8357894736842106,
"grad_norm": 0.5056635737419128,
"learning_rate": 8.45384090130752e-08,
"loss": 0.6408,
"step": 898
},
{
"epoch": 2.838947368421053,
"grad_norm": 0.4924507141113281,
"learning_rate": 8.11997793349667e-08,
"loss": 0.622,
"step": 899
},
{
"epoch": 2.8421052631578947,
"grad_norm": 0.4771850109100342,
"learning_rate": 7.792787032335657e-08,
"loss": 0.6321,
"step": 900
},
{
"epoch": 2.845263157894737,
"grad_norm": 0.5036438703536987,
"learning_rate": 7.472272635978995e-08,
"loss": 0.6403,
"step": 901
},
{
"epoch": 2.848421052631579,
"grad_norm": 0.5470741987228394,
"learning_rate": 7.158439092018077e-08,
"loss": 0.6147,
"step": 902
},
{
"epoch": 2.851578947368421,
"grad_norm": 0.4736262261867523,
"learning_rate": 6.851290657422627e-08,
"loss": 0.6591,
"step": 903
},
{
"epoch": 2.8547368421052632,
"grad_norm": 0.4981045424938202,
"learning_rate": 6.550831498482679e-08,
"loss": 0.5957,
"step": 904
},
{
"epoch": 2.8578947368421055,
"grad_norm": 0.5055891275405884,
"learning_rate": 6.257065690752129e-08,
"loss": 0.6222,
"step": 905
},
{
"epoch": 2.8610526315789473,
"grad_norm": 0.5165517330169678,
"learning_rate": 5.969997218993328e-08,
"loss": 0.6335,
"step": 906
},
{
"epoch": 2.8642105263157895,
"grad_norm": 0.5130405426025391,
"learning_rate": 5.689629977123412e-08,
"loss": 0.602,
"step": 907
},
{
"epoch": 2.867368421052632,
"grad_norm": 0.5092564821243286,
"learning_rate": 5.415967768160946e-08,
"loss": 0.6221,
"step": 908
},
{
"epoch": 2.8705263157894736,
"grad_norm": 0.5033932328224182,
"learning_rate": 5.149014304174915e-08,
"loss": 0.632,
"step": 909
},
{
"epoch": 2.873684210526316,
"grad_norm": 0.5016679167747498,
"learning_rate": 4.8887732062337656e-08,
"loss": 0.61,
"step": 910
},
{
"epoch": 2.876842105263158,
"grad_norm": 0.5133535265922546,
"learning_rate": 4.635248004356885e-08,
"loss": 0.6702,
"step": 911
},
{
"epoch": 2.88,
"grad_norm": 0.496550977230072,
"learning_rate": 4.388442137466198e-08,
"loss": 0.6243,
"step": 912
},
{
"epoch": 2.883157894736842,
"grad_norm": 0.48129701614379883,
"learning_rate": 4.148358953339926e-08,
"loss": 0.6565,
"step": 913
},
{
"epoch": 2.8863157894736844,
"grad_norm": 0.5186068415641785,
"learning_rate": 3.9150017085669566e-08,
"loss": 0.63,
"step": 914
},
{
"epoch": 2.889473684210526,
"grad_norm": 0.5132560133934021,
"learning_rate": 3.688373568502601e-08,
"loss": 0.6137,
"step": 915
},
{
"epoch": 2.8926315789473684,
"grad_norm": 0.5437251329421997,
"learning_rate": 3.468477607226017e-08,
"loss": 0.6646,
"step": 916
},
{
"epoch": 2.8957894736842107,
"grad_norm": 0.5179529786109924,
"learning_rate": 3.255316807498077e-08,
"loss": 0.5848,
"step": 917
},
{
"epoch": 2.8989473684210525,
"grad_norm": 0.5003638863563538,
"learning_rate": 3.048894060721064e-08,
"loss": 0.6227,
"step": 918
},
{
"epoch": 2.9021052631578947,
"grad_norm": 0.5409918427467346,
"learning_rate": 2.8492121668997064e-08,
"loss": 0.6075,
"step": 919
},
{
"epoch": 2.905263157894737,
"grad_norm": 0.4994703531265259,
"learning_rate": 2.6562738346027627e-08,
"loss": 0.6214,
"step": 920
},
{
"epoch": 2.908421052631579,
"grad_norm": 0.5022985935211182,
"learning_rate": 2.4700816809266615e-08,
"loss": 0.6123,
"step": 921
},
{
"epoch": 2.911578947368421,
"grad_norm": 0.5189907550811768,
"learning_rate": 2.290638231459641e-08,
"loss": 0.5687,
"step": 922
},
{
"epoch": 2.9147368421052633,
"grad_norm": 0.4975310266017914,
"learning_rate": 2.1179459202479436e-08,
"loss": 0.6152,
"step": 923
},
{
"epoch": 2.917894736842105,
"grad_norm": 0.5323106050491333,
"learning_rate": 1.9520070897623976e-08,
"loss": 0.6638,
"step": 924
},
{
"epoch": 2.9210526315789473,
"grad_norm": 0.48443955183029175,
"learning_rate": 1.792823990866721e-08,
"loss": 0.6238,
"step": 925
},
{
"epoch": 2.9242105263157896,
"grad_norm": 0.48831799626350403,
"learning_rate": 1.640398782787267e-08,
"loss": 0.5758,
"step": 926
},
{
"epoch": 2.9273684210526314,
"grad_norm": 0.49946728348731995,
"learning_rate": 1.49473353308327e-08,
"loss": 0.6192,
"step": 927
},
{
"epoch": 2.9305263157894736,
"grad_norm": 0.5276715159416199,
"learning_rate": 1.3558302176192584e-08,
"loss": 0.6171,
"step": 928
},
{
"epoch": 2.933684210526316,
"grad_norm": 0.5295907855033875,
"learning_rate": 1.2236907205379623e-08,
"loss": 0.6198,
"step": 929
},
{
"epoch": 2.9368421052631577,
"grad_norm": 0.4946899712085724,
"learning_rate": 1.0983168342348915e-08,
"loss": 0.6253,
"step": 930
},
{
"epoch": 2.94,
"grad_norm": 0.5253959894180298,
"learning_rate": 9.797102593339659e-09,
"loss": 0.5993,
"step": 931
},
{
"epoch": 2.943157894736842,
"grad_norm": 0.5255020260810852,
"learning_rate": 8.678726046644215e-09,
"loss": 0.6384,
"step": 932
},
{
"epoch": 2.946315789473684,
"grad_norm": 0.5470498204231262,
"learning_rate": 7.628053872390517e-09,
"loss": 0.6002,
"step": 933
},
{
"epoch": 2.9494736842105262,
"grad_norm": 0.5109714865684509,
"learning_rate": 6.645100322336118e-09,
"loss": 0.6359,
"step": 934
},
{
"epoch": 2.9526315789473685,
"grad_norm": 0.517336368560791,
"learning_rate": 5.7298787296750094e-09,
"loss": 0.6493,
"step": 935
},
{
"epoch": 2.9557894736842103,
"grad_norm": 0.5175721049308777,
"learning_rate": 4.88240150885555e-09,
"loss": 0.6275,
"step": 936
},
{
"epoch": 2.9589473684210525,
"grad_norm": 0.52158522605896,
"learning_rate": 4.1026801554139296e-09,
"loss": 0.613,
"step": 937
},
{
"epoch": 2.962105263157895,
"grad_norm": 0.5218417048454285,
"learning_rate": 3.3907252458176277e-09,
"loss": 0.6094,
"step": 938
},
{
"epoch": 2.9652631578947366,
"grad_norm": 0.5241231322288513,
"learning_rate": 2.7465464373205296e-09,
"loss": 0.6365,
"step": 939
},
{
"epoch": 2.968421052631579,
"grad_norm": 0.5028926730155945,
"learning_rate": 2.1701524678346964e-09,
"loss": 0.5889,
"step": 940
},
{
"epoch": 2.971578947368421,
"grad_norm": 0.5517033934593201,
"learning_rate": 1.6615511558082386e-09,
"loss": 0.559,
"step": 941
},
{
"epoch": 2.974736842105263,
"grad_norm": 0.49616527557373047,
"learning_rate": 1.220749400123733e-09,
"loss": 0.6422,
"step": 942
},
{
"epoch": 2.977894736842105,
"grad_norm": 0.5219214558601379,
"learning_rate": 8.477531799999661e-10,
"loss": 0.6043,
"step": 943
},
{
"epoch": 2.9810526315789474,
"grad_norm": 0.5003503561019897,
"learning_rate": 5.425675549136645e-10,
"loss": 0.6027,
"step": 944
},
{
"epoch": 2.984210526315789,
"grad_norm": 0.5330973863601685,
"learning_rate": 3.051966645312154e-10,
"loss": 0.5806,
"step": 945
},
{
"epoch": 2.9873684210526315,
"grad_norm": 0.5317094326019287,
"learning_rate": 1.3564372865038088e-10,
"loss": 0.607,
"step": 946
},
{
"epoch": 2.9905263157894737,
"grad_norm": 0.5060402750968933,
"learning_rate": 3.3911047158663445e-11,
"loss": 0.6126,
"step": 947
},
{
"epoch": 2.993684210526316,
"grad_norm": 0.5056828260421753,
"learning_rate": 0.0,
"loss": 0.6224,
"step": 948
},
{
"epoch": 2.993684210526316,
"step": 948,
"total_flos": 8.371731904995656e+17,
"train_loss": 0.7066250008998541,
"train_runtime": 24790.5767,
"train_samples_per_second": 3.677,
"train_steps_per_second": 0.038
}
],
"logging_steps": 1,
"max_steps": 948,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 8.371731904995656e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}