GAE-Retriever / trainer_state.json
magicgh's picture
First commit
632dd61
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.01953125,
"eval_steps": 500,
"global_step": 256,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.00390625,
"grad_norm": 988.0331420898438,
"learning_rate": 3.846153846153847e-06,
"loss": 23.0443,
"step": 1
},
{
"epoch": 0.0078125,
"grad_norm": 1018.0032958984375,
"learning_rate": 7.692307692307694e-06,
"loss": 22.2415,
"step": 2
},
{
"epoch": 0.01171875,
"grad_norm": 905.46435546875,
"learning_rate": 1.153846153846154e-05,
"loss": 23.9407,
"step": 3
},
{
"epoch": 0.015625,
"grad_norm": 772.3015747070312,
"learning_rate": 1.5384615384615387e-05,
"loss": 20.3585,
"step": 4
},
{
"epoch": 0.01953125,
"grad_norm": 1083.967529296875,
"learning_rate": 1.923076923076923e-05,
"loss": 17.7004,
"step": 5
},
{
"epoch": 0.0234375,
"grad_norm": 518.4392700195312,
"learning_rate": 2.307692307692308e-05,
"loss": 14.8905,
"step": 6
},
{
"epoch": 0.02734375,
"grad_norm": 588.9624633789062,
"learning_rate": 2.6923076923076923e-05,
"loss": 15.46,
"step": 7
},
{
"epoch": 0.03125,
"grad_norm": 291.337646484375,
"learning_rate": 3.0769230769230774e-05,
"loss": 11.7233,
"step": 8
},
{
"epoch": 0.03515625,
"grad_norm": 243.8762969970703,
"learning_rate": 3.461538461538462e-05,
"loss": 9.8939,
"step": 9
},
{
"epoch": 0.0390625,
"grad_norm": 306.4288635253906,
"learning_rate": 3.846153846153846e-05,
"loss": 10.5548,
"step": 10
},
{
"epoch": 0.04296875,
"grad_norm": 138.4855194091797,
"learning_rate": 4.230769230769231e-05,
"loss": 7.1429,
"step": 11
},
{
"epoch": 0.046875,
"grad_norm": 79.80815887451172,
"learning_rate": 4.615384615384616e-05,
"loss": 7.0453,
"step": 12
},
{
"epoch": 0.05078125,
"grad_norm": 95.14498901367188,
"learning_rate": 5e-05,
"loss": 6.2978,
"step": 13
},
{
"epoch": 0.0546875,
"grad_norm": 76.86641693115234,
"learning_rate": 4.9794238683127575e-05,
"loss": 6.9177,
"step": 14
},
{
"epoch": 0.05859375,
"grad_norm": 90.26302337646484,
"learning_rate": 4.958847736625515e-05,
"loss": 5.7372,
"step": 15
},
{
"epoch": 0.0625,
"grad_norm": 51.85117721557617,
"learning_rate": 4.938271604938271e-05,
"loss": 5.7694,
"step": 16
},
{
"epoch": 0.06640625,
"grad_norm": 28.343177795410156,
"learning_rate": 4.9176954732510286e-05,
"loss": 5.8857,
"step": 17
},
{
"epoch": 0.0703125,
"grad_norm": 28.591285705566406,
"learning_rate": 4.8971193415637865e-05,
"loss": 5.2027,
"step": 18
},
{
"epoch": 0.07421875,
"grad_norm": 23.616016387939453,
"learning_rate": 4.876543209876544e-05,
"loss": 4.8946,
"step": 19
},
{
"epoch": 0.078125,
"grad_norm": 20.2352294921875,
"learning_rate": 4.855967078189301e-05,
"loss": 5.0189,
"step": 20
},
{
"epoch": 0.08203125,
"grad_norm": 16.59236717224121,
"learning_rate": 4.835390946502058e-05,
"loss": 4.9116,
"step": 21
},
{
"epoch": 0.0859375,
"grad_norm": 15.008003234863281,
"learning_rate": 4.814814814814815e-05,
"loss": 4.8674,
"step": 22
},
{
"epoch": 0.08984375,
"grad_norm": 11.656904220581055,
"learning_rate": 4.794238683127572e-05,
"loss": 4.96,
"step": 23
},
{
"epoch": 0.09375,
"grad_norm": 17.49643898010254,
"learning_rate": 4.773662551440329e-05,
"loss": 4.7911,
"step": 24
},
{
"epoch": 0.09765625,
"grad_norm": 12.967889785766602,
"learning_rate": 4.7530864197530866e-05,
"loss": 4.7574,
"step": 25
},
{
"epoch": 0.1015625,
"grad_norm": 13.78781795501709,
"learning_rate": 4.732510288065844e-05,
"loss": 4.5257,
"step": 26
},
{
"epoch": 0.10546875,
"grad_norm": 14.40069580078125,
"learning_rate": 4.711934156378601e-05,
"loss": 4.6842,
"step": 27
},
{
"epoch": 0.109375,
"grad_norm": 8.558459281921387,
"learning_rate": 4.691358024691358e-05,
"loss": 4.4912,
"step": 28
},
{
"epoch": 0.11328125,
"grad_norm": 9.563591003417969,
"learning_rate": 4.6707818930041156e-05,
"loss": 4.3927,
"step": 29
},
{
"epoch": 0.1171875,
"grad_norm": 13.471641540527344,
"learning_rate": 4.650205761316873e-05,
"loss": 4.446,
"step": 30
},
{
"epoch": 0.12109375,
"grad_norm": 8.875003814697266,
"learning_rate": 4.62962962962963e-05,
"loss": 4.4367,
"step": 31
},
{
"epoch": 0.125,
"grad_norm": 9.281575202941895,
"learning_rate": 4.609053497942387e-05,
"loss": 4.4643,
"step": 32
},
{
"epoch": 0.12890625,
"grad_norm": 9.402353286743164,
"learning_rate": 4.5884773662551446e-05,
"loss": 4.4036,
"step": 33
},
{
"epoch": 0.1328125,
"grad_norm": 8.509848594665527,
"learning_rate": 4.567901234567901e-05,
"loss": 4.3953,
"step": 34
},
{
"epoch": 0.13671875,
"grad_norm": 9.05785846710205,
"learning_rate": 4.5473251028806584e-05,
"loss": 4.4259,
"step": 35
},
{
"epoch": 0.140625,
"grad_norm": 7.090909481048584,
"learning_rate": 4.5267489711934157e-05,
"loss": 4.2375,
"step": 36
},
{
"epoch": 0.14453125,
"grad_norm": 10.807791709899902,
"learning_rate": 4.506172839506173e-05,
"loss": 4.3719,
"step": 37
},
{
"epoch": 0.1484375,
"grad_norm": 13.562170028686523,
"learning_rate": 4.48559670781893e-05,
"loss": 4.6153,
"step": 38
},
{
"epoch": 0.15234375,
"grad_norm": 11.16930103302002,
"learning_rate": 4.4650205761316874e-05,
"loss": 4.4007,
"step": 39
},
{
"epoch": 0.15625,
"grad_norm": 12.779727935791016,
"learning_rate": 4.4444444444444447e-05,
"loss": 4.5876,
"step": 40
},
{
"epoch": 0.16015625,
"grad_norm": 14.613138198852539,
"learning_rate": 4.423868312757202e-05,
"loss": 4.3842,
"step": 41
},
{
"epoch": 0.1640625,
"grad_norm": 16.920902252197266,
"learning_rate": 4.403292181069959e-05,
"loss": 4.2888,
"step": 42
},
{
"epoch": 0.16796875,
"grad_norm": 23.178911209106445,
"learning_rate": 4.3827160493827164e-05,
"loss": 4.4708,
"step": 43
},
{
"epoch": 0.171875,
"grad_norm": 28.53046989440918,
"learning_rate": 4.3621399176954737e-05,
"loss": 4.5108,
"step": 44
},
{
"epoch": 0.17578125,
"grad_norm": 19.8005428314209,
"learning_rate": 4.341563786008231e-05,
"loss": 4.4592,
"step": 45
},
{
"epoch": 0.1796875,
"grad_norm": 12.320777893066406,
"learning_rate": 4.3209876543209875e-05,
"loss": 4.5041,
"step": 46
},
{
"epoch": 0.18359375,
"grad_norm": 8.350976943969727,
"learning_rate": 4.300411522633745e-05,
"loss": 4.3769,
"step": 47
},
{
"epoch": 0.1875,
"grad_norm": 14.553681373596191,
"learning_rate": 4.279835390946502e-05,
"loss": 4.683,
"step": 48
},
{
"epoch": 0.19140625,
"grad_norm": 13.202040672302246,
"learning_rate": 4.259259259259259e-05,
"loss": 4.4271,
"step": 49
},
{
"epoch": 0.1953125,
"grad_norm": 12.252931594848633,
"learning_rate": 4.2386831275720165e-05,
"loss": 4.202,
"step": 50
},
{
"epoch": 0.19921875,
"grad_norm": 9.770834922790527,
"learning_rate": 4.2181069958847744e-05,
"loss": 4.4271,
"step": 51
},
{
"epoch": 0.203125,
"grad_norm": 10.324782371520996,
"learning_rate": 4.197530864197531e-05,
"loss": 4.3385,
"step": 52
},
{
"epoch": 0.20703125,
"grad_norm": 8.983942031860352,
"learning_rate": 4.176954732510288e-05,
"loss": 4.2885,
"step": 53
},
{
"epoch": 0.2109375,
"grad_norm": 8.984896659851074,
"learning_rate": 4.1563786008230455e-05,
"loss": 4.446,
"step": 54
},
{
"epoch": 0.21484375,
"grad_norm": 10.578001022338867,
"learning_rate": 4.135802469135803e-05,
"loss": 4.3787,
"step": 55
},
{
"epoch": 0.21875,
"grad_norm": 12.245546340942383,
"learning_rate": 4.11522633744856e-05,
"loss": 4.1896,
"step": 56
},
{
"epoch": 0.22265625,
"grad_norm": 12.414233207702637,
"learning_rate": 4.094650205761317e-05,
"loss": 4.4919,
"step": 57
},
{
"epoch": 0.2265625,
"grad_norm": 10.887884140014648,
"learning_rate": 4.074074074074074e-05,
"loss": 4.2051,
"step": 58
},
{
"epoch": 0.23046875,
"grad_norm": 11.619287490844727,
"learning_rate": 4.053497942386831e-05,
"loss": 4.015,
"step": 59
},
{
"epoch": 0.234375,
"grad_norm": 16.54395294189453,
"learning_rate": 4.032921810699588e-05,
"loss": 3.9459,
"step": 60
},
{
"epoch": 0.23828125,
"grad_norm": 22.401296615600586,
"learning_rate": 4.012345679012346e-05,
"loss": 3.8599,
"step": 61
},
{
"epoch": 0.2421875,
"grad_norm": 30.748231887817383,
"learning_rate": 3.9917695473251035e-05,
"loss": 3.8411,
"step": 62
},
{
"epoch": 0.24609375,
"grad_norm": 48.33995056152344,
"learning_rate": 3.971193415637861e-05,
"loss": 3.8338,
"step": 63
},
{
"epoch": 0.25,
"grad_norm": 40.25141906738281,
"learning_rate": 3.950617283950617e-05,
"loss": 3.7575,
"step": 64
},
{
"epoch": 0.25390625,
"grad_norm": 70.64269256591797,
"learning_rate": 3.9300411522633746e-05,
"loss": 3.7886,
"step": 65
},
{
"epoch": 0.2578125,
"grad_norm": 46.51031494140625,
"learning_rate": 3.909465020576132e-05,
"loss": 3.7727,
"step": 66
},
{
"epoch": 0.26171875,
"grad_norm": 44.5860481262207,
"learning_rate": 3.888888888888889e-05,
"loss": 3.7284,
"step": 67
},
{
"epoch": 0.265625,
"grad_norm": 37.13584899902344,
"learning_rate": 3.868312757201646e-05,
"loss": 3.7974,
"step": 68
},
{
"epoch": 0.26953125,
"grad_norm": 17.04970932006836,
"learning_rate": 3.8477366255144036e-05,
"loss": 3.7399,
"step": 69
},
{
"epoch": 0.2734375,
"grad_norm": 25.48470687866211,
"learning_rate": 3.82716049382716e-05,
"loss": 3.6582,
"step": 70
},
{
"epoch": 0.27734375,
"grad_norm": 15.296791076660156,
"learning_rate": 3.806584362139918e-05,
"loss": 3.759,
"step": 71
},
{
"epoch": 0.28125,
"grad_norm": 18.959625244140625,
"learning_rate": 3.786008230452675e-05,
"loss": 3.6962,
"step": 72
},
{
"epoch": 0.28515625,
"grad_norm": 16.249454498291016,
"learning_rate": 3.7654320987654326e-05,
"loss": 3.5659,
"step": 73
},
{
"epoch": 0.2890625,
"grad_norm": 17.6076602935791,
"learning_rate": 3.74485596707819e-05,
"loss": 3.6214,
"step": 74
},
{
"epoch": 0.29296875,
"grad_norm": 15.325469017028809,
"learning_rate": 3.724279835390947e-05,
"loss": 3.5469,
"step": 75
},
{
"epoch": 0.296875,
"grad_norm": 14.967309951782227,
"learning_rate": 3.7037037037037037e-05,
"loss": 3.5844,
"step": 76
},
{
"epoch": 0.30078125,
"grad_norm": 16.454774856567383,
"learning_rate": 3.683127572016461e-05,
"loss": 3.4725,
"step": 77
},
{
"epoch": 0.3046875,
"grad_norm": 15.80962085723877,
"learning_rate": 3.662551440329218e-05,
"loss": 3.4781,
"step": 78
},
{
"epoch": 0.30859375,
"grad_norm": 12.834327697753906,
"learning_rate": 3.6419753086419754e-05,
"loss": 3.6498,
"step": 79
},
{
"epoch": 0.3125,
"grad_norm": 12.81863784790039,
"learning_rate": 3.6213991769547327e-05,
"loss": 3.4208,
"step": 80
},
{
"epoch": 0.31640625,
"grad_norm": 16.791345596313477,
"learning_rate": 3.60082304526749e-05,
"loss": 3.4432,
"step": 81
},
{
"epoch": 0.3203125,
"grad_norm": 20.827123641967773,
"learning_rate": 3.580246913580247e-05,
"loss": 3.4386,
"step": 82
},
{
"epoch": 0.32421875,
"grad_norm": 14.668889999389648,
"learning_rate": 3.5596707818930044e-05,
"loss": 3.3887,
"step": 83
},
{
"epoch": 0.328125,
"grad_norm": 17.22427749633789,
"learning_rate": 3.539094650205762e-05,
"loss": 3.6215,
"step": 84
},
{
"epoch": 0.33203125,
"grad_norm": 16.612550735473633,
"learning_rate": 3.518518518518519e-05,
"loss": 3.4215,
"step": 85
},
{
"epoch": 0.3359375,
"grad_norm": 15.173816680908203,
"learning_rate": 3.497942386831276e-05,
"loss": 3.3305,
"step": 86
},
{
"epoch": 0.33984375,
"grad_norm": 18.891874313354492,
"learning_rate": 3.4773662551440334e-05,
"loss": 3.4129,
"step": 87
},
{
"epoch": 0.34375,
"grad_norm": 19.012630462646484,
"learning_rate": 3.45679012345679e-05,
"loss": 3.4039,
"step": 88
},
{
"epoch": 0.34765625,
"grad_norm": 25.285091400146484,
"learning_rate": 3.436213991769547e-05,
"loss": 3.7148,
"step": 89
},
{
"epoch": 0.3515625,
"grad_norm": 23.138742446899414,
"learning_rate": 3.4156378600823045e-05,
"loss": 3.3594,
"step": 90
},
{
"epoch": 0.35546875,
"grad_norm": 24.770191192626953,
"learning_rate": 3.395061728395062e-05,
"loss": 3.3106,
"step": 91
},
{
"epoch": 0.359375,
"grad_norm": 17.597614288330078,
"learning_rate": 3.374485596707819e-05,
"loss": 3.2024,
"step": 92
},
{
"epoch": 0.36328125,
"grad_norm": 13.329920768737793,
"learning_rate": 3.353909465020576e-05,
"loss": 3.196,
"step": 93
},
{
"epoch": 0.3671875,
"grad_norm": 20.01732635498047,
"learning_rate": 3.3333333333333335e-05,
"loss": 3.3282,
"step": 94
},
{
"epoch": 0.37109375,
"grad_norm": 21.501489639282227,
"learning_rate": 3.312757201646091e-05,
"loss": 3.2842,
"step": 95
},
{
"epoch": 0.375,
"grad_norm": 15.159469604492188,
"learning_rate": 3.292181069958848e-05,
"loss": 3.149,
"step": 96
},
{
"epoch": 0.37890625,
"grad_norm": 15.139326095581055,
"learning_rate": 3.271604938271605e-05,
"loss": 3.2335,
"step": 97
},
{
"epoch": 0.3828125,
"grad_norm": 14.6196870803833,
"learning_rate": 3.2510288065843625e-05,
"loss": 3.104,
"step": 98
},
{
"epoch": 0.38671875,
"grad_norm": 12.317699432373047,
"learning_rate": 3.230452674897119e-05,
"loss": 3.1088,
"step": 99
},
{
"epoch": 0.390625,
"grad_norm": 17.009883880615234,
"learning_rate": 3.209876543209876e-05,
"loss": 3.1713,
"step": 100
},
{
"epoch": 0.39453125,
"grad_norm": 22.170177459716797,
"learning_rate": 3.1893004115226336e-05,
"loss": 3.1348,
"step": 101
},
{
"epoch": 0.3984375,
"grad_norm": 18.459367752075195,
"learning_rate": 3.168724279835391e-05,
"loss": 3.1625,
"step": 102
},
{
"epoch": 0.40234375,
"grad_norm": 11.8716402053833,
"learning_rate": 3.148148148148148e-05,
"loss": 3.0979,
"step": 103
},
{
"epoch": 0.40625,
"grad_norm": 16.9968204498291,
"learning_rate": 3.127572016460906e-05,
"loss": 3.1705,
"step": 104
},
{
"epoch": 0.41015625,
"grad_norm": 18.078325271606445,
"learning_rate": 3.1069958847736626e-05,
"loss": 3.1945,
"step": 105
},
{
"epoch": 0.4140625,
"grad_norm": 16.14826202392578,
"learning_rate": 3.08641975308642e-05,
"loss": 3.2458,
"step": 106
},
{
"epoch": 0.41796875,
"grad_norm": 14.381550788879395,
"learning_rate": 3.065843621399177e-05,
"loss": 3.1335,
"step": 107
},
{
"epoch": 0.421875,
"grad_norm": 11.909482955932617,
"learning_rate": 3.0452674897119343e-05,
"loss": 3.09,
"step": 108
},
{
"epoch": 0.42578125,
"grad_norm": 17.585634231567383,
"learning_rate": 3.0246913580246916e-05,
"loss": 3.0548,
"step": 109
},
{
"epoch": 0.4296875,
"grad_norm": 15.303757667541504,
"learning_rate": 3.0041152263374488e-05,
"loss": 2.9545,
"step": 110
},
{
"epoch": 0.43359375,
"grad_norm": 15.899727821350098,
"learning_rate": 2.9835390946502057e-05,
"loss": 3.0344,
"step": 111
},
{
"epoch": 0.4375,
"grad_norm": 12.18794059753418,
"learning_rate": 2.962962962962963e-05,
"loss": 3.1163,
"step": 112
},
{
"epoch": 0.44140625,
"grad_norm": 15.14059066772461,
"learning_rate": 2.9423868312757202e-05,
"loss": 3.0572,
"step": 113
},
{
"epoch": 0.4453125,
"grad_norm": 16.42298698425293,
"learning_rate": 2.9218106995884775e-05,
"loss": 2.9781,
"step": 114
},
{
"epoch": 0.44921875,
"grad_norm": 16.090763092041016,
"learning_rate": 2.9012345679012347e-05,
"loss": 3.1644,
"step": 115
},
{
"epoch": 0.453125,
"grad_norm": 18.76362419128418,
"learning_rate": 2.880658436213992e-05,
"loss": 3.0987,
"step": 116
},
{
"epoch": 0.45703125,
"grad_norm": 17.354793548583984,
"learning_rate": 2.860082304526749e-05,
"loss": 2.9983,
"step": 117
},
{
"epoch": 0.4609375,
"grad_norm": 13.282272338867188,
"learning_rate": 2.839506172839506e-05,
"loss": 3.0733,
"step": 118
},
{
"epoch": 0.46484375,
"grad_norm": 21.229665756225586,
"learning_rate": 2.8189300411522634e-05,
"loss": 3.0484,
"step": 119
},
{
"epoch": 0.46875,
"grad_norm": 16.58381462097168,
"learning_rate": 2.7983539094650207e-05,
"loss": 3.1174,
"step": 120
},
{
"epoch": 0.47265625,
"grad_norm": 15.544482231140137,
"learning_rate": 2.777777777777778e-05,
"loss": 3.1705,
"step": 121
},
{
"epoch": 0.4765625,
"grad_norm": 15.846022605895996,
"learning_rate": 2.757201646090535e-05,
"loss": 3.0341,
"step": 122
},
{
"epoch": 0.48046875,
"grad_norm": 21.867094039916992,
"learning_rate": 2.736625514403292e-05,
"loss": 2.9884,
"step": 123
},
{
"epoch": 0.484375,
"grad_norm": 13.162125587463379,
"learning_rate": 2.7160493827160493e-05,
"loss": 3.0612,
"step": 124
},
{
"epoch": 0.48828125,
"grad_norm": 11.88615608215332,
"learning_rate": 2.6954732510288066e-05,
"loss": 2.9845,
"step": 125
},
{
"epoch": 0.4921875,
"grad_norm": 11.516526222229004,
"learning_rate": 2.6748971193415638e-05,
"loss": 2.9974,
"step": 126
},
{
"epoch": 0.49609375,
"grad_norm": 13.762920379638672,
"learning_rate": 2.654320987654321e-05,
"loss": 3.0515,
"step": 127
},
{
"epoch": 0.5,
"grad_norm": 13.856005668640137,
"learning_rate": 2.6337448559670787e-05,
"loss": 3.0096,
"step": 128
},
{
"epoch": 0.50390625,
"grad_norm": 12.380202293395996,
"learning_rate": 2.6131687242798352e-05,
"loss": 3.0274,
"step": 129
},
{
"epoch": 0.5078125,
"grad_norm": 15.164392471313477,
"learning_rate": 2.5925925925925925e-05,
"loss": 2.9838,
"step": 130
},
{
"epoch": 0.51171875,
"grad_norm": 15.10387134552002,
"learning_rate": 2.5720164609053497e-05,
"loss": 3.1058,
"step": 131
},
{
"epoch": 0.515625,
"grad_norm": 11.87817096710205,
"learning_rate": 2.551440329218107e-05,
"loss": 3.0164,
"step": 132
},
{
"epoch": 0.51953125,
"grad_norm": 17.242656707763672,
"learning_rate": 2.5308641975308646e-05,
"loss": 2.9376,
"step": 133
},
{
"epoch": 0.5234375,
"grad_norm": 20.221240997314453,
"learning_rate": 2.510288065843622e-05,
"loss": 3.0894,
"step": 134
},
{
"epoch": 0.52734375,
"grad_norm": 12.36820125579834,
"learning_rate": 2.4897119341563787e-05,
"loss": 3.0347,
"step": 135
},
{
"epoch": 0.53125,
"grad_norm": 14.862237930297852,
"learning_rate": 2.4691358024691357e-05,
"loss": 3.0826,
"step": 136
},
{
"epoch": 0.53515625,
"grad_norm": 14.29667854309082,
"learning_rate": 2.4485596707818932e-05,
"loss": 3.0467,
"step": 137
},
{
"epoch": 0.5390625,
"grad_norm": 15.347952842712402,
"learning_rate": 2.4279835390946505e-05,
"loss": 2.9934,
"step": 138
},
{
"epoch": 0.54296875,
"grad_norm": 19.236717224121094,
"learning_rate": 2.4074074074074074e-05,
"loss": 2.8921,
"step": 139
},
{
"epoch": 0.546875,
"grad_norm": 13.943548202514648,
"learning_rate": 2.3868312757201647e-05,
"loss": 3.1285,
"step": 140
},
{
"epoch": 0.55078125,
"grad_norm": 15.554057121276855,
"learning_rate": 2.366255144032922e-05,
"loss": 2.9709,
"step": 141
},
{
"epoch": 0.5546875,
"grad_norm": 15.124194145202637,
"learning_rate": 2.345679012345679e-05,
"loss": 2.9845,
"step": 142
},
{
"epoch": 0.55859375,
"grad_norm": 15.458159446716309,
"learning_rate": 2.3251028806584364e-05,
"loss": 2.9581,
"step": 143
},
{
"epoch": 0.5625,
"grad_norm": 11.531893730163574,
"learning_rate": 2.3045267489711937e-05,
"loss": 3.0545,
"step": 144
},
{
"epoch": 0.56640625,
"grad_norm": 16.634984970092773,
"learning_rate": 2.2839506172839506e-05,
"loss": 2.9772,
"step": 145
},
{
"epoch": 0.5703125,
"grad_norm": 13.670973777770996,
"learning_rate": 2.2633744855967078e-05,
"loss": 2.9478,
"step": 146
},
{
"epoch": 0.57421875,
"grad_norm": 13.582576751708984,
"learning_rate": 2.242798353909465e-05,
"loss": 3.1053,
"step": 147
},
{
"epoch": 0.578125,
"grad_norm": 17.05980110168457,
"learning_rate": 2.2222222222222223e-05,
"loss": 2.9789,
"step": 148
},
{
"epoch": 0.58203125,
"grad_norm": 13.586709976196289,
"learning_rate": 2.2016460905349796e-05,
"loss": 2.9268,
"step": 149
},
{
"epoch": 0.5859375,
"grad_norm": 14.885200500488281,
"learning_rate": 2.1810699588477368e-05,
"loss": 3.0281,
"step": 150
},
{
"epoch": 0.58984375,
"grad_norm": 14.411493301391602,
"learning_rate": 2.1604938271604937e-05,
"loss": 2.9164,
"step": 151
},
{
"epoch": 0.59375,
"grad_norm": 16.72563934326172,
"learning_rate": 2.139917695473251e-05,
"loss": 2.9586,
"step": 152
},
{
"epoch": 0.59765625,
"grad_norm": 11.668434143066406,
"learning_rate": 2.1193415637860082e-05,
"loss": 3.0239,
"step": 153
},
{
"epoch": 0.6015625,
"grad_norm": 11.849235534667969,
"learning_rate": 2.0987654320987655e-05,
"loss": 2.9665,
"step": 154
},
{
"epoch": 0.60546875,
"grad_norm": 17.961620330810547,
"learning_rate": 2.0781893004115227e-05,
"loss": 3.0192,
"step": 155
},
{
"epoch": 0.609375,
"grad_norm": 13.321170806884766,
"learning_rate": 2.05761316872428e-05,
"loss": 2.9901,
"step": 156
},
{
"epoch": 0.61328125,
"grad_norm": 12.714600563049316,
"learning_rate": 2.037037037037037e-05,
"loss": 2.9933,
"step": 157
},
{
"epoch": 0.6171875,
"grad_norm": 11.27708625793457,
"learning_rate": 2.016460905349794e-05,
"loss": 2.9186,
"step": 158
},
{
"epoch": 0.62109375,
"grad_norm": 11.366385459899902,
"learning_rate": 1.9958847736625517e-05,
"loss": 2.9615,
"step": 159
},
{
"epoch": 0.625,
"grad_norm": 14.28786849975586,
"learning_rate": 1.9753086419753087e-05,
"loss": 3.0847,
"step": 160
},
{
"epoch": 0.62890625,
"grad_norm": 17.186941146850586,
"learning_rate": 1.954732510288066e-05,
"loss": 3.1361,
"step": 161
},
{
"epoch": 0.6328125,
"grad_norm": 12.188273429870605,
"learning_rate": 1.934156378600823e-05,
"loss": 3.0134,
"step": 162
},
{
"epoch": 0.63671875,
"grad_norm": 18.98809051513672,
"learning_rate": 1.91358024691358e-05,
"loss": 2.904,
"step": 163
},
{
"epoch": 0.640625,
"grad_norm": 12.578585624694824,
"learning_rate": 1.8930041152263377e-05,
"loss": 2.9072,
"step": 164
},
{
"epoch": 0.64453125,
"grad_norm": 13.295378684997559,
"learning_rate": 1.872427983539095e-05,
"loss": 3.0163,
"step": 165
},
{
"epoch": 0.6484375,
"grad_norm": 16.138151168823242,
"learning_rate": 1.8518518518518518e-05,
"loss": 2.9509,
"step": 166
},
{
"epoch": 0.65234375,
"grad_norm": 18.39076805114746,
"learning_rate": 1.831275720164609e-05,
"loss": 2.9098,
"step": 167
},
{
"epoch": 0.65625,
"grad_norm": 19.46346664428711,
"learning_rate": 1.8106995884773663e-05,
"loss": 2.9096,
"step": 168
},
{
"epoch": 0.66015625,
"grad_norm": 13.3604154586792,
"learning_rate": 1.7901234567901236e-05,
"loss": 2.8992,
"step": 169
},
{
"epoch": 0.6640625,
"grad_norm": 12.542205810546875,
"learning_rate": 1.769547325102881e-05,
"loss": 2.9293,
"step": 170
},
{
"epoch": 0.66796875,
"grad_norm": 12.55959415435791,
"learning_rate": 1.748971193415638e-05,
"loss": 2.8809,
"step": 171
},
{
"epoch": 0.671875,
"grad_norm": 15.57677936553955,
"learning_rate": 1.728395061728395e-05,
"loss": 2.7552,
"step": 172
},
{
"epoch": 0.67578125,
"grad_norm": 11.344679832458496,
"learning_rate": 1.7078189300411522e-05,
"loss": 2.9377,
"step": 173
},
{
"epoch": 0.6796875,
"grad_norm": 15.26870059967041,
"learning_rate": 1.6872427983539095e-05,
"loss": 2.9145,
"step": 174
},
{
"epoch": 0.68359375,
"grad_norm": 12.774370193481445,
"learning_rate": 1.6666666666666667e-05,
"loss": 2.9775,
"step": 175
},
{
"epoch": 0.6875,
"grad_norm": 19.02437400817871,
"learning_rate": 1.646090534979424e-05,
"loss": 2.9581,
"step": 176
},
{
"epoch": 0.69140625,
"grad_norm": 14.445993423461914,
"learning_rate": 1.6255144032921812e-05,
"loss": 2.9639,
"step": 177
},
{
"epoch": 0.6953125,
"grad_norm": 13.563401222229004,
"learning_rate": 1.604938271604938e-05,
"loss": 2.8964,
"step": 178
},
{
"epoch": 0.69921875,
"grad_norm": 17.80751609802246,
"learning_rate": 1.5843621399176954e-05,
"loss": 3.0603,
"step": 179
},
{
"epoch": 0.703125,
"grad_norm": 13.13770866394043,
"learning_rate": 1.563786008230453e-05,
"loss": 3.0473,
"step": 180
},
{
"epoch": 0.70703125,
"grad_norm": 15.236124038696289,
"learning_rate": 1.54320987654321e-05,
"loss": 2.8392,
"step": 181
},
{
"epoch": 0.7109375,
"grad_norm": 17.386003494262695,
"learning_rate": 1.5226337448559672e-05,
"loss": 2.8805,
"step": 182
},
{
"epoch": 0.71484375,
"grad_norm": 16.233539581298828,
"learning_rate": 1.5020576131687244e-05,
"loss": 2.9692,
"step": 183
},
{
"epoch": 0.71875,
"grad_norm": 19.721954345703125,
"learning_rate": 1.4814814814814815e-05,
"loss": 3.0988,
"step": 184
},
{
"epoch": 0.72265625,
"grad_norm": 12.52316951751709,
"learning_rate": 1.4609053497942387e-05,
"loss": 2.933,
"step": 185
},
{
"epoch": 0.7265625,
"grad_norm": 12.95042610168457,
"learning_rate": 1.440329218106996e-05,
"loss": 2.927,
"step": 186
},
{
"epoch": 0.73046875,
"grad_norm": 12.077346801757812,
"learning_rate": 1.419753086419753e-05,
"loss": 2.8952,
"step": 187
},
{
"epoch": 0.734375,
"grad_norm": 11.764900207519531,
"learning_rate": 1.3991769547325103e-05,
"loss": 2.8488,
"step": 188
},
{
"epoch": 0.73828125,
"grad_norm": 13.644368171691895,
"learning_rate": 1.3786008230452676e-05,
"loss": 2.8838,
"step": 189
},
{
"epoch": 0.7421875,
"grad_norm": 12.35655689239502,
"learning_rate": 1.3580246913580247e-05,
"loss": 2.9183,
"step": 190
},
{
"epoch": 0.74609375,
"grad_norm": 13.980887413024902,
"learning_rate": 1.3374485596707819e-05,
"loss": 2.8803,
"step": 191
},
{
"epoch": 0.75,
"grad_norm": 18.003271102905273,
"learning_rate": 1.3168724279835393e-05,
"loss": 3.0357,
"step": 192
},
{
"epoch": 0.75390625,
"grad_norm": 14.126523971557617,
"learning_rate": 1.2962962962962962e-05,
"loss": 2.9313,
"step": 193
},
{
"epoch": 0.7578125,
"grad_norm": 12.55939769744873,
"learning_rate": 1.2757201646090535e-05,
"loss": 2.9224,
"step": 194
},
{
"epoch": 0.76171875,
"grad_norm": 13.05750846862793,
"learning_rate": 1.255144032921811e-05,
"loss": 2.9167,
"step": 195
},
{
"epoch": 0.765625,
"grad_norm": 16.52781867980957,
"learning_rate": 1.2345679012345678e-05,
"loss": 2.8645,
"step": 196
},
{
"epoch": 0.76953125,
"grad_norm": 15.458019256591797,
"learning_rate": 1.2139917695473252e-05,
"loss": 2.9848,
"step": 197
},
{
"epoch": 0.7734375,
"grad_norm": 12.694581985473633,
"learning_rate": 1.1934156378600823e-05,
"loss": 3.0691,
"step": 198
},
{
"epoch": 0.77734375,
"grad_norm": 11.771615982055664,
"learning_rate": 1.1728395061728396e-05,
"loss": 2.9316,
"step": 199
},
{
"epoch": 0.78125,
"grad_norm": 15.438037872314453,
"learning_rate": 1.1522633744855968e-05,
"loss": 2.9295,
"step": 200
},
{
"epoch": 0.78515625,
"grad_norm": 13.371684074401855,
"learning_rate": 1.1316872427983539e-05,
"loss": 2.8849,
"step": 201
},
{
"epoch": 0.7890625,
"grad_norm": 11.904099464416504,
"learning_rate": 1.1111111111111112e-05,
"loss": 2.862,
"step": 202
},
{
"epoch": 0.79296875,
"grad_norm": 11.638395309448242,
"learning_rate": 1.0905349794238684e-05,
"loss": 2.949,
"step": 203
},
{
"epoch": 0.796875,
"grad_norm": 14.490175247192383,
"learning_rate": 1.0699588477366255e-05,
"loss": 2.9131,
"step": 204
},
{
"epoch": 0.80078125,
"grad_norm": 11.613717079162598,
"learning_rate": 1.0493827160493827e-05,
"loss": 2.9497,
"step": 205
},
{
"epoch": 0.8046875,
"grad_norm": 13.931456565856934,
"learning_rate": 1.02880658436214e-05,
"loss": 2.917,
"step": 206
},
{
"epoch": 0.80859375,
"grad_norm": 12.256906509399414,
"learning_rate": 1.008230452674897e-05,
"loss": 2.9048,
"step": 207
},
{
"epoch": 0.8125,
"grad_norm": 12.851861953735352,
"learning_rate": 9.876543209876543e-06,
"loss": 2.9084,
"step": 208
},
{
"epoch": 0.81640625,
"grad_norm": 18.05097770690918,
"learning_rate": 9.670781893004116e-06,
"loss": 2.802,
"step": 209
},
{
"epoch": 0.8203125,
"grad_norm": 12.831151962280273,
"learning_rate": 9.465020576131688e-06,
"loss": 2.8659,
"step": 210
},
{
"epoch": 0.82421875,
"grad_norm": 11.60468864440918,
"learning_rate": 9.259259259259259e-06,
"loss": 2.8285,
"step": 211
},
{
"epoch": 0.828125,
"grad_norm": 14.392627716064453,
"learning_rate": 9.053497942386832e-06,
"loss": 2.8605,
"step": 212
},
{
"epoch": 0.83203125,
"grad_norm": 17.23533821105957,
"learning_rate": 8.847736625514404e-06,
"loss": 2.9186,
"step": 213
},
{
"epoch": 0.8359375,
"grad_norm": 13.937773704528809,
"learning_rate": 8.641975308641975e-06,
"loss": 2.9164,
"step": 214
},
{
"epoch": 0.83984375,
"grad_norm": 16.57691764831543,
"learning_rate": 8.436213991769547e-06,
"loss": 2.8829,
"step": 215
},
{
"epoch": 0.84375,
"grad_norm": 12.391244888305664,
"learning_rate": 8.23045267489712e-06,
"loss": 2.8417,
"step": 216
},
{
"epoch": 0.84765625,
"grad_norm": 15.762882232666016,
"learning_rate": 8.02469135802469e-06,
"loss": 2.871,
"step": 217
},
{
"epoch": 0.8515625,
"grad_norm": 16.37859344482422,
"learning_rate": 7.818930041152265e-06,
"loss": 2.9265,
"step": 218
},
{
"epoch": 0.85546875,
"grad_norm": 21.5294132232666,
"learning_rate": 7.613168724279836e-06,
"loss": 2.9628,
"step": 219
},
{
"epoch": 0.859375,
"grad_norm": 18.281295776367188,
"learning_rate": 7.4074074074074075e-06,
"loss": 2.8301,
"step": 220
},
{
"epoch": 0.86328125,
"grad_norm": 16.0869140625,
"learning_rate": 7.20164609053498e-06,
"loss": 2.8179,
"step": 221
},
{
"epoch": 0.8671875,
"grad_norm": 21.662134170532227,
"learning_rate": 6.995884773662552e-06,
"loss": 2.8893,
"step": 222
},
{
"epoch": 0.87109375,
"grad_norm": 12.4130277633667,
"learning_rate": 6.790123456790123e-06,
"loss": 2.8628,
"step": 223
},
{
"epoch": 0.875,
"grad_norm": 15.154963493347168,
"learning_rate": 6.584362139917697e-06,
"loss": 2.8991,
"step": 224
},
{
"epoch": 0.87890625,
"grad_norm": 12.677474021911621,
"learning_rate": 6.3786008230452675e-06,
"loss": 2.9428,
"step": 225
},
{
"epoch": 0.8828125,
"grad_norm": 12.973712921142578,
"learning_rate": 6.172839506172839e-06,
"loss": 2.8694,
"step": 226
},
{
"epoch": 0.88671875,
"grad_norm": 13.162025451660156,
"learning_rate": 5.967078189300412e-06,
"loss": 2.8119,
"step": 227
},
{
"epoch": 0.890625,
"grad_norm": 13.606464385986328,
"learning_rate": 5.761316872427984e-06,
"loss": 2.8455,
"step": 228
},
{
"epoch": 0.89453125,
"grad_norm": 17.518091201782227,
"learning_rate": 5.555555555555556e-06,
"loss": 2.8708,
"step": 229
},
{
"epoch": 0.8984375,
"grad_norm": 13.689889907836914,
"learning_rate": 5.3497942386831275e-06,
"loss": 2.8384,
"step": 230
},
{
"epoch": 0.90234375,
"grad_norm": 14.69385051727295,
"learning_rate": 5.1440329218107e-06,
"loss": 2.8291,
"step": 231
},
{
"epoch": 0.90625,
"grad_norm": 17.468791961669922,
"learning_rate": 4.938271604938272e-06,
"loss": 2.8588,
"step": 232
},
{
"epoch": 0.91015625,
"grad_norm": 14.33683967590332,
"learning_rate": 4.732510288065844e-06,
"loss": 2.7762,
"step": 233
},
{
"epoch": 0.9140625,
"grad_norm": 17.65381622314453,
"learning_rate": 4.526748971193416e-06,
"loss": 2.8697,
"step": 234
},
{
"epoch": 0.91796875,
"grad_norm": 17.74317169189453,
"learning_rate": 4.3209876543209875e-06,
"loss": 2.8151,
"step": 235
},
{
"epoch": 0.921875,
"grad_norm": 20.99629020690918,
"learning_rate": 4.11522633744856e-06,
"loss": 2.893,
"step": 236
},
{
"epoch": 0.92578125,
"grad_norm": 17.19089698791504,
"learning_rate": 3.9094650205761325e-06,
"loss": 2.8025,
"step": 237
},
{
"epoch": 0.9296875,
"grad_norm": 12.898162841796875,
"learning_rate": 3.7037037037037037e-06,
"loss": 2.9405,
"step": 238
},
{
"epoch": 0.93359375,
"grad_norm": 13.902021408081055,
"learning_rate": 3.497942386831276e-06,
"loss": 2.8937,
"step": 239
},
{
"epoch": 0.9375,
"grad_norm": 13.255682945251465,
"learning_rate": 3.2921810699588483e-06,
"loss": 2.7604,
"step": 240
},
{
"epoch": 0.94140625,
"grad_norm": 14.98218822479248,
"learning_rate": 3.0864197530864196e-06,
"loss": 2.8931,
"step": 241
},
{
"epoch": 0.9453125,
"grad_norm": 15.084565162658691,
"learning_rate": 2.880658436213992e-06,
"loss": 2.8565,
"step": 242
},
{
"epoch": 0.94921875,
"grad_norm": 12.473812103271484,
"learning_rate": 2.6748971193415637e-06,
"loss": 2.795,
"step": 243
},
{
"epoch": 0.953125,
"grad_norm": 13.222654342651367,
"learning_rate": 2.469135802469136e-06,
"loss": 2.8734,
"step": 244
},
{
"epoch": 0.95703125,
"grad_norm": 13.204935073852539,
"learning_rate": 2.263374485596708e-06,
"loss": 2.8013,
"step": 245
},
{
"epoch": 0.9609375,
"grad_norm": 15.458930015563965,
"learning_rate": 2.05761316872428e-06,
"loss": 2.8541,
"step": 246
},
{
"epoch": 0.96484375,
"grad_norm": 17.476573944091797,
"learning_rate": 1.8518518518518519e-06,
"loss": 2.7538,
"step": 247
},
{
"epoch": 0.96875,
"grad_norm": 13.751199722290039,
"learning_rate": 1.6460905349794242e-06,
"loss": 2.8324,
"step": 248
},
{
"epoch": 0.97265625,
"grad_norm": 15.517616271972656,
"learning_rate": 1.440329218106996e-06,
"loss": 2.8402,
"step": 249
},
{
"epoch": 0.9765625,
"grad_norm": 16.016067504882812,
"learning_rate": 1.234567901234568e-06,
"loss": 2.895,
"step": 250
},
{
"epoch": 0.98046875,
"grad_norm": 11.978571891784668,
"learning_rate": 1.02880658436214e-06,
"loss": 2.9139,
"step": 251
},
{
"epoch": 1.00390625,
"grad_norm": 58.774688720703125,
"learning_rate": 8.230452674897121e-07,
"loss": 3.7051,
"step": 252
},
{
"epoch": 1.0078125,
"grad_norm": 52.733154296875,
"learning_rate": 6.17283950617284e-07,
"loss": 3.7482,
"step": 253
},
{
"epoch": 1.01171875,
"grad_norm": 68.14625549316406,
"learning_rate": 4.1152263374485604e-07,
"loss": 3.8065,
"step": 254
},
{
"epoch": 1.015625,
"grad_norm": 59.991546630859375,
"learning_rate": 2.0576131687242802e-07,
"loss": 3.6531,
"step": 255
},
{
"epoch": 1.01953125,
"grad_norm": 40.77336883544922,
"learning_rate": 0.0,
"loss": 3.4788,
"step": 256
}
],
"logging_steps": 1,
"max_steps": 256,
"num_input_tokens_seen": 0,
"num_train_epochs": 9223372036854775807,
"save_steps": 1.0,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 0.0,
"train_batch_size": 128,
"trial_name": null,
"trial_params": null
}