Robometer-4B-LIBERO / trainer_state.json
aliangdw's picture
Duplicate from aliangdw/libero_ablation_prog_pref_with_fail_lora_ft_4frames
e179783
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 24.137931034482758,
"eval_steps": 50,
"global_step": 700,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.034482758620689655,
"grad_norm": 4.955975532531738,
"learning_rate": 0.0,
"loss": 1.5752,
"step": 1
},
{
"epoch": 0.06896551724137931,
"grad_norm": 3.4605445861816406,
"learning_rate": 1.2500000000000002e-07,
"loss": 1.6145,
"step": 2
},
{
"epoch": 0.10344827586206896,
"grad_norm": 3.622607946395874,
"learning_rate": 2.5000000000000004e-07,
"loss": 1.6033,
"step": 3
},
{
"epoch": 0.13793103448275862,
"grad_norm": 6.3884453773498535,
"learning_rate": 3.75e-07,
"loss": 1.6219,
"step": 4
},
{
"epoch": 0.1724137931034483,
"grad_norm": 3.212383270263672,
"learning_rate": 5.000000000000001e-07,
"loss": 1.5958,
"step": 5
},
{
"epoch": 0.20689655172413793,
"grad_norm": 6.6238508224487305,
"learning_rate": 6.25e-07,
"loss": 1.6268,
"step": 6
},
{
"epoch": 0.2413793103448276,
"grad_norm": 7.565281391143799,
"learning_rate": 7.5e-07,
"loss": 1.6133,
"step": 7
},
{
"epoch": 0.27586206896551724,
"grad_norm": 11.01657772064209,
"learning_rate": 8.750000000000001e-07,
"loss": 1.6061,
"step": 8
},
{
"epoch": 0.3103448275862069,
"grad_norm": 8.828049659729004,
"learning_rate": 1.0000000000000002e-06,
"loss": 1.6352,
"step": 9
},
{
"epoch": 0.3448275862068966,
"grad_norm": 3.2700514793395996,
"learning_rate": 1.125e-06,
"loss": 1.5985,
"step": 10
},
{
"epoch": 0.3793103448275862,
"grad_norm": 7.4744086265563965,
"learning_rate": 1.25e-06,
"loss": 1.6075,
"step": 11
},
{
"epoch": 0.41379310344827586,
"grad_norm": 5.462789535522461,
"learning_rate": 1.3750000000000002e-06,
"loss": 1.6174,
"step": 12
},
{
"epoch": 0.4482758620689655,
"grad_norm": 3.593034267425537,
"learning_rate": 1.5e-06,
"loss": 1.6083,
"step": 13
},
{
"epoch": 0.4827586206896552,
"grad_norm": 4.305792331695557,
"learning_rate": 1.6250000000000001e-06,
"loss": 1.5997,
"step": 14
},
{
"epoch": 0.5172413793103449,
"grad_norm": 8.656575202941895,
"learning_rate": 1.7500000000000002e-06,
"loss": 1.6283,
"step": 15
},
{
"epoch": 0.5517241379310345,
"grad_norm": 6.6580491065979,
"learning_rate": 1.875e-06,
"loss": 1.5906,
"step": 16
},
{
"epoch": 0.5862068965517241,
"grad_norm": 1.8648769855499268,
"learning_rate": 2.0000000000000003e-06,
"loss": 1.5667,
"step": 17
},
{
"epoch": 0.6206896551724138,
"grad_norm": 8.883339881896973,
"learning_rate": 2.1250000000000004e-06,
"loss": 1.5999,
"step": 18
},
{
"epoch": 0.6551724137931034,
"grad_norm": 3.959343194961548,
"learning_rate": 2.25e-06,
"loss": 1.5751,
"step": 19
},
{
"epoch": 0.6896551724137931,
"grad_norm": 5.278771877288818,
"learning_rate": 2.375e-06,
"loss": 1.6216,
"step": 20
},
{
"epoch": 0.7241379310344828,
"grad_norm": 3.716949224472046,
"learning_rate": 2.5e-06,
"loss": 1.613,
"step": 21
},
{
"epoch": 0.7586206896551724,
"grad_norm": 1.6308438777923584,
"learning_rate": 2.625e-06,
"loss": 1.5822,
"step": 22
},
{
"epoch": 0.7931034482758621,
"grad_norm": 1.8407272100448608,
"learning_rate": 2.7500000000000004e-06,
"loss": 1.5369,
"step": 23
},
{
"epoch": 0.8275862068965517,
"grad_norm": 7.528758525848389,
"learning_rate": 2.8750000000000004e-06,
"loss": 1.584,
"step": 24
},
{
"epoch": 0.8620689655172413,
"grad_norm": 3.860356569290161,
"learning_rate": 3e-06,
"loss": 1.58,
"step": 25
},
{
"epoch": 0.896551724137931,
"grad_norm": 1.7416785955429077,
"learning_rate": 3.125e-06,
"loss": 1.5714,
"step": 26
},
{
"epoch": 0.9310344827586207,
"grad_norm": 4.024614334106445,
"learning_rate": 3.2500000000000002e-06,
"loss": 1.5905,
"step": 27
},
{
"epoch": 0.9655172413793104,
"grad_norm": 5.027279376983643,
"learning_rate": 3.3750000000000003e-06,
"loss": 1.5906,
"step": 28
},
{
"epoch": 1.0,
"grad_norm": 7.354084014892578,
"learning_rate": 3.5000000000000004e-06,
"loss": 1.5781,
"step": 29
},
{
"epoch": 1.0344827586206897,
"grad_norm": 1.950278401374817,
"learning_rate": 3.625e-06,
"loss": 1.5747,
"step": 30
},
{
"epoch": 1.0689655172413792,
"grad_norm": 3.3933162689208984,
"learning_rate": 3.75e-06,
"loss": 1.5642,
"step": 31
},
{
"epoch": 1.103448275862069,
"grad_norm": 4.551718235015869,
"learning_rate": 3.875e-06,
"loss": 1.5872,
"step": 32
},
{
"epoch": 1.1379310344827587,
"grad_norm": 3.4449899196624756,
"learning_rate": 4.000000000000001e-06,
"loss": 1.5755,
"step": 33
},
{
"epoch": 1.1724137931034484,
"grad_norm": 2.4862890243530273,
"learning_rate": 4.125e-06,
"loss": 1.5661,
"step": 34
},
{
"epoch": 1.206896551724138,
"grad_norm": 3.3560233116149902,
"learning_rate": 4.250000000000001e-06,
"loss": 1.5813,
"step": 35
},
{
"epoch": 1.2413793103448276,
"grad_norm": 1.7421871423721313,
"learning_rate": 4.375e-06,
"loss": 1.5653,
"step": 36
},
{
"epoch": 1.2758620689655173,
"grad_norm": 1.8098750114440918,
"learning_rate": 4.5e-06,
"loss": 1.5686,
"step": 37
},
{
"epoch": 1.3103448275862069,
"grad_norm": 5.115236759185791,
"learning_rate": 4.625e-06,
"loss": 1.5678,
"step": 38
},
{
"epoch": 1.3448275862068966,
"grad_norm": 3.044386625289917,
"learning_rate": 4.75e-06,
"loss": 1.5634,
"step": 39
},
{
"epoch": 1.3793103448275863,
"grad_norm": 5.8161139488220215,
"learning_rate": 4.875000000000001e-06,
"loss": 1.5846,
"step": 40
},
{
"epoch": 1.4137931034482758,
"grad_norm": 1.6947190761566162,
"learning_rate": 5e-06,
"loss": 1.5512,
"step": 41
},
{
"epoch": 1.4482758620689655,
"grad_norm": 3.2046549320220947,
"learning_rate": 5.125e-06,
"loss": 1.5579,
"step": 42
},
{
"epoch": 1.4827586206896552,
"grad_norm": 2.9457404613494873,
"learning_rate": 5.25e-06,
"loss": 1.5661,
"step": 43
},
{
"epoch": 1.5172413793103448,
"grad_norm": 1.604246735572815,
"learning_rate": 5.375e-06,
"loss": 1.5584,
"step": 44
},
{
"epoch": 1.5517241379310345,
"grad_norm": 1.433769702911377,
"learning_rate": 5.500000000000001e-06,
"loss": 1.5398,
"step": 45
},
{
"epoch": 1.5862068965517242,
"grad_norm": 1.4531933069229126,
"learning_rate": 5.625e-06,
"loss": 1.5459,
"step": 46
},
{
"epoch": 1.6206896551724137,
"grad_norm": 3.0887269973754883,
"learning_rate": 5.750000000000001e-06,
"loss": 1.5475,
"step": 47
},
{
"epoch": 1.6551724137931034,
"grad_norm": 1.7311522960662842,
"learning_rate": 5.875e-06,
"loss": 1.5395,
"step": 48
},
{
"epoch": 1.6896551724137931,
"grad_norm": 1.57200026512146,
"learning_rate": 6e-06,
"loss": 1.5355,
"step": 49
},
{
"epoch": 1.7241379310344827,
"grad_norm": 6.282898426055908,
"learning_rate": 6.125e-06,
"loss": 1.5479,
"step": 50
},
{
"epoch": 1.7586206896551724,
"grad_norm": 5.877266883850098,
"learning_rate": 6.25e-06,
"loss": 1.535,
"step": 51
},
{
"epoch": 1.793103448275862,
"grad_norm": 2.238494396209717,
"learning_rate": 6.375000000000001e-06,
"loss": 1.5439,
"step": 52
},
{
"epoch": 1.8275862068965516,
"grad_norm": 3.176295518875122,
"learning_rate": 6.5000000000000004e-06,
"loss": 1.5411,
"step": 53
},
{
"epoch": 1.8620689655172413,
"grad_norm": 1.5491082668304443,
"learning_rate": 6.625000000000001e-06,
"loss": 1.5279,
"step": 54
},
{
"epoch": 1.896551724137931,
"grad_norm": 2.956325054168701,
"learning_rate": 6.750000000000001e-06,
"loss": 1.5381,
"step": 55
},
{
"epoch": 1.9310344827586206,
"grad_norm": 5.207773685455322,
"learning_rate": 6.875000000000001e-06,
"loss": 1.5499,
"step": 56
},
{
"epoch": 1.9655172413793105,
"grad_norm": 1.360984444618225,
"learning_rate": 7.000000000000001e-06,
"loss": 1.5294,
"step": 57
},
{
"epoch": 2.0,
"grad_norm": 1.9358643293380737,
"learning_rate": 7.1249999999999995e-06,
"loss": 1.5492,
"step": 58
},
{
"epoch": 2.0344827586206895,
"grad_norm": 1.5760064125061035,
"learning_rate": 7.25e-06,
"loss": 1.5279,
"step": 59
},
{
"epoch": 2.0689655172413794,
"grad_norm": 2.5182933807373047,
"learning_rate": 7.375e-06,
"loss": 1.5272,
"step": 60
},
{
"epoch": 2.103448275862069,
"grad_norm": 1.9873909950256348,
"learning_rate": 7.5e-06,
"loss": 1.5298,
"step": 61
},
{
"epoch": 2.1379310344827585,
"grad_norm": 1.4847761392593384,
"learning_rate": 7.625e-06,
"loss": 1.5129,
"step": 62
},
{
"epoch": 2.1724137931034484,
"grad_norm": 1.4767472743988037,
"learning_rate": 7.75e-06,
"loss": 1.5255,
"step": 63
},
{
"epoch": 2.206896551724138,
"grad_norm": 4.426845550537109,
"learning_rate": 7.875e-06,
"loss": 1.5277,
"step": 64
},
{
"epoch": 2.2413793103448274,
"grad_norm": 2.6555216312408447,
"learning_rate": 8.000000000000001e-06,
"loss": 1.5278,
"step": 65
},
{
"epoch": 2.2758620689655173,
"grad_norm": 2.148672580718994,
"learning_rate": 8.125000000000001e-06,
"loss": 1.5202,
"step": 66
},
{
"epoch": 2.310344827586207,
"grad_norm": 2.529113531112671,
"learning_rate": 8.25e-06,
"loss": 1.533,
"step": 67
},
{
"epoch": 2.344827586206897,
"grad_norm": 2.6022942066192627,
"learning_rate": 8.375e-06,
"loss": 1.5221,
"step": 68
},
{
"epoch": 2.3793103448275863,
"grad_norm": 4.527515888214111,
"learning_rate": 8.500000000000002e-06,
"loss": 1.5143,
"step": 69
},
{
"epoch": 2.413793103448276,
"grad_norm": 3.554746627807617,
"learning_rate": 8.625e-06,
"loss": 1.5105,
"step": 70
},
{
"epoch": 2.4482758620689653,
"grad_norm": 2.6532673835754395,
"learning_rate": 8.75e-06,
"loss": 1.4933,
"step": 71
},
{
"epoch": 2.4827586206896552,
"grad_norm": 3.799637794494629,
"learning_rate": 8.875e-06,
"loss": 1.5067,
"step": 72
},
{
"epoch": 2.5172413793103448,
"grad_norm": 3.7445719242095947,
"learning_rate": 9e-06,
"loss": 1.5193,
"step": 73
},
{
"epoch": 2.5517241379310347,
"grad_norm": 1.5109844207763672,
"learning_rate": 9.125e-06,
"loss": 1.5219,
"step": 74
},
{
"epoch": 2.586206896551724,
"grad_norm": 1.2224950790405273,
"learning_rate": 9.25e-06,
"loss": 1.4971,
"step": 75
},
{
"epoch": 2.6206896551724137,
"grad_norm": 5.30598783493042,
"learning_rate": 9.375000000000001e-06,
"loss": 1.4901,
"step": 76
},
{
"epoch": 2.655172413793103,
"grad_norm": 1.4342738389968872,
"learning_rate": 9.5e-06,
"loss": 1.5118,
"step": 77
},
{
"epoch": 2.689655172413793,
"grad_norm": 4.92336368560791,
"learning_rate": 9.625e-06,
"loss": 1.4859,
"step": 78
},
{
"epoch": 2.7241379310344827,
"grad_norm": 10.244889259338379,
"learning_rate": 9.750000000000002e-06,
"loss": 1.5147,
"step": 79
},
{
"epoch": 2.7586206896551726,
"grad_norm": 5.509783744812012,
"learning_rate": 9.875000000000001e-06,
"loss": 1.5108,
"step": 80
},
{
"epoch": 2.793103448275862,
"grad_norm": 1.6187537908554077,
"learning_rate": 1e-05,
"loss": 1.4753,
"step": 81
},
{
"epoch": 2.8275862068965516,
"grad_norm": 1.2228670120239258,
"learning_rate": 1.0125e-05,
"loss": 1.5034,
"step": 82
},
{
"epoch": 2.862068965517241,
"grad_norm": 2.5575168132781982,
"learning_rate": 1.025e-05,
"loss": 1.515,
"step": 83
},
{
"epoch": 2.896551724137931,
"grad_norm": 2.906822919845581,
"learning_rate": 1.0375e-05,
"loss": 1.5243,
"step": 84
},
{
"epoch": 2.9310344827586206,
"grad_norm": 2.380941152572632,
"learning_rate": 1.05e-05,
"loss": 1.4967,
"step": 85
},
{
"epoch": 2.9655172413793105,
"grad_norm": 2.5720791816711426,
"learning_rate": 1.0625e-05,
"loss": 1.5065,
"step": 86
},
{
"epoch": 3.0,
"grad_norm": 2.6703929901123047,
"learning_rate": 1.075e-05,
"loss": 1.4984,
"step": 87
},
{
"epoch": 3.0344827586206895,
"grad_norm": 3.5005722045898438,
"learning_rate": 1.0875e-05,
"loss": 1.5122,
"step": 88
},
{
"epoch": 3.0689655172413794,
"grad_norm": 4.101701736450195,
"learning_rate": 1.1000000000000001e-05,
"loss": 1.5057,
"step": 89
},
{
"epoch": 3.103448275862069,
"grad_norm": 1.8032540082931519,
"learning_rate": 1.1125000000000001e-05,
"loss": 1.4953,
"step": 90
},
{
"epoch": 3.1379310344827585,
"grad_norm": 1.8906192779541016,
"learning_rate": 1.125e-05,
"loss": 1.4885,
"step": 91
},
{
"epoch": 3.1724137931034484,
"grad_norm": 4.843202114105225,
"learning_rate": 1.1375e-05,
"loss": 1.4819,
"step": 92
},
{
"epoch": 3.206896551724138,
"grad_norm": 2.6482224464416504,
"learning_rate": 1.1500000000000002e-05,
"loss": 1.4766,
"step": 93
},
{
"epoch": 3.2413793103448274,
"grad_norm": 2.58847975730896,
"learning_rate": 1.1625000000000001e-05,
"loss": 1.4883,
"step": 94
},
{
"epoch": 3.2758620689655173,
"grad_norm": 2.6886796951293945,
"learning_rate": 1.175e-05,
"loss": 1.4912,
"step": 95
},
{
"epoch": 3.310344827586207,
"grad_norm": 4.70560359954834,
"learning_rate": 1.1875e-05,
"loss": 1.4835,
"step": 96
},
{
"epoch": 3.344827586206897,
"grad_norm": 1.3306467533111572,
"learning_rate": 1.2e-05,
"loss": 1.4993,
"step": 97
},
{
"epoch": 3.3793103448275863,
"grad_norm": 6.546566009521484,
"learning_rate": 1.2125e-05,
"loss": 1.4564,
"step": 98
},
{
"epoch": 3.413793103448276,
"grad_norm": 2.113272190093994,
"learning_rate": 1.225e-05,
"loss": 1.4483,
"step": 99
},
{
"epoch": 3.4482758620689653,
"grad_norm": 3.0866124629974365,
"learning_rate": 1.2375000000000001e-05,
"loss": 1.4824,
"step": 100
},
{
"epoch": 3.4827586206896552,
"grad_norm": 1.7086529731750488,
"learning_rate": 1.25e-05,
"loss": 1.4845,
"step": 101
},
{
"epoch": 3.5172413793103448,
"grad_norm": 1.3498625755310059,
"learning_rate": 1.2625e-05,
"loss": 1.4632,
"step": 102
},
{
"epoch": 3.5517241379310347,
"grad_norm": 2.0302608013153076,
"learning_rate": 1.2750000000000002e-05,
"loss": 1.4623,
"step": 103
},
{
"epoch": 3.586206896551724,
"grad_norm": 1.4253138303756714,
"learning_rate": 1.2875000000000001e-05,
"loss": 1.4459,
"step": 104
},
{
"epoch": 3.6206896551724137,
"grad_norm": 2.8165297508239746,
"learning_rate": 1.3000000000000001e-05,
"loss": 1.448,
"step": 105
},
{
"epoch": 3.655172413793103,
"grad_norm": 2.0474743843078613,
"learning_rate": 1.3125e-05,
"loss": 1.4324,
"step": 106
},
{
"epoch": 3.689655172413793,
"grad_norm": 1.4272180795669556,
"learning_rate": 1.3250000000000002e-05,
"loss": 1.4328,
"step": 107
},
{
"epoch": 3.7241379310344827,
"grad_norm": 5.992876052856445,
"learning_rate": 1.3375000000000002e-05,
"loss": 1.4314,
"step": 108
},
{
"epoch": 3.7586206896551726,
"grad_norm": 3.8637657165527344,
"learning_rate": 1.3500000000000001e-05,
"loss": 1.4809,
"step": 109
},
{
"epoch": 3.793103448275862,
"grad_norm": 4.207685947418213,
"learning_rate": 1.3625e-05,
"loss": 1.4811,
"step": 110
},
{
"epoch": 3.8275862068965516,
"grad_norm": 1.545461654663086,
"learning_rate": 1.3750000000000002e-05,
"loss": 1.4942,
"step": 111
},
{
"epoch": 3.862068965517241,
"grad_norm": 1.6922414302825928,
"learning_rate": 1.3875000000000002e-05,
"loss": 1.4439,
"step": 112
},
{
"epoch": 3.896551724137931,
"grad_norm": 1.6125301122665405,
"learning_rate": 1.4000000000000001e-05,
"loss": 1.4745,
"step": 113
},
{
"epoch": 3.9310344827586206,
"grad_norm": 3.392108678817749,
"learning_rate": 1.4125e-05,
"loss": 1.4744,
"step": 114
},
{
"epoch": 3.9655172413793105,
"grad_norm": 2.7309863567352295,
"learning_rate": 1.4249999999999999e-05,
"loss": 1.4439,
"step": 115
},
{
"epoch": 4.0,
"grad_norm": 6.303447246551514,
"learning_rate": 1.4374999999999999e-05,
"loss": 1.461,
"step": 116
},
{
"epoch": 4.0344827586206895,
"grad_norm": 1.5163921117782593,
"learning_rate": 1.45e-05,
"loss": 1.4461,
"step": 117
},
{
"epoch": 4.068965517241379,
"grad_norm": 2.4618422985076904,
"learning_rate": 1.4625e-05,
"loss": 1.4361,
"step": 118
},
{
"epoch": 4.103448275862069,
"grad_norm": 2.786285877227783,
"learning_rate": 1.475e-05,
"loss": 1.4183,
"step": 119
},
{
"epoch": 4.137931034482759,
"grad_norm": 5.21203088760376,
"learning_rate": 1.4875e-05,
"loss": 1.4407,
"step": 120
},
{
"epoch": 4.172413793103448,
"grad_norm": 1.920630931854248,
"learning_rate": 1.5e-05,
"loss": 1.3928,
"step": 121
},
{
"epoch": 4.206896551724138,
"grad_norm": 2.916647434234619,
"learning_rate": 1.5125e-05,
"loss": 1.409,
"step": 122
},
{
"epoch": 4.241379310344827,
"grad_norm": 4.050491809844971,
"learning_rate": 1.525e-05,
"loss": 1.4555,
"step": 123
},
{
"epoch": 4.275862068965517,
"grad_norm": 3.479999303817749,
"learning_rate": 1.5375e-05,
"loss": 1.396,
"step": 124
},
{
"epoch": 4.310344827586207,
"grad_norm": 5.414281368255615,
"learning_rate": 1.55e-05,
"loss": 1.3829,
"step": 125
},
{
"epoch": 4.344827586206897,
"grad_norm": 4.517972946166992,
"learning_rate": 1.5625e-05,
"loss": 1.3182,
"step": 126
},
{
"epoch": 4.379310344827586,
"grad_norm": 10.26574993133545,
"learning_rate": 1.575e-05,
"loss": 1.4205,
"step": 127
},
{
"epoch": 4.413793103448276,
"grad_norm": 2.3254988193511963,
"learning_rate": 1.5875e-05,
"loss": 1.2768,
"step": 128
},
{
"epoch": 4.448275862068965,
"grad_norm": 5.00213098526001,
"learning_rate": 1.6000000000000003e-05,
"loss": 1.3419,
"step": 129
},
{
"epoch": 4.482758620689655,
"grad_norm": 11.136003494262695,
"learning_rate": 1.6125000000000002e-05,
"loss": 1.2928,
"step": 130
},
{
"epoch": 4.517241379310345,
"grad_norm": 6.113260269165039,
"learning_rate": 1.6250000000000002e-05,
"loss": 1.2989,
"step": 131
},
{
"epoch": 4.551724137931035,
"grad_norm": 5.029887676239014,
"learning_rate": 1.6375e-05,
"loss": 1.3367,
"step": 132
},
{
"epoch": 4.586206896551724,
"grad_norm": 15.060640335083008,
"learning_rate": 1.65e-05,
"loss": 1.4661,
"step": 133
},
{
"epoch": 4.620689655172414,
"grad_norm": 18.83147430419922,
"learning_rate": 1.6625e-05,
"loss": 1.5725,
"step": 134
},
{
"epoch": 4.655172413793103,
"grad_norm": 8.334407806396484,
"learning_rate": 1.675e-05,
"loss": 1.3389,
"step": 135
},
{
"epoch": 4.689655172413794,
"grad_norm": 5.727024555206299,
"learning_rate": 1.6875000000000004e-05,
"loss": 1.3305,
"step": 136
},
{
"epoch": 4.724137931034483,
"grad_norm": 6.0667009353637695,
"learning_rate": 1.7000000000000003e-05,
"loss": 1.2844,
"step": 137
},
{
"epoch": 4.758620689655173,
"grad_norm": 8.634448051452637,
"learning_rate": 1.7125000000000003e-05,
"loss": 1.2871,
"step": 138
},
{
"epoch": 4.793103448275862,
"grad_norm": 6.196165561676025,
"learning_rate": 1.725e-05,
"loss": 1.2925,
"step": 139
},
{
"epoch": 4.827586206896552,
"grad_norm": 4.187297821044922,
"learning_rate": 1.7375e-05,
"loss": 1.2614,
"step": 140
},
{
"epoch": 4.862068965517241,
"grad_norm": 4.917987823486328,
"learning_rate": 1.75e-05,
"loss": 1.2843,
"step": 141
},
{
"epoch": 4.896551724137931,
"grad_norm": 3.2452642917633057,
"learning_rate": 1.7625e-05,
"loss": 1.2741,
"step": 142
},
{
"epoch": 4.931034482758621,
"grad_norm": 6.927144527435303,
"learning_rate": 1.775e-05,
"loss": 1.2967,
"step": 143
},
{
"epoch": 4.9655172413793105,
"grad_norm": 4.876591682434082,
"learning_rate": 1.7875e-05,
"loss": 1.1916,
"step": 144
},
{
"epoch": 5.0,
"grad_norm": 3.315114736557007,
"learning_rate": 1.8e-05,
"loss": 1.1966,
"step": 145
},
{
"epoch": 5.0344827586206895,
"grad_norm": 4.320967674255371,
"learning_rate": 1.8125e-05,
"loss": 1.2583,
"step": 146
},
{
"epoch": 5.068965517241379,
"grad_norm": 3.7428784370422363,
"learning_rate": 1.825e-05,
"loss": 1.2523,
"step": 147
},
{
"epoch": 5.103448275862069,
"grad_norm": 4.975188255310059,
"learning_rate": 1.8375e-05,
"loss": 1.2224,
"step": 148
},
{
"epoch": 5.137931034482759,
"grad_norm": 5.255687713623047,
"learning_rate": 1.85e-05,
"loss": 1.1885,
"step": 149
},
{
"epoch": 5.172413793103448,
"grad_norm": 7.562132835388184,
"learning_rate": 1.8625000000000002e-05,
"loss": 1.1543,
"step": 150
},
{
"epoch": 5.206896551724138,
"grad_norm": 4.356451988220215,
"learning_rate": 1.8750000000000002e-05,
"loss": 1.2019,
"step": 151
},
{
"epoch": 5.241379310344827,
"grad_norm": 8.989137649536133,
"learning_rate": 1.8875e-05,
"loss": 1.378,
"step": 152
},
{
"epoch": 5.275862068965517,
"grad_norm": 2.729421615600586,
"learning_rate": 1.9e-05,
"loss": 1.1547,
"step": 153
},
{
"epoch": 5.310344827586207,
"grad_norm": 2.701746702194214,
"learning_rate": 1.9125e-05,
"loss": 1.184,
"step": 154
},
{
"epoch": 5.344827586206897,
"grad_norm": 4.610611915588379,
"learning_rate": 1.925e-05,
"loss": 1.2343,
"step": 155
},
{
"epoch": 5.379310344827586,
"grad_norm": 5.182763576507568,
"learning_rate": 1.9375e-05,
"loss": 1.189,
"step": 156
},
{
"epoch": 5.413793103448276,
"grad_norm": 5.774525165557861,
"learning_rate": 1.9500000000000003e-05,
"loss": 1.2143,
"step": 157
},
{
"epoch": 5.448275862068965,
"grad_norm": 3.1842565536499023,
"learning_rate": 1.9625000000000003e-05,
"loss": 1.1613,
"step": 158
},
{
"epoch": 5.482758620689655,
"grad_norm": 5.885169982910156,
"learning_rate": 1.9750000000000002e-05,
"loss": 1.24,
"step": 159
},
{
"epoch": 5.517241379310345,
"grad_norm": 8.06775951385498,
"learning_rate": 1.9875000000000002e-05,
"loss": 1.1973,
"step": 160
},
{
"epoch": 5.551724137931035,
"grad_norm": 3.398749589920044,
"learning_rate": 2e-05,
"loss": 1.1485,
"step": 161
},
{
"epoch": 5.586206896551724,
"grad_norm": 3.8745975494384766,
"learning_rate": 2.0125e-05,
"loss": 1.1218,
"step": 162
},
{
"epoch": 5.620689655172414,
"grad_norm": 6.615961074829102,
"learning_rate": 2.025e-05,
"loss": 1.1056,
"step": 163
},
{
"epoch": 5.655172413793103,
"grad_norm": 4.613058090209961,
"learning_rate": 2.0375e-05,
"loss": 1.0493,
"step": 164
},
{
"epoch": 5.689655172413794,
"grad_norm": 4.730791091918945,
"learning_rate": 2.05e-05,
"loss": 1.0686,
"step": 165
},
{
"epoch": 5.724137931034483,
"grad_norm": 6.086816310882568,
"learning_rate": 2.0625e-05,
"loss": 1.084,
"step": 166
},
{
"epoch": 5.758620689655173,
"grad_norm": 5.453216552734375,
"learning_rate": 2.075e-05,
"loss": 1.0637,
"step": 167
},
{
"epoch": 5.793103448275862,
"grad_norm": 4.013542652130127,
"learning_rate": 2.0875e-05,
"loss": 1.0962,
"step": 168
},
{
"epoch": 5.827586206896552,
"grad_norm": 3.824021816253662,
"learning_rate": 2.1e-05,
"loss": 1.0632,
"step": 169
},
{
"epoch": 5.862068965517241,
"grad_norm": 3.00637149810791,
"learning_rate": 2.1125000000000002e-05,
"loss": 1.0575,
"step": 170
},
{
"epoch": 5.896551724137931,
"grad_norm": 4.230837821960449,
"learning_rate": 2.125e-05,
"loss": 1.0095,
"step": 171
},
{
"epoch": 5.931034482758621,
"grad_norm": 4.6836371421813965,
"learning_rate": 2.1375e-05,
"loss": 1.1063,
"step": 172
},
{
"epoch": 5.9655172413793105,
"grad_norm": 3.750352382659912,
"learning_rate": 2.15e-05,
"loss": 1.1189,
"step": 173
},
{
"epoch": 6.0,
"grad_norm": 4.330487251281738,
"learning_rate": 2.1625e-05,
"loss": 1.0574,
"step": 174
},
{
"epoch": 6.0344827586206895,
"grad_norm": 5.845319747924805,
"learning_rate": 2.175e-05,
"loss": 1.08,
"step": 175
},
{
"epoch": 6.068965517241379,
"grad_norm": 3.027203321456909,
"learning_rate": 2.1875e-05,
"loss": 0.9774,
"step": 176
},
{
"epoch": 6.103448275862069,
"grad_norm": 3.455016851425171,
"learning_rate": 2.2000000000000003e-05,
"loss": 1.0531,
"step": 177
},
{
"epoch": 6.137931034482759,
"grad_norm": 3.488523483276367,
"learning_rate": 2.2125000000000002e-05,
"loss": 0.9497,
"step": 178
},
{
"epoch": 6.172413793103448,
"grad_norm": 3.371523141860962,
"learning_rate": 2.2250000000000002e-05,
"loss": 1.0177,
"step": 179
},
{
"epoch": 6.206896551724138,
"grad_norm": 5.724728107452393,
"learning_rate": 2.2375000000000002e-05,
"loss": 1.1333,
"step": 180
},
{
"epoch": 6.241379310344827,
"grad_norm": 3.2824714183807373,
"learning_rate": 2.25e-05,
"loss": 1.0047,
"step": 181
},
{
"epoch": 6.275862068965517,
"grad_norm": 7.460841178894043,
"learning_rate": 2.2625e-05,
"loss": 1.1217,
"step": 182
},
{
"epoch": 6.310344827586207,
"grad_norm": 4.130681037902832,
"learning_rate": 2.275e-05,
"loss": 1.066,
"step": 183
},
{
"epoch": 6.344827586206897,
"grad_norm": 3.287168025970459,
"learning_rate": 2.2875e-05,
"loss": 0.9983,
"step": 184
},
{
"epoch": 6.379310344827586,
"grad_norm": 4.445807456970215,
"learning_rate": 2.3000000000000003e-05,
"loss": 1.1108,
"step": 185
},
{
"epoch": 6.413793103448276,
"grad_norm": 4.80116081237793,
"learning_rate": 2.3125000000000003e-05,
"loss": 1.0195,
"step": 186
},
{
"epoch": 6.448275862068965,
"grad_norm": 3.579498529434204,
"learning_rate": 2.3250000000000003e-05,
"loss": 1.0323,
"step": 187
},
{
"epoch": 6.482758620689655,
"grad_norm": 3.3527798652648926,
"learning_rate": 2.3375000000000002e-05,
"loss": 0.976,
"step": 188
},
{
"epoch": 6.517241379310345,
"grad_norm": 2.929718017578125,
"learning_rate": 2.35e-05,
"loss": 0.9717,
"step": 189
},
{
"epoch": 6.551724137931035,
"grad_norm": 4.6633758544921875,
"learning_rate": 2.3624999999999998e-05,
"loss": 1.1071,
"step": 190
},
{
"epoch": 6.586206896551724,
"grad_norm": 4.373714923858643,
"learning_rate": 2.375e-05,
"loss": 0.9349,
"step": 191
},
{
"epoch": 6.620689655172414,
"grad_norm": 4.665045738220215,
"learning_rate": 2.3875e-05,
"loss": 0.9615,
"step": 192
},
{
"epoch": 6.655172413793103,
"grad_norm": 7.3372907638549805,
"learning_rate": 2.4e-05,
"loss": 0.9949,
"step": 193
},
{
"epoch": 6.689655172413794,
"grad_norm": 4.059077739715576,
"learning_rate": 2.4125e-05,
"loss": 0.9306,
"step": 194
},
{
"epoch": 6.724137931034483,
"grad_norm": 5.908609390258789,
"learning_rate": 2.425e-05,
"loss": 0.7885,
"step": 195
},
{
"epoch": 6.758620689655173,
"grad_norm": 6.472893238067627,
"learning_rate": 2.4375e-05,
"loss": 1.1013,
"step": 196
},
{
"epoch": 6.793103448275862,
"grad_norm": 4.9746575355529785,
"learning_rate": 2.45e-05,
"loss": 1.0222,
"step": 197
},
{
"epoch": 6.827586206896552,
"grad_norm": 5.179764270782471,
"learning_rate": 2.4625000000000002e-05,
"loss": 0.9958,
"step": 198
},
{
"epoch": 6.862068965517241,
"grad_norm": 7.71660041809082,
"learning_rate": 2.4750000000000002e-05,
"loss": 1.05,
"step": 199
},
{
"epoch": 6.896551724137931,
"grad_norm": 6.319756507873535,
"learning_rate": 2.4875e-05,
"loss": 1.0519,
"step": 200
},
{
"epoch": 6.931034482758621,
"grad_norm": 4.988058567047119,
"learning_rate": 2.5e-05,
"loss": 0.998,
"step": 201
},
{
"epoch": 6.9655172413793105,
"grad_norm": 5.239701747894287,
"learning_rate": 2.4999980961416097e-05,
"loss": 1.0276,
"step": 202
},
{
"epoch": 7.0,
"grad_norm": 9.979659080505371,
"learning_rate": 2.499992384572238e-05,
"loss": 1.005,
"step": 203
},
{
"epoch": 7.0344827586206895,
"grad_norm": 5.444819927215576,
"learning_rate": 2.4999828653092835e-05,
"loss": 1.0034,
"step": 204
},
{
"epoch": 7.068965517241379,
"grad_norm": 5.544732570648193,
"learning_rate": 2.4999695383817435e-05,
"loss": 1.0074,
"step": 205
},
{
"epoch": 7.103448275862069,
"grad_norm": 7.6183762550354,
"learning_rate": 2.499952403830214e-05,
"loss": 0.9798,
"step": 206
},
{
"epoch": 7.137931034482759,
"grad_norm": 4.80704927444458,
"learning_rate": 2.4999314617068904e-05,
"loss": 0.9429,
"step": 207
},
{
"epoch": 7.172413793103448,
"grad_norm": 5.107629776000977,
"learning_rate": 2.4999067120755652e-05,
"loss": 1.0054,
"step": 208
},
{
"epoch": 7.206896551724138,
"grad_norm": 5.443073272705078,
"learning_rate": 2.4998781550116305e-05,
"loss": 1.0603,
"step": 209
},
{
"epoch": 7.241379310344827,
"grad_norm": 2.814347267150879,
"learning_rate": 2.499845790602076e-05,
"loss": 1.0818,
"step": 210
},
{
"epoch": 7.275862068965517,
"grad_norm": 4.260725021362305,
"learning_rate": 2.4998096189454893e-05,
"loss": 0.9751,
"step": 211
},
{
"epoch": 7.310344827586207,
"grad_norm": 3.0248708724975586,
"learning_rate": 2.4997696401520555e-05,
"loss": 0.9156,
"step": 212
},
{
"epoch": 7.344827586206897,
"grad_norm": 6.022355079650879,
"learning_rate": 2.499725854343557e-05,
"loss": 0.9621,
"step": 213
},
{
"epoch": 7.379310344827586,
"grad_norm": 2.480070114135742,
"learning_rate": 2.4996782616533732e-05,
"loss": 0.9835,
"step": 214
},
{
"epoch": 7.413793103448276,
"grad_norm": 2.478194236755371,
"learning_rate": 2.499626862226479e-05,
"loss": 0.8196,
"step": 215
},
{
"epoch": 7.448275862068965,
"grad_norm": 2.9012348651885986,
"learning_rate": 2.4995716562194465e-05,
"loss": 0.9136,
"step": 216
},
{
"epoch": 7.482758620689655,
"grad_norm": 2.776332378387451,
"learning_rate": 2.499512643800443e-05,
"loss": 0.874,
"step": 217
},
{
"epoch": 7.517241379310345,
"grad_norm": 3.0665347576141357,
"learning_rate": 2.4994498251492302e-05,
"loss": 0.8381,
"step": 218
},
{
"epoch": 7.551724137931035,
"grad_norm": 3.6337053775787354,
"learning_rate": 2.4993832004571646e-05,
"loss": 0.9145,
"step": 219
},
{
"epoch": 7.586206896551724,
"grad_norm": 3.960299015045166,
"learning_rate": 2.4993127699271966e-05,
"loss": 0.8541,
"step": 220
},
{
"epoch": 7.620689655172414,
"grad_norm": 4.1685967445373535,
"learning_rate": 2.49923853377387e-05,
"loss": 0.9458,
"step": 221
},
{
"epoch": 7.655172413793103,
"grad_norm": 3.217474937438965,
"learning_rate": 2.4991604922233204e-05,
"loss": 0.842,
"step": 222
},
{
"epoch": 7.689655172413794,
"grad_norm": 3.9884676933288574,
"learning_rate": 2.4990786455132764e-05,
"loss": 0.8746,
"step": 223
},
{
"epoch": 7.724137931034483,
"grad_norm": 6.753876209259033,
"learning_rate": 2.4989929938930576e-05,
"loss": 0.8353,
"step": 224
},
{
"epoch": 7.758620689655173,
"grad_norm": 4.734915256500244,
"learning_rate": 2.498903537623573e-05,
"loss": 0.8488,
"step": 225
},
{
"epoch": 7.793103448275862,
"grad_norm": 2.8307461738586426,
"learning_rate": 2.4988102769773227e-05,
"loss": 0.8886,
"step": 226
},
{
"epoch": 7.827586206896552,
"grad_norm": 4.811133861541748,
"learning_rate": 2.4987132122383936e-05,
"loss": 0.8869,
"step": 227
},
{
"epoch": 7.862068965517241,
"grad_norm": 6.6021728515625,
"learning_rate": 2.4986123437024627e-05,
"loss": 1.0391,
"step": 228
},
{
"epoch": 7.896551724137931,
"grad_norm": 8.46661376953125,
"learning_rate": 2.4985076716767927e-05,
"loss": 0.9857,
"step": 229
},
{
"epoch": 7.931034482758621,
"grad_norm": 7.210445880889893,
"learning_rate": 2.4983991964802327e-05,
"loss": 0.8793,
"step": 230
},
{
"epoch": 7.9655172413793105,
"grad_norm": 4.752058506011963,
"learning_rate": 2.4982869184432174e-05,
"loss": 0.8634,
"step": 231
},
{
"epoch": 8.0,
"grad_norm": 8.460209846496582,
"learning_rate": 2.498170837907765e-05,
"loss": 0.9404,
"step": 232
},
{
"epoch": 8.03448275862069,
"grad_norm": 4.385580539703369,
"learning_rate": 2.4980509552274765e-05,
"loss": 0.8939,
"step": 233
},
{
"epoch": 8.068965517241379,
"grad_norm": 4.554064750671387,
"learning_rate": 2.4979272707675356e-05,
"loss": 0.8854,
"step": 234
},
{
"epoch": 8.10344827586207,
"grad_norm": 5.013895034790039,
"learning_rate": 2.497799784904707e-05,
"loss": 0.8524,
"step": 235
},
{
"epoch": 8.137931034482758,
"grad_norm": 2.559938669204712,
"learning_rate": 2.4976684980273338e-05,
"loss": 0.8763,
"step": 236
},
{
"epoch": 8.172413793103448,
"grad_norm": 2.896611452102661,
"learning_rate": 2.4975334105353396e-05,
"loss": 0.8625,
"step": 237
},
{
"epoch": 8.206896551724139,
"grad_norm": 3.1764471530914307,
"learning_rate": 2.497394522840224e-05,
"loss": 0.8599,
"step": 238
},
{
"epoch": 8.241379310344827,
"grad_norm": 2.926391363143921,
"learning_rate": 2.4972518353650626e-05,
"loss": 0.8583,
"step": 239
},
{
"epoch": 8.275862068965518,
"grad_norm": 3.4869906902313232,
"learning_rate": 2.497105348544507e-05,
"loss": 0.8417,
"step": 240
},
{
"epoch": 8.310344827586206,
"grad_norm": 5.803765296936035,
"learning_rate": 2.4969550628247805e-05,
"loss": 0.8914,
"step": 241
},
{
"epoch": 8.344827586206897,
"grad_norm": 3.910322904586792,
"learning_rate": 2.49680097866368e-05,
"loss": 0.8737,
"step": 242
},
{
"epoch": 8.379310344827585,
"grad_norm": 3.2522225379943848,
"learning_rate": 2.4966430965305727e-05,
"loss": 0.8915,
"step": 243
},
{
"epoch": 8.413793103448276,
"grad_norm": 3.7491557598114014,
"learning_rate": 2.4964814169063948e-05,
"loss": 0.9399,
"step": 244
},
{
"epoch": 8.448275862068966,
"grad_norm": 2.59580397605896,
"learning_rate": 2.4963159402836506e-05,
"loss": 0.9132,
"step": 245
},
{
"epoch": 8.482758620689655,
"grad_norm": 2.840510368347168,
"learning_rate": 2.49614666716641e-05,
"loss": 0.8135,
"step": 246
},
{
"epoch": 8.517241379310345,
"grad_norm": 6.813579082489014,
"learning_rate": 2.495973598070309e-05,
"loss": 0.8306,
"step": 247
},
{
"epoch": 8.551724137931034,
"grad_norm": 5.595628261566162,
"learning_rate": 2.4957967335225456e-05,
"loss": 0.9031,
"step": 248
},
{
"epoch": 8.586206896551724,
"grad_norm": 3.344548463821411,
"learning_rate": 2.4956160740618806e-05,
"loss": 0.8205,
"step": 249
},
{
"epoch": 8.620689655172415,
"grad_norm": 2.66579270362854,
"learning_rate": 2.495431620238633e-05,
"loss": 0.8087,
"step": 250
},
{
"epoch": 8.655172413793103,
"grad_norm": 3.5464556217193604,
"learning_rate": 2.495243372614682e-05,
"loss": 0.7636,
"step": 251
},
{
"epoch": 8.689655172413794,
"grad_norm": 5.313910484313965,
"learning_rate": 2.495051331763462e-05,
"loss": 0.8267,
"step": 252
},
{
"epoch": 8.724137931034482,
"grad_norm": 5.469503879547119,
"learning_rate": 2.494855498269963e-05,
"loss": 0.7766,
"step": 253
},
{
"epoch": 8.758620689655173,
"grad_norm": 4.186700820922852,
"learning_rate": 2.4946558727307277e-05,
"loss": 0.8521,
"step": 254
},
{
"epoch": 8.793103448275861,
"grad_norm": 3.4536843299865723,
"learning_rate": 2.4944524557538503e-05,
"loss": 0.7532,
"step": 255
},
{
"epoch": 8.827586206896552,
"grad_norm": 3.1947126388549805,
"learning_rate": 2.4942452479589735e-05,
"loss": 0.716,
"step": 256
},
{
"epoch": 8.862068965517242,
"grad_norm": 4.5636138916015625,
"learning_rate": 2.494034249977289e-05,
"loss": 0.9678,
"step": 257
},
{
"epoch": 8.89655172413793,
"grad_norm": 4.235171794891357,
"learning_rate": 2.4938194624515333e-05,
"loss": 0.9419,
"step": 258
},
{
"epoch": 8.931034482758621,
"grad_norm": 4.488134860992432,
"learning_rate": 2.4936008860359854e-05,
"loss": 0.9512,
"step": 259
},
{
"epoch": 8.96551724137931,
"grad_norm": 5.502348899841309,
"learning_rate": 2.4933785213964677e-05,
"loss": 0.8107,
"step": 260
},
{
"epoch": 9.0,
"grad_norm": 3.3806426525115967,
"learning_rate": 2.4931523692103418e-05,
"loss": 0.8593,
"step": 261
},
{
"epoch": 9.03448275862069,
"grad_norm": 5.362583160400391,
"learning_rate": 2.492922430166506e-05,
"loss": 0.9019,
"step": 262
},
{
"epoch": 9.068965517241379,
"grad_norm": 6.098759651184082,
"learning_rate": 2.4926887049653943e-05,
"loss": 0.9845,
"step": 263
},
{
"epoch": 9.10344827586207,
"grad_norm": 3.2926342487335205,
"learning_rate": 2.492451194318975e-05,
"loss": 0.8941,
"step": 264
},
{
"epoch": 9.137931034482758,
"grad_norm": 3.6239383220672607,
"learning_rate": 2.4922098989507454e-05,
"loss": 0.9016,
"step": 265
},
{
"epoch": 9.172413793103448,
"grad_norm": 2.945098400115967,
"learning_rate": 2.4919648195957344e-05,
"loss": 0.8249,
"step": 266
},
{
"epoch": 9.206896551724139,
"grad_norm": 5.63911247253418,
"learning_rate": 2.4917159570004954e-05,
"loss": 0.8786,
"step": 267
},
{
"epoch": 9.241379310344827,
"grad_norm": 4.707492351531982,
"learning_rate": 2.491463311923108e-05,
"loss": 0.9624,
"step": 268
},
{
"epoch": 9.275862068965518,
"grad_norm": 3.5989744663238525,
"learning_rate": 2.491206885133171e-05,
"loss": 0.7872,
"step": 269
},
{
"epoch": 9.310344827586206,
"grad_norm": 3.1965887546539307,
"learning_rate": 2.490946677411807e-05,
"loss": 0.8428,
"step": 270
},
{
"epoch": 9.344827586206897,
"grad_norm": 3.745168924331665,
"learning_rate": 2.4906826895516528e-05,
"loss": 0.8836,
"step": 271
},
{
"epoch": 9.379310344827585,
"grad_norm": 4.135939598083496,
"learning_rate": 2.490414922356861e-05,
"loss": 0.8569,
"step": 272
},
{
"epoch": 9.413793103448276,
"grad_norm": 3.0350985527038574,
"learning_rate": 2.4901433766430975e-05,
"loss": 0.8092,
"step": 273
},
{
"epoch": 9.448275862068966,
"grad_norm": 2.5709338188171387,
"learning_rate": 2.4898680532375374e-05,
"loss": 0.8073,
"step": 274
},
{
"epoch": 9.482758620689655,
"grad_norm": 3.360161542892456,
"learning_rate": 2.489588952978863e-05,
"loss": 0.7794,
"step": 275
},
{
"epoch": 9.517241379310345,
"grad_norm": 3.463740587234497,
"learning_rate": 2.4893060767172632e-05,
"loss": 0.8216,
"step": 276
},
{
"epoch": 9.551724137931034,
"grad_norm": 4.866174221038818,
"learning_rate": 2.489019425314427e-05,
"loss": 0.821,
"step": 277
},
{
"epoch": 9.586206896551724,
"grad_norm": 4.300073146820068,
"learning_rate": 2.4887289996435452e-05,
"loss": 0.8998,
"step": 278
},
{
"epoch": 9.620689655172415,
"grad_norm": 7.732598304748535,
"learning_rate": 2.4884348005893045e-05,
"loss": 0.8335,
"step": 279
},
{
"epoch": 9.655172413793103,
"grad_norm": 6.1583027839660645,
"learning_rate": 2.488136829047886e-05,
"loss": 0.8181,
"step": 280
},
{
"epoch": 9.689655172413794,
"grad_norm": 3.6511595249176025,
"learning_rate": 2.487835085926963e-05,
"loss": 0.8059,
"step": 281
},
{
"epoch": 9.724137931034482,
"grad_norm": 3.68314790725708,
"learning_rate": 2.487529572145697e-05,
"loss": 0.8394,
"step": 282
},
{
"epoch": 9.758620689655173,
"grad_norm": 3.6726181507110596,
"learning_rate": 2.4872202886347362e-05,
"loss": 0.7453,
"step": 283
},
{
"epoch": 9.793103448275861,
"grad_norm": 2.719984531402588,
"learning_rate": 2.486907236336212e-05,
"loss": 0.8272,
"step": 284
},
{
"epoch": 9.827586206896552,
"grad_norm": 3.7963995933532715,
"learning_rate": 2.4865904162037358e-05,
"loss": 0.8991,
"step": 285
},
{
"epoch": 9.862068965517242,
"grad_norm": 2.754535436630249,
"learning_rate": 2.4862698292023963e-05,
"loss": 0.7064,
"step": 286
},
{
"epoch": 9.89655172413793,
"grad_norm": 3.0903878211975098,
"learning_rate": 2.4859454763087577e-05,
"loss": 0.735,
"step": 287
},
{
"epoch": 9.931034482758621,
"grad_norm": 4.279144287109375,
"learning_rate": 2.4856173585108544e-05,
"loss": 0.8074,
"step": 288
},
{
"epoch": 9.96551724137931,
"grad_norm": 4.23798942565918,
"learning_rate": 2.4852854768081912e-05,
"loss": 0.7802,
"step": 289
},
{
"epoch": 10.0,
"grad_norm": 3.5624747276306152,
"learning_rate": 2.4849498322117364e-05,
"loss": 0.8029,
"step": 290
},
{
"epoch": 10.03448275862069,
"grad_norm": 2.764233350753784,
"learning_rate": 2.4846104257439222e-05,
"loss": 0.7497,
"step": 291
},
{
"epoch": 10.068965517241379,
"grad_norm": 4.359961032867432,
"learning_rate": 2.4842672584386396e-05,
"loss": 0.7794,
"step": 292
},
{
"epoch": 10.10344827586207,
"grad_norm": 5.876353740692139,
"learning_rate": 2.483920331341235e-05,
"loss": 0.8483,
"step": 293
},
{
"epoch": 10.137931034482758,
"grad_norm": 4.351016998291016,
"learning_rate": 2.4835696455085093e-05,
"loss": 0.7924,
"step": 294
},
{
"epoch": 10.172413793103448,
"grad_norm": 3.5949246883392334,
"learning_rate": 2.483215202008712e-05,
"loss": 0.8293,
"step": 295
},
{
"epoch": 10.206896551724139,
"grad_norm": 3.226196765899658,
"learning_rate": 2.4828570019215396e-05,
"loss": 0.7934,
"step": 296
},
{
"epoch": 10.241379310344827,
"grad_norm": 5.124762535095215,
"learning_rate": 2.4824950463381314e-05,
"loss": 0.8055,
"step": 297
},
{
"epoch": 10.275862068965518,
"grad_norm": 3.658576011657715,
"learning_rate": 2.482129336361067e-05,
"loss": 0.6952,
"step": 298
},
{
"epoch": 10.310344827586206,
"grad_norm": 3.146904945373535,
"learning_rate": 2.481759873104363e-05,
"loss": 0.714,
"step": 299
},
{
"epoch": 10.344827586206897,
"grad_norm": 2.9892313480377197,
"learning_rate": 2.4813866576934676e-05,
"loss": 0.7746,
"step": 300
},
{
"epoch": 10.379310344827585,
"grad_norm": 2.505178928375244,
"learning_rate": 2.4810096912652604e-05,
"loss": 0.6921,
"step": 301
},
{
"epoch": 10.413793103448276,
"grad_norm": 2.44994854927063,
"learning_rate": 2.480628974968046e-05,
"loss": 0.7294,
"step": 302
},
{
"epoch": 10.448275862068966,
"grad_norm": 3.606544256210327,
"learning_rate": 2.4802445099615525e-05,
"loss": 0.8161,
"step": 303
},
{
"epoch": 10.482758620689655,
"grad_norm": 3.289849042892456,
"learning_rate": 2.479856297416927e-05,
"loss": 0.7754,
"step": 304
},
{
"epoch": 10.517241379310345,
"grad_norm": 3.646949529647827,
"learning_rate": 2.4794643385167327e-05,
"loss": 0.8014,
"step": 305
},
{
"epoch": 10.551724137931034,
"grad_norm": 6.4262847900390625,
"learning_rate": 2.4790686344549436e-05,
"loss": 0.8158,
"step": 306
},
{
"epoch": 10.586206896551724,
"grad_norm": 3.1850662231445312,
"learning_rate": 2.478669186436943e-05,
"loss": 0.7479,
"step": 307
},
{
"epoch": 10.620689655172415,
"grad_norm": 3.0070602893829346,
"learning_rate": 2.478265995679519e-05,
"loss": 0.769,
"step": 308
},
{
"epoch": 10.655172413793103,
"grad_norm": 6.722799777984619,
"learning_rate": 2.4778590634108613e-05,
"loss": 0.7574,
"step": 309
},
{
"epoch": 10.689655172413794,
"grad_norm": 3.4408349990844727,
"learning_rate": 2.4774483908705546e-05,
"loss": 0.6743,
"step": 310
},
{
"epoch": 10.724137931034482,
"grad_norm": 2.932551383972168,
"learning_rate": 2.4770339793095802e-05,
"loss": 0.7759,
"step": 311
},
{
"epoch": 10.758620689655173,
"grad_norm": 3.669299602508545,
"learning_rate": 2.4766158299903062e-05,
"loss": 0.8693,
"step": 312
},
{
"epoch": 10.793103448275861,
"grad_norm": 3.167391061782837,
"learning_rate": 2.4761939441864895e-05,
"loss": 0.7359,
"step": 313
},
{
"epoch": 10.827586206896552,
"grad_norm": 2.894587993621826,
"learning_rate": 2.4757683231832662e-05,
"loss": 0.6821,
"step": 314
},
{
"epoch": 10.862068965517242,
"grad_norm": 3.0310933589935303,
"learning_rate": 2.4753389682771523e-05,
"loss": 0.7143,
"step": 315
},
{
"epoch": 10.89655172413793,
"grad_norm": 3.1652767658233643,
"learning_rate": 2.474905880776037e-05,
"loss": 0.7949,
"step": 316
},
{
"epoch": 10.931034482758621,
"grad_norm": 5.710330009460449,
"learning_rate": 2.47446906199918e-05,
"loss": 0.7518,
"step": 317
},
{
"epoch": 10.96551724137931,
"grad_norm": 5.877900123596191,
"learning_rate": 2.4740285132772072e-05,
"loss": 0.8513,
"step": 318
},
{
"epoch": 11.0,
"grad_norm": 3.645622730255127,
"learning_rate": 2.4735842359521064e-05,
"loss": 0.8153,
"step": 319
},
{
"epoch": 11.03448275862069,
"grad_norm": 3.9911906719207764,
"learning_rate": 2.4731362313772233e-05,
"loss": 0.7651,
"step": 320
},
{
"epoch": 11.068965517241379,
"grad_norm": 4.0215935707092285,
"learning_rate": 2.4726845009172572e-05,
"loss": 0.7985,
"step": 321
},
{
"epoch": 11.10344827586207,
"grad_norm": 2.402211904525757,
"learning_rate": 2.4722290459482578e-05,
"loss": 0.7054,
"step": 322
},
{
"epoch": 11.137931034482758,
"grad_norm": 2.782041549682617,
"learning_rate": 2.47176986785762e-05,
"loss": 0.7909,
"step": 323
},
{
"epoch": 11.172413793103448,
"grad_norm": 4.322781085968018,
"learning_rate": 2.47130696804408e-05,
"loss": 0.747,
"step": 324
},
{
"epoch": 11.206896551724139,
"grad_norm": 5.149921417236328,
"learning_rate": 2.47084034791771e-05,
"loss": 0.8314,
"step": 325
},
{
"epoch": 11.241379310344827,
"grad_norm": 3.920372486114502,
"learning_rate": 2.4703700088999167e-05,
"loss": 0.8623,
"step": 326
},
{
"epoch": 11.275862068965518,
"grad_norm": 2.6054768562316895,
"learning_rate": 2.4698959524234346e-05,
"loss": 0.7899,
"step": 327
},
{
"epoch": 11.310344827586206,
"grad_norm": 4.032412528991699,
"learning_rate": 2.4694181799323206e-05,
"loss": 0.8142,
"step": 328
},
{
"epoch": 11.344827586206897,
"grad_norm": 3.573333978652954,
"learning_rate": 2.468936692881954e-05,
"loss": 0.7309,
"step": 329
},
{
"epoch": 11.379310344827585,
"grad_norm": 7.383831977844238,
"learning_rate": 2.4684514927390274e-05,
"loss": 0.8518,
"step": 330
},
{
"epoch": 11.413793103448276,
"grad_norm": 3.8350508213043213,
"learning_rate": 2.4679625809815443e-05,
"loss": 0.7778,
"step": 331
},
{
"epoch": 11.448275862068966,
"grad_norm": 5.659679889678955,
"learning_rate": 2.467469959098815e-05,
"loss": 0.7818,
"step": 332
},
{
"epoch": 11.482758620689655,
"grad_norm": 6.893414497375488,
"learning_rate": 2.4669736285914505e-05,
"loss": 0.7871,
"step": 333
},
{
"epoch": 11.517241379310345,
"grad_norm": 3.4787838459014893,
"learning_rate": 2.4664735909713606e-05,
"loss": 0.6967,
"step": 334
},
{
"epoch": 11.551724137931034,
"grad_norm": 3.2118844985961914,
"learning_rate": 2.465969847761746e-05,
"loss": 0.6598,
"step": 335
},
{
"epoch": 11.586206896551724,
"grad_norm": 3.6841318607330322,
"learning_rate": 2.4654624004970957e-05,
"loss": 0.7323,
"step": 336
},
{
"epoch": 11.620689655172415,
"grad_norm": 4.46773624420166,
"learning_rate": 2.464951250723183e-05,
"loss": 0.6713,
"step": 337
},
{
"epoch": 11.655172413793103,
"grad_norm": 2.266669988632202,
"learning_rate": 2.4644363999970576e-05,
"loss": 0.6425,
"step": 338
},
{
"epoch": 11.689655172413794,
"grad_norm": 5.7526140213012695,
"learning_rate": 2.4639178498870452e-05,
"loss": 0.6836,
"step": 339
},
{
"epoch": 11.724137931034482,
"grad_norm": 5.038420677185059,
"learning_rate": 2.4633956019727385e-05,
"loss": 0.7308,
"step": 340
},
{
"epoch": 11.758620689655173,
"grad_norm": 5.116992473602295,
"learning_rate": 2.4628696578449956e-05,
"loss": 0.7882,
"step": 341
},
{
"epoch": 11.793103448275861,
"grad_norm": 4.496801376342773,
"learning_rate": 2.4623400191059335e-05,
"loss": 0.7327,
"step": 342
},
{
"epoch": 11.827586206896552,
"grad_norm": 5.261680603027344,
"learning_rate": 2.4618066873689238e-05,
"loss": 0.8085,
"step": 343
},
{
"epoch": 11.862068965517242,
"grad_norm": 4.4358367919921875,
"learning_rate": 2.461269664258587e-05,
"loss": 0.6914,
"step": 344
},
{
"epoch": 11.89655172413793,
"grad_norm": 3.8224196434020996,
"learning_rate": 2.4607289514107888e-05,
"loss": 0.6859,
"step": 345
},
{
"epoch": 11.931034482758621,
"grad_norm": 3.285635471343994,
"learning_rate": 2.460184550472635e-05,
"loss": 0.8106,
"step": 346
},
{
"epoch": 11.96551724137931,
"grad_norm": 4.5762939453125,
"learning_rate": 2.4596364631024643e-05,
"loss": 0.7377,
"step": 347
},
{
"epoch": 12.0,
"grad_norm": 3.780836820602417,
"learning_rate": 2.459084690969846e-05,
"loss": 0.707,
"step": 348
},
{
"epoch": 12.03448275862069,
"grad_norm": 4.315356731414795,
"learning_rate": 2.4585292357555746e-05,
"loss": 0.7106,
"step": 349
},
{
"epoch": 12.068965517241379,
"grad_norm": 3.3244175910949707,
"learning_rate": 2.457970099151662e-05,
"loss": 0.699,
"step": 350
},
{
"epoch": 12.10344827586207,
"grad_norm": 2.7675392627716064,
"learning_rate": 2.4574072828613354e-05,
"loss": 0.6546,
"step": 351
},
{
"epoch": 12.137931034482758,
"grad_norm": 3.6822872161865234,
"learning_rate": 2.4568407885990313e-05,
"loss": 0.6953,
"step": 352
},
{
"epoch": 12.172413793103448,
"grad_norm": 4.7930731773376465,
"learning_rate": 2.4562706180903894e-05,
"loss": 0.7863,
"step": 353
},
{
"epoch": 12.206896551724139,
"grad_norm": 3.7039031982421875,
"learning_rate": 2.4556967730722478e-05,
"loss": 0.7581,
"step": 354
},
{
"epoch": 12.241379310344827,
"grad_norm": 5.0022053718566895,
"learning_rate": 2.455119255292638e-05,
"loss": 0.8091,
"step": 355
},
{
"epoch": 12.275862068965518,
"grad_norm": 3.0869357585906982,
"learning_rate": 2.4545380665107786e-05,
"loss": 0.7677,
"step": 356
},
{
"epoch": 12.310344827586206,
"grad_norm": 3.7048325538635254,
"learning_rate": 2.453953208497073e-05,
"loss": 0.7886,
"step": 357
},
{
"epoch": 12.344827586206897,
"grad_norm": 4.7608561515808105,
"learning_rate": 2.4533646830330986e-05,
"loss": 0.9105,
"step": 358
},
{
"epoch": 12.379310344827585,
"grad_norm": 5.352680683135986,
"learning_rate": 2.452772491911607e-05,
"loss": 0.8114,
"step": 359
},
{
"epoch": 12.413793103448276,
"grad_norm": 8.703157424926758,
"learning_rate": 2.4521766369365142e-05,
"loss": 0.8285,
"step": 360
},
{
"epoch": 12.448275862068966,
"grad_norm": 3.2992141246795654,
"learning_rate": 2.4515771199228987e-05,
"loss": 0.7644,
"step": 361
},
{
"epoch": 12.482758620689655,
"grad_norm": 6.761632919311523,
"learning_rate": 2.450973942696993e-05,
"loss": 0.8596,
"step": 362
},
{
"epoch": 12.517241379310345,
"grad_norm": 4.267969608306885,
"learning_rate": 2.450367107096179e-05,
"loss": 0.7883,
"step": 363
},
{
"epoch": 12.551724137931034,
"grad_norm": 3.685096025466919,
"learning_rate": 2.449756614968984e-05,
"loss": 0.6954,
"step": 364
},
{
"epoch": 12.586206896551724,
"grad_norm": 3.5344960689544678,
"learning_rate": 2.449142468175072e-05,
"loss": 0.6647,
"step": 365
},
{
"epoch": 12.620689655172415,
"grad_norm": 4.839688777923584,
"learning_rate": 2.4485246685852413e-05,
"loss": 0.7058,
"step": 366
},
{
"epoch": 12.655172413793103,
"grad_norm": 3.455937623977661,
"learning_rate": 2.4479032180814166e-05,
"loss": 0.6476,
"step": 367
},
{
"epoch": 12.689655172413794,
"grad_norm": 4.819273948669434,
"learning_rate": 2.447278118556644e-05,
"loss": 0.6687,
"step": 368
},
{
"epoch": 12.724137931034482,
"grad_norm": 3.5306191444396973,
"learning_rate": 2.446649371915084e-05,
"loss": 0.709,
"step": 369
},
{
"epoch": 12.758620689655173,
"grad_norm": 5.589993953704834,
"learning_rate": 2.4460169800720095e-05,
"loss": 0.8271,
"step": 370
},
{
"epoch": 12.793103448275861,
"grad_norm": 2.941460132598877,
"learning_rate": 2.4453809449537947e-05,
"loss": 0.7747,
"step": 371
},
{
"epoch": 12.827586206896552,
"grad_norm": 2.6910247802734375,
"learning_rate": 2.4447412684979127e-05,
"loss": 0.6994,
"step": 372
},
{
"epoch": 12.862068965517242,
"grad_norm": 4.3834662437438965,
"learning_rate": 2.4440979526529295e-05,
"loss": 0.7782,
"step": 373
},
{
"epoch": 12.89655172413793,
"grad_norm": 5.008590221405029,
"learning_rate": 2.4434509993784972e-05,
"loss": 0.8679,
"step": 374
},
{
"epoch": 12.931034482758621,
"grad_norm": 3.8566648960113525,
"learning_rate": 2.4428004106453462e-05,
"loss": 0.8427,
"step": 375
},
{
"epoch": 12.96551724137931,
"grad_norm": 4.028674125671387,
"learning_rate": 2.4421461884352836e-05,
"loss": 0.7371,
"step": 376
},
{
"epoch": 13.0,
"grad_norm": 5.727315902709961,
"learning_rate": 2.4414883347411836e-05,
"loss": 0.7163,
"step": 377
},
{
"epoch": 13.03448275862069,
"grad_norm": 2.517099618911743,
"learning_rate": 2.440826851566983e-05,
"loss": 0.6465,
"step": 378
},
{
"epoch": 13.068965517241379,
"grad_norm": 2.401848316192627,
"learning_rate": 2.4401617409276735e-05,
"loss": 0.6307,
"step": 379
},
{
"epoch": 13.10344827586207,
"grad_norm": 3.5928995609283447,
"learning_rate": 2.439493004849298e-05,
"loss": 0.665,
"step": 380
},
{
"epoch": 13.137931034482758,
"grad_norm": 3.128108024597168,
"learning_rate": 2.438820645368942e-05,
"loss": 0.6785,
"step": 381
},
{
"epoch": 13.172413793103448,
"grad_norm": 2.5367393493652344,
"learning_rate": 2.4381446645347297e-05,
"loss": 0.6353,
"step": 382
},
{
"epoch": 13.206896551724139,
"grad_norm": 3.4342269897460938,
"learning_rate": 2.4374650644058156e-05,
"loss": 0.6067,
"step": 383
},
{
"epoch": 13.241379310344827,
"grad_norm": 2.9581339359283447,
"learning_rate": 2.43678184705238e-05,
"loss": 0.5938,
"step": 384
},
{
"epoch": 13.275862068965518,
"grad_norm": 4.147696018218994,
"learning_rate": 2.4360950145556208e-05,
"loss": 0.6053,
"step": 385
},
{
"epoch": 13.310344827586206,
"grad_norm": 3.097829580307007,
"learning_rate": 2.4354045690077492e-05,
"loss": 0.7284,
"step": 386
},
{
"epoch": 13.344827586206897,
"grad_norm": 3.5135653018951416,
"learning_rate": 2.4347105125119824e-05,
"loss": 0.6605,
"step": 387
},
{
"epoch": 13.379310344827585,
"grad_norm": 3.659688949584961,
"learning_rate": 2.4340128471825362e-05,
"loss": 0.6748,
"step": 388
},
{
"epoch": 13.413793103448276,
"grad_norm": 2.355313539505005,
"learning_rate": 2.4333115751446208e-05,
"loss": 0.6609,
"step": 389
},
{
"epoch": 13.448275862068966,
"grad_norm": 4.517484664916992,
"learning_rate": 2.4326066985344318e-05,
"loss": 0.6709,
"step": 390
},
{
"epoch": 13.482758620689655,
"grad_norm": 5.471557140350342,
"learning_rate": 2.4318982194991463e-05,
"loss": 0.7848,
"step": 391
},
{
"epoch": 13.517241379310345,
"grad_norm": 3.947722911834717,
"learning_rate": 2.4311861401969138e-05,
"loss": 0.7373,
"step": 392
},
{
"epoch": 13.551724137931034,
"grad_norm": 3.1544623374938965,
"learning_rate": 2.4304704627968515e-05,
"loss": 0.688,
"step": 393
},
{
"epoch": 13.586206896551724,
"grad_norm": 3.4366607666015625,
"learning_rate": 2.429751189479037e-05,
"loss": 0.6537,
"step": 394
},
{
"epoch": 13.620689655172415,
"grad_norm": 2.357426404953003,
"learning_rate": 2.429028322434501e-05,
"loss": 0.5864,
"step": 395
},
{
"epoch": 13.655172413793103,
"grad_norm": 2.7362890243530273,
"learning_rate": 2.4283018638652234e-05,
"loss": 0.6294,
"step": 396
},
{
"epoch": 13.689655172413794,
"grad_norm": 2.7302050590515137,
"learning_rate": 2.427571815984121e-05,
"loss": 0.6314,
"step": 397
},
{
"epoch": 13.724137931034482,
"grad_norm": 3.3525278568267822,
"learning_rate": 2.4268381810150474e-05,
"loss": 0.5765,
"step": 398
},
{
"epoch": 13.758620689655173,
"grad_norm": 5.29491662979126,
"learning_rate": 2.426100961192782e-05,
"loss": 0.6679,
"step": 399
},
{
"epoch": 13.793103448275861,
"grad_norm": 3.3575775623321533,
"learning_rate": 2.4253601587630236e-05,
"loss": 0.6177,
"step": 400
},
{
"epoch": 13.827586206896552,
"grad_norm": 3.106152057647705,
"learning_rate": 2.4246157759823855e-05,
"loss": 0.6268,
"step": 401
},
{
"epoch": 13.862068965517242,
"grad_norm": 3.3294615745544434,
"learning_rate": 2.4238678151183863e-05,
"loss": 0.6302,
"step": 402
},
{
"epoch": 13.89655172413793,
"grad_norm": 3.1822283267974854,
"learning_rate": 2.423116278449445e-05,
"loss": 0.626,
"step": 403
},
{
"epoch": 13.931034482758621,
"grad_norm": 2.9840409755706787,
"learning_rate": 2.4223611682648724e-05,
"loss": 0.6787,
"step": 404
},
{
"epoch": 13.96551724137931,
"grad_norm": 2.9849910736083984,
"learning_rate": 2.4216024868648644e-05,
"loss": 0.6664,
"step": 405
},
{
"epoch": 14.0,
"grad_norm": 4.7554731369018555,
"learning_rate": 2.4208402365604972e-05,
"loss": 0.6893,
"step": 406
},
{
"epoch": 14.03448275862069,
"grad_norm": 4.694082260131836,
"learning_rate": 2.420074419673717e-05,
"loss": 0.6614,
"step": 407
},
{
"epoch": 14.068965517241379,
"grad_norm": 4.404138088226318,
"learning_rate": 2.4193050385373344e-05,
"loss": 0.6874,
"step": 408
},
{
"epoch": 14.10344827586207,
"grad_norm": 3.568782091140747,
"learning_rate": 2.418532095495018e-05,
"loss": 0.6397,
"step": 409
},
{
"epoch": 14.137931034482758,
"grad_norm": 3.321381092071533,
"learning_rate": 2.417755592901287e-05,
"loss": 0.7479,
"step": 410
},
{
"epoch": 14.172413793103448,
"grad_norm": 2.899473190307617,
"learning_rate": 2.4169755331215023e-05,
"loss": 0.5792,
"step": 411
},
{
"epoch": 14.206896551724139,
"grad_norm": 5.696684837341309,
"learning_rate": 2.4161919185318617e-05,
"loss": 0.6417,
"step": 412
},
{
"epoch": 14.241379310344827,
"grad_norm": 3.3947043418884277,
"learning_rate": 2.4154047515193904e-05,
"loss": 0.596,
"step": 413
},
{
"epoch": 14.275862068965518,
"grad_norm": 4.2483229637146,
"learning_rate": 2.4146140344819363e-05,
"loss": 0.6651,
"step": 414
},
{
"epoch": 14.310344827586206,
"grad_norm": 4.062285423278809,
"learning_rate": 2.4138197698281606e-05,
"loss": 0.6853,
"step": 415
},
{
"epoch": 14.344827586206897,
"grad_norm": 3.0232627391815186,
"learning_rate": 2.413021959977531e-05,
"loss": 0.6696,
"step": 416
},
{
"epoch": 14.379310344827585,
"grad_norm": 2.5746848583221436,
"learning_rate": 2.4122206073603142e-05,
"loss": 0.7158,
"step": 417
},
{
"epoch": 14.413793103448276,
"grad_norm": 5.224025249481201,
"learning_rate": 2.4114157144175703e-05,
"loss": 0.7015,
"step": 418
},
{
"epoch": 14.448275862068966,
"grad_norm": 3.374368906021118,
"learning_rate": 2.4106072836011422e-05,
"loss": 0.6896,
"step": 419
},
{
"epoch": 14.482758620689655,
"grad_norm": 2.7048842906951904,
"learning_rate": 2.40979531737365e-05,
"loss": 0.6628,
"step": 420
},
{
"epoch": 14.517241379310345,
"grad_norm": 3.953042984008789,
"learning_rate": 2.4089798182084845e-05,
"loss": 0.6238,
"step": 421
},
{
"epoch": 14.551724137931034,
"grad_norm": 2.6758952140808105,
"learning_rate": 2.4081607885897966e-05,
"loss": 0.6396,
"step": 422
},
{
"epoch": 14.586206896551724,
"grad_norm": 3.827179193496704,
"learning_rate": 2.407338231012494e-05,
"loss": 0.6959,
"step": 423
},
{
"epoch": 14.620689655172415,
"grad_norm": 6.4605607986450195,
"learning_rate": 2.406512147982228e-05,
"loss": 0.7085,
"step": 424
},
{
"epoch": 14.655172413793103,
"grad_norm": 6.189979076385498,
"learning_rate": 2.4056825420153917e-05,
"loss": 0.7617,
"step": 425
},
{
"epoch": 14.689655172413794,
"grad_norm": 5.246057987213135,
"learning_rate": 2.4048494156391087e-05,
"loss": 0.7481,
"step": 426
},
{
"epoch": 14.724137931034482,
"grad_norm": 2.7838568687438965,
"learning_rate": 2.4040127713912264e-05,
"loss": 0.634,
"step": 427
},
{
"epoch": 14.758620689655173,
"grad_norm": 2.8083300590515137,
"learning_rate": 2.403172611820308e-05,
"loss": 0.6767,
"step": 428
},
{
"epoch": 14.793103448275861,
"grad_norm": 3.925990343093872,
"learning_rate": 2.4023289394856257e-05,
"loss": 0.7427,
"step": 429
},
{
"epoch": 14.827586206896552,
"grad_norm": 2.5433623790740967,
"learning_rate": 2.401481756957152e-05,
"loss": 0.6591,
"step": 430
},
{
"epoch": 14.862068965517242,
"grad_norm": 4.543763160705566,
"learning_rate": 2.4006310668155508e-05,
"loss": 0.6728,
"step": 431
},
{
"epoch": 14.89655172413793,
"grad_norm": 4.649848461151123,
"learning_rate": 2.3997768716521723e-05,
"loss": 0.741,
"step": 432
},
{
"epoch": 14.931034482758621,
"grad_norm": 3.5682265758514404,
"learning_rate": 2.398919174069043e-05,
"loss": 0.6855,
"step": 433
},
{
"epoch": 14.96551724137931,
"grad_norm": 2.9493958950042725,
"learning_rate": 2.398057976678859e-05,
"loss": 0.6561,
"step": 434
},
{
"epoch": 15.0,
"grad_norm": 2.569502115249634,
"learning_rate": 2.3971932821049765e-05,
"loss": 0.6723,
"step": 435
},
{
"epoch": 15.03448275862069,
"grad_norm": 2.2379672527313232,
"learning_rate": 2.396325092981405e-05,
"loss": 0.6323,
"step": 436
},
{
"epoch": 15.068965517241379,
"grad_norm": 3.066236972808838,
"learning_rate": 2.3954534119527996e-05,
"loss": 0.7536,
"step": 437
},
{
"epoch": 15.10344827586207,
"grad_norm": 3.0116641521453857,
"learning_rate": 2.3945782416744517e-05,
"loss": 0.7323,
"step": 438
},
{
"epoch": 15.137931034482758,
"grad_norm": 4.455348491668701,
"learning_rate": 2.3936995848122812e-05,
"loss": 0.6905,
"step": 439
},
{
"epoch": 15.172413793103448,
"grad_norm": 3.218564033508301,
"learning_rate": 2.3928174440428297e-05,
"loss": 0.663,
"step": 440
},
{
"epoch": 15.206896551724139,
"grad_norm": 3.6186459064483643,
"learning_rate": 2.391931822053251e-05,
"loss": 0.6548,
"step": 441
},
{
"epoch": 15.241379310344827,
"grad_norm": 2.9447264671325684,
"learning_rate": 2.3910427215413036e-05,
"loss": 0.6127,
"step": 442
},
{
"epoch": 15.275862068965518,
"grad_norm": 2.7375073432922363,
"learning_rate": 2.390150145215341e-05,
"loss": 0.6404,
"step": 443
},
{
"epoch": 15.310344827586206,
"grad_norm": 2.91684889793396,
"learning_rate": 2.3892540957943067e-05,
"loss": 0.629,
"step": 444
},
{
"epoch": 15.344827586206897,
"grad_norm": 4.3465399742126465,
"learning_rate": 2.3883545760077215e-05,
"loss": 0.7172,
"step": 445
},
{
"epoch": 15.379310344827585,
"grad_norm": 6.414719104766846,
"learning_rate": 2.3874515885956792e-05,
"loss": 0.6642,
"step": 446
},
{
"epoch": 15.413793103448276,
"grad_norm": 2.770498752593994,
"learning_rate": 2.386545136308836e-05,
"loss": 0.7017,
"step": 447
},
{
"epoch": 15.448275862068966,
"grad_norm": 3.6498284339904785,
"learning_rate": 2.3856352219084024e-05,
"loss": 0.6538,
"step": 448
},
{
"epoch": 15.482758620689655,
"grad_norm": 2.958104372024536,
"learning_rate": 2.384721848166136e-05,
"loss": 0.5846,
"step": 449
},
{
"epoch": 15.517241379310345,
"grad_norm": 2.3549644947052,
"learning_rate": 2.3838050178643312e-05,
"loss": 0.5435,
"step": 450
},
{
"epoch": 15.551724137931034,
"grad_norm": 3.659111261367798,
"learning_rate": 2.3828847337958126e-05,
"loss": 0.6394,
"step": 451
},
{
"epoch": 15.586206896551724,
"grad_norm": 2.7539238929748535,
"learning_rate": 2.3819609987639247e-05,
"loss": 0.6072,
"step": 452
},
{
"epoch": 15.620689655172415,
"grad_norm": 3.189476728439331,
"learning_rate": 2.3810338155825245e-05,
"loss": 0.6523,
"step": 453
},
{
"epoch": 15.655172413793103,
"grad_norm": 2.3295481204986572,
"learning_rate": 2.3801031870759732e-05,
"loss": 0.6303,
"step": 454
},
{
"epoch": 15.689655172413794,
"grad_norm": 2.7341103553771973,
"learning_rate": 2.379169116079126e-05,
"loss": 0.6822,
"step": 455
},
{
"epoch": 15.724137931034482,
"grad_norm": 2.9363160133361816,
"learning_rate": 2.378231605437326e-05,
"loss": 0.6329,
"step": 456
},
{
"epoch": 15.758620689655173,
"grad_norm": 5.169081211090088,
"learning_rate": 2.3772906580063924e-05,
"loss": 0.644,
"step": 457
},
{
"epoch": 15.793103448275861,
"grad_norm": 4.046895503997803,
"learning_rate": 2.3763462766526145e-05,
"loss": 0.6442,
"step": 458
},
{
"epoch": 15.827586206896552,
"grad_norm": 5.600619316101074,
"learning_rate": 2.3753984642527423e-05,
"loss": 0.6732,
"step": 459
},
{
"epoch": 15.862068965517242,
"grad_norm": 2.420551061630249,
"learning_rate": 2.3744472236939753e-05,
"loss": 0.5782,
"step": 460
},
{
"epoch": 15.89655172413793,
"grad_norm": 4.496841907501221,
"learning_rate": 2.3734925578739588e-05,
"loss": 0.6463,
"step": 461
},
{
"epoch": 15.931034482758621,
"grad_norm": 5.044814586639404,
"learning_rate": 2.3725344697007696e-05,
"loss": 0.5673,
"step": 462
},
{
"epoch": 15.96551724137931,
"grad_norm": 4.280508518218994,
"learning_rate": 2.3715729620929106e-05,
"loss": 0.5989,
"step": 463
},
{
"epoch": 16.0,
"grad_norm": 3.7099342346191406,
"learning_rate": 2.3706080379793007e-05,
"loss": 0.641,
"step": 464
},
{
"epoch": 16.03448275862069,
"grad_norm": 4.294964790344238,
"learning_rate": 2.3696397002992663e-05,
"loss": 0.6666,
"step": 465
},
{
"epoch": 16.06896551724138,
"grad_norm": 3.043022632598877,
"learning_rate": 2.3686679520025314e-05,
"loss": 0.6193,
"step": 466
},
{
"epoch": 16.103448275862068,
"grad_norm": 2.8048079013824463,
"learning_rate": 2.36769279604921e-05,
"loss": 0.6389,
"step": 467
},
{
"epoch": 16.137931034482758,
"grad_norm": 4.923339366912842,
"learning_rate": 2.366714235409797e-05,
"loss": 0.6676,
"step": 468
},
{
"epoch": 16.17241379310345,
"grad_norm": 2.6823081970214844,
"learning_rate": 2.365732273065157e-05,
"loss": 0.6144,
"step": 469
},
{
"epoch": 16.20689655172414,
"grad_norm": 2.74808406829834,
"learning_rate": 2.3647469120065177e-05,
"loss": 0.5502,
"step": 470
},
{
"epoch": 16.24137931034483,
"grad_norm": 4.341047286987305,
"learning_rate": 2.36375815523546e-05,
"loss": 0.6018,
"step": 471
},
{
"epoch": 16.275862068965516,
"grad_norm": 3.325765371322632,
"learning_rate": 2.3627660057639082e-05,
"loss": 0.6211,
"step": 472
},
{
"epoch": 16.310344827586206,
"grad_norm": 4.053094387054443,
"learning_rate": 2.361770466614122e-05,
"loss": 0.7121,
"step": 473
},
{
"epoch": 16.344827586206897,
"grad_norm": 3.4117484092712402,
"learning_rate": 2.3607715408186863e-05,
"loss": 0.6648,
"step": 474
},
{
"epoch": 16.379310344827587,
"grad_norm": 5.172271251678467,
"learning_rate": 2.3597692314205016e-05,
"loss": 0.659,
"step": 475
},
{
"epoch": 16.413793103448278,
"grad_norm": 3.7227883338928223,
"learning_rate": 2.358763541472777e-05,
"loss": 0.6244,
"step": 476
},
{
"epoch": 16.448275862068964,
"grad_norm": 2.213212013244629,
"learning_rate": 2.3577544740390184e-05,
"loss": 0.6075,
"step": 477
},
{
"epoch": 16.482758620689655,
"grad_norm": 2.4992291927337646,
"learning_rate": 2.35674203219302e-05,
"loss": 0.6339,
"step": 478
},
{
"epoch": 16.517241379310345,
"grad_norm": 5.619305610656738,
"learning_rate": 2.355726219018855e-05,
"loss": 0.6946,
"step": 479
},
{
"epoch": 16.551724137931036,
"grad_norm": 3.0107035636901855,
"learning_rate": 2.354707037610867e-05,
"loss": 0.6593,
"step": 480
},
{
"epoch": 16.586206896551722,
"grad_norm": 2.2236149311065674,
"learning_rate": 2.353684491073659e-05,
"loss": 0.615,
"step": 481
},
{
"epoch": 16.620689655172413,
"grad_norm": 2.139848470687866,
"learning_rate": 2.3526585825220848e-05,
"loss": 0.6075,
"step": 482
},
{
"epoch": 16.655172413793103,
"grad_norm": 2.2389163970947266,
"learning_rate": 2.35162931508124e-05,
"loss": 0.6533,
"step": 483
},
{
"epoch": 16.689655172413794,
"grad_norm": 3.55932879447937,
"learning_rate": 2.3505966918864525e-05,
"loss": 0.7435,
"step": 484
},
{
"epoch": 16.724137931034484,
"grad_norm": 2.86104416847229,
"learning_rate": 2.3495607160832707e-05,
"loss": 0.6867,
"step": 485
},
{
"epoch": 16.75862068965517,
"grad_norm": 3.4844839572906494,
"learning_rate": 2.3485213908274567e-05,
"loss": 0.571,
"step": 486
},
{
"epoch": 16.79310344827586,
"grad_norm": 2.288844108581543,
"learning_rate": 2.3474787192849756e-05,
"loss": 0.6348,
"step": 487
},
{
"epoch": 16.82758620689655,
"grad_norm": 2.6627936363220215,
"learning_rate": 2.346432704631986e-05,
"loss": 0.5483,
"step": 488
},
{
"epoch": 16.862068965517242,
"grad_norm": 3.744588851928711,
"learning_rate": 2.3453833500548295e-05,
"loss": 0.6306,
"step": 489
},
{
"epoch": 16.896551724137932,
"grad_norm": 2.5902559757232666,
"learning_rate": 2.3443306587500225e-05,
"loss": 0.6857,
"step": 490
},
{
"epoch": 16.93103448275862,
"grad_norm": 3.003737449645996,
"learning_rate": 2.3432746339242448e-05,
"loss": 0.6444,
"step": 491
},
{
"epoch": 16.96551724137931,
"grad_norm": 2.010629653930664,
"learning_rate": 2.342215278794332e-05,
"loss": 0.5778,
"step": 492
},
{
"epoch": 17.0,
"grad_norm": 3.528261423110962,
"learning_rate": 2.341152596587262e-05,
"loss": 0.6003,
"step": 493
},
{
"epoch": 17.03448275862069,
"grad_norm": 4.4283223152160645,
"learning_rate": 2.340086590540151e-05,
"loss": 0.6058,
"step": 494
},
{
"epoch": 17.06896551724138,
"grad_norm": 2.131556272506714,
"learning_rate": 2.339017263900237e-05,
"loss": 0.6174,
"step": 495
},
{
"epoch": 17.103448275862068,
"grad_norm": 3.118192672729492,
"learning_rate": 2.3379446199248747e-05,
"loss": 0.5973,
"step": 496
},
{
"epoch": 17.137931034482758,
"grad_norm": 4.915523052215576,
"learning_rate": 2.3368686618815238e-05,
"loss": 0.6529,
"step": 497
},
{
"epoch": 17.17241379310345,
"grad_norm": 3.0185019969940186,
"learning_rate": 2.335789393047739e-05,
"loss": 0.5984,
"step": 498
},
{
"epoch": 17.20689655172414,
"grad_norm": 2.78558087348938,
"learning_rate": 2.334706816711161e-05,
"loss": 0.6387,
"step": 499
},
{
"epoch": 17.24137931034483,
"grad_norm": 6.381213665008545,
"learning_rate": 2.3336209361695035e-05,
"loss": 0.6722,
"step": 500
},
{
"epoch": 17.275862068965516,
"grad_norm": 3.779954195022583,
"learning_rate": 2.3325317547305485e-05,
"loss": 0.6373,
"step": 501
},
{
"epoch": 17.310344827586206,
"grad_norm": 4.086780548095703,
"learning_rate": 2.3314392757121308e-05,
"loss": 0.6152,
"step": 502
},
{
"epoch": 17.344827586206897,
"grad_norm": 4.311843395233154,
"learning_rate": 2.3303435024421312e-05,
"loss": 0.6094,
"step": 503
},
{
"epoch": 17.379310344827587,
"grad_norm": 4.192368507385254,
"learning_rate": 2.3292444382584648e-05,
"loss": 0.6398,
"step": 504
},
{
"epoch": 17.413793103448278,
"grad_norm": 1.923520565032959,
"learning_rate": 2.328142086509072e-05,
"loss": 0.6108,
"step": 505
},
{
"epoch": 17.448275862068964,
"grad_norm": 3.6873779296875,
"learning_rate": 2.3270364505519073e-05,
"loss": 0.6146,
"step": 506
},
{
"epoch": 17.482758620689655,
"grad_norm": 2.5435547828674316,
"learning_rate": 2.32592753375493e-05,
"loss": 0.5429,
"step": 507
},
{
"epoch": 17.517241379310345,
"grad_norm": 4.662963390350342,
"learning_rate": 2.324815339496092e-05,
"loss": 0.5818,
"step": 508
},
{
"epoch": 17.551724137931036,
"grad_norm": 3.448092222213745,
"learning_rate": 2.3236998711633307e-05,
"loss": 0.5538,
"step": 509
},
{
"epoch": 17.586206896551722,
"grad_norm": 2.9241552352905273,
"learning_rate": 2.322581132154556e-05,
"loss": 0.5671,
"step": 510
},
{
"epoch": 17.620689655172413,
"grad_norm": 3.708261489868164,
"learning_rate": 2.3214591258776404e-05,
"loss": 0.5787,
"step": 511
},
{
"epoch": 17.655172413793103,
"grad_norm": 3.2863380908966064,
"learning_rate": 2.3203338557504105e-05,
"loss": 0.5985,
"step": 512
},
{
"epoch": 17.689655172413794,
"grad_norm": 2.276576042175293,
"learning_rate": 2.3192053252006335e-05,
"loss": 0.5104,
"step": 513
},
{
"epoch": 17.724137931034484,
"grad_norm": 4.1456780433654785,
"learning_rate": 2.3180735376660094e-05,
"loss": 0.5687,
"step": 514
},
{
"epoch": 17.75862068965517,
"grad_norm": 4.008676528930664,
"learning_rate": 2.3169384965941592e-05,
"loss": 0.6583,
"step": 515
},
{
"epoch": 17.79310344827586,
"grad_norm": 3.410106658935547,
"learning_rate": 2.3158002054426153e-05,
"loss": 0.6284,
"step": 516
},
{
"epoch": 17.82758620689655,
"grad_norm": 3.3623507022857666,
"learning_rate": 2.3146586676788095e-05,
"loss": 0.5826,
"step": 517
},
{
"epoch": 17.862068965517242,
"grad_norm": 2.8540868759155273,
"learning_rate": 2.3135138867800642e-05,
"loss": 0.6098,
"step": 518
},
{
"epoch": 17.896551724137932,
"grad_norm": 5.287265777587891,
"learning_rate": 2.3123658662335802e-05,
"loss": 0.6344,
"step": 519
},
{
"epoch": 17.93103448275862,
"grad_norm": 2.628873586654663,
"learning_rate": 2.311214609536427e-05,
"loss": 0.5716,
"step": 520
},
{
"epoch": 17.96551724137931,
"grad_norm": 3.3482089042663574,
"learning_rate": 2.3100601201955324e-05,
"loss": 0.6347,
"step": 521
},
{
"epoch": 18.0,
"grad_norm": 2.661550521850586,
"learning_rate": 2.308902401727672e-05,
"loss": 0.6066,
"step": 522
},
{
"epoch": 18.03448275862069,
"grad_norm": 3.0608065128326416,
"learning_rate": 2.3077414576594553e-05,
"loss": 0.639,
"step": 523
},
{
"epoch": 18.06896551724138,
"grad_norm": 5.210854530334473,
"learning_rate": 2.3065772915273203e-05,
"loss": 0.6772,
"step": 524
},
{
"epoch": 18.103448275862068,
"grad_norm": 3.0067954063415527,
"learning_rate": 2.305409906877519e-05,
"loss": 0.6215,
"step": 525
},
{
"epoch": 18.137931034482758,
"grad_norm": 3.596611261367798,
"learning_rate": 2.3042393072661074e-05,
"loss": 0.6174,
"step": 526
},
{
"epoch": 18.17241379310345,
"grad_norm": 2.643958806991577,
"learning_rate": 2.3030654962589346e-05,
"loss": 0.6218,
"step": 527
},
{
"epoch": 18.20689655172414,
"grad_norm": 2.740034818649292,
"learning_rate": 2.3018884774316328e-05,
"loss": 0.5306,
"step": 528
},
{
"epoch": 18.24137931034483,
"grad_norm": 2.9877166748046875,
"learning_rate": 2.3007082543696055e-05,
"loss": 0.573,
"step": 529
},
{
"epoch": 18.275862068965516,
"grad_norm": 2.206543445587158,
"learning_rate": 2.299524830668017e-05,
"loss": 0.5803,
"step": 530
},
{
"epoch": 18.310344827586206,
"grad_norm": 3.766697883605957,
"learning_rate": 2.2983382099317803e-05,
"loss": 0.7642,
"step": 531
},
{
"epoch": 18.344827586206897,
"grad_norm": 2.516557455062866,
"learning_rate": 2.2971483957755487e-05,
"loss": 0.588,
"step": 532
},
{
"epoch": 18.379310344827587,
"grad_norm": 5.181385517120361,
"learning_rate": 2.295955391823702e-05,
"loss": 0.7196,
"step": 533
},
{
"epoch": 18.413793103448278,
"grad_norm": 6.308480739593506,
"learning_rate": 2.2947592017103383e-05,
"loss": 0.7492,
"step": 534
},
{
"epoch": 18.448275862068964,
"grad_norm": 5.336180210113525,
"learning_rate": 2.2935598290792583e-05,
"loss": 0.6505,
"step": 535
},
{
"epoch": 18.482758620689655,
"grad_norm": 3.4817123413085938,
"learning_rate": 2.2923572775839603e-05,
"loss": 0.624,
"step": 536
},
{
"epoch": 18.517241379310345,
"grad_norm": 3.327723503112793,
"learning_rate": 2.2911515508876243e-05,
"loss": 0.5971,
"step": 537
},
{
"epoch": 18.551724137931036,
"grad_norm": 4.9538350105285645,
"learning_rate": 2.2899426526631033e-05,
"loss": 0.6098,
"step": 538
},
{
"epoch": 18.586206896551722,
"grad_norm": 4.590574264526367,
"learning_rate": 2.2887305865929104e-05,
"loss": 0.6799,
"step": 539
},
{
"epoch": 18.620689655172413,
"grad_norm": 3.821331262588501,
"learning_rate": 2.2875153563692094e-05,
"loss": 0.6897,
"step": 540
},
{
"epoch": 18.655172413793103,
"grad_norm": 3.8579297065734863,
"learning_rate": 2.286296965693802e-05,
"loss": 0.6482,
"step": 541
},
{
"epoch": 18.689655172413794,
"grad_norm": 6.096618175506592,
"learning_rate": 2.285075418278118e-05,
"loss": 0.6636,
"step": 542
},
{
"epoch": 18.724137931034484,
"grad_norm": 5.061996936798096,
"learning_rate": 2.283850717843202e-05,
"loss": 0.6584,
"step": 543
},
{
"epoch": 18.75862068965517,
"grad_norm": 3.448591709136963,
"learning_rate": 2.2826228681197047e-05,
"loss": 0.6224,
"step": 544
},
{
"epoch": 18.79310344827586,
"grad_norm": 5.197028160095215,
"learning_rate": 2.281391872847869e-05,
"loss": 0.5659,
"step": 545
},
{
"epoch": 18.82758620689655,
"grad_norm": 4.6534318923950195,
"learning_rate": 2.2801577357775193e-05,
"loss": 0.6029,
"step": 546
},
{
"epoch": 18.862068965517242,
"grad_norm": 4.554355144500732,
"learning_rate": 2.2789204606680524e-05,
"loss": 0.681,
"step": 547
},
{
"epoch": 18.896551724137932,
"grad_norm": 6.15927791595459,
"learning_rate": 2.2776800512884218e-05,
"loss": 0.7005,
"step": 548
},
{
"epoch": 18.93103448275862,
"grad_norm": 4.163325786590576,
"learning_rate": 2.2764365114171303e-05,
"loss": 0.6504,
"step": 549
},
{
"epoch": 18.96551724137931,
"grad_norm": 5.173964500427246,
"learning_rate": 2.2751898448422155e-05,
"loss": 0.6507,
"step": 550
},
{
"epoch": 19.0,
"grad_norm": 3.4791982173919678,
"learning_rate": 2.27394005536124e-05,
"loss": 0.6203,
"step": 551
},
{
"epoch": 19.03448275862069,
"grad_norm": 2.631436347961426,
"learning_rate": 2.2726871467812795e-05,
"loss": 0.6327,
"step": 552
},
{
"epoch": 19.06896551724138,
"grad_norm": 4.092041969299316,
"learning_rate": 2.2714311229189105e-05,
"loss": 0.649,
"step": 553
},
{
"epoch": 19.103448275862068,
"grad_norm": 3.6334736347198486,
"learning_rate": 2.2701719876002004e-05,
"loss": 0.6024,
"step": 554
},
{
"epoch": 19.137931034482758,
"grad_norm": 3.8364288806915283,
"learning_rate": 2.268909744660693e-05,
"loss": 0.5991,
"step": 555
},
{
"epoch": 19.17241379310345,
"grad_norm": 4.198896408081055,
"learning_rate": 2.267644397945399e-05,
"loss": 0.6068,
"step": 556
},
{
"epoch": 19.20689655172414,
"grad_norm": 3.1704394817352295,
"learning_rate": 2.2663759513087846e-05,
"loss": 0.6155,
"step": 557
},
{
"epoch": 19.24137931034483,
"grad_norm": 5.422239303588867,
"learning_rate": 2.2651044086147578e-05,
"loss": 0.6836,
"step": 558
},
{
"epoch": 19.275862068965516,
"grad_norm": 3.0339040756225586,
"learning_rate": 2.2638297737366583e-05,
"loss": 0.5523,
"step": 559
},
{
"epoch": 19.310344827586206,
"grad_norm": 2.4835903644561768,
"learning_rate": 2.2625520505572452e-05,
"loss": 0.6802,
"step": 560
},
{
"epoch": 19.344827586206897,
"grad_norm": 4.562428951263428,
"learning_rate": 2.2612712429686845e-05,
"loss": 0.7499,
"step": 561
},
{
"epoch": 19.379310344827587,
"grad_norm": 4.1088409423828125,
"learning_rate": 2.259987354872538e-05,
"loss": 0.6736,
"step": 562
},
{
"epoch": 19.413793103448278,
"grad_norm": 3.155371904373169,
"learning_rate": 2.2587003901797528e-05,
"loss": 0.6435,
"step": 563
},
{
"epoch": 19.448275862068964,
"grad_norm": 4.05208158493042,
"learning_rate": 2.2574103528106448e-05,
"loss": 0.6947,
"step": 564
},
{
"epoch": 19.482758620689655,
"grad_norm": 2.9079976081848145,
"learning_rate": 2.2561172466948925e-05,
"loss": 0.6085,
"step": 565
},
{
"epoch": 19.517241379310345,
"grad_norm": 2.7722418308258057,
"learning_rate": 2.2548210757715216e-05,
"loss": 0.6012,
"step": 566
},
{
"epoch": 19.551724137931036,
"grad_norm": 3.1863250732421875,
"learning_rate": 2.2535218439888933e-05,
"loss": 0.6131,
"step": 567
},
{
"epoch": 19.586206896551722,
"grad_norm": 4.646592140197754,
"learning_rate": 2.2522195553046926e-05,
"loss": 0.6314,
"step": 568
},
{
"epoch": 19.620689655172413,
"grad_norm": 5.8424577713012695,
"learning_rate": 2.2509142136859168e-05,
"loss": 0.6064,
"step": 569
},
{
"epoch": 19.655172413793103,
"grad_norm": 3.509737968444824,
"learning_rate": 2.2496058231088635e-05,
"loss": 0.5589,
"step": 570
},
{
"epoch": 19.689655172413794,
"grad_norm": 6.115649700164795,
"learning_rate": 2.248294387559116e-05,
"loss": 0.646,
"step": 571
},
{
"epoch": 19.724137931034484,
"grad_norm": 2.8004534244537354,
"learning_rate": 2.246979911031536e-05,
"loss": 0.5837,
"step": 572
},
{
"epoch": 19.75862068965517,
"grad_norm": 2.3154537677764893,
"learning_rate": 2.2456623975302454e-05,
"loss": 0.5972,
"step": 573
},
{
"epoch": 19.79310344827586,
"grad_norm": 2.849769115447998,
"learning_rate": 2.2443418510686198e-05,
"loss": 0.6159,
"step": 574
},
{
"epoch": 19.82758620689655,
"grad_norm": 2.4394686222076416,
"learning_rate": 2.2430182756692728e-05,
"loss": 0.5879,
"step": 575
},
{
"epoch": 19.862068965517242,
"grad_norm": 1.768813133239746,
"learning_rate": 2.2416916753640442e-05,
"loss": 0.5802,
"step": 576
},
{
"epoch": 19.896551724137932,
"grad_norm": 2.7482104301452637,
"learning_rate": 2.2403620541939884e-05,
"loss": 0.5606,
"step": 577
},
{
"epoch": 19.93103448275862,
"grad_norm": 1.9933055639266968,
"learning_rate": 2.2390294162093627e-05,
"loss": 0.5557,
"step": 578
},
{
"epoch": 19.96551724137931,
"grad_norm": 3.114577293395996,
"learning_rate": 2.2376937654696133e-05,
"loss": 0.5388,
"step": 579
},
{
"epoch": 20.0,
"grad_norm": 2.786837339401245,
"learning_rate": 2.2363551060433636e-05,
"loss": 0.5458,
"step": 580
},
{
"epoch": 20.03448275862069,
"grad_norm": 3.6182146072387695,
"learning_rate": 2.2350134420084023e-05,
"loss": 0.6075,
"step": 581
},
{
"epoch": 20.06896551724138,
"grad_norm": 4.18286657333374,
"learning_rate": 2.233668777451672e-05,
"loss": 0.6064,
"step": 582
},
{
"epoch": 20.103448275862068,
"grad_norm": 2.0372018814086914,
"learning_rate": 2.2323211164692526e-05,
"loss": 0.575,
"step": 583
},
{
"epoch": 20.137931034482758,
"grad_norm": 2.5157814025878906,
"learning_rate": 2.2309704631663536e-05,
"loss": 0.5397,
"step": 584
},
{
"epoch": 20.17241379310345,
"grad_norm": 2.8599259853363037,
"learning_rate": 2.2296168216573e-05,
"loss": 0.6183,
"step": 585
},
{
"epoch": 20.20689655172414,
"grad_norm": 2.6183788776397705,
"learning_rate": 2.2282601960655175e-05,
"loss": 0.5912,
"step": 586
},
{
"epoch": 20.24137931034483,
"grad_norm": 2.3617968559265137,
"learning_rate": 2.2269005905235234e-05,
"loss": 0.6095,
"step": 587
},
{
"epoch": 20.275862068965516,
"grad_norm": 2.041109323501587,
"learning_rate": 2.2255380091729124e-05,
"loss": 0.5709,
"step": 588
},
{
"epoch": 20.310344827586206,
"grad_norm": 2.473818778991699,
"learning_rate": 2.224172456164343e-05,
"loss": 0.622,
"step": 589
},
{
"epoch": 20.344827586206897,
"grad_norm": 4.0362324714660645,
"learning_rate": 2.2228039356575265e-05,
"loss": 0.6281,
"step": 590
},
{
"epoch": 20.379310344827587,
"grad_norm": 4.358520984649658,
"learning_rate": 2.221432451821214e-05,
"loss": 0.6563,
"step": 591
},
{
"epoch": 20.413793103448278,
"grad_norm": 2.3213906288146973,
"learning_rate": 2.2200580088331825e-05,
"loss": 0.577,
"step": 592
},
{
"epoch": 20.448275862068964,
"grad_norm": 4.955238342285156,
"learning_rate": 2.2186806108802248e-05,
"loss": 0.6418,
"step": 593
},
{
"epoch": 20.482758620689655,
"grad_norm": 5.53913688659668,
"learning_rate": 2.217300262158133e-05,
"loss": 0.5817,
"step": 594
},
{
"epoch": 20.517241379310345,
"grad_norm": 3.345353603363037,
"learning_rate": 2.215916966871689e-05,
"loss": 0.5873,
"step": 595
},
{
"epoch": 20.551724137931036,
"grad_norm": 3.257923126220703,
"learning_rate": 2.2145307292346502e-05,
"loss": 0.5968,
"step": 596
},
{
"epoch": 20.586206896551722,
"grad_norm": 3.00583553314209,
"learning_rate": 2.213141553469737e-05,
"loss": 0.5637,
"step": 597
},
{
"epoch": 20.620689655172413,
"grad_norm": 3.975980043411255,
"learning_rate": 2.211749443808619e-05,
"loss": 0.6064,
"step": 598
},
{
"epoch": 20.655172413793103,
"grad_norm": 4.438958644866943,
"learning_rate": 2.2103544044919045e-05,
"loss": 0.6115,
"step": 599
},
{
"epoch": 20.689655172413794,
"grad_norm": 5.422476768493652,
"learning_rate": 2.208956439769125e-05,
"loss": 0.6534,
"step": 600
},
{
"epoch": 20.724137931034484,
"grad_norm": 2.1388237476348877,
"learning_rate": 2.2075555538987227e-05,
"loss": 0.5648,
"step": 601
},
{
"epoch": 20.75862068965517,
"grad_norm": 3.457695245742798,
"learning_rate": 2.20615175114804e-05,
"loss": 0.53,
"step": 602
},
{
"epoch": 20.79310344827586,
"grad_norm": 4.380722999572754,
"learning_rate": 2.2047450357933032e-05,
"loss": 0.6238,
"step": 603
},
{
"epoch": 20.82758620689655,
"grad_norm": 3.4675967693328857,
"learning_rate": 2.2033354121196102e-05,
"loss": 0.594,
"step": 604
},
{
"epoch": 20.862068965517242,
"grad_norm": 6.335079669952393,
"learning_rate": 2.201922884420921e-05,
"loss": 0.5905,
"step": 605
},
{
"epoch": 20.896551724137932,
"grad_norm": 4.242478370666504,
"learning_rate": 2.200507457000039e-05,
"loss": 0.5376,
"step": 606
},
{
"epoch": 20.93103448275862,
"grad_norm": 2.4837803840637207,
"learning_rate": 2.1990891341686008e-05,
"loss": 0.5069,
"step": 607
},
{
"epoch": 20.96551724137931,
"grad_norm": 2.1897082328796387,
"learning_rate": 2.1976679202470654e-05,
"loss": 0.5472,
"step": 608
},
{
"epoch": 21.0,
"grad_norm": 2.4400973320007324,
"learning_rate": 2.1962438195646958e-05,
"loss": 0.547,
"step": 609
},
{
"epoch": 21.03448275862069,
"grad_norm": 4.650160789489746,
"learning_rate": 2.1948168364595497e-05,
"loss": 0.6903,
"step": 610
},
{
"epoch": 21.06896551724138,
"grad_norm": 3.0969042778015137,
"learning_rate": 2.1933869752784654e-05,
"loss": 0.5143,
"step": 611
},
{
"epoch": 21.103448275862068,
"grad_norm": 4.73034143447876,
"learning_rate": 2.1919542403770476e-05,
"loss": 0.6413,
"step": 612
},
{
"epoch": 21.137931034482758,
"grad_norm": 2.4023869037628174,
"learning_rate": 2.1905186361196556e-05,
"loss": 0.5607,
"step": 613
},
{
"epoch": 21.17241379310345,
"grad_norm": 3.85475754737854,
"learning_rate": 2.189080166879389e-05,
"loss": 0.6169,
"step": 614
},
{
"epoch": 21.20689655172414,
"grad_norm": 2.8691437244415283,
"learning_rate": 2.1876388370380745e-05,
"loss": 0.4884,
"step": 615
},
{
"epoch": 21.24137931034483,
"grad_norm": 2.3422751426696777,
"learning_rate": 2.186194650986253e-05,
"loss": 0.5269,
"step": 616
},
{
"epoch": 21.275862068965516,
"grad_norm": 3.4676578044891357,
"learning_rate": 2.184747613123165e-05,
"loss": 0.5517,
"step": 617
},
{
"epoch": 21.310344827586206,
"grad_norm": 3.613835573196411,
"learning_rate": 2.1832977278567394e-05,
"loss": 0.5919,
"step": 618
},
{
"epoch": 21.344827586206897,
"grad_norm": 3.2445290088653564,
"learning_rate": 2.181844999603578e-05,
"loss": 0.5624,
"step": 619
},
{
"epoch": 21.379310344827587,
"grad_norm": 3.5118792057037354,
"learning_rate": 2.1803894327889425e-05,
"loss": 0.587,
"step": 620
},
{
"epoch": 21.413793103448278,
"grad_norm": 5.919454574584961,
"learning_rate": 2.178931031846743e-05,
"loss": 0.6432,
"step": 621
},
{
"epoch": 21.448275862068964,
"grad_norm": 3.322352409362793,
"learning_rate": 2.1774698012195206e-05,
"loss": 0.6003,
"step": 622
},
{
"epoch": 21.482758620689655,
"grad_norm": 6.281351566314697,
"learning_rate": 2.1760057453584376e-05,
"loss": 0.5805,
"step": 623
},
{
"epoch": 21.517241379310345,
"grad_norm": 5.132048606872559,
"learning_rate": 2.1745388687232624e-05,
"loss": 0.5572,
"step": 624
},
{
"epoch": 21.551724137931036,
"grad_norm": 2.332777976989746,
"learning_rate": 2.1730691757823553e-05,
"loss": 0.5633,
"step": 625
},
{
"epoch": 21.586206896551722,
"grad_norm": 3.328850030899048,
"learning_rate": 2.171596671012655e-05,
"loss": 0.631,
"step": 626
},
{
"epoch": 21.620689655172413,
"grad_norm": 2.524472713470459,
"learning_rate": 2.1701213588996683e-05,
"loss": 0.5458,
"step": 627
},
{
"epoch": 21.655172413793103,
"grad_norm": 2.7530570030212402,
"learning_rate": 2.16864324393745e-05,
"loss": 0.5397,
"step": 628
},
{
"epoch": 21.689655172413794,
"grad_norm": 4.604249954223633,
"learning_rate": 2.1671623306285956e-05,
"loss": 0.5842,
"step": 629
},
{
"epoch": 21.724137931034484,
"grad_norm": 2.306530714035034,
"learning_rate": 2.1656786234842237e-05,
"loss": 0.5489,
"step": 630
},
{
"epoch": 21.75862068965517,
"grad_norm": 3.085984230041504,
"learning_rate": 2.1641921270239632e-05,
"loss": 0.715,
"step": 631
},
{
"epoch": 21.79310344827586,
"grad_norm": 3.6664793491363525,
"learning_rate": 2.1627028457759408e-05,
"loss": 0.5976,
"step": 632
},
{
"epoch": 21.82758620689655,
"grad_norm": 3.8308627605438232,
"learning_rate": 2.1612107842767647e-05,
"loss": 0.5413,
"step": 633
},
{
"epoch": 21.862068965517242,
"grad_norm": 3.766814947128296,
"learning_rate": 2.1597159470715133e-05,
"loss": 0.5849,
"step": 634
},
{
"epoch": 21.896551724137932,
"grad_norm": 4.085217475891113,
"learning_rate": 2.15821833871372e-05,
"loss": 0.6535,
"step": 635
},
{
"epoch": 21.93103448275862,
"grad_norm": 3.2848877906799316,
"learning_rate": 2.1567179637653594e-05,
"loss": 0.6572,
"step": 636
},
{
"epoch": 21.96551724137931,
"grad_norm": 4.021419048309326,
"learning_rate": 2.1552148267968347e-05,
"loss": 0.5827,
"step": 637
},
{
"epoch": 22.0,
"grad_norm": 3.385075569152832,
"learning_rate": 2.1537089323869604e-05,
"loss": 0.575,
"step": 638
},
{
"epoch": 22.03448275862069,
"grad_norm": 3.6210548877716064,
"learning_rate": 2.152200285122953e-05,
"loss": 0.5036,
"step": 639
},
{
"epoch": 22.06896551724138,
"grad_norm": 4.066530704498291,
"learning_rate": 2.1506888896004133e-05,
"loss": 0.5167,
"step": 640
},
{
"epoch": 22.103448275862068,
"grad_norm": 2.326347589492798,
"learning_rate": 2.1491747504233138e-05,
"loss": 0.5364,
"step": 641
},
{
"epoch": 22.137931034482758,
"grad_norm": 3.226869583129883,
"learning_rate": 2.147657872203986e-05,
"loss": 0.5291,
"step": 642
},
{
"epoch": 22.17241379310345,
"grad_norm": 2.8205795288085938,
"learning_rate": 2.1461382595631036e-05,
"loss": 0.5034,
"step": 643
},
{
"epoch": 22.20689655172414,
"grad_norm": 5.307386875152588,
"learning_rate": 2.14461591712967e-05,
"loss": 0.6833,
"step": 644
},
{
"epoch": 22.24137931034483,
"grad_norm": 3.1979541778564453,
"learning_rate": 2.1430908495410042e-05,
"loss": 0.5307,
"step": 645
},
{
"epoch": 22.275862068965516,
"grad_norm": 3.491663694381714,
"learning_rate": 2.1415630614427272e-05,
"loss": 0.5779,
"step": 646
},
{
"epoch": 22.310344827586206,
"grad_norm": 2.6055986881256104,
"learning_rate": 2.140032557488746e-05,
"loss": 0.5429,
"step": 647
},
{
"epoch": 22.344827586206897,
"grad_norm": 1.5949605703353882,
"learning_rate": 2.1384993423412407e-05,
"loss": 0.5109,
"step": 648
},
{
"epoch": 22.379310344827587,
"grad_norm": 3.324326515197754,
"learning_rate": 2.136963420670651e-05,
"loss": 0.5277,
"step": 649
},
{
"epoch": 22.413793103448278,
"grad_norm": 3.71690034866333,
"learning_rate": 2.135424797155661e-05,
"loss": 0.5901,
"step": 650
},
{
"epoch": 22.448275862068964,
"grad_norm": 3.789421319961548,
"learning_rate": 2.1338834764831845e-05,
"loss": 0.5424,
"step": 651
},
{
"epoch": 22.482758620689655,
"grad_norm": 3.0680930614471436,
"learning_rate": 2.1323394633483514e-05,
"loss": 0.6194,
"step": 652
},
{
"epoch": 22.517241379310345,
"grad_norm": 2.454967975616455,
"learning_rate": 2.1307927624544934e-05,
"loss": 0.5856,
"step": 653
},
{
"epoch": 22.551724137931036,
"grad_norm": 4.197128772735596,
"learning_rate": 2.1292433785131298e-05,
"loss": 0.6164,
"step": 654
},
{
"epoch": 22.586206896551722,
"grad_norm": 4.163417339324951,
"learning_rate": 2.1276913162439532e-05,
"loss": 0.5499,
"step": 655
},
{
"epoch": 22.620689655172413,
"grad_norm": 3.207831621170044,
"learning_rate": 2.1261365803748138e-05,
"loss": 0.5379,
"step": 656
},
{
"epoch": 22.655172413793103,
"grad_norm": 2.1474342346191406,
"learning_rate": 2.124579175641707e-05,
"loss": 0.5198,
"step": 657
},
{
"epoch": 22.689655172413794,
"grad_norm": 2.0802290439605713,
"learning_rate": 2.1230191067887574e-05,
"loss": 0.5988,
"step": 658
},
{
"epoch": 22.724137931034484,
"grad_norm": 2.383214235305786,
"learning_rate": 2.121456378568206e-05,
"loss": 0.5644,
"step": 659
},
{
"epoch": 22.75862068965517,
"grad_norm": 3.3427810668945312,
"learning_rate": 2.1198909957403928e-05,
"loss": 0.5382,
"step": 660
},
{
"epoch": 22.79310344827586,
"grad_norm": 2.118605136871338,
"learning_rate": 2.1183229630737467e-05,
"loss": 0.5364,
"step": 661
},
{
"epoch": 22.82758620689655,
"grad_norm": 2.1082427501678467,
"learning_rate": 2.1167522853447664e-05,
"loss": 0.5631,
"step": 662
},
{
"epoch": 22.862068965517242,
"grad_norm": 2.0310721397399902,
"learning_rate": 2.1151789673380086e-05,
"loss": 0.583,
"step": 663
},
{
"epoch": 22.896551724137932,
"grad_norm": 3.352940082550049,
"learning_rate": 2.113603013846073e-05,
"loss": 0.5694,
"step": 664
},
{
"epoch": 22.93103448275862,
"grad_norm": 2.5591657161712646,
"learning_rate": 2.1120244296695874e-05,
"loss": 0.5605,
"step": 665
},
{
"epoch": 22.96551724137931,
"grad_norm": 3.0828518867492676,
"learning_rate": 2.1104432196171924e-05,
"loss": 0.5105,
"step": 666
},
{
"epoch": 23.0,
"grad_norm": 4.050001621246338,
"learning_rate": 2.1088593885055288e-05,
"loss": 0.5483,
"step": 667
},
{
"epoch": 23.03448275862069,
"grad_norm": 3.4078168869018555,
"learning_rate": 2.1072729411592206e-05,
"loss": 0.5672,
"step": 668
},
{
"epoch": 23.06896551724138,
"grad_norm": 4.16298246383667,
"learning_rate": 2.105683882410861e-05,
"loss": 0.6305,
"step": 669
},
{
"epoch": 23.103448275862068,
"grad_norm": 4.402109146118164,
"learning_rate": 2.1040922171009993e-05,
"loss": 0.6278,
"step": 670
},
{
"epoch": 23.137931034482758,
"grad_norm": 2.614933729171753,
"learning_rate": 2.1024979500781232e-05,
"loss": 0.5573,
"step": 671
},
{
"epoch": 23.17241379310345,
"grad_norm": 5.566406726837158,
"learning_rate": 2.1009010861986476e-05,
"loss": 0.5987,
"step": 672
},
{
"epoch": 23.20689655172414,
"grad_norm": 3.687830686569214,
"learning_rate": 2.099301630326896e-05,
"loss": 0.5332,
"step": 673
},
{
"epoch": 23.24137931034483,
"grad_norm": 1.985662579536438,
"learning_rate": 2.0976995873350887e-05,
"loss": 0.5032,
"step": 674
},
{
"epoch": 23.275862068965516,
"grad_norm": 3.1905019283294678,
"learning_rate": 2.096094962103326e-05,
"loss": 0.5556,
"step": 675
},
{
"epoch": 23.310344827586206,
"grad_norm": 2.235889434814453,
"learning_rate": 2.0944877595195755e-05,
"loss": 0.5636,
"step": 676
},
{
"epoch": 23.344827586206897,
"grad_norm": 4.098913669586182,
"learning_rate": 2.092877984479654e-05,
"loss": 0.5974,
"step": 677
},
{
"epoch": 23.379310344827587,
"grad_norm": 5.365559101104736,
"learning_rate": 2.091265641887217e-05,
"loss": 0.5036,
"step": 678
},
{
"epoch": 23.413793103448278,
"grad_norm": 3.486875534057617,
"learning_rate": 2.089650736653738e-05,
"loss": 0.5835,
"step": 679
},
{
"epoch": 23.448275862068964,
"grad_norm": 6.258121967315674,
"learning_rate": 2.088033273698499e-05,
"loss": 0.6289,
"step": 680
},
{
"epoch": 23.482758620689655,
"grad_norm": 3.1548774242401123,
"learning_rate": 2.086413257948573e-05,
"loss": 0.4888,
"step": 681
},
{
"epoch": 23.517241379310345,
"grad_norm": 2.209894895553589,
"learning_rate": 2.0847906943388085e-05,
"loss": 0.4746,
"step": 682
},
{
"epoch": 23.551724137931036,
"grad_norm": 2.9352316856384277,
"learning_rate": 2.0831655878118155e-05,
"loss": 0.525,
"step": 683
},
{
"epoch": 23.586206896551722,
"grad_norm": 2.3592867851257324,
"learning_rate": 2.081537943317951e-05,
"loss": 0.4875,
"step": 684
},
{
"epoch": 23.620689655172413,
"grad_norm": 3.657504081726074,
"learning_rate": 2.0799077658153022e-05,
"loss": 0.554,
"step": 685
},
{
"epoch": 23.655172413793103,
"grad_norm": 6.461456298828125,
"learning_rate": 2.0782750602696722e-05,
"loss": 0.5205,
"step": 686
},
{
"epoch": 23.689655172413794,
"grad_norm": 4.881630897521973,
"learning_rate": 2.0766398316545648e-05,
"loss": 0.5587,
"step": 687
},
{
"epoch": 23.724137931034484,
"grad_norm": 2.7277519702911377,
"learning_rate": 2.0750020849511712e-05,
"loss": 0.546,
"step": 688
},
{
"epoch": 23.75862068965517,
"grad_norm": 4.6229071617126465,
"learning_rate": 2.0733618251483506e-05,
"loss": 0.5404,
"step": 689
},
{
"epoch": 23.79310344827586,
"grad_norm": 5.046594619750977,
"learning_rate": 2.07171905724262e-05,
"loss": 0.59,
"step": 690
},
{
"epoch": 23.82758620689655,
"grad_norm": 4.2928853034973145,
"learning_rate": 2.070073786238134e-05,
"loss": 0.5288,
"step": 691
},
{
"epoch": 23.862068965517242,
"grad_norm": 6.604857444763184,
"learning_rate": 2.0684260171466745e-05,
"loss": 0.6729,
"step": 692
},
{
"epoch": 23.896551724137932,
"grad_norm": 7.314076900482178,
"learning_rate": 2.066775754987632e-05,
"loss": 0.6542,
"step": 693
},
{
"epoch": 23.93103448275862,
"grad_norm": 3.7828738689422607,
"learning_rate": 2.0651230047879905e-05,
"loss": 0.6574,
"step": 694
},
{
"epoch": 23.96551724137931,
"grad_norm": 3.6089084148406982,
"learning_rate": 2.0634677715823137e-05,
"loss": 0.6718,
"step": 695
},
{
"epoch": 24.0,
"grad_norm": 3.206212043762207,
"learning_rate": 2.0618100604127295e-05,
"loss": 0.6149,
"step": 696
},
{
"epoch": 24.03448275862069,
"grad_norm": 2.9614017009735107,
"learning_rate": 2.0601498763289138e-05,
"loss": 0.6372,
"step": 697
},
{
"epoch": 24.06896551724138,
"grad_norm": 3.116159439086914,
"learning_rate": 2.058487224388075e-05,
"loss": 0.6268,
"step": 698
},
{
"epoch": 24.103448275862068,
"grad_norm": 3.259319543838501,
"learning_rate": 2.0568221096549384e-05,
"loss": 0.5631,
"step": 699
},
{
"epoch": 24.137931034482758,
"grad_norm": 2.3190183639526367,
"learning_rate": 2.0551545372017332e-05,
"loss": 0.5782,
"step": 700
}
],
"logging_steps": 1,
"max_steps": 2000,
"num_input_tokens_seen": 0,
"num_train_epochs": 69,
"save_steps": 200,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": false,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 0.0,
"train_batch_size": 64,
"trial_name": null,
"trial_params": null
}