Dr-SQL-Miner-Architect-Phase1-27B / trainer_state.json
dan-text2sql's picture
Upload folder using huggingface_hub
6854348 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.1765536723163842,
"eval_steps": 500,
"global_step": 500,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.00035310734463276836,
"grad_norm": 0.45211324095726013,
"learning_rate": 5e-05,
"loss": 3.273,
"step": 1
},
{
"epoch": 0.0007062146892655367,
"grad_norm": 0.30982279777526855,
"learning_rate": 4.99e-05,
"loss": 2.5125,
"step": 2
},
{
"epoch": 0.001059322033898305,
"grad_norm": 0.39342790842056274,
"learning_rate": 4.9800000000000004e-05,
"loss": 2.8347,
"step": 3
},
{
"epoch": 0.0014124293785310734,
"grad_norm": 0.3912785053253174,
"learning_rate": 4.97e-05,
"loss": 2.6189,
"step": 4
},
{
"epoch": 0.0017655367231638418,
"grad_norm": 0.43560200929641724,
"learning_rate": 4.96e-05,
"loss": 2.8784,
"step": 5
},
{
"epoch": 0.00211864406779661,
"grad_norm": 0.5163903832435608,
"learning_rate": 4.9500000000000004e-05,
"loss": 3.0496,
"step": 6
},
{
"epoch": 0.0024717514124293787,
"grad_norm": 0.3364083468914032,
"learning_rate": 4.94e-05,
"loss": 2.0403,
"step": 7
},
{
"epoch": 0.002824858757062147,
"grad_norm": 0.4723369777202606,
"learning_rate": 4.93e-05,
"loss": 2.5065,
"step": 8
},
{
"epoch": 0.0031779661016949155,
"grad_norm": 0.5686460137367249,
"learning_rate": 4.92e-05,
"loss": 2.8529,
"step": 9
},
{
"epoch": 0.0035310734463276836,
"grad_norm": 0.7666249871253967,
"learning_rate": 4.91e-05,
"loss": 3.2163,
"step": 10
},
{
"epoch": 0.003884180790960452,
"grad_norm": 0.5175825953483582,
"learning_rate": 4.9e-05,
"loss": 2.3817,
"step": 11
},
{
"epoch": 0.00423728813559322,
"grad_norm": 0.7693667411804199,
"learning_rate": 4.89e-05,
"loss": 3.1173,
"step": 12
},
{
"epoch": 0.004590395480225989,
"grad_norm": 0.6010498404502869,
"learning_rate": 4.88e-05,
"loss": 2.5264,
"step": 13
},
{
"epoch": 0.0049435028248587575,
"grad_norm": 0.8724003434181213,
"learning_rate": 4.87e-05,
"loss": 3.1545,
"step": 14
},
{
"epoch": 0.005296610169491525,
"grad_norm": 0.6222332119941711,
"learning_rate": 4.86e-05,
"loss": 2.4859,
"step": 15
},
{
"epoch": 0.005649717514124294,
"grad_norm": 0.5855898261070251,
"learning_rate": 4.85e-05,
"loss": 2.2748,
"step": 16
},
{
"epoch": 0.006002824858757062,
"grad_norm": 0.8227508068084717,
"learning_rate": 4.8400000000000004e-05,
"loss": 2.777,
"step": 17
},
{
"epoch": 0.006355932203389831,
"grad_norm": 0.747832179069519,
"learning_rate": 4.83e-05,
"loss": 2.6429,
"step": 18
},
{
"epoch": 0.006709039548022599,
"grad_norm": 0.8245562314987183,
"learning_rate": 4.82e-05,
"loss": 2.8865,
"step": 19
},
{
"epoch": 0.007062146892655367,
"grad_norm": 0.8959988951683044,
"learning_rate": 4.8100000000000004e-05,
"loss": 2.8727,
"step": 20
},
{
"epoch": 0.007415254237288136,
"grad_norm": 0.7936757802963257,
"learning_rate": 4.8e-05,
"loss": 2.5376,
"step": 21
},
{
"epoch": 0.007768361581920904,
"grad_norm": 0.8037199378013611,
"learning_rate": 4.79e-05,
"loss": 2.6522,
"step": 22
},
{
"epoch": 0.008121468926553672,
"grad_norm": 0.7104085087776184,
"learning_rate": 4.78e-05,
"loss": 2.3427,
"step": 23
},
{
"epoch": 0.00847457627118644,
"grad_norm": 0.8215932846069336,
"learning_rate": 4.77e-05,
"loss": 2.7918,
"step": 24
},
{
"epoch": 0.00882768361581921,
"grad_norm": 1.0282623767852783,
"learning_rate": 4.76e-05,
"loss": 2.9389,
"step": 25
},
{
"epoch": 0.009180790960451978,
"grad_norm": 0.7932501435279846,
"learning_rate": 4.75e-05,
"loss": 2.7691,
"step": 26
},
{
"epoch": 0.009533898305084746,
"grad_norm": 0.6620106101036072,
"learning_rate": 4.74e-05,
"loss": 2.1335,
"step": 27
},
{
"epoch": 0.009887005649717515,
"grad_norm": 0.8058229684829712,
"learning_rate": 4.73e-05,
"loss": 2.4778,
"step": 28
},
{
"epoch": 0.010240112994350282,
"grad_norm": 0.7049869894981384,
"learning_rate": 4.72e-05,
"loss": 2.3379,
"step": 29
},
{
"epoch": 0.01059322033898305,
"grad_norm": 0.6863004565238953,
"learning_rate": 4.71e-05,
"loss": 2.0131,
"step": 30
},
{
"epoch": 0.010946327683615819,
"grad_norm": 0.9368977546691895,
"learning_rate": 4.7e-05,
"loss": 2.4664,
"step": 31
},
{
"epoch": 0.011299435028248588,
"grad_norm": 0.7291913628578186,
"learning_rate": 4.69e-05,
"loss": 2.2541,
"step": 32
},
{
"epoch": 0.011652542372881356,
"grad_norm": 0.6416012048721313,
"learning_rate": 4.6800000000000006e-05,
"loss": 1.8727,
"step": 33
},
{
"epoch": 0.012005649717514125,
"grad_norm": 0.6020755767822266,
"learning_rate": 4.6700000000000003e-05,
"loss": 1.8596,
"step": 34
},
{
"epoch": 0.012358757062146893,
"grad_norm": 0.635847270488739,
"learning_rate": 4.660000000000001e-05,
"loss": 1.9688,
"step": 35
},
{
"epoch": 0.012711864406779662,
"grad_norm": 0.6946876645088196,
"learning_rate": 4.6500000000000005e-05,
"loss": 2.6249,
"step": 36
},
{
"epoch": 0.013064971751412429,
"grad_norm": 0.7043133974075317,
"learning_rate": 4.64e-05,
"loss": 2.5754,
"step": 37
},
{
"epoch": 0.013418079096045197,
"grad_norm": 0.5472213625907898,
"learning_rate": 4.630000000000001e-05,
"loss": 1.91,
"step": 38
},
{
"epoch": 0.013771186440677966,
"grad_norm": 0.5201235413551331,
"learning_rate": 4.6200000000000005e-05,
"loss": 2.0296,
"step": 39
},
{
"epoch": 0.014124293785310734,
"grad_norm": 0.5103974342346191,
"learning_rate": 4.61e-05,
"loss": 1.7567,
"step": 40
},
{
"epoch": 0.014477401129943503,
"grad_norm": 0.42137110233306885,
"learning_rate": 4.600000000000001e-05,
"loss": 1.7719,
"step": 41
},
{
"epoch": 0.014830508474576272,
"grad_norm": 0.6730819940567017,
"learning_rate": 4.5900000000000004e-05,
"loss": 2.4245,
"step": 42
},
{
"epoch": 0.01518361581920904,
"grad_norm": 0.6370269060134888,
"learning_rate": 4.58e-05,
"loss": 2.501,
"step": 43
},
{
"epoch": 0.015536723163841809,
"grad_norm": 0.5084403157234192,
"learning_rate": 4.5700000000000006e-05,
"loss": 2.0893,
"step": 44
},
{
"epoch": 0.015889830508474576,
"grad_norm": 0.6537813544273376,
"learning_rate": 4.5600000000000004e-05,
"loss": 2.3681,
"step": 45
},
{
"epoch": 0.016242937853107344,
"grad_norm": 0.5805037617683411,
"learning_rate": 4.55e-05,
"loss": 2.3028,
"step": 46
},
{
"epoch": 0.016596045197740113,
"grad_norm": 0.5214577913284302,
"learning_rate": 4.5400000000000006e-05,
"loss": 2.1601,
"step": 47
},
{
"epoch": 0.01694915254237288,
"grad_norm": 0.7660169005393982,
"learning_rate": 4.53e-05,
"loss": 2.4195,
"step": 48
},
{
"epoch": 0.01730225988700565,
"grad_norm": 0.537522554397583,
"learning_rate": 4.52e-05,
"loss": 2.3828,
"step": 49
},
{
"epoch": 0.01765536723163842,
"grad_norm": 0.5875993371009827,
"learning_rate": 4.5100000000000005e-05,
"loss": 2.4529,
"step": 50
},
{
"epoch": 0.018008474576271187,
"grad_norm": 0.480882465839386,
"learning_rate": 4.5e-05,
"loss": 1.7377,
"step": 51
},
{
"epoch": 0.018361581920903956,
"grad_norm": 0.4598454236984253,
"learning_rate": 4.49e-05,
"loss": 1.8277,
"step": 52
},
{
"epoch": 0.018714689265536724,
"grad_norm": 0.4913174510002136,
"learning_rate": 4.4800000000000005e-05,
"loss": 2.0882,
"step": 53
},
{
"epoch": 0.019067796610169493,
"grad_norm": 0.4395110011100769,
"learning_rate": 4.47e-05,
"loss": 1.8825,
"step": 54
},
{
"epoch": 0.01942090395480226,
"grad_norm": 0.5200910568237305,
"learning_rate": 4.46e-05,
"loss": 2.1883,
"step": 55
},
{
"epoch": 0.01977401129943503,
"grad_norm": 0.7119576930999756,
"learning_rate": 4.4500000000000004e-05,
"loss": 2.3147,
"step": 56
},
{
"epoch": 0.020127118644067795,
"grad_norm": 0.5865232348442078,
"learning_rate": 4.44e-05,
"loss": 2.4823,
"step": 57
},
{
"epoch": 0.020480225988700564,
"grad_norm": 0.5342106223106384,
"learning_rate": 4.43e-05,
"loss": 2.3112,
"step": 58
},
{
"epoch": 0.020833333333333332,
"grad_norm": 0.63221275806427,
"learning_rate": 4.4200000000000004e-05,
"loss": 2.4642,
"step": 59
},
{
"epoch": 0.0211864406779661,
"grad_norm": 0.5381503701210022,
"learning_rate": 4.41e-05,
"loss": 2.1702,
"step": 60
},
{
"epoch": 0.02153954802259887,
"grad_norm": 0.34933093190193176,
"learning_rate": 4.4000000000000006e-05,
"loss": 1.6217,
"step": 61
},
{
"epoch": 0.021892655367231638,
"grad_norm": 0.4486427307128906,
"learning_rate": 4.39e-05,
"loss": 1.9545,
"step": 62
},
{
"epoch": 0.022245762711864406,
"grad_norm": 0.39275842905044556,
"learning_rate": 4.38e-05,
"loss": 1.692,
"step": 63
},
{
"epoch": 0.022598870056497175,
"grad_norm": 0.5706101655960083,
"learning_rate": 4.3700000000000005e-05,
"loss": 2.3182,
"step": 64
},
{
"epoch": 0.022951977401129944,
"grad_norm": 0.43103381991386414,
"learning_rate": 4.36e-05,
"loss": 1.8674,
"step": 65
},
{
"epoch": 0.023305084745762712,
"grad_norm": 0.6326637268066406,
"learning_rate": 4.35e-05,
"loss": 2.3281,
"step": 66
},
{
"epoch": 0.02365819209039548,
"grad_norm": 0.5602039694786072,
"learning_rate": 4.3400000000000005e-05,
"loss": 2.24,
"step": 67
},
{
"epoch": 0.02401129943502825,
"grad_norm": 0.42276912927627563,
"learning_rate": 4.33e-05,
"loss": 1.9933,
"step": 68
},
{
"epoch": 0.024364406779661018,
"grad_norm": 0.5240011811256409,
"learning_rate": 4.32e-05,
"loss": 2.1384,
"step": 69
},
{
"epoch": 0.024717514124293787,
"grad_norm": 0.579456627368927,
"learning_rate": 4.3100000000000004e-05,
"loss": 2.2305,
"step": 70
},
{
"epoch": 0.025070621468926555,
"grad_norm": 0.5782936215400696,
"learning_rate": 4.3e-05,
"loss": 2.2021,
"step": 71
},
{
"epoch": 0.025423728813559324,
"grad_norm": 0.6299965977668762,
"learning_rate": 4.29e-05,
"loss": 2.3214,
"step": 72
},
{
"epoch": 0.02577683615819209,
"grad_norm": 0.634605884552002,
"learning_rate": 4.2800000000000004e-05,
"loss": 2.3198,
"step": 73
},
{
"epoch": 0.026129943502824857,
"grad_norm": 0.5833197832107544,
"learning_rate": 4.27e-05,
"loss": 2.1705,
"step": 74
},
{
"epoch": 0.026483050847457626,
"grad_norm": 0.5400241017341614,
"learning_rate": 4.26e-05,
"loss": 1.8534,
"step": 75
},
{
"epoch": 0.026836158192090395,
"grad_norm": 0.6428772807121277,
"learning_rate": 4.25e-05,
"loss": 2.3044,
"step": 76
},
{
"epoch": 0.027189265536723163,
"grad_norm": 0.611422061920166,
"learning_rate": 4.24e-05,
"loss": 2.1591,
"step": 77
},
{
"epoch": 0.02754237288135593,
"grad_norm": 0.602570652961731,
"learning_rate": 4.23e-05,
"loss": 2.1249,
"step": 78
},
{
"epoch": 0.0278954802259887,
"grad_norm": 0.48029282689094543,
"learning_rate": 4.22e-05,
"loss": 1.905,
"step": 79
},
{
"epoch": 0.02824858757062147,
"grad_norm": 0.6572628617286682,
"learning_rate": 4.21e-05,
"loss": 2.2358,
"step": 80
},
{
"epoch": 0.028601694915254237,
"grad_norm": 0.5220134258270264,
"learning_rate": 4.2e-05,
"loss": 1.9954,
"step": 81
},
{
"epoch": 0.028954802259887006,
"grad_norm": 0.4906415641307831,
"learning_rate": 4.19e-05,
"loss": 1.8642,
"step": 82
},
{
"epoch": 0.029307909604519775,
"grad_norm": 0.6776078343391418,
"learning_rate": 4.18e-05,
"loss": 2.2061,
"step": 83
},
{
"epoch": 0.029661016949152543,
"grad_norm": 0.496427059173584,
"learning_rate": 4.17e-05,
"loss": 1.9494,
"step": 84
},
{
"epoch": 0.03001412429378531,
"grad_norm": 0.4432898759841919,
"learning_rate": 4.16e-05,
"loss": 1.6809,
"step": 85
},
{
"epoch": 0.03036723163841808,
"grad_norm": 0.6649871468544006,
"learning_rate": 4.15e-05,
"loss": 2.0615,
"step": 86
},
{
"epoch": 0.03072033898305085,
"grad_norm": 0.6379185318946838,
"learning_rate": 4.14e-05,
"loss": 1.9796,
"step": 87
},
{
"epoch": 0.031073446327683617,
"grad_norm": 0.5551559925079346,
"learning_rate": 4.13e-05,
"loss": 1.874,
"step": 88
},
{
"epoch": 0.031426553672316386,
"grad_norm": 0.7269681096076965,
"learning_rate": 4.12e-05,
"loss": 2.1407,
"step": 89
},
{
"epoch": 0.03177966101694915,
"grad_norm": 0.6741663217544556,
"learning_rate": 4.11e-05,
"loss": 1.9975,
"step": 90
},
{
"epoch": 0.03213276836158192,
"grad_norm": 0.5726102590560913,
"learning_rate": 4.1e-05,
"loss": 1.8897,
"step": 91
},
{
"epoch": 0.03248587570621469,
"grad_norm": 0.7033076286315918,
"learning_rate": 4.09e-05,
"loss": 2.0044,
"step": 92
},
{
"epoch": 0.03283898305084746,
"grad_norm": 0.6314510107040405,
"learning_rate": 4.08e-05,
"loss": 2.1114,
"step": 93
},
{
"epoch": 0.033192090395480225,
"grad_norm": 0.49348896741867065,
"learning_rate": 4.07e-05,
"loss": 1.5142,
"step": 94
},
{
"epoch": 0.033545197740113,
"grad_norm": 0.5636611580848694,
"learning_rate": 4.0600000000000004e-05,
"loss": 1.7442,
"step": 95
},
{
"epoch": 0.03389830508474576,
"grad_norm": 0.5502078533172607,
"learning_rate": 4.05e-05,
"loss": 1.744,
"step": 96
},
{
"epoch": 0.03425141242937853,
"grad_norm": 0.7853062152862549,
"learning_rate": 4.0400000000000006e-05,
"loss": 2.0471,
"step": 97
},
{
"epoch": 0.0346045197740113,
"grad_norm": 0.7374521493911743,
"learning_rate": 4.0300000000000004e-05,
"loss": 1.9658,
"step": 98
},
{
"epoch": 0.034957627118644065,
"grad_norm": 0.5843427181243896,
"learning_rate": 4.02e-05,
"loss": 1.698,
"step": 99
},
{
"epoch": 0.03531073446327684,
"grad_norm": 0.7103281617164612,
"learning_rate": 4.0100000000000006e-05,
"loss": 1.8531,
"step": 100
},
{
"epoch": 0.0356638418079096,
"grad_norm": 0.547626256942749,
"learning_rate": 4e-05,
"loss": 1.6113,
"step": 101
},
{
"epoch": 0.036016949152542374,
"grad_norm": 0.7051151990890503,
"learning_rate": 3.99e-05,
"loss": 1.8774,
"step": 102
},
{
"epoch": 0.03637005649717514,
"grad_norm": 0.5148686766624451,
"learning_rate": 3.9800000000000005e-05,
"loss": 1.6082,
"step": 103
},
{
"epoch": 0.03672316384180791,
"grad_norm": 0.5013210773468018,
"learning_rate": 3.97e-05,
"loss": 1.6964,
"step": 104
},
{
"epoch": 0.037076271186440676,
"grad_norm": 0.7677639126777649,
"learning_rate": 3.960000000000001e-05,
"loss": 1.8114,
"step": 105
},
{
"epoch": 0.03742937853107345,
"grad_norm": 0.6067560911178589,
"learning_rate": 3.9500000000000005e-05,
"loss": 1.6947,
"step": 106
},
{
"epoch": 0.037782485875706213,
"grad_norm": 0.5241538286209106,
"learning_rate": 3.94e-05,
"loss": 1.4953,
"step": 107
},
{
"epoch": 0.038135593220338986,
"grad_norm": 0.5037181377410889,
"learning_rate": 3.9300000000000007e-05,
"loss": 1.4166,
"step": 108
},
{
"epoch": 0.03848870056497175,
"grad_norm": 0.7840654253959656,
"learning_rate": 3.9200000000000004e-05,
"loss": 1.8501,
"step": 109
},
{
"epoch": 0.03884180790960452,
"grad_norm": 0.6080065369606018,
"learning_rate": 3.91e-05,
"loss": 1.6206,
"step": 110
},
{
"epoch": 0.03919491525423729,
"grad_norm": 0.616134762763977,
"learning_rate": 3.9000000000000006e-05,
"loss": 1.7034,
"step": 111
},
{
"epoch": 0.03954802259887006,
"grad_norm": 0.7034599781036377,
"learning_rate": 3.8900000000000004e-05,
"loss": 1.9318,
"step": 112
},
{
"epoch": 0.039901129943502825,
"grad_norm": 0.8047641515731812,
"learning_rate": 3.88e-05,
"loss": 1.7351,
"step": 113
},
{
"epoch": 0.04025423728813559,
"grad_norm": 0.8917104005813599,
"learning_rate": 3.8700000000000006e-05,
"loss": 1.8398,
"step": 114
},
{
"epoch": 0.04060734463276836,
"grad_norm": 0.8026954531669617,
"learning_rate": 3.86e-05,
"loss": 1.7842,
"step": 115
},
{
"epoch": 0.04096045197740113,
"grad_norm": 0.5217316150665283,
"learning_rate": 3.85e-05,
"loss": 1.5885,
"step": 116
},
{
"epoch": 0.0413135593220339,
"grad_norm": 0.6288496851921082,
"learning_rate": 3.8400000000000005e-05,
"loss": 1.5425,
"step": 117
},
{
"epoch": 0.041666666666666664,
"grad_norm": 0.77554851770401,
"learning_rate": 3.83e-05,
"loss": 1.6508,
"step": 118
},
{
"epoch": 0.042019774011299436,
"grad_norm": 0.5694034695625305,
"learning_rate": 3.82e-05,
"loss": 1.2894,
"step": 119
},
{
"epoch": 0.0423728813559322,
"grad_norm": 0.6931347846984863,
"learning_rate": 3.8100000000000005e-05,
"loss": 1.7874,
"step": 120
},
{
"epoch": 0.042725988700564974,
"grad_norm": 0.7180691957473755,
"learning_rate": 3.8e-05,
"loss": 1.5679,
"step": 121
},
{
"epoch": 0.04307909604519774,
"grad_norm": 0.5494984984397888,
"learning_rate": 3.79e-05,
"loss": 1.2885,
"step": 122
},
{
"epoch": 0.04343220338983051,
"grad_norm": 0.8779198527336121,
"learning_rate": 3.7800000000000004e-05,
"loss": 1.6815,
"step": 123
},
{
"epoch": 0.043785310734463276,
"grad_norm": 0.7112749218940735,
"learning_rate": 3.77e-05,
"loss": 1.4889,
"step": 124
},
{
"epoch": 0.04413841807909605,
"grad_norm": 0.6239631772041321,
"learning_rate": 3.76e-05,
"loss": 1.3156,
"step": 125
},
{
"epoch": 0.04449152542372881,
"grad_norm": 0.8934807181358337,
"learning_rate": 3.7500000000000003e-05,
"loss": 1.7629,
"step": 126
},
{
"epoch": 0.044844632768361585,
"grad_norm": 0.9005821943283081,
"learning_rate": 3.74e-05,
"loss": 1.5775,
"step": 127
},
{
"epoch": 0.04519774011299435,
"grad_norm": 0.7520875334739685,
"learning_rate": 3.73e-05,
"loss": 1.6556,
"step": 128
},
{
"epoch": 0.045550847457627115,
"grad_norm": 0.8343580961227417,
"learning_rate": 3.72e-05,
"loss": 1.5429,
"step": 129
},
{
"epoch": 0.04590395480225989,
"grad_norm": 0.7819499373435974,
"learning_rate": 3.71e-05,
"loss": 1.5313,
"step": 130
},
{
"epoch": 0.04625706214689265,
"grad_norm": 0.6105003952980042,
"learning_rate": 3.7e-05,
"loss": 1.3124,
"step": 131
},
{
"epoch": 0.046610169491525424,
"grad_norm": 0.6769589185714722,
"learning_rate": 3.69e-05,
"loss": 1.4305,
"step": 132
},
{
"epoch": 0.04696327683615819,
"grad_norm": 0.7654910683631897,
"learning_rate": 3.68e-05,
"loss": 1.4972,
"step": 133
},
{
"epoch": 0.04731638418079096,
"grad_norm": 0.7610757946968079,
"learning_rate": 3.6700000000000004e-05,
"loss": 1.4652,
"step": 134
},
{
"epoch": 0.04766949152542373,
"grad_norm": 0.6594064831733704,
"learning_rate": 3.66e-05,
"loss": 1.2301,
"step": 135
},
{
"epoch": 0.0480225988700565,
"grad_norm": 0.719152569770813,
"learning_rate": 3.65e-05,
"loss": 1.3399,
"step": 136
},
{
"epoch": 0.048375706214689264,
"grad_norm": 0.9311054348945618,
"learning_rate": 3.6400000000000004e-05,
"loss": 1.5786,
"step": 137
},
{
"epoch": 0.048728813559322036,
"grad_norm": 0.7845869660377502,
"learning_rate": 3.63e-05,
"loss": 1.3293,
"step": 138
},
{
"epoch": 0.0490819209039548,
"grad_norm": 0.6555350422859192,
"learning_rate": 3.62e-05,
"loss": 1.1677,
"step": 139
},
{
"epoch": 0.04943502824858757,
"grad_norm": 0.863808810710907,
"learning_rate": 3.61e-05,
"loss": 1.3453,
"step": 140
},
{
"epoch": 0.04978813559322034,
"grad_norm": 1.1489522457122803,
"learning_rate": 3.6e-05,
"loss": 1.5277,
"step": 141
},
{
"epoch": 0.05014124293785311,
"grad_norm": 0.8987497091293335,
"learning_rate": 3.59e-05,
"loss": 1.3435,
"step": 142
},
{
"epoch": 0.050494350282485875,
"grad_norm": 1.1033629179000854,
"learning_rate": 3.58e-05,
"loss": 1.5565,
"step": 143
},
{
"epoch": 0.05084745762711865,
"grad_norm": 0.9892997145652771,
"learning_rate": 3.57e-05,
"loss": 1.5017,
"step": 144
},
{
"epoch": 0.05120056497175141,
"grad_norm": 0.8673040866851807,
"learning_rate": 3.56e-05,
"loss": 1.3484,
"step": 145
},
{
"epoch": 0.05155367231638418,
"grad_norm": 0.8514487147331238,
"learning_rate": 3.55e-05,
"loss": 1.4329,
"step": 146
},
{
"epoch": 0.05190677966101695,
"grad_norm": 0.7865382432937622,
"learning_rate": 3.54e-05,
"loss": 1.3205,
"step": 147
},
{
"epoch": 0.052259887005649715,
"grad_norm": 0.9893730878829956,
"learning_rate": 3.53e-05,
"loss": 1.2557,
"step": 148
},
{
"epoch": 0.05261299435028249,
"grad_norm": 0.7459471225738525,
"learning_rate": 3.52e-05,
"loss": 1.398,
"step": 149
},
{
"epoch": 0.05296610169491525,
"grad_norm": 1.2132583856582642,
"learning_rate": 3.51e-05,
"loss": 1.396,
"step": 150
},
{
"epoch": 0.053319209039548024,
"grad_norm": 1.1160212755203247,
"learning_rate": 3.5e-05,
"loss": 1.3013,
"step": 151
},
{
"epoch": 0.05367231638418079,
"grad_norm": 1.058889389038086,
"learning_rate": 3.49e-05,
"loss": 1.4149,
"step": 152
},
{
"epoch": 0.05402542372881356,
"grad_norm": 1.065194845199585,
"learning_rate": 3.48e-05,
"loss": 1.3913,
"step": 153
},
{
"epoch": 0.054378531073446326,
"grad_norm": 0.8159513473510742,
"learning_rate": 3.4699999999999996e-05,
"loss": 1.1171,
"step": 154
},
{
"epoch": 0.0547316384180791,
"grad_norm": 1.2298649549484253,
"learning_rate": 3.46e-05,
"loss": 1.2559,
"step": 155
},
{
"epoch": 0.05508474576271186,
"grad_norm": 0.845851480960846,
"learning_rate": 3.45e-05,
"loss": 1.1559,
"step": 156
},
{
"epoch": 0.055437853107344635,
"grad_norm": 0.8105614185333252,
"learning_rate": 3.4399999999999996e-05,
"loss": 1.092,
"step": 157
},
{
"epoch": 0.0557909604519774,
"grad_norm": 0.9624556303024292,
"learning_rate": 3.430000000000001e-05,
"loss": 1.1862,
"step": 158
},
{
"epoch": 0.05614406779661017,
"grad_norm": 1.2695869207382202,
"learning_rate": 3.4200000000000005e-05,
"loss": 1.1886,
"step": 159
},
{
"epoch": 0.05649717514124294,
"grad_norm": 0.9481382369995117,
"learning_rate": 3.41e-05,
"loss": 1.2366,
"step": 160
},
{
"epoch": 0.0568502824858757,
"grad_norm": 1.1612368822097778,
"learning_rate": 3.4000000000000007e-05,
"loss": 1.3102,
"step": 161
},
{
"epoch": 0.057203389830508475,
"grad_norm": 1.296866536140442,
"learning_rate": 3.3900000000000004e-05,
"loss": 1.2689,
"step": 162
},
{
"epoch": 0.05755649717514124,
"grad_norm": 1.0605868101119995,
"learning_rate": 3.38e-05,
"loss": 1.1056,
"step": 163
},
{
"epoch": 0.05790960451977401,
"grad_norm": 1.06467604637146,
"learning_rate": 3.3700000000000006e-05,
"loss": 1.2329,
"step": 164
},
{
"epoch": 0.05826271186440678,
"grad_norm": 0.9635260105133057,
"learning_rate": 3.3600000000000004e-05,
"loss": 1.082,
"step": 165
},
{
"epoch": 0.05861581920903955,
"grad_norm": 1.2294679880142212,
"learning_rate": 3.35e-05,
"loss": 1.3892,
"step": 166
},
{
"epoch": 0.058968926553672314,
"grad_norm": 1.4516520500183105,
"learning_rate": 3.3400000000000005e-05,
"loss": 1.0772,
"step": 167
},
{
"epoch": 0.059322033898305086,
"grad_norm": 1.08195960521698,
"learning_rate": 3.33e-05,
"loss": 1.0811,
"step": 168
},
{
"epoch": 0.05967514124293785,
"grad_norm": 1.1554380655288696,
"learning_rate": 3.32e-05,
"loss": 1.1417,
"step": 169
},
{
"epoch": 0.06002824858757062,
"grad_norm": 1.117101788520813,
"learning_rate": 3.3100000000000005e-05,
"loss": 1.2163,
"step": 170
},
{
"epoch": 0.06038135593220339,
"grad_norm": 1.2141193151474,
"learning_rate": 3.3e-05,
"loss": 1.2469,
"step": 171
},
{
"epoch": 0.06073446327683616,
"grad_norm": 1.29922616481781,
"learning_rate": 3.29e-05,
"loss": 1.2541,
"step": 172
},
{
"epoch": 0.061087570621468926,
"grad_norm": 1.3265364170074463,
"learning_rate": 3.2800000000000004e-05,
"loss": 1.1884,
"step": 173
},
{
"epoch": 0.0614406779661017,
"grad_norm": 1.4586164951324463,
"learning_rate": 3.27e-05,
"loss": 1.1861,
"step": 174
},
{
"epoch": 0.06179378531073446,
"grad_norm": 1.3976298570632935,
"learning_rate": 3.26e-05,
"loss": 1.0568,
"step": 175
},
{
"epoch": 0.062146892655367235,
"grad_norm": 1.0947189331054688,
"learning_rate": 3.2500000000000004e-05,
"loss": 1.0583,
"step": 176
},
{
"epoch": 0.0625,
"grad_norm": 1.0739431381225586,
"learning_rate": 3.24e-05,
"loss": 0.973,
"step": 177
},
{
"epoch": 0.06285310734463277,
"grad_norm": 1.0500564575195312,
"learning_rate": 3.2300000000000006e-05,
"loss": 0.961,
"step": 178
},
{
"epoch": 0.06320621468926553,
"grad_norm": 1.3906699419021606,
"learning_rate": 3.2200000000000003e-05,
"loss": 1.1366,
"step": 179
},
{
"epoch": 0.0635593220338983,
"grad_norm": 1.2436065673828125,
"learning_rate": 3.21e-05,
"loss": 1.0769,
"step": 180
},
{
"epoch": 0.06391242937853107,
"grad_norm": 1.137102484703064,
"learning_rate": 3.2000000000000005e-05,
"loss": 0.9609,
"step": 181
},
{
"epoch": 0.06426553672316385,
"grad_norm": 1.7101986408233643,
"learning_rate": 3.19e-05,
"loss": 1.138,
"step": 182
},
{
"epoch": 0.0646186440677966,
"grad_norm": 1.1683790683746338,
"learning_rate": 3.18e-05,
"loss": 1.1872,
"step": 183
},
{
"epoch": 0.06497175141242938,
"grad_norm": 1.6510800123214722,
"learning_rate": 3.1700000000000005e-05,
"loss": 1.0618,
"step": 184
},
{
"epoch": 0.06532485875706215,
"grad_norm": 1.5926707983016968,
"learning_rate": 3.16e-05,
"loss": 1.0558,
"step": 185
},
{
"epoch": 0.06567796610169492,
"grad_norm": 1.502793312072754,
"learning_rate": 3.15e-05,
"loss": 0.9155,
"step": 186
},
{
"epoch": 0.06603107344632768,
"grad_norm": 1.5513490438461304,
"learning_rate": 3.1400000000000004e-05,
"loss": 1.1119,
"step": 187
},
{
"epoch": 0.06638418079096045,
"grad_norm": 1.7766902446746826,
"learning_rate": 3.13e-05,
"loss": 1.0724,
"step": 188
},
{
"epoch": 0.06673728813559322,
"grad_norm": 1.6195155382156372,
"learning_rate": 3.12e-05,
"loss": 1.0377,
"step": 189
},
{
"epoch": 0.067090395480226,
"grad_norm": 1.5613833665847778,
"learning_rate": 3.1100000000000004e-05,
"loss": 1.0057,
"step": 190
},
{
"epoch": 0.06744350282485875,
"grad_norm": 1.5648159980773926,
"learning_rate": 3.1e-05,
"loss": 0.9758,
"step": 191
},
{
"epoch": 0.06779661016949153,
"grad_norm": 2.3179798126220703,
"learning_rate": 3.09e-05,
"loss": 1.076,
"step": 192
},
{
"epoch": 0.0681497175141243,
"grad_norm": 1.6775708198547363,
"learning_rate": 3.08e-05,
"loss": 0.8133,
"step": 193
},
{
"epoch": 0.06850282485875706,
"grad_norm": 1.6693114042282104,
"learning_rate": 3.07e-05,
"loss": 0.8958,
"step": 194
},
{
"epoch": 0.06885593220338983,
"grad_norm": 1.623583436012268,
"learning_rate": 3.06e-05,
"loss": 0.9226,
"step": 195
},
{
"epoch": 0.0692090395480226,
"grad_norm": 1.6135281324386597,
"learning_rate": 3.05e-05,
"loss": 0.8707,
"step": 196
},
{
"epoch": 0.06956214689265537,
"grad_norm": 1.6629116535186768,
"learning_rate": 3.04e-05,
"loss": 0.8883,
"step": 197
},
{
"epoch": 0.06991525423728813,
"grad_norm": 1.9516435861587524,
"learning_rate": 3.03e-05,
"loss": 0.8909,
"step": 198
},
{
"epoch": 0.0702683615819209,
"grad_norm": 1.9378143548965454,
"learning_rate": 3.02e-05,
"loss": 0.7697,
"step": 199
},
{
"epoch": 0.07062146892655367,
"grad_norm": 2.0332353115081787,
"learning_rate": 3.01e-05,
"loss": 0.9999,
"step": 200
},
{
"epoch": 0.07097457627118645,
"grad_norm": 1.4929075241088867,
"learning_rate": 3e-05,
"loss": 0.8301,
"step": 201
},
{
"epoch": 0.0713276836158192,
"grad_norm": 1.9575985670089722,
"learning_rate": 2.9900000000000002e-05,
"loss": 1.0613,
"step": 202
},
{
"epoch": 0.07168079096045198,
"grad_norm": 1.3801037073135376,
"learning_rate": 2.98e-05,
"loss": 1.0363,
"step": 203
},
{
"epoch": 0.07203389830508475,
"grad_norm": 1.8629417419433594,
"learning_rate": 2.97e-05,
"loss": 0.8268,
"step": 204
},
{
"epoch": 0.07238700564971752,
"grad_norm": 1.5077078342437744,
"learning_rate": 2.96e-05,
"loss": 0.9769,
"step": 205
},
{
"epoch": 0.07274011299435028,
"grad_norm": 1.8674854040145874,
"learning_rate": 2.95e-05,
"loss": 1.011,
"step": 206
},
{
"epoch": 0.07309322033898305,
"grad_norm": 2.605397939682007,
"learning_rate": 2.94e-05,
"loss": 1.2138,
"step": 207
},
{
"epoch": 0.07344632768361582,
"grad_norm": 1.3890961408615112,
"learning_rate": 2.93e-05,
"loss": 0.9088,
"step": 208
},
{
"epoch": 0.0737994350282486,
"grad_norm": 2.1312317848205566,
"learning_rate": 2.9199999999999998e-05,
"loss": 1.2366,
"step": 209
},
{
"epoch": 0.07415254237288135,
"grad_norm": 1.817219853401184,
"learning_rate": 2.91e-05,
"loss": 0.7755,
"step": 210
},
{
"epoch": 0.07450564971751412,
"grad_norm": 1.8591907024383545,
"learning_rate": 2.9e-05,
"loss": 0.9953,
"step": 211
},
{
"epoch": 0.0748587570621469,
"grad_norm": 1.7456353902816772,
"learning_rate": 2.8899999999999998e-05,
"loss": 0.7605,
"step": 212
},
{
"epoch": 0.07521186440677965,
"grad_norm": 1.6678094863891602,
"learning_rate": 2.88e-05,
"loss": 0.779,
"step": 213
},
{
"epoch": 0.07556497175141243,
"grad_norm": 1.7547659873962402,
"learning_rate": 2.87e-05,
"loss": 0.7277,
"step": 214
},
{
"epoch": 0.0759180790960452,
"grad_norm": 2.3935739994049072,
"learning_rate": 2.86e-05,
"loss": 1.0243,
"step": 215
},
{
"epoch": 0.07627118644067797,
"grad_norm": 2.1737802028656006,
"learning_rate": 2.8499999999999998e-05,
"loss": 0.8159,
"step": 216
},
{
"epoch": 0.07662429378531073,
"grad_norm": 2.551464557647705,
"learning_rate": 2.84e-05,
"loss": 0.7808,
"step": 217
},
{
"epoch": 0.0769774011299435,
"grad_norm": 2.0867929458618164,
"learning_rate": 2.83e-05,
"loss": 0.7764,
"step": 218
},
{
"epoch": 0.07733050847457627,
"grad_norm": 2.001920700073242,
"learning_rate": 2.8199999999999998e-05,
"loss": 0.7499,
"step": 219
},
{
"epoch": 0.07768361581920905,
"grad_norm": 2.4141225814819336,
"learning_rate": 2.8100000000000005e-05,
"loss": 0.7964,
"step": 220
},
{
"epoch": 0.0780367231638418,
"grad_norm": 2.371626615524292,
"learning_rate": 2.8000000000000003e-05,
"loss": 0.8997,
"step": 221
},
{
"epoch": 0.07838983050847458,
"grad_norm": 1.6987398862838745,
"learning_rate": 2.7900000000000004e-05,
"loss": 0.9769,
"step": 222
},
{
"epoch": 0.07874293785310735,
"grad_norm": 1.768283486366272,
"learning_rate": 2.7800000000000005e-05,
"loss": 0.6651,
"step": 223
},
{
"epoch": 0.07909604519774012,
"grad_norm": 1.827958106994629,
"learning_rate": 2.7700000000000002e-05,
"loss": 0.6805,
"step": 224
},
{
"epoch": 0.07944915254237288,
"grad_norm": 2.7550013065338135,
"learning_rate": 2.7600000000000003e-05,
"loss": 0.8034,
"step": 225
},
{
"epoch": 0.07980225988700565,
"grad_norm": 1.7423927783966064,
"learning_rate": 2.7500000000000004e-05,
"loss": 0.7345,
"step": 226
},
{
"epoch": 0.08015536723163842,
"grad_norm": 1.9607951641082764,
"learning_rate": 2.7400000000000002e-05,
"loss": 0.6771,
"step": 227
},
{
"epoch": 0.08050847457627118,
"grad_norm": 2.737659215927124,
"learning_rate": 2.7300000000000003e-05,
"loss": 1.1056,
"step": 228
},
{
"epoch": 0.08086158192090395,
"grad_norm": 3.3492250442504883,
"learning_rate": 2.7200000000000004e-05,
"loss": 1.2373,
"step": 229
},
{
"epoch": 0.08121468926553672,
"grad_norm": 2.10353946685791,
"learning_rate": 2.7100000000000005e-05,
"loss": 0.8509,
"step": 230
},
{
"epoch": 0.0815677966101695,
"grad_norm": 2.1462435722351074,
"learning_rate": 2.7000000000000002e-05,
"loss": 0.8893,
"step": 231
},
{
"epoch": 0.08192090395480225,
"grad_norm": 2.2392938137054443,
"learning_rate": 2.6900000000000003e-05,
"loss": 0.8682,
"step": 232
},
{
"epoch": 0.08227401129943503,
"grad_norm": 2.242868661880493,
"learning_rate": 2.6800000000000004e-05,
"loss": 0.812,
"step": 233
},
{
"epoch": 0.0826271186440678,
"grad_norm": 1.9220484495162964,
"learning_rate": 2.6700000000000002e-05,
"loss": 0.7693,
"step": 234
},
{
"epoch": 0.08298022598870057,
"grad_norm": 1.5621178150177002,
"learning_rate": 2.6600000000000003e-05,
"loss": 0.875,
"step": 235
},
{
"epoch": 0.08333333333333333,
"grad_norm": 1.8653405904769897,
"learning_rate": 2.6500000000000004e-05,
"loss": 0.6442,
"step": 236
},
{
"epoch": 0.0836864406779661,
"grad_norm": 2.035147190093994,
"learning_rate": 2.64e-05,
"loss": 0.8103,
"step": 237
},
{
"epoch": 0.08403954802259887,
"grad_norm": 2.0776431560516357,
"learning_rate": 2.6300000000000002e-05,
"loss": 0.8276,
"step": 238
},
{
"epoch": 0.08439265536723164,
"grad_norm": 1.745665192604065,
"learning_rate": 2.6200000000000003e-05,
"loss": 0.6145,
"step": 239
},
{
"epoch": 0.0847457627118644,
"grad_norm": 1.9675729274749756,
"learning_rate": 2.61e-05,
"loss": 0.7005,
"step": 240
},
{
"epoch": 0.08509887005649718,
"grad_norm": 2.2235677242279053,
"learning_rate": 2.6000000000000002e-05,
"loss": 0.5884,
"step": 241
},
{
"epoch": 0.08545197740112995,
"grad_norm": 1.8204203844070435,
"learning_rate": 2.5900000000000003e-05,
"loss": 0.6029,
"step": 242
},
{
"epoch": 0.0858050847457627,
"grad_norm": 1.6106022596359253,
"learning_rate": 2.58e-05,
"loss": 0.5933,
"step": 243
},
{
"epoch": 0.08615819209039548,
"grad_norm": 1.8598597049713135,
"learning_rate": 2.57e-05,
"loss": 0.6792,
"step": 244
},
{
"epoch": 0.08651129943502825,
"grad_norm": 2.186007499694824,
"learning_rate": 2.5600000000000002e-05,
"loss": 0.8079,
"step": 245
},
{
"epoch": 0.08686440677966102,
"grad_norm": 2.206789255142212,
"learning_rate": 2.5500000000000003e-05,
"loss": 0.936,
"step": 246
},
{
"epoch": 0.08721751412429378,
"grad_norm": 1.865030288696289,
"learning_rate": 2.54e-05,
"loss": 0.8269,
"step": 247
},
{
"epoch": 0.08757062146892655,
"grad_norm": 3.2279410362243652,
"learning_rate": 2.5300000000000002e-05,
"loss": 0.9072,
"step": 248
},
{
"epoch": 0.08792372881355932,
"grad_norm": 1.5364686250686646,
"learning_rate": 2.5200000000000003e-05,
"loss": 0.5486,
"step": 249
},
{
"epoch": 0.0882768361581921,
"grad_norm": 1.7777973413467407,
"learning_rate": 2.51e-05,
"loss": 0.5444,
"step": 250
},
{
"epoch": 0.08862994350282485,
"grad_norm": 2.1404268741607666,
"learning_rate": 2.5e-05,
"loss": 0.6368,
"step": 251
},
{
"epoch": 0.08898305084745763,
"grad_norm": 1.80747389793396,
"learning_rate": 2.4900000000000002e-05,
"loss": 0.6483,
"step": 252
},
{
"epoch": 0.0893361581920904,
"grad_norm": 1.780600666999817,
"learning_rate": 2.48e-05,
"loss": 0.7834,
"step": 253
},
{
"epoch": 0.08968926553672317,
"grad_norm": 2.232827663421631,
"learning_rate": 2.47e-05,
"loss": 0.7617,
"step": 254
},
{
"epoch": 0.09004237288135593,
"grad_norm": 1.9321049451828003,
"learning_rate": 2.46e-05,
"loss": 0.5936,
"step": 255
},
{
"epoch": 0.0903954802259887,
"grad_norm": 1.7767024040222168,
"learning_rate": 2.45e-05,
"loss": 0.5475,
"step": 256
},
{
"epoch": 0.09074858757062147,
"grad_norm": 1.7460103034973145,
"learning_rate": 2.44e-05,
"loss": 0.7268,
"step": 257
},
{
"epoch": 0.09110169491525423,
"grad_norm": 2.010267734527588,
"learning_rate": 2.43e-05,
"loss": 0.5198,
"step": 258
},
{
"epoch": 0.091454802259887,
"grad_norm": 2.2814278602600098,
"learning_rate": 2.4200000000000002e-05,
"loss": 0.8076,
"step": 259
},
{
"epoch": 0.09180790960451977,
"grad_norm": 2.043018341064453,
"learning_rate": 2.41e-05,
"loss": 0.7212,
"step": 260
},
{
"epoch": 0.09216101694915255,
"grad_norm": 1.5442239046096802,
"learning_rate": 2.4e-05,
"loss": 0.6238,
"step": 261
},
{
"epoch": 0.0925141242937853,
"grad_norm": 2.0493950843811035,
"learning_rate": 2.39e-05,
"loss": 0.6279,
"step": 262
},
{
"epoch": 0.09286723163841808,
"grad_norm": 1.9058549404144287,
"learning_rate": 2.38e-05,
"loss": 0.6636,
"step": 263
},
{
"epoch": 0.09322033898305085,
"grad_norm": 1.766516089439392,
"learning_rate": 2.37e-05,
"loss": 0.7014,
"step": 264
},
{
"epoch": 0.09357344632768362,
"grad_norm": 1.7538968324661255,
"learning_rate": 2.36e-05,
"loss": 0.7425,
"step": 265
},
{
"epoch": 0.09392655367231638,
"grad_norm": 1.9766415357589722,
"learning_rate": 2.35e-05,
"loss": 0.6857,
"step": 266
},
{
"epoch": 0.09427966101694915,
"grad_norm": 1.8969347476959229,
"learning_rate": 2.3400000000000003e-05,
"loss": 0.509,
"step": 267
},
{
"epoch": 0.09463276836158192,
"grad_norm": 2.556140422821045,
"learning_rate": 2.3300000000000004e-05,
"loss": 0.9246,
"step": 268
},
{
"epoch": 0.0949858757062147,
"grad_norm": 2.192314624786377,
"learning_rate": 2.32e-05,
"loss": 0.8475,
"step": 269
},
{
"epoch": 0.09533898305084745,
"grad_norm": 1.6340404748916626,
"learning_rate": 2.3100000000000002e-05,
"loss": 0.8045,
"step": 270
},
{
"epoch": 0.09569209039548023,
"grad_norm": 1.849310040473938,
"learning_rate": 2.3000000000000003e-05,
"loss": 0.6461,
"step": 271
},
{
"epoch": 0.096045197740113,
"grad_norm": 1.7643516063690186,
"learning_rate": 2.29e-05,
"loss": 0.8039,
"step": 272
},
{
"epoch": 0.09639830508474577,
"grad_norm": 2.249030351638794,
"learning_rate": 2.2800000000000002e-05,
"loss": 0.6257,
"step": 273
},
{
"epoch": 0.09675141242937853,
"grad_norm": 1.8955878019332886,
"learning_rate": 2.2700000000000003e-05,
"loss": 0.4904,
"step": 274
},
{
"epoch": 0.0971045197740113,
"grad_norm": 2.037266254425049,
"learning_rate": 2.26e-05,
"loss": 0.9272,
"step": 275
},
{
"epoch": 0.09745762711864407,
"grad_norm": 2.061370611190796,
"learning_rate": 2.25e-05,
"loss": 0.5984,
"step": 276
},
{
"epoch": 0.09781073446327683,
"grad_norm": 3.4211597442626953,
"learning_rate": 2.2400000000000002e-05,
"loss": 0.9706,
"step": 277
},
{
"epoch": 0.0981638418079096,
"grad_norm": 2.7527401447296143,
"learning_rate": 2.23e-05,
"loss": 0.7066,
"step": 278
},
{
"epoch": 0.09851694915254237,
"grad_norm": 1.8547669649124146,
"learning_rate": 2.22e-05,
"loss": 0.458,
"step": 279
},
{
"epoch": 0.09887005649717515,
"grad_norm": 1.9559056758880615,
"learning_rate": 2.2100000000000002e-05,
"loss": 0.453,
"step": 280
},
{
"epoch": 0.0992231638418079,
"grad_norm": 1.7714954614639282,
"learning_rate": 2.2000000000000003e-05,
"loss": 0.5063,
"step": 281
},
{
"epoch": 0.09957627118644068,
"grad_norm": 3.084165334701538,
"learning_rate": 2.19e-05,
"loss": 0.8611,
"step": 282
},
{
"epoch": 0.09992937853107345,
"grad_norm": 1.8296337127685547,
"learning_rate": 2.18e-05,
"loss": 0.6884,
"step": 283
},
{
"epoch": 0.10028248587570622,
"grad_norm": 2.5133588314056396,
"learning_rate": 2.1700000000000002e-05,
"loss": 0.6772,
"step": 284
},
{
"epoch": 0.10063559322033898,
"grad_norm": 1.612042784690857,
"learning_rate": 2.16e-05,
"loss": 0.4726,
"step": 285
},
{
"epoch": 0.10098870056497175,
"grad_norm": 1.9312748908996582,
"learning_rate": 2.15e-05,
"loss": 0.6412,
"step": 286
},
{
"epoch": 0.10134180790960452,
"grad_norm": 2.4446702003479004,
"learning_rate": 2.1400000000000002e-05,
"loss": 0.8463,
"step": 287
},
{
"epoch": 0.1016949152542373,
"grad_norm": 4.253321647644043,
"learning_rate": 2.13e-05,
"loss": 0.6177,
"step": 288
},
{
"epoch": 0.10204802259887005,
"grad_norm": 2.584301471710205,
"learning_rate": 2.12e-05,
"loss": 0.7182,
"step": 289
},
{
"epoch": 0.10240112994350282,
"grad_norm": 1.7066371440887451,
"learning_rate": 2.11e-05,
"loss": 0.5563,
"step": 290
},
{
"epoch": 0.1027542372881356,
"grad_norm": 3.1515045166015625,
"learning_rate": 2.1e-05,
"loss": 0.8663,
"step": 291
},
{
"epoch": 0.10310734463276836,
"grad_norm": 2.3074467182159424,
"learning_rate": 2.09e-05,
"loss": 0.8209,
"step": 292
},
{
"epoch": 0.10346045197740113,
"grad_norm": 2.0748472213745117,
"learning_rate": 2.08e-05,
"loss": 0.6029,
"step": 293
},
{
"epoch": 0.1038135593220339,
"grad_norm": 1.8115911483764648,
"learning_rate": 2.07e-05,
"loss": 0.4101,
"step": 294
},
{
"epoch": 0.10416666666666667,
"grad_norm": 4.644301891326904,
"learning_rate": 2.06e-05,
"loss": 0.8823,
"step": 295
},
{
"epoch": 0.10451977401129943,
"grad_norm": 1.8356761932373047,
"learning_rate": 2.05e-05,
"loss": 0.6229,
"step": 296
},
{
"epoch": 0.1048728813559322,
"grad_norm": 2.584825038909912,
"learning_rate": 2.04e-05,
"loss": 0.7403,
"step": 297
},
{
"epoch": 0.10522598870056497,
"grad_norm": 2.1190598011016846,
"learning_rate": 2.0300000000000002e-05,
"loss": 0.5843,
"step": 298
},
{
"epoch": 0.10557909604519775,
"grad_norm": 2.0052149295806885,
"learning_rate": 2.0200000000000003e-05,
"loss": 0.5476,
"step": 299
},
{
"epoch": 0.1059322033898305,
"grad_norm": 2.5021302700042725,
"learning_rate": 2.01e-05,
"loss": 0.6322,
"step": 300
},
{
"epoch": 0.10628531073446328,
"grad_norm": 2.790252208709717,
"learning_rate": 2e-05,
"loss": 0.7356,
"step": 301
},
{
"epoch": 0.10663841807909605,
"grad_norm": 1.896161437034607,
"learning_rate": 1.9900000000000003e-05,
"loss": 0.6414,
"step": 302
},
{
"epoch": 0.10699152542372882,
"grad_norm": 2.121238946914673,
"learning_rate": 1.9800000000000004e-05,
"loss": 0.4369,
"step": 303
},
{
"epoch": 0.10734463276836158,
"grad_norm": 2.465498685836792,
"learning_rate": 1.97e-05,
"loss": 0.7176,
"step": 304
},
{
"epoch": 0.10769774011299435,
"grad_norm": 2.068796396255493,
"learning_rate": 1.9600000000000002e-05,
"loss": 0.4301,
"step": 305
},
{
"epoch": 0.10805084745762712,
"grad_norm": 1.9807084798812866,
"learning_rate": 1.9500000000000003e-05,
"loss": 0.4117,
"step": 306
},
{
"epoch": 0.10840395480225988,
"grad_norm": 2.1290664672851562,
"learning_rate": 1.94e-05,
"loss": 0.636,
"step": 307
},
{
"epoch": 0.10875706214689265,
"grad_norm": 1.9659547805786133,
"learning_rate": 1.93e-05,
"loss": 0.51,
"step": 308
},
{
"epoch": 0.10911016949152542,
"grad_norm": 2.1764960289001465,
"learning_rate": 1.9200000000000003e-05,
"loss": 0.6201,
"step": 309
},
{
"epoch": 0.1094632768361582,
"grad_norm": 2.654648542404175,
"learning_rate": 1.91e-05,
"loss": 0.7892,
"step": 310
},
{
"epoch": 0.10981638418079095,
"grad_norm": 1.8828105926513672,
"learning_rate": 1.9e-05,
"loss": 0.6453,
"step": 311
},
{
"epoch": 0.11016949152542373,
"grad_norm": 2.0999748706817627,
"learning_rate": 1.8900000000000002e-05,
"loss": 0.7547,
"step": 312
},
{
"epoch": 0.1105225988700565,
"grad_norm": 1.9529651403427124,
"learning_rate": 1.88e-05,
"loss": 0.3852,
"step": 313
},
{
"epoch": 0.11087570621468927,
"grad_norm": 2.1099841594696045,
"learning_rate": 1.87e-05,
"loss": 0.8429,
"step": 314
},
{
"epoch": 0.11122881355932203,
"grad_norm": 2.6168792247772217,
"learning_rate": 1.86e-05,
"loss": 0.8384,
"step": 315
},
{
"epoch": 0.1115819209039548,
"grad_norm": 1.8378658294677734,
"learning_rate": 1.85e-05,
"loss": 0.5949,
"step": 316
},
{
"epoch": 0.11193502824858757,
"grad_norm": 2.333801746368408,
"learning_rate": 1.84e-05,
"loss": 0.4418,
"step": 317
},
{
"epoch": 0.11228813559322035,
"grad_norm": 2.6359400749206543,
"learning_rate": 1.83e-05,
"loss": 0.7196,
"step": 318
},
{
"epoch": 0.1126412429378531,
"grad_norm": 1.830614447593689,
"learning_rate": 1.8200000000000002e-05,
"loss": 0.5038,
"step": 319
},
{
"epoch": 0.11299435028248588,
"grad_norm": 1.8708122968673706,
"learning_rate": 1.81e-05,
"loss": 0.356,
"step": 320
},
{
"epoch": 0.11334745762711865,
"grad_norm": 1.8595423698425293,
"learning_rate": 1.8e-05,
"loss": 0.5855,
"step": 321
},
{
"epoch": 0.1137005649717514,
"grad_norm": 2.0622501373291016,
"learning_rate": 1.79e-05,
"loss": 0.4762,
"step": 322
},
{
"epoch": 0.11405367231638418,
"grad_norm": 1.8576161861419678,
"learning_rate": 1.78e-05,
"loss": 0.6245,
"step": 323
},
{
"epoch": 0.11440677966101695,
"grad_norm": 2.014206886291504,
"learning_rate": 1.77e-05,
"loss": 0.4627,
"step": 324
},
{
"epoch": 0.11475988700564972,
"grad_norm": 1.8058968782424927,
"learning_rate": 1.76e-05,
"loss": 0.3285,
"step": 325
},
{
"epoch": 0.11511299435028248,
"grad_norm": 1.7399612665176392,
"learning_rate": 1.75e-05,
"loss": 0.4845,
"step": 326
},
{
"epoch": 0.11546610169491525,
"grad_norm": 2.8591692447662354,
"learning_rate": 1.74e-05,
"loss": 0.7819,
"step": 327
},
{
"epoch": 0.11581920903954802,
"grad_norm": 2.9054338932037354,
"learning_rate": 1.73e-05,
"loss": 0.6961,
"step": 328
},
{
"epoch": 0.1161723163841808,
"grad_norm": 2.231865167617798,
"learning_rate": 1.7199999999999998e-05,
"loss": 0.532,
"step": 329
},
{
"epoch": 0.11652542372881355,
"grad_norm": 3.112537145614624,
"learning_rate": 1.7100000000000002e-05,
"loss": 0.7317,
"step": 330
},
{
"epoch": 0.11687853107344633,
"grad_norm": 1.886364221572876,
"learning_rate": 1.7000000000000003e-05,
"loss": 0.4278,
"step": 331
},
{
"epoch": 0.1172316384180791,
"grad_norm": 2.2548704147338867,
"learning_rate": 1.69e-05,
"loss": 0.4959,
"step": 332
},
{
"epoch": 0.11758474576271187,
"grad_norm": 1.778956651687622,
"learning_rate": 1.6800000000000002e-05,
"loss": 0.4093,
"step": 333
},
{
"epoch": 0.11793785310734463,
"grad_norm": 3.506840229034424,
"learning_rate": 1.6700000000000003e-05,
"loss": 0.889,
"step": 334
},
{
"epoch": 0.1182909604519774,
"grad_norm": 1.7313032150268555,
"learning_rate": 1.66e-05,
"loss": 0.3059,
"step": 335
},
{
"epoch": 0.11864406779661017,
"grad_norm": 2.914469003677368,
"learning_rate": 1.65e-05,
"loss": 0.6975,
"step": 336
},
{
"epoch": 0.11899717514124294,
"grad_norm": 2.3646645545959473,
"learning_rate": 1.6400000000000002e-05,
"loss": 0.3799,
"step": 337
},
{
"epoch": 0.1193502824858757,
"grad_norm": 2.708645820617676,
"learning_rate": 1.63e-05,
"loss": 0.6651,
"step": 338
},
{
"epoch": 0.11970338983050847,
"grad_norm": 2.4956207275390625,
"learning_rate": 1.62e-05,
"loss": 0.5777,
"step": 339
},
{
"epoch": 0.12005649717514125,
"grad_norm": 2.371713161468506,
"learning_rate": 1.6100000000000002e-05,
"loss": 0.477,
"step": 340
},
{
"epoch": 0.120409604519774,
"grad_norm": 1.81998872756958,
"learning_rate": 1.6000000000000003e-05,
"loss": 0.3214,
"step": 341
},
{
"epoch": 0.12076271186440678,
"grad_norm": 2.2463088035583496,
"learning_rate": 1.59e-05,
"loss": 0.602,
"step": 342
},
{
"epoch": 0.12111581920903955,
"grad_norm": 2.2001595497131348,
"learning_rate": 1.58e-05,
"loss": 0.6024,
"step": 343
},
{
"epoch": 0.12146892655367232,
"grad_norm": 2.6866791248321533,
"learning_rate": 1.5700000000000002e-05,
"loss": 0.693,
"step": 344
},
{
"epoch": 0.12182203389830508,
"grad_norm": 2.6667373180389404,
"learning_rate": 1.56e-05,
"loss": 0.6857,
"step": 345
},
{
"epoch": 0.12217514124293785,
"grad_norm": 2.1822969913482666,
"learning_rate": 1.55e-05,
"loss": 0.51,
"step": 346
},
{
"epoch": 0.12252824858757062,
"grad_norm": 2.881263256072998,
"learning_rate": 1.54e-05,
"loss": 0.6404,
"step": 347
},
{
"epoch": 0.1228813559322034,
"grad_norm": 2.1057329177856445,
"learning_rate": 1.53e-05,
"loss": 0.495,
"step": 348
},
{
"epoch": 0.12323446327683615,
"grad_norm": 1.6811401844024658,
"learning_rate": 1.52e-05,
"loss": 0.2848,
"step": 349
},
{
"epoch": 0.12358757062146893,
"grad_norm": 2.3314602375030518,
"learning_rate": 1.51e-05,
"loss": 0.5686,
"step": 350
},
{
"epoch": 0.1239406779661017,
"grad_norm": 2.8521921634674072,
"learning_rate": 1.5e-05,
"loss": 0.7763,
"step": 351
},
{
"epoch": 0.12429378531073447,
"grad_norm": 2.379612922668457,
"learning_rate": 1.49e-05,
"loss": 0.6625,
"step": 352
},
{
"epoch": 0.12464689265536723,
"grad_norm": 1.7267616987228394,
"learning_rate": 1.48e-05,
"loss": 0.3089,
"step": 353
},
{
"epoch": 0.125,
"grad_norm": 2.078075647354126,
"learning_rate": 1.47e-05,
"loss": 0.5671,
"step": 354
},
{
"epoch": 0.12535310734463276,
"grad_norm": 2.491132974624634,
"learning_rate": 1.4599999999999999e-05,
"loss": 0.5108,
"step": 355
},
{
"epoch": 0.12570621468926554,
"grad_norm": 2.2118701934814453,
"learning_rate": 1.45e-05,
"loss": 0.4687,
"step": 356
},
{
"epoch": 0.1260593220338983,
"grad_norm": 1.556470274925232,
"learning_rate": 1.44e-05,
"loss": 0.2878,
"step": 357
},
{
"epoch": 0.12641242937853106,
"grad_norm": 3.0352628231048584,
"learning_rate": 1.43e-05,
"loss": 0.7796,
"step": 358
},
{
"epoch": 0.12676553672316385,
"grad_norm": 2.498322010040283,
"learning_rate": 1.42e-05,
"loss": 0.6919,
"step": 359
},
{
"epoch": 0.1271186440677966,
"grad_norm": 2.7435500621795654,
"learning_rate": 1.4099999999999999e-05,
"loss": 0.6379,
"step": 360
},
{
"epoch": 0.1274717514124294,
"grad_norm": 2.1981303691864014,
"learning_rate": 1.4000000000000001e-05,
"loss": 0.4841,
"step": 361
},
{
"epoch": 0.12782485875706215,
"grad_norm": 4.253507137298584,
"learning_rate": 1.3900000000000002e-05,
"loss": 0.8417,
"step": 362
},
{
"epoch": 0.1281779661016949,
"grad_norm": 2.2498624324798584,
"learning_rate": 1.3800000000000002e-05,
"loss": 0.3849,
"step": 363
},
{
"epoch": 0.1285310734463277,
"grad_norm": 2.5313210487365723,
"learning_rate": 1.3700000000000001e-05,
"loss": 0.7152,
"step": 364
},
{
"epoch": 0.12888418079096045,
"grad_norm": 2.427729368209839,
"learning_rate": 1.3600000000000002e-05,
"loss": 0.613,
"step": 365
},
{
"epoch": 0.1292372881355932,
"grad_norm": 2.1720805168151855,
"learning_rate": 1.3500000000000001e-05,
"loss": 0.527,
"step": 366
},
{
"epoch": 0.129590395480226,
"grad_norm": 2.054292678833008,
"learning_rate": 1.3400000000000002e-05,
"loss": 0.4659,
"step": 367
},
{
"epoch": 0.12994350282485875,
"grad_norm": 2.235578775405884,
"learning_rate": 1.3300000000000001e-05,
"loss": 0.4747,
"step": 368
},
{
"epoch": 0.13029661016949154,
"grad_norm": 1.5781618356704712,
"learning_rate": 1.32e-05,
"loss": 0.2759,
"step": 369
},
{
"epoch": 0.1306497175141243,
"grad_norm": 1.8499678373336792,
"learning_rate": 1.3100000000000002e-05,
"loss": 0.4529,
"step": 370
},
{
"epoch": 0.13100282485875706,
"grad_norm": 1.901394248008728,
"learning_rate": 1.3000000000000001e-05,
"loss": 0.5179,
"step": 371
},
{
"epoch": 0.13135593220338984,
"grad_norm": 2.1701114177703857,
"learning_rate": 1.29e-05,
"loss": 0.5104,
"step": 372
},
{
"epoch": 0.1317090395480226,
"grad_norm": 2.1720824241638184,
"learning_rate": 1.2800000000000001e-05,
"loss": 0.3809,
"step": 373
},
{
"epoch": 0.13206214689265536,
"grad_norm": 2.1540091037750244,
"learning_rate": 1.27e-05,
"loss": 0.5311,
"step": 374
},
{
"epoch": 0.13241525423728814,
"grad_norm": 2.0049026012420654,
"learning_rate": 1.2600000000000001e-05,
"loss": 0.363,
"step": 375
},
{
"epoch": 0.1327683615819209,
"grad_norm": 2.1072335243225098,
"learning_rate": 1.25e-05,
"loss": 0.4218,
"step": 376
},
{
"epoch": 0.13312146892655366,
"grad_norm": 1.9690706729888916,
"learning_rate": 1.24e-05,
"loss": 0.4304,
"step": 377
},
{
"epoch": 0.13347457627118645,
"grad_norm": 1.7669113874435425,
"learning_rate": 1.23e-05,
"loss": 0.3686,
"step": 378
},
{
"epoch": 0.1338276836158192,
"grad_norm": 1.9775711297988892,
"learning_rate": 1.22e-05,
"loss": 0.503,
"step": 379
},
{
"epoch": 0.134180790960452,
"grad_norm": 2.458343744277954,
"learning_rate": 1.2100000000000001e-05,
"loss": 0.4038,
"step": 380
},
{
"epoch": 0.13453389830508475,
"grad_norm": 1.9653310775756836,
"learning_rate": 1.2e-05,
"loss": 0.5042,
"step": 381
},
{
"epoch": 0.1348870056497175,
"grad_norm": 1.832046627998352,
"learning_rate": 1.19e-05,
"loss": 0.3696,
"step": 382
},
{
"epoch": 0.1352401129943503,
"grad_norm": 4.0539374351501465,
"learning_rate": 1.18e-05,
"loss": 0.8044,
"step": 383
},
{
"epoch": 0.13559322033898305,
"grad_norm": 4.322567462921143,
"learning_rate": 1.1700000000000001e-05,
"loss": 0.9251,
"step": 384
},
{
"epoch": 0.1359463276836158,
"grad_norm": 1.9108883142471313,
"learning_rate": 1.16e-05,
"loss": 0.4827,
"step": 385
},
{
"epoch": 0.1362994350282486,
"grad_norm": 1.4768140316009521,
"learning_rate": 1.1500000000000002e-05,
"loss": 0.2541,
"step": 386
},
{
"epoch": 0.13665254237288135,
"grad_norm": 2.005392551422119,
"learning_rate": 1.1400000000000001e-05,
"loss": 0.5762,
"step": 387
},
{
"epoch": 0.1370056497175141,
"grad_norm": 3.0048270225524902,
"learning_rate": 1.13e-05,
"loss": 0.6806,
"step": 388
},
{
"epoch": 0.1373587570621469,
"grad_norm": 2.6098363399505615,
"learning_rate": 1.1200000000000001e-05,
"loss": 0.5409,
"step": 389
},
{
"epoch": 0.13771186440677965,
"grad_norm": 2.0291426181793213,
"learning_rate": 1.11e-05,
"loss": 0.4466,
"step": 390
},
{
"epoch": 0.13806497175141244,
"grad_norm": 1.4195945262908936,
"learning_rate": 1.1000000000000001e-05,
"loss": 0.2425,
"step": 391
},
{
"epoch": 0.1384180790960452,
"grad_norm": 3.1550965309143066,
"learning_rate": 1.09e-05,
"loss": 0.6254,
"step": 392
},
{
"epoch": 0.13877118644067796,
"grad_norm": 1.7930920124053955,
"learning_rate": 1.08e-05,
"loss": 0.3466,
"step": 393
},
{
"epoch": 0.13912429378531074,
"grad_norm": 1.9969501495361328,
"learning_rate": 1.0700000000000001e-05,
"loss": 0.3996,
"step": 394
},
{
"epoch": 0.1394774011299435,
"grad_norm": 2.3298864364624023,
"learning_rate": 1.06e-05,
"loss": 0.4196,
"step": 395
},
{
"epoch": 0.13983050847457626,
"grad_norm": 4.122328281402588,
"learning_rate": 1.05e-05,
"loss": 0.3735,
"step": 396
},
{
"epoch": 0.14018361581920905,
"grad_norm": 2.8163208961486816,
"learning_rate": 1.04e-05,
"loss": 0.4261,
"step": 397
},
{
"epoch": 0.1405367231638418,
"grad_norm": 1.7612152099609375,
"learning_rate": 1.03e-05,
"loss": 0.4001,
"step": 398
},
{
"epoch": 0.1408898305084746,
"grad_norm": 2.542123317718506,
"learning_rate": 1.02e-05,
"loss": 0.5853,
"step": 399
},
{
"epoch": 0.14124293785310735,
"grad_norm": 2.013018846511841,
"learning_rate": 1.0100000000000002e-05,
"loss": 0.344,
"step": 400
},
{
"epoch": 0.1415960451977401,
"grad_norm": 2.6476850509643555,
"learning_rate": 1e-05,
"loss": 0.5595,
"step": 401
},
{
"epoch": 0.1419491525423729,
"grad_norm": 2.0753297805786133,
"learning_rate": 9.900000000000002e-06,
"loss": 0.3958,
"step": 402
},
{
"epoch": 0.14230225988700565,
"grad_norm": 1.726521372795105,
"learning_rate": 9.800000000000001e-06,
"loss": 0.3896,
"step": 403
},
{
"epoch": 0.1426553672316384,
"grad_norm": 2.0191128253936768,
"learning_rate": 9.7e-06,
"loss": 0.4541,
"step": 404
},
{
"epoch": 0.1430084745762712,
"grad_norm": 2.52801775932312,
"learning_rate": 9.600000000000001e-06,
"loss": 0.4609,
"step": 405
},
{
"epoch": 0.14336158192090395,
"grad_norm": 2.738588809967041,
"learning_rate": 9.5e-06,
"loss": 0.4313,
"step": 406
},
{
"epoch": 0.1437146892655367,
"grad_norm": 2.0913681983947754,
"learning_rate": 9.4e-06,
"loss": 0.4176,
"step": 407
},
{
"epoch": 0.1440677966101695,
"grad_norm": 1.8325021266937256,
"learning_rate": 9.3e-06,
"loss": 0.3419,
"step": 408
},
{
"epoch": 0.14442090395480225,
"grad_norm": 2.245701551437378,
"learning_rate": 9.2e-06,
"loss": 0.4799,
"step": 409
},
{
"epoch": 0.14477401129943504,
"grad_norm": 3.9585647583007812,
"learning_rate": 9.100000000000001e-06,
"loss": 0.6476,
"step": 410
},
{
"epoch": 0.1451271186440678,
"grad_norm": 3.2401275634765625,
"learning_rate": 9e-06,
"loss": 0.6274,
"step": 411
},
{
"epoch": 0.14548022598870056,
"grad_norm": 2.1975040435791016,
"learning_rate": 8.9e-06,
"loss": 0.4277,
"step": 412
},
{
"epoch": 0.14583333333333334,
"grad_norm": 1.8795210123062134,
"learning_rate": 8.8e-06,
"loss": 0.2486,
"step": 413
},
{
"epoch": 0.1461864406779661,
"grad_norm": 2.162196159362793,
"learning_rate": 8.7e-06,
"loss": 0.3372,
"step": 414
},
{
"epoch": 0.14653954802259886,
"grad_norm": 2.1754472255706787,
"learning_rate": 8.599999999999999e-06,
"loss": 0.4053,
"step": 415
},
{
"epoch": 0.14689265536723164,
"grad_norm": 2.0052478313446045,
"learning_rate": 8.500000000000002e-06,
"loss": 0.3775,
"step": 416
},
{
"epoch": 0.1472457627118644,
"grad_norm": 2.1078007221221924,
"learning_rate": 8.400000000000001e-06,
"loss": 0.5518,
"step": 417
},
{
"epoch": 0.1475988700564972,
"grad_norm": 2.85091495513916,
"learning_rate": 8.3e-06,
"loss": 0.8025,
"step": 418
},
{
"epoch": 0.14795197740112995,
"grad_norm": 1.9279358386993408,
"learning_rate": 8.200000000000001e-06,
"loss": 0.4753,
"step": 419
},
{
"epoch": 0.1483050847457627,
"grad_norm": 2.3065125942230225,
"learning_rate": 8.1e-06,
"loss": 0.3965,
"step": 420
},
{
"epoch": 0.1486581920903955,
"grad_norm": 3.294403076171875,
"learning_rate": 8.000000000000001e-06,
"loss": 0.6205,
"step": 421
},
{
"epoch": 0.14901129943502825,
"grad_norm": 1.7436306476593018,
"learning_rate": 7.9e-06,
"loss": 0.239,
"step": 422
},
{
"epoch": 0.149364406779661,
"grad_norm": 1.420858383178711,
"learning_rate": 7.8e-06,
"loss": 0.2265,
"step": 423
},
{
"epoch": 0.1497175141242938,
"grad_norm": 2.940077066421509,
"learning_rate": 7.7e-06,
"loss": 0.6006,
"step": 424
},
{
"epoch": 0.15007062146892655,
"grad_norm": 2.1452486515045166,
"learning_rate": 7.6e-06,
"loss": 0.3333,
"step": 425
},
{
"epoch": 0.1504237288135593,
"grad_norm": 4.092085838317871,
"learning_rate": 7.5e-06,
"loss": 0.7523,
"step": 426
},
{
"epoch": 0.1507768361581921,
"grad_norm": 1.8198660612106323,
"learning_rate": 7.4e-06,
"loss": 0.3283,
"step": 427
},
{
"epoch": 0.15112994350282485,
"grad_norm": 1.3757944107055664,
"learning_rate": 7.2999999999999996e-06,
"loss": 0.2199,
"step": 428
},
{
"epoch": 0.15148305084745764,
"grad_norm": 2.0032095909118652,
"learning_rate": 7.2e-06,
"loss": 0.3994,
"step": 429
},
{
"epoch": 0.1518361581920904,
"grad_norm": 2.9988086223602295,
"learning_rate": 7.1e-06,
"loss": 0.3558,
"step": 430
},
{
"epoch": 0.15218926553672316,
"grad_norm": 2.7509193420410156,
"learning_rate": 7.000000000000001e-06,
"loss": 0.5825,
"step": 431
},
{
"epoch": 0.15254237288135594,
"grad_norm": 2.523036003112793,
"learning_rate": 6.900000000000001e-06,
"loss": 0.484,
"step": 432
},
{
"epoch": 0.1528954802259887,
"grad_norm": 1.8027594089508057,
"learning_rate": 6.800000000000001e-06,
"loss": 0.382,
"step": 433
},
{
"epoch": 0.15324858757062146,
"grad_norm": 2.55008602142334,
"learning_rate": 6.700000000000001e-06,
"loss": 0.4458,
"step": 434
},
{
"epoch": 0.15360169491525424,
"grad_norm": 2.159984827041626,
"learning_rate": 6.6e-06,
"loss": 0.3597,
"step": 435
},
{
"epoch": 0.153954802259887,
"grad_norm": 2.2070958614349365,
"learning_rate": 6.5000000000000004e-06,
"loss": 0.483,
"step": 436
},
{
"epoch": 0.15430790960451976,
"grad_norm": 2.283949851989746,
"learning_rate": 6.4000000000000006e-06,
"loss": 0.3999,
"step": 437
},
{
"epoch": 0.15466101694915255,
"grad_norm": 2.086303472518921,
"learning_rate": 6.300000000000001e-06,
"loss": 0.4877,
"step": 438
},
{
"epoch": 0.1550141242937853,
"grad_norm": 2.2829809188842773,
"learning_rate": 6.2e-06,
"loss": 0.4839,
"step": 439
},
{
"epoch": 0.1553672316384181,
"grad_norm": 2.9696552753448486,
"learning_rate": 6.1e-06,
"loss": 0.5502,
"step": 440
},
{
"epoch": 0.15572033898305085,
"grad_norm": 2.7994322776794434,
"learning_rate": 6e-06,
"loss": 0.5695,
"step": 441
},
{
"epoch": 0.1560734463276836,
"grad_norm": 1.9340434074401855,
"learning_rate": 5.9e-06,
"loss": 0.3296,
"step": 442
},
{
"epoch": 0.1564265536723164,
"grad_norm": 2.5606064796447754,
"learning_rate": 5.8e-06,
"loss": 0.411,
"step": 443
},
{
"epoch": 0.15677966101694915,
"grad_norm": 2.6650333404541016,
"learning_rate": 5.7000000000000005e-06,
"loss": 0.3179,
"step": 444
},
{
"epoch": 0.1571327683615819,
"grad_norm": 4.912551403045654,
"learning_rate": 5.600000000000001e-06,
"loss": 0.8535,
"step": 445
},
{
"epoch": 0.1574858757062147,
"grad_norm": 2.2671737670898438,
"learning_rate": 5.500000000000001e-06,
"loss": 0.3959,
"step": 446
},
{
"epoch": 0.15783898305084745,
"grad_norm": 2.4728825092315674,
"learning_rate": 5.4e-06,
"loss": 0.3118,
"step": 447
},
{
"epoch": 0.15819209039548024,
"grad_norm": 1.7785649299621582,
"learning_rate": 5.3e-06,
"loss": 0.228,
"step": 448
},
{
"epoch": 0.158545197740113,
"grad_norm": 2.288245916366577,
"learning_rate": 5.2e-06,
"loss": 0.3586,
"step": 449
},
{
"epoch": 0.15889830508474576,
"grad_norm": 1.821871042251587,
"learning_rate": 5.1e-06,
"loss": 0.362,
"step": 450
},
{
"epoch": 0.15925141242937854,
"grad_norm": 1.9763644933700562,
"learning_rate": 5e-06,
"loss": 0.3057,
"step": 451
},
{
"epoch": 0.1596045197740113,
"grad_norm": 1.8915050029754639,
"learning_rate": 4.9000000000000005e-06,
"loss": 0.3815,
"step": 452
},
{
"epoch": 0.15995762711864406,
"grad_norm": 1.787156343460083,
"learning_rate": 4.800000000000001e-06,
"loss": 0.2223,
"step": 453
},
{
"epoch": 0.16031073446327684,
"grad_norm": 2.0852158069610596,
"learning_rate": 4.7e-06,
"loss": 0.4158,
"step": 454
},
{
"epoch": 0.1606638418079096,
"grad_norm": 1.9223723411560059,
"learning_rate": 4.6e-06,
"loss": 0.3043,
"step": 455
},
{
"epoch": 0.16101694915254236,
"grad_norm": 1.8598310947418213,
"learning_rate": 4.5e-06,
"loss": 0.3886,
"step": 456
},
{
"epoch": 0.16137005649717515,
"grad_norm": 2.242039918899536,
"learning_rate": 4.4e-06,
"loss": 0.4703,
"step": 457
},
{
"epoch": 0.1617231638418079,
"grad_norm": 1.7315804958343506,
"learning_rate": 4.2999999999999995e-06,
"loss": 0.2178,
"step": 458
},
{
"epoch": 0.1620762711864407,
"grad_norm": 2.204817295074463,
"learning_rate": 4.2000000000000004e-06,
"loss": 0.3898,
"step": 459
},
{
"epoch": 0.16242937853107345,
"grad_norm": 2.1929430961608887,
"learning_rate": 4.1000000000000006e-06,
"loss": 0.3549,
"step": 460
},
{
"epoch": 0.1627824858757062,
"grad_norm": 1.7955076694488525,
"learning_rate": 4.000000000000001e-06,
"loss": 0.3815,
"step": 461
},
{
"epoch": 0.163135593220339,
"grad_norm": 4.15421724319458,
"learning_rate": 3.9e-06,
"loss": 0.6433,
"step": 462
},
{
"epoch": 0.16348870056497175,
"grad_norm": 1.9674700498580933,
"learning_rate": 3.8e-06,
"loss": 0.3425,
"step": 463
},
{
"epoch": 0.1638418079096045,
"grad_norm": 1.909616231918335,
"learning_rate": 3.7e-06,
"loss": 0.343,
"step": 464
},
{
"epoch": 0.1641949152542373,
"grad_norm": 2.229336977005005,
"learning_rate": 3.6e-06,
"loss": 0.4534,
"step": 465
},
{
"epoch": 0.16454802259887005,
"grad_norm": 1.3924556970596313,
"learning_rate": 3.5000000000000004e-06,
"loss": 0.2077,
"step": 466
},
{
"epoch": 0.1649011299435028,
"grad_norm": 1.3659465312957764,
"learning_rate": 3.4000000000000005e-06,
"loss": 0.2068,
"step": 467
},
{
"epoch": 0.1652542372881356,
"grad_norm": 1.867235779762268,
"learning_rate": 3.3e-06,
"loss": 0.2963,
"step": 468
},
{
"epoch": 0.16560734463276836,
"grad_norm": 1.332623839378357,
"learning_rate": 3.2000000000000003e-06,
"loss": 0.2054,
"step": 469
},
{
"epoch": 0.16596045197740114,
"grad_norm": 1.6542198657989502,
"learning_rate": 3.1e-06,
"loss": 0.2103,
"step": 470
},
{
"epoch": 0.1663135593220339,
"grad_norm": 2.120319128036499,
"learning_rate": 3e-06,
"loss": 0.4215,
"step": 471
},
{
"epoch": 0.16666666666666666,
"grad_norm": 1.8779698610305786,
"learning_rate": 2.9e-06,
"loss": 0.298,
"step": 472
},
{
"epoch": 0.16701977401129944,
"grad_norm": 2.830299139022827,
"learning_rate": 2.8000000000000003e-06,
"loss": 0.542,
"step": 473
},
{
"epoch": 0.1673728813559322,
"grad_norm": 2.909973621368408,
"learning_rate": 2.7e-06,
"loss": 0.4126,
"step": 474
},
{
"epoch": 0.16772598870056496,
"grad_norm": 4.498685359954834,
"learning_rate": 2.6e-06,
"loss": 0.5541,
"step": 475
},
{
"epoch": 0.16807909604519775,
"grad_norm": 2.452082395553589,
"learning_rate": 2.5e-06,
"loss": 0.3653,
"step": 476
},
{
"epoch": 0.1684322033898305,
"grad_norm": 1.9502593278884888,
"learning_rate": 2.4000000000000003e-06,
"loss": 0.3149,
"step": 477
},
{
"epoch": 0.1687853107344633,
"grad_norm": 2.2951347827911377,
"learning_rate": 2.3e-06,
"loss": 0.3794,
"step": 478
},
{
"epoch": 0.16913841807909605,
"grad_norm": 3.0120177268981934,
"learning_rate": 2.2e-06,
"loss": 0.5245,
"step": 479
},
{
"epoch": 0.1694915254237288,
"grad_norm": 4.143613815307617,
"learning_rate": 2.1000000000000002e-06,
"loss": 0.6444,
"step": 480
},
{
"epoch": 0.1698446327683616,
"grad_norm": 1.838753342628479,
"learning_rate": 2.0000000000000003e-06,
"loss": 0.3379,
"step": 481
},
{
"epoch": 0.17019774011299435,
"grad_norm": 1.9792391061782837,
"learning_rate": 1.9e-06,
"loss": 0.3329,
"step": 482
},
{
"epoch": 0.1705508474576271,
"grad_norm": 1.8502469062805176,
"learning_rate": 1.8e-06,
"loss": 0.3703,
"step": 483
},
{
"epoch": 0.1709039548022599,
"grad_norm": 1.618812918663025,
"learning_rate": 1.7000000000000002e-06,
"loss": 0.2051,
"step": 484
},
{
"epoch": 0.17125706214689265,
"grad_norm": 2.346790075302124,
"learning_rate": 1.6000000000000001e-06,
"loss": 0.49,
"step": 485
},
{
"epoch": 0.1716101694915254,
"grad_norm": 1.8763465881347656,
"learning_rate": 1.5e-06,
"loss": 0.2948,
"step": 486
},
{
"epoch": 0.1719632768361582,
"grad_norm": 1.7063709497451782,
"learning_rate": 1.4000000000000001e-06,
"loss": 0.3513,
"step": 487
},
{
"epoch": 0.17231638418079095,
"grad_norm": 2.963041067123413,
"learning_rate": 1.3e-06,
"loss": 0.6783,
"step": 488
},
{
"epoch": 0.17266949152542374,
"grad_norm": 2.9806196689605713,
"learning_rate": 1.2000000000000002e-06,
"loss": 0.5123,
"step": 489
},
{
"epoch": 0.1730225988700565,
"grad_norm": 1.6302564144134521,
"learning_rate": 1.1e-06,
"loss": 0.204,
"step": 490
},
{
"epoch": 0.17337570621468926,
"grad_norm": 1.9233272075653076,
"learning_rate": 1.0000000000000002e-06,
"loss": 0.3037,
"step": 491
},
{
"epoch": 0.17372881355932204,
"grad_norm": 1.8432778120040894,
"learning_rate": 9e-07,
"loss": 0.3368,
"step": 492
},
{
"epoch": 0.1740819209039548,
"grad_norm": 2.5116734504699707,
"learning_rate": 8.000000000000001e-07,
"loss": 0.5263,
"step": 493
},
{
"epoch": 0.17443502824858756,
"grad_norm": 3.0080406665802,
"learning_rate": 7.000000000000001e-07,
"loss": 0.3415,
"step": 494
},
{
"epoch": 0.17478813559322035,
"grad_norm": 1.982960820198059,
"learning_rate": 6.000000000000001e-07,
"loss": 0.459,
"step": 495
},
{
"epoch": 0.1751412429378531,
"grad_norm": 2.259504795074463,
"learning_rate": 5.000000000000001e-07,
"loss": 0.4088,
"step": 496
},
{
"epoch": 0.1754943502824859,
"grad_norm": 1.937008261680603,
"learning_rate": 4.0000000000000003e-07,
"loss": 0.2903,
"step": 497
},
{
"epoch": 0.17584745762711865,
"grad_norm": 1.8799011707305908,
"learning_rate": 3.0000000000000004e-07,
"loss": 0.2905,
"step": 498
},
{
"epoch": 0.1762005649717514,
"grad_norm": 2.9519193172454834,
"learning_rate": 2.0000000000000002e-07,
"loss": 0.525,
"step": 499
},
{
"epoch": 0.1765536723163842,
"grad_norm": 1.2929198741912842,
"learning_rate": 1.0000000000000001e-07,
"loss": 0.1987,
"step": 500
}
],
"logging_steps": 1,
"max_steps": 500,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 4.099784700491894e+16,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}