MedAgentReasoner-3B / trainer_state.json
tossowski's picture
Upload folder using huggingface_hub
717eeae verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 4.0,
"eval_steps": 500,
"global_step": 932,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.004308023694130318,
"grad_norm": 46.39875030517578,
"learning_rate": 9.9999954157983e-06,
"loss": 2.2558,
"step": 1
},
{
"epoch": 0.008616047388260635,
"grad_norm": 15.635394096374512,
"learning_rate": 9.999981663201606e-06,
"loss": 1.9294,
"step": 2
},
{
"epoch": 0.012924071082390954,
"grad_norm": 11.906808853149414,
"learning_rate": 9.999958742235133e-06,
"loss": 1.6914,
"step": 3
},
{
"epoch": 0.01723209477652127,
"grad_norm": 8.960531234741211,
"learning_rate": 9.999926652940914e-06,
"loss": 1.5586,
"step": 4
},
{
"epoch": 0.021540118470651588,
"grad_norm": 6.32525634765625,
"learning_rate": 9.999885395377788e-06,
"loss": 1.3009,
"step": 5
},
{
"epoch": 0.025848142164781908,
"grad_norm": 4.644639015197754,
"learning_rate": 9.999834969621408e-06,
"loss": 1.1522,
"step": 6
},
{
"epoch": 0.030156165858912225,
"grad_norm": 3.9122426509857178,
"learning_rate": 9.999775375764244e-06,
"loss": 1.061,
"step": 7
},
{
"epoch": 0.03446418955304254,
"grad_norm": 4.616422653198242,
"learning_rate": 9.999706613915567e-06,
"loss": 0.9941,
"step": 8
},
{
"epoch": 0.03877221324717286,
"grad_norm": 3.8503878116607666,
"learning_rate": 9.999628684201464e-06,
"loss": 0.8931,
"step": 9
},
{
"epoch": 0.043080236941303175,
"grad_norm": 3.5663602352142334,
"learning_rate": 9.999541586764836e-06,
"loss": 0.943,
"step": 10
},
{
"epoch": 0.04738826063543349,
"grad_norm": 3.536245584487915,
"learning_rate": 9.999445321765392e-06,
"loss": 0.8607,
"step": 11
},
{
"epoch": 0.051696284329563816,
"grad_norm": 3.5765976905822754,
"learning_rate": 9.999339889379647e-06,
"loss": 0.8233,
"step": 12
},
{
"epoch": 0.05600430802369413,
"grad_norm": 3.3080978393554688,
"learning_rate": 9.999225289800935e-06,
"loss": 0.7818,
"step": 13
},
{
"epoch": 0.06031233171782445,
"grad_norm": 3.949313163757324,
"learning_rate": 9.999101523239392e-06,
"loss": 0.8687,
"step": 14
},
{
"epoch": 0.06462035541195477,
"grad_norm": 3.9806888103485107,
"learning_rate": 9.998968589921969e-06,
"loss": 0.7793,
"step": 15
},
{
"epoch": 0.06892837910608508,
"grad_norm": 3.6018688678741455,
"learning_rate": 9.99882649009242e-06,
"loss": 0.7317,
"step": 16
},
{
"epoch": 0.0732364028002154,
"grad_norm": 3.872166872024536,
"learning_rate": 9.998675224011317e-06,
"loss": 0.7262,
"step": 17
},
{
"epoch": 0.07754442649434572,
"grad_norm": 5.003402233123779,
"learning_rate": 9.998514791956025e-06,
"loss": 0.7164,
"step": 18
},
{
"epoch": 0.08185245018847603,
"grad_norm": 3.848304033279419,
"learning_rate": 9.998345194220732e-06,
"loss": 0.6768,
"step": 19
},
{
"epoch": 0.08616047388260635,
"grad_norm": 3.976318359375,
"learning_rate": 9.998166431116421e-06,
"loss": 0.6572,
"step": 20
},
{
"epoch": 0.09046849757673667,
"grad_norm": 4.318434238433838,
"learning_rate": 9.99797850297089e-06,
"loss": 0.6814,
"step": 21
},
{
"epoch": 0.09477652127086698,
"grad_norm": 4.235775947570801,
"learning_rate": 9.997781410128737e-06,
"loss": 0.6579,
"step": 22
},
{
"epoch": 0.0990845449649973,
"grad_norm": 4.045382022857666,
"learning_rate": 9.99757515295137e-06,
"loss": 0.6532,
"step": 23
},
{
"epoch": 0.10339256865912763,
"grad_norm": 3.9923672676086426,
"learning_rate": 9.997359731816998e-06,
"loss": 0.6583,
"step": 24
},
{
"epoch": 0.10770059235325795,
"grad_norm": 4.043659687042236,
"learning_rate": 9.997135147120633e-06,
"loss": 0.6421,
"step": 25
},
{
"epoch": 0.11200861604738827,
"grad_norm": 4.166319370269775,
"learning_rate": 9.996901399274093e-06,
"loss": 0.6647,
"step": 26
},
{
"epoch": 0.11631663974151858,
"grad_norm": 4.195486545562744,
"learning_rate": 9.996658488705997e-06,
"loss": 0.6038,
"step": 27
},
{
"epoch": 0.1206246634356489,
"grad_norm": 3.893251657485962,
"learning_rate": 9.996406415861763e-06,
"loss": 0.5954,
"step": 28
},
{
"epoch": 0.12493268712977922,
"grad_norm": 3.857978582382202,
"learning_rate": 9.996145181203616e-06,
"loss": 0.5698,
"step": 29
},
{
"epoch": 0.12924071082390953,
"grad_norm": 3.830559492111206,
"learning_rate": 9.995874785210573e-06,
"loss": 0.5761,
"step": 30
},
{
"epoch": 0.13354873451803984,
"grad_norm": 3.8703525066375732,
"learning_rate": 9.995595228378456e-06,
"loss": 0.5811,
"step": 31
},
{
"epoch": 0.13785675821217017,
"grad_norm": 3.8460729122161865,
"learning_rate": 9.995306511219885e-06,
"loss": 0.5873,
"step": 32
},
{
"epoch": 0.1421647819063005,
"grad_norm": 3.9502267837524414,
"learning_rate": 9.995008634264272e-06,
"loss": 0.5795,
"step": 33
},
{
"epoch": 0.1464728056004308,
"grad_norm": 3.718348979949951,
"learning_rate": 9.994701598057828e-06,
"loss": 0.5406,
"step": 34
},
{
"epoch": 0.15078082929456113,
"grad_norm": 3.751901865005493,
"learning_rate": 9.99438540316356e-06,
"loss": 0.5619,
"step": 35
},
{
"epoch": 0.15508885298869143,
"grad_norm": 3.7187840938568115,
"learning_rate": 9.99406005016127e-06,
"loss": 0.5569,
"step": 36
},
{
"epoch": 0.15939687668282176,
"grad_norm": 3.647937536239624,
"learning_rate": 9.99372553964755e-06,
"loss": 0.5342,
"step": 37
},
{
"epoch": 0.16370490037695207,
"grad_norm": 3.6027166843414307,
"learning_rate": 9.993381872235785e-06,
"loss": 0.5279,
"step": 38
},
{
"epoch": 0.1680129240710824,
"grad_norm": 3.61478590965271,
"learning_rate": 9.993029048556154e-06,
"loss": 0.5131,
"step": 39
},
{
"epoch": 0.1723209477652127,
"grad_norm": 3.574810743331909,
"learning_rate": 9.99266706925562e-06,
"loss": 0.5165,
"step": 40
},
{
"epoch": 0.17662897145934303,
"grad_norm": 3.5639119148254395,
"learning_rate": 9.99229593499794e-06,
"loss": 0.5093,
"step": 41
},
{
"epoch": 0.18093699515347333,
"grad_norm": 3.4759559631347656,
"learning_rate": 9.991915646463652e-06,
"loss": 0.4864,
"step": 42
},
{
"epoch": 0.18524501884760367,
"grad_norm": 3.4125938415527344,
"learning_rate": 9.991526204350087e-06,
"loss": 0.4915,
"step": 43
},
{
"epoch": 0.18955304254173397,
"grad_norm": 3.4541664123535156,
"learning_rate": 9.991127609371357e-06,
"loss": 0.4891,
"step": 44
},
{
"epoch": 0.1938610662358643,
"grad_norm": 3.4495904445648193,
"learning_rate": 9.990719862258357e-06,
"loss": 0.4676,
"step": 45
},
{
"epoch": 0.1981690899299946,
"grad_norm": 3.2870962619781494,
"learning_rate": 9.990302963758765e-06,
"loss": 0.4679,
"step": 46
},
{
"epoch": 0.20247711362412493,
"grad_norm": 3.3008880615234375,
"learning_rate": 9.989876914637042e-06,
"loss": 0.4688,
"step": 47
},
{
"epoch": 0.20678513731825526,
"grad_norm": 3.086277723312378,
"learning_rate": 9.989441715674422e-06,
"loss": 0.4565,
"step": 48
},
{
"epoch": 0.21109316101238557,
"grad_norm": 3.048494338989258,
"learning_rate": 9.988997367668924e-06,
"loss": 0.4358,
"step": 49
},
{
"epoch": 0.2154011847065159,
"grad_norm": 2.9959912300109863,
"learning_rate": 9.988543871435342e-06,
"loss": 0.439,
"step": 50
},
{
"epoch": 0.2197092084006462,
"grad_norm": 2.891278028488159,
"learning_rate": 9.988081227805237e-06,
"loss": 0.4534,
"step": 51
},
{
"epoch": 0.22401723209477653,
"grad_norm": 3.0530009269714355,
"learning_rate": 9.987609437626955e-06,
"loss": 0.42,
"step": 52
},
{
"epoch": 0.22832525578890683,
"grad_norm": 2.8397328853607178,
"learning_rate": 9.987128501765606e-06,
"loss": 0.4071,
"step": 53
},
{
"epoch": 0.23263327948303716,
"grad_norm": 2.5578064918518066,
"learning_rate": 9.986638421103074e-06,
"loss": 0.3947,
"step": 54
},
{
"epoch": 0.23694130317716747,
"grad_norm": 2.627624750137329,
"learning_rate": 9.986139196538011e-06,
"loss": 0.4039,
"step": 55
},
{
"epoch": 0.2412493268712978,
"grad_norm": 2.515475273132324,
"learning_rate": 9.985630828985835e-06,
"loss": 0.4188,
"step": 56
},
{
"epoch": 0.2455573505654281,
"grad_norm": 2.286128282546997,
"learning_rate": 9.98511331937873e-06,
"loss": 0.3922,
"step": 57
},
{
"epoch": 0.24986537425955843,
"grad_norm": 2.2815139293670654,
"learning_rate": 9.984586668665641e-06,
"loss": 0.4452,
"step": 58
},
{
"epoch": 0.25417339795368876,
"grad_norm": 2.0464518070220947,
"learning_rate": 9.98405087781228e-06,
"loss": 0.391,
"step": 59
},
{
"epoch": 0.25848142164781907,
"grad_norm": 2.0818228721618652,
"learning_rate": 9.983505947801115e-06,
"loss": 0.4185,
"step": 60
},
{
"epoch": 0.26278944534194937,
"grad_norm": 1.843110203742981,
"learning_rate": 9.982951879631373e-06,
"loss": 0.3706,
"step": 61
},
{
"epoch": 0.26709746903607967,
"grad_norm": 1.8725769519805908,
"learning_rate": 9.982388674319041e-06,
"loss": 0.3723,
"step": 62
},
{
"epoch": 0.27140549273021003,
"grad_norm": 1.7885315418243408,
"learning_rate": 9.981816332896854e-06,
"loss": 0.3903,
"step": 63
},
{
"epoch": 0.27571351642434033,
"grad_norm": 1.5986205339431763,
"learning_rate": 9.981234856414306e-06,
"loss": 0.3759,
"step": 64
},
{
"epoch": 0.28002154011847064,
"grad_norm": 1.5188963413238525,
"learning_rate": 9.98064424593764e-06,
"loss": 0.3641,
"step": 65
},
{
"epoch": 0.284329563812601,
"grad_norm": 1.4895678758621216,
"learning_rate": 9.980044502549843e-06,
"loss": 0.3722,
"step": 66
},
{
"epoch": 0.2886375875067313,
"grad_norm": 1.517785668373108,
"learning_rate": 9.979435627350658e-06,
"loss": 0.4034,
"step": 67
},
{
"epoch": 0.2929456112008616,
"grad_norm": 1.5465859174728394,
"learning_rate": 9.978817621456562e-06,
"loss": 0.404,
"step": 68
},
{
"epoch": 0.2972536348949919,
"grad_norm": 1.4948019981384277,
"learning_rate": 9.978190486000784e-06,
"loss": 0.409,
"step": 69
},
{
"epoch": 0.30156165858912226,
"grad_norm": 1.349805474281311,
"learning_rate": 9.977554222133293e-06,
"loss": 0.3809,
"step": 70
},
{
"epoch": 0.30586968228325256,
"grad_norm": 1.4803236722946167,
"learning_rate": 9.976908831020787e-06,
"loss": 0.3709,
"step": 71
},
{
"epoch": 0.31017770597738287,
"grad_norm": 1.500108003616333,
"learning_rate": 9.97625431384671e-06,
"loss": 0.3875,
"step": 72
},
{
"epoch": 0.31448572967151317,
"grad_norm": 1.431912899017334,
"learning_rate": 9.975590671811239e-06,
"loss": 0.3897,
"step": 73
},
{
"epoch": 0.31879375336564353,
"grad_norm": 1.3888511657714844,
"learning_rate": 9.974917906131283e-06,
"loss": 0.3558,
"step": 74
},
{
"epoch": 0.32310177705977383,
"grad_norm": 1.3320142030715942,
"learning_rate": 9.974236018040476e-06,
"loss": 0.3619,
"step": 75
},
{
"epoch": 0.32740980075390413,
"grad_norm": 1.3935425281524658,
"learning_rate": 9.973545008789182e-06,
"loss": 0.3759,
"step": 76
},
{
"epoch": 0.33171782444803444,
"grad_norm": 1.359113335609436,
"learning_rate": 9.972844879644494e-06,
"loss": 0.3532,
"step": 77
},
{
"epoch": 0.3360258481421648,
"grad_norm": 1.4055691957473755,
"learning_rate": 9.972135631890226e-06,
"loss": 0.366,
"step": 78
},
{
"epoch": 0.3403338718362951,
"grad_norm": 1.51666259765625,
"learning_rate": 9.97141726682691e-06,
"loss": 0.3551,
"step": 79
},
{
"epoch": 0.3446418955304254,
"grad_norm": 1.3641939163208008,
"learning_rate": 9.970689785771798e-06,
"loss": 0.3552,
"step": 80
},
{
"epoch": 0.34894991922455576,
"grad_norm": 1.476531744003296,
"learning_rate": 9.969953190058861e-06,
"loss": 0.3918,
"step": 81
},
{
"epoch": 0.35325794291868606,
"grad_norm": 1.4188109636306763,
"learning_rate": 9.969207481038776e-06,
"loss": 0.3829,
"step": 82
},
{
"epoch": 0.35756596661281637,
"grad_norm": 1.3417792320251465,
"learning_rate": 9.968452660078939e-06,
"loss": 0.355,
"step": 83
},
{
"epoch": 0.36187399030694667,
"grad_norm": 1.4782617092132568,
"learning_rate": 9.967688728563446e-06,
"loss": 0.3721,
"step": 84
},
{
"epoch": 0.366182014001077,
"grad_norm": 1.6000183820724487,
"learning_rate": 9.966915687893109e-06,
"loss": 0.3935,
"step": 85
},
{
"epoch": 0.37049003769520733,
"grad_norm": 1.4256422519683838,
"learning_rate": 9.966133539485435e-06,
"loss": 0.3697,
"step": 86
},
{
"epoch": 0.37479806138933763,
"grad_norm": 1.3764766454696655,
"learning_rate": 9.965342284774633e-06,
"loss": 0.3521,
"step": 87
},
{
"epoch": 0.37910608508346794,
"grad_norm": 1.4846636056900024,
"learning_rate": 9.964541925211613e-06,
"loss": 0.3394,
"step": 88
},
{
"epoch": 0.3834141087775983,
"grad_norm": 1.4550178050994873,
"learning_rate": 9.963732462263979e-06,
"loss": 0.3502,
"step": 89
},
{
"epoch": 0.3877221324717286,
"grad_norm": 1.374940037727356,
"learning_rate": 9.962913897416029e-06,
"loss": 0.3333,
"step": 90
},
{
"epoch": 0.3920301561658589,
"grad_norm": 1.384493350982666,
"learning_rate": 9.962086232168747e-06,
"loss": 0.3291,
"step": 91
},
{
"epoch": 0.3963381798599892,
"grad_norm": 1.4366718530654907,
"learning_rate": 9.961249468039806e-06,
"loss": 0.3322,
"step": 92
},
{
"epoch": 0.40064620355411956,
"grad_norm": 1.5821336507797241,
"learning_rate": 9.960403606563568e-06,
"loss": 0.346,
"step": 93
},
{
"epoch": 0.40495422724824987,
"grad_norm": 1.5977967977523804,
"learning_rate": 9.959548649291071e-06,
"loss": 0.372,
"step": 94
},
{
"epoch": 0.40926225094238017,
"grad_norm": 1.7642663717269897,
"learning_rate": 9.958684597790031e-06,
"loss": 0.3854,
"step": 95
},
{
"epoch": 0.4135702746365105,
"grad_norm": 1.5853767395019531,
"learning_rate": 9.957811453644848e-06,
"loss": 0.354,
"step": 96
},
{
"epoch": 0.41787829833064083,
"grad_norm": 1.445274829864502,
"learning_rate": 9.956929218456586e-06,
"loss": 0.3402,
"step": 97
},
{
"epoch": 0.42218632202477113,
"grad_norm": 1.5509393215179443,
"learning_rate": 9.956037893842982e-06,
"loss": 0.3361,
"step": 98
},
{
"epoch": 0.42649434571890144,
"grad_norm": 1.512851595878601,
"learning_rate": 9.955137481438442e-06,
"loss": 0.3451,
"step": 99
},
{
"epoch": 0.4308023694130318,
"grad_norm": 1.4434539079666138,
"learning_rate": 9.954227982894034e-06,
"loss": 0.3113,
"step": 100
},
{
"epoch": 0.4351103931071621,
"grad_norm": 1.4094306230545044,
"learning_rate": 9.953309399877491e-06,
"loss": 0.3417,
"step": 101
},
{
"epoch": 0.4394184168012924,
"grad_norm": 1.3972957134246826,
"learning_rate": 9.952381734073197e-06,
"loss": 0.3258,
"step": 102
},
{
"epoch": 0.4437264404954227,
"grad_norm": 1.4385653734207153,
"learning_rate": 9.951444987182195e-06,
"loss": 0.3299,
"step": 103
},
{
"epoch": 0.44803446418955306,
"grad_norm": 1.4804537296295166,
"learning_rate": 9.950499160922184e-06,
"loss": 0.3011,
"step": 104
},
{
"epoch": 0.45234248788368336,
"grad_norm": 1.501686453819275,
"learning_rate": 9.949544257027503e-06,
"loss": 0.3309,
"step": 105
},
{
"epoch": 0.45665051157781367,
"grad_norm": 1.44529128074646,
"learning_rate": 9.948580277249142e-06,
"loss": 0.3138,
"step": 106
},
{
"epoch": 0.46095853527194397,
"grad_norm": 1.5322788953781128,
"learning_rate": 9.947607223354731e-06,
"loss": 0.3349,
"step": 107
},
{
"epoch": 0.46526655896607433,
"grad_norm": 1.4645901918411255,
"learning_rate": 9.946625097128544e-06,
"loss": 0.3242,
"step": 108
},
{
"epoch": 0.46957458266020463,
"grad_norm": 1.4635151624679565,
"learning_rate": 9.945633900371483e-06,
"loss": 0.3153,
"step": 109
},
{
"epoch": 0.47388260635433493,
"grad_norm": 1.5546009540557861,
"learning_rate": 9.94463363490109e-06,
"loss": 0.3121,
"step": 110
},
{
"epoch": 0.4781906300484653,
"grad_norm": 1.4418449401855469,
"learning_rate": 9.943624302551527e-06,
"loss": 0.2974,
"step": 111
},
{
"epoch": 0.4824986537425956,
"grad_norm": 1.519238829612732,
"learning_rate": 9.942605905173593e-06,
"loss": 0.3301,
"step": 112
},
{
"epoch": 0.4868066774367259,
"grad_norm": 1.6555852890014648,
"learning_rate": 9.941578444634699e-06,
"loss": 0.2964,
"step": 113
},
{
"epoch": 0.4911147011308562,
"grad_norm": 1.690470814704895,
"learning_rate": 9.940541922818882e-06,
"loss": 0.3006,
"step": 114
},
{
"epoch": 0.49542272482498656,
"grad_norm": 1.4557969570159912,
"learning_rate": 9.939496341626791e-06,
"loss": 0.3061,
"step": 115
},
{
"epoch": 0.49973074851911686,
"grad_norm": 1.5956724882125854,
"learning_rate": 9.938441702975689e-06,
"loss": 0.3177,
"step": 116
},
{
"epoch": 0.5040387722132472,
"grad_norm": 1.742788314819336,
"learning_rate": 9.937378008799448e-06,
"loss": 0.2813,
"step": 117
},
{
"epoch": 0.5083467959073775,
"grad_norm": 1.4342483282089233,
"learning_rate": 9.93630526104854e-06,
"loss": 0.284,
"step": 118
},
{
"epoch": 0.5126548196015078,
"grad_norm": 1.427372932434082,
"learning_rate": 9.935223461690042e-06,
"loss": 0.3,
"step": 119
},
{
"epoch": 0.5169628432956381,
"grad_norm": 1.4646867513656616,
"learning_rate": 9.934132612707631e-06,
"loss": 0.2841,
"step": 120
},
{
"epoch": 0.5212708669897684,
"grad_norm": 1.6393767595291138,
"learning_rate": 9.933032716101576e-06,
"loss": 0.3194,
"step": 121
},
{
"epoch": 0.5255788906838987,
"grad_norm": 1.4437857866287231,
"learning_rate": 9.931923773888734e-06,
"loss": 0.3089,
"step": 122
},
{
"epoch": 0.529886914378029,
"grad_norm": 1.4526920318603516,
"learning_rate": 9.930805788102551e-06,
"loss": 0.3044,
"step": 123
},
{
"epoch": 0.5341949380721593,
"grad_norm": 1.5132489204406738,
"learning_rate": 9.929678760793057e-06,
"loss": 0.265,
"step": 124
},
{
"epoch": 0.5385029617662898,
"grad_norm": 1.6274410486221313,
"learning_rate": 9.928542694026862e-06,
"loss": 0.3252,
"step": 125
},
{
"epoch": 0.5428109854604201,
"grad_norm": 1.4639244079589844,
"learning_rate": 9.927397589887144e-06,
"loss": 0.288,
"step": 126
},
{
"epoch": 0.5471190091545504,
"grad_norm": 1.4215894937515259,
"learning_rate": 9.926243450473664e-06,
"loss": 0.2738,
"step": 127
},
{
"epoch": 0.5514270328486807,
"grad_norm": 1.3172849416732788,
"learning_rate": 9.925080277902743e-06,
"loss": 0.2847,
"step": 128
},
{
"epoch": 0.555735056542811,
"grad_norm": 1.414466381072998,
"learning_rate": 9.923908074307267e-06,
"loss": 0.3024,
"step": 129
},
{
"epoch": 0.5600430802369413,
"grad_norm": 1.322507619857788,
"learning_rate": 9.922726841836685e-06,
"loss": 0.2826,
"step": 130
},
{
"epoch": 0.5643511039310716,
"grad_norm": 1.3307005167007446,
"learning_rate": 9.921536582657002e-06,
"loss": 0.2865,
"step": 131
},
{
"epoch": 0.568659127625202,
"grad_norm": 1.303130865097046,
"learning_rate": 9.920337298950767e-06,
"loss": 0.275,
"step": 132
},
{
"epoch": 0.5729671513193323,
"grad_norm": 1.3872160911560059,
"learning_rate": 9.919128992917086e-06,
"loss": 0.3168,
"step": 133
},
{
"epoch": 0.5772751750134626,
"grad_norm": 1.3143068552017212,
"learning_rate": 9.91791166677161e-06,
"loss": 0.2831,
"step": 134
},
{
"epoch": 0.5815831987075929,
"grad_norm": 1.4942936897277832,
"learning_rate": 9.916685322746524e-06,
"loss": 0.339,
"step": 135
},
{
"epoch": 0.5858912224017232,
"grad_norm": 1.243410587310791,
"learning_rate": 9.915449963090551e-06,
"loss": 0.2798,
"step": 136
},
{
"epoch": 0.5901992460958535,
"grad_norm": 1.2295470237731934,
"learning_rate": 9.914205590068946e-06,
"loss": 0.275,
"step": 137
},
{
"epoch": 0.5945072697899838,
"grad_norm": 1.251297950744629,
"learning_rate": 9.912952205963491e-06,
"loss": 0.2714,
"step": 138
},
{
"epoch": 0.5988152934841141,
"grad_norm": 1.2663097381591797,
"learning_rate": 9.911689813072495e-06,
"loss": 0.2733,
"step": 139
},
{
"epoch": 0.6031233171782445,
"grad_norm": 1.2405517101287842,
"learning_rate": 9.91041841371078e-06,
"loss": 0.2659,
"step": 140
},
{
"epoch": 0.6074313408723748,
"grad_norm": 1.178120732307434,
"learning_rate": 9.90913801020969e-06,
"loss": 0.264,
"step": 141
},
{
"epoch": 0.6117393645665051,
"grad_norm": 1.292300820350647,
"learning_rate": 9.907848604917075e-06,
"loss": 0.2857,
"step": 142
},
{
"epoch": 0.6160473882606354,
"grad_norm": 1.1422131061553955,
"learning_rate": 9.906550200197288e-06,
"loss": 0.2547,
"step": 143
},
{
"epoch": 0.6203554119547657,
"grad_norm": 1.1891106367111206,
"learning_rate": 9.905242798431196e-06,
"loss": 0.2644,
"step": 144
},
{
"epoch": 0.624663435648896,
"grad_norm": 1.1191891431808472,
"learning_rate": 9.903926402016153e-06,
"loss": 0.2602,
"step": 145
},
{
"epoch": 0.6289714593430263,
"grad_norm": 1.260938286781311,
"learning_rate": 9.902601013366009e-06,
"loss": 0.2954,
"step": 146
},
{
"epoch": 0.6332794830371568,
"grad_norm": 1.033594012260437,
"learning_rate": 9.901266634911104e-06,
"loss": 0.2624,
"step": 147
},
{
"epoch": 0.6375875067312871,
"grad_norm": 1.0682276487350464,
"learning_rate": 9.899923269098262e-06,
"loss": 0.2524,
"step": 148
},
{
"epoch": 0.6418955304254174,
"grad_norm": 1.1801897287368774,
"learning_rate": 9.898570918390789e-06,
"loss": 0.2599,
"step": 149
},
{
"epoch": 0.6462035541195477,
"grad_norm": 0.9911927580833435,
"learning_rate": 9.897209585268459e-06,
"loss": 0.2388,
"step": 150
},
{
"epoch": 0.650511577813678,
"grad_norm": 1.055557131767273,
"learning_rate": 9.895839272227529e-06,
"loss": 0.2476,
"step": 151
},
{
"epoch": 0.6548196015078083,
"grad_norm": 1.1163235902786255,
"learning_rate": 9.894459981780711e-06,
"loss": 0.2506,
"step": 152
},
{
"epoch": 0.6591276252019386,
"grad_norm": 1.1236166954040527,
"learning_rate": 9.893071716457183e-06,
"loss": 0.2765,
"step": 153
},
{
"epoch": 0.6634356488960689,
"grad_norm": 0.9692741632461548,
"learning_rate": 9.891674478802585e-06,
"loss": 0.237,
"step": 154
},
{
"epoch": 0.6677436725901993,
"grad_norm": 1.79551362991333,
"learning_rate": 9.890268271379e-06,
"loss": 0.2653,
"step": 155
},
{
"epoch": 0.6720516962843296,
"grad_norm": 1.0855523347854614,
"learning_rate": 9.888853096764963e-06,
"loss": 0.2556,
"step": 156
},
{
"epoch": 0.6763597199784599,
"grad_norm": 1.1746084690093994,
"learning_rate": 9.887428957555457e-06,
"loss": 0.2717,
"step": 157
},
{
"epoch": 0.6806677436725902,
"grad_norm": 1.03372061252594,
"learning_rate": 9.885995856361895e-06,
"loss": 0.2558,
"step": 158
},
{
"epoch": 0.6849757673667205,
"grad_norm": 0.9783212542533875,
"learning_rate": 9.884553795812128e-06,
"loss": 0.2484,
"step": 159
},
{
"epoch": 0.6892837910608508,
"grad_norm": 0.8822944760322571,
"learning_rate": 9.883102778550434e-06,
"loss": 0.2471,
"step": 160
},
{
"epoch": 0.6935918147549811,
"grad_norm": 1.0045849084854126,
"learning_rate": 9.881642807237515e-06,
"loss": 0.2597,
"step": 161
},
{
"epoch": 0.6978998384491115,
"grad_norm": 0.912653923034668,
"learning_rate": 9.880173884550495e-06,
"loss": 0.2428,
"step": 162
},
{
"epoch": 0.7022078621432418,
"grad_norm": 1.0020110607147217,
"learning_rate": 9.878696013182906e-06,
"loss": 0.2714,
"step": 163
},
{
"epoch": 0.7065158858373721,
"grad_norm": 0.8890970945358276,
"learning_rate": 9.877209195844692e-06,
"loss": 0.2346,
"step": 164
},
{
"epoch": 0.7108239095315024,
"grad_norm": 1.1626794338226318,
"learning_rate": 9.875713435262205e-06,
"loss": 0.2839,
"step": 165
},
{
"epoch": 0.7151319332256327,
"grad_norm": 0.8313829898834229,
"learning_rate": 9.874208734178187e-06,
"loss": 0.2376,
"step": 166
},
{
"epoch": 0.719439956919763,
"grad_norm": 0.9518410563468933,
"learning_rate": 9.872695095351784e-06,
"loss": 0.2536,
"step": 167
},
{
"epoch": 0.7237479806138933,
"grad_norm": 0.9485108852386475,
"learning_rate": 9.871172521558523e-06,
"loss": 0.2592,
"step": 168
},
{
"epoch": 0.7280560043080236,
"grad_norm": 0.9191961884498596,
"learning_rate": 9.869641015590319e-06,
"loss": 0.2466,
"step": 169
},
{
"epoch": 0.732364028002154,
"grad_norm": 0.9200321435928345,
"learning_rate": 9.868100580255466e-06,
"loss": 0.2453,
"step": 170
},
{
"epoch": 0.7366720516962844,
"grad_norm": 0.9513406753540039,
"learning_rate": 9.86655121837863e-06,
"loss": 0.2628,
"step": 171
},
{
"epoch": 0.7409800753904147,
"grad_norm": 0.8816955089569092,
"learning_rate": 9.864992932800845e-06,
"loss": 0.2514,
"step": 172
},
{
"epoch": 0.745288099084545,
"grad_norm": 0.9091234803199768,
"learning_rate": 9.863425726379512e-06,
"loss": 0.2377,
"step": 173
},
{
"epoch": 0.7495961227786753,
"grad_norm": 1.138776183128357,
"learning_rate": 9.861849601988384e-06,
"loss": 0.2772,
"step": 174
},
{
"epoch": 0.7539041464728056,
"grad_norm": 0.9423293471336365,
"learning_rate": 9.860264562517571e-06,
"loss": 0.2369,
"step": 175
},
{
"epoch": 0.7582121701669359,
"grad_norm": 0.8946828842163086,
"learning_rate": 9.858670610873528e-06,
"loss": 0.2485,
"step": 176
},
{
"epoch": 0.7625201938610663,
"grad_norm": 1.0042487382888794,
"learning_rate": 9.857067749979057e-06,
"loss": 0.2707,
"step": 177
},
{
"epoch": 0.7668282175551966,
"grad_norm": 0.9007334113121033,
"learning_rate": 9.855455982773288e-06,
"loss": 0.2721,
"step": 178
},
{
"epoch": 0.7711362412493269,
"grad_norm": 0.9035285115242004,
"learning_rate": 9.853835312211692e-06,
"loss": 0.2471,
"step": 179
},
{
"epoch": 0.7754442649434572,
"grad_norm": 1.0299876928329468,
"learning_rate": 9.852205741266058e-06,
"loss": 0.2767,
"step": 180
},
{
"epoch": 0.7797522886375875,
"grad_norm": 1.1170578002929688,
"learning_rate": 9.8505672729245e-06,
"loss": 0.2659,
"step": 181
},
{
"epoch": 0.7840603123317178,
"grad_norm": 0.9261436462402344,
"learning_rate": 9.848919910191446e-06,
"loss": 0.2569,
"step": 182
},
{
"epoch": 0.7883683360258481,
"grad_norm": 0.9240891933441162,
"learning_rate": 9.847263656087633e-06,
"loss": 0.2384,
"step": 183
},
{
"epoch": 0.7926763597199784,
"grad_norm": 1.0352424383163452,
"learning_rate": 9.845598513650104e-06,
"loss": 0.2422,
"step": 184
},
{
"epoch": 0.7969843834141088,
"grad_norm": 0.9237746596336365,
"learning_rate": 9.843924485932195e-06,
"loss": 0.2417,
"step": 185
},
{
"epoch": 0.8012924071082391,
"grad_norm": 0.9721134901046753,
"learning_rate": 9.84224157600354e-06,
"loss": 0.2504,
"step": 186
},
{
"epoch": 0.8056004308023694,
"grad_norm": 1.0062624216079712,
"learning_rate": 9.840549786950058e-06,
"loss": 0.2513,
"step": 187
},
{
"epoch": 0.8099084544964997,
"grad_norm": 0.9284637570381165,
"learning_rate": 9.83884912187395e-06,
"loss": 0.2466,
"step": 188
},
{
"epoch": 0.81421647819063,
"grad_norm": 0.9498764872550964,
"learning_rate": 9.837139583893693e-06,
"loss": 0.2568,
"step": 189
},
{
"epoch": 0.8185245018847603,
"grad_norm": 0.9171614050865173,
"learning_rate": 9.835421176144035e-06,
"loss": 0.2595,
"step": 190
},
{
"epoch": 0.8228325255788906,
"grad_norm": 0.914812445640564,
"learning_rate": 9.833693901775985e-06,
"loss": 0.2518,
"step": 191
},
{
"epoch": 0.827140549273021,
"grad_norm": 1.0692369937896729,
"learning_rate": 9.831957763956814e-06,
"loss": 0.2395,
"step": 192
},
{
"epoch": 0.8314485729671514,
"grad_norm": 0.9172776341438293,
"learning_rate": 9.830212765870043e-06,
"loss": 0.2523,
"step": 193
},
{
"epoch": 0.8357565966612817,
"grad_norm": 1.1387518644332886,
"learning_rate": 9.828458910715442e-06,
"loss": 0.2571,
"step": 194
},
{
"epoch": 0.840064620355412,
"grad_norm": 0.9793438911437988,
"learning_rate": 9.826696201709022e-06,
"loss": 0.2421,
"step": 195
},
{
"epoch": 0.8443726440495423,
"grad_norm": 0.9396490454673767,
"learning_rate": 9.824924642083026e-06,
"loss": 0.2442,
"step": 196
},
{
"epoch": 0.8486806677436726,
"grad_norm": 0.889569103717804,
"learning_rate": 9.823144235085934e-06,
"loss": 0.2389,
"step": 197
},
{
"epoch": 0.8529886914378029,
"grad_norm": 0.931921124458313,
"learning_rate": 9.821354983982438e-06,
"loss": 0.259,
"step": 198
},
{
"epoch": 0.8572967151319332,
"grad_norm": 0.7875272035598755,
"learning_rate": 9.819556892053456e-06,
"loss": 0.2132,
"step": 199
},
{
"epoch": 0.8616047388260636,
"grad_norm": 0.9318849444389343,
"learning_rate": 9.817749962596115e-06,
"loss": 0.2516,
"step": 200
},
{
"epoch": 0.8659127625201939,
"grad_norm": 0.948217511177063,
"learning_rate": 9.815934198923746e-06,
"loss": 0.2577,
"step": 201
},
{
"epoch": 0.8702207862143242,
"grad_norm": 0.8567696213722229,
"learning_rate": 9.814109604365878e-06,
"loss": 0.2431,
"step": 202
},
{
"epoch": 0.8745288099084545,
"grad_norm": 0.8806264996528625,
"learning_rate": 9.812276182268236e-06,
"loss": 0.2342,
"step": 203
},
{
"epoch": 0.8788368336025848,
"grad_norm": 0.968190610408783,
"learning_rate": 9.810433935992734e-06,
"loss": 0.2636,
"step": 204
},
{
"epoch": 0.8831448572967151,
"grad_norm": 0.9591284394264221,
"learning_rate": 9.808582868917458e-06,
"loss": 0.2364,
"step": 205
},
{
"epoch": 0.8874528809908454,
"grad_norm": 0.930848240852356,
"learning_rate": 9.806722984436676e-06,
"loss": 0.2372,
"step": 206
},
{
"epoch": 0.8917609046849758,
"grad_norm": 0.9203491806983948,
"learning_rate": 9.804854285960823e-06,
"loss": 0.2441,
"step": 207
},
{
"epoch": 0.8960689283791061,
"grad_norm": 0.8258495926856995,
"learning_rate": 9.802976776916493e-06,
"loss": 0.2254,
"step": 208
},
{
"epoch": 0.9003769520732364,
"grad_norm": 0.936922550201416,
"learning_rate": 9.801090460746442e-06,
"loss": 0.2519,
"step": 209
},
{
"epoch": 0.9046849757673667,
"grad_norm": 0.8387218117713928,
"learning_rate": 9.799195340909569e-06,
"loss": 0.2433,
"step": 210
},
{
"epoch": 0.908992999461497,
"grad_norm": 0.9370262026786804,
"learning_rate": 9.79729142088092e-06,
"loss": 0.2384,
"step": 211
},
{
"epoch": 0.9133010231556273,
"grad_norm": 0.8370246291160583,
"learning_rate": 9.795378704151675e-06,
"loss": 0.2554,
"step": 212
},
{
"epoch": 0.9176090468497576,
"grad_norm": 0.8435834050178528,
"learning_rate": 9.793457194229145e-06,
"loss": 0.2372,
"step": 213
},
{
"epoch": 0.9219170705438879,
"grad_norm": 1.1667269468307495,
"learning_rate": 9.791526894636767e-06,
"loss": 0.2446,
"step": 214
},
{
"epoch": 0.9262250942380184,
"grad_norm": 0.9069737792015076,
"learning_rate": 9.789587808914094e-06,
"loss": 0.2571,
"step": 215
},
{
"epoch": 0.9305331179321487,
"grad_norm": 0.844599187374115,
"learning_rate": 9.787639940616789e-06,
"loss": 0.2547,
"step": 216
},
{
"epoch": 0.934841141626279,
"grad_norm": 0.8763352632522583,
"learning_rate": 9.785683293316622e-06,
"loss": 0.2562,
"step": 217
},
{
"epoch": 0.9391491653204093,
"grad_norm": 0.7414149045944214,
"learning_rate": 9.783717870601458e-06,
"loss": 0.2158,
"step": 218
},
{
"epoch": 0.9434571890145396,
"grad_norm": 0.8541145324707031,
"learning_rate": 9.781743676075257e-06,
"loss": 0.2425,
"step": 219
},
{
"epoch": 0.9477652127086699,
"grad_norm": 0.869875431060791,
"learning_rate": 9.77976071335806e-06,
"loss": 0.2241,
"step": 220
},
{
"epoch": 0.9520732364028002,
"grad_norm": 0.8272488713264465,
"learning_rate": 9.777768986085985e-06,
"loss": 0.2454,
"step": 221
},
{
"epoch": 0.9563812600969306,
"grad_norm": 0.8880266547203064,
"learning_rate": 9.775768497911226e-06,
"loss": 0.2531,
"step": 222
},
{
"epoch": 0.9606892837910609,
"grad_norm": 0.9570372700691223,
"learning_rate": 9.77375925250204e-06,
"loss": 0.2691,
"step": 223
},
{
"epoch": 0.9649973074851912,
"grad_norm": 0.8235428333282471,
"learning_rate": 9.771741253542742e-06,
"loss": 0.257,
"step": 224
},
{
"epoch": 0.9693053311793215,
"grad_norm": 1.0435497760772705,
"learning_rate": 9.769714504733695e-06,
"loss": 0.2724,
"step": 225
},
{
"epoch": 0.9736133548734518,
"grad_norm": 0.8413583040237427,
"learning_rate": 9.767679009791312e-06,
"loss": 0.2519,
"step": 226
},
{
"epoch": 0.9779213785675821,
"grad_norm": 0.8566842079162598,
"learning_rate": 9.765634772448039e-06,
"loss": 0.24,
"step": 227
},
{
"epoch": 0.9822294022617124,
"grad_norm": 0.9400589466094971,
"learning_rate": 9.763581796452353e-06,
"loss": 0.2264,
"step": 228
},
{
"epoch": 0.9865374259558427,
"grad_norm": 0.7903165221214294,
"learning_rate": 9.76152008556876e-06,
"loss": 0.2555,
"step": 229
},
{
"epoch": 0.9908454496499731,
"grad_norm": 0.9579477310180664,
"learning_rate": 9.759449643577779e-06,
"loss": 0.2428,
"step": 230
},
{
"epoch": 0.9951534733441034,
"grad_norm": 0.8865399360656738,
"learning_rate": 9.757370474275938e-06,
"loss": 0.2425,
"step": 231
},
{
"epoch": 0.9994614970382337,
"grad_norm": 0.8118222951889038,
"learning_rate": 9.755282581475769e-06,
"loss": 0.2327,
"step": 232
},
{
"epoch": 1.0,
"grad_norm": 0.8118222951889038,
"learning_rate": 9.753185969005802e-06,
"loss": 0.3039,
"step": 233
},
{
"epoch": 1.0043080236941304,
"grad_norm": 3.10906982421875,
"learning_rate": 9.751080640710554e-06,
"loss": 0.2138,
"step": 234
},
{
"epoch": 1.0086160473882606,
"grad_norm": 0.8318678140640259,
"learning_rate": 9.748966600450526e-06,
"loss": 0.2287,
"step": 235
},
{
"epoch": 1.012924071082391,
"grad_norm": 0.8306625485420227,
"learning_rate": 9.746843852102191e-06,
"loss": 0.2144,
"step": 236
},
{
"epoch": 1.0172320947765212,
"grad_norm": 0.7923266291618347,
"learning_rate": 9.744712399557992e-06,
"loss": 0.2211,
"step": 237
},
{
"epoch": 1.0215401184706516,
"grad_norm": 0.9020610451698303,
"learning_rate": 9.742572246726336e-06,
"loss": 0.2431,
"step": 238
},
{
"epoch": 1.0258481421647818,
"grad_norm": 0.6959359645843506,
"learning_rate": 9.740423397531573e-06,
"loss": 0.2142,
"step": 239
},
{
"epoch": 1.0301561658589122,
"grad_norm": 0.725628137588501,
"learning_rate": 9.738265855914014e-06,
"loss": 0.2194,
"step": 240
},
{
"epoch": 1.0344641895530426,
"grad_norm": 0.8057601451873779,
"learning_rate": 9.736099625829894e-06,
"loss": 0.2358,
"step": 241
},
{
"epoch": 1.0387722132471728,
"grad_norm": 0.8351283669471741,
"learning_rate": 9.733924711251393e-06,
"loss": 0.2412,
"step": 242
},
{
"epoch": 1.0430802369413033,
"grad_norm": 0.8563336730003357,
"learning_rate": 9.731741116166607e-06,
"loss": 0.2138,
"step": 243
},
{
"epoch": 1.0473882606354334,
"grad_norm": 0.8097681999206543,
"learning_rate": 9.729548844579552e-06,
"loss": 0.2256,
"step": 244
},
{
"epoch": 1.0516962843295639,
"grad_norm": 0.7601490616798401,
"learning_rate": 9.727347900510155e-06,
"loss": 0.208,
"step": 245
},
{
"epoch": 1.056004308023694,
"grad_norm": 0.7128692269325256,
"learning_rate": 9.725138287994246e-06,
"loss": 0.2121,
"step": 246
},
{
"epoch": 1.0603123317178245,
"grad_norm": 0.7977615594863892,
"learning_rate": 9.722920011083546e-06,
"loss": 0.2319,
"step": 247
},
{
"epoch": 1.0646203554119547,
"grad_norm": 0.874777615070343,
"learning_rate": 9.720693073845668e-06,
"loss": 0.2166,
"step": 248
},
{
"epoch": 1.068928379106085,
"grad_norm": 0.8400263786315918,
"learning_rate": 9.718457480364103e-06,
"loss": 0.2517,
"step": 249
},
{
"epoch": 1.0732364028002155,
"grad_norm": 0.8173924684524536,
"learning_rate": 9.716213234738216e-06,
"loss": 0.2227,
"step": 250
},
{
"epoch": 1.0775444264943457,
"grad_norm": 0.7754281759262085,
"learning_rate": 9.713960341083237e-06,
"loss": 0.1943,
"step": 251
},
{
"epoch": 1.081852450188476,
"grad_norm": 0.8374117612838745,
"learning_rate": 9.711698803530253e-06,
"loss": 0.237,
"step": 252
},
{
"epoch": 1.0861604738826063,
"grad_norm": 0.7473996877670288,
"learning_rate": 9.709428626226204e-06,
"loss": 0.208,
"step": 253
},
{
"epoch": 1.0904684975767367,
"grad_norm": 0.7273510694503784,
"learning_rate": 9.707149813333866e-06,
"loss": 0.2177,
"step": 254
},
{
"epoch": 1.0947765212708669,
"grad_norm": 0.8145310878753662,
"learning_rate": 9.704862369031857e-06,
"loss": 0.24,
"step": 255
},
{
"epoch": 1.0990845449649973,
"grad_norm": 0.8657172322273254,
"learning_rate": 9.70256629751462e-06,
"loss": 0.2252,
"step": 256
},
{
"epoch": 1.1033925686591277,
"grad_norm": 0.9188759922981262,
"learning_rate": 9.700261602992417e-06,
"loss": 0.241,
"step": 257
},
{
"epoch": 1.107700592353258,
"grad_norm": 0.7244163751602173,
"learning_rate": 9.69794828969132e-06,
"loss": 0.2139,
"step": 258
},
{
"epoch": 1.1120086160473883,
"grad_norm": 0.9003731608390808,
"learning_rate": 9.695626361853207e-06,
"loss": 0.2256,
"step": 259
},
{
"epoch": 1.1163166397415185,
"grad_norm": 0.7939245700836182,
"learning_rate": 9.693295823735754e-06,
"loss": 0.2345,
"step": 260
},
{
"epoch": 1.120624663435649,
"grad_norm": 0.7665999531745911,
"learning_rate": 9.690956679612422e-06,
"loss": 0.2204,
"step": 261
},
{
"epoch": 1.1249326871297791,
"grad_norm": 0.758368194103241,
"learning_rate": 9.688608933772454e-06,
"loss": 0.2057,
"step": 262
},
{
"epoch": 1.1292407108239095,
"grad_norm": 0.8348474502563477,
"learning_rate": 9.686252590520869e-06,
"loss": 0.2158,
"step": 263
},
{
"epoch": 1.13354873451804,
"grad_norm": 0.7352973818778992,
"learning_rate": 9.683887654178446e-06,
"loss": 0.2097,
"step": 264
},
{
"epoch": 1.1378567582121701,
"grad_norm": 0.8477746844291687,
"learning_rate": 9.681514129081725e-06,
"loss": 0.2288,
"step": 265
},
{
"epoch": 1.1421647819063006,
"grad_norm": 0.7330527305603027,
"learning_rate": 9.679132019582988e-06,
"loss": 0.2179,
"step": 266
},
{
"epoch": 1.1464728056004307,
"grad_norm": 0.7249947190284729,
"learning_rate": 9.67674133005027e-06,
"loss": 0.2062,
"step": 267
},
{
"epoch": 1.1507808292945612,
"grad_norm": 0.8508214354515076,
"learning_rate": 9.674342064867326e-06,
"loss": 0.2142,
"step": 268
},
{
"epoch": 1.1550888529886914,
"grad_norm": 0.9992220401763916,
"learning_rate": 9.671934228433647e-06,
"loss": 0.2201,
"step": 269
},
{
"epoch": 1.1593968766828218,
"grad_norm": 0.722809910774231,
"learning_rate": 9.669517825164435e-06,
"loss": 0.215,
"step": 270
},
{
"epoch": 1.1637049003769522,
"grad_norm": 0.770807683467865,
"learning_rate": 9.667092859490599e-06,
"loss": 0.2112,
"step": 271
},
{
"epoch": 1.1680129240710824,
"grad_norm": 0.7955392003059387,
"learning_rate": 9.664659335858755e-06,
"loss": 0.211,
"step": 272
},
{
"epoch": 1.1723209477652128,
"grad_norm": 0.7562293410301208,
"learning_rate": 9.662217258731208e-06,
"loss": 0.2166,
"step": 273
},
{
"epoch": 1.176628971459343,
"grad_norm": 0.7826163172721863,
"learning_rate": 9.659766632585946e-06,
"loss": 0.2333,
"step": 274
},
{
"epoch": 1.1809369951534734,
"grad_norm": 0.7867729663848877,
"learning_rate": 9.657307461916637e-06,
"loss": 0.2181,
"step": 275
},
{
"epoch": 1.1852450188476036,
"grad_norm": 0.7663466334342957,
"learning_rate": 9.654839751232612e-06,
"loss": 0.2177,
"step": 276
},
{
"epoch": 1.189553042541734,
"grad_norm": 0.9194732308387756,
"learning_rate": 9.652363505058866e-06,
"loss": 0.2374,
"step": 277
},
{
"epoch": 1.1938610662358644,
"grad_norm": 0.8086763620376587,
"learning_rate": 9.649878727936044e-06,
"loss": 0.225,
"step": 278
},
{
"epoch": 1.1981690899299946,
"grad_norm": 0.8902127146720886,
"learning_rate": 9.647385424420435e-06,
"loss": 0.2024,
"step": 279
},
{
"epoch": 1.202477113624125,
"grad_norm": 0.8926809430122375,
"learning_rate": 9.644883599083959e-06,
"loss": 0.2201,
"step": 280
},
{
"epoch": 1.2067851373182552,
"grad_norm": 0.8426185846328735,
"learning_rate": 9.642373256514164e-06,
"loss": 0.2371,
"step": 281
},
{
"epoch": 1.2110931610123856,
"grad_norm": 0.7640630006790161,
"learning_rate": 9.639854401314219e-06,
"loss": 0.2148,
"step": 282
},
{
"epoch": 1.2154011847065158,
"grad_norm": 0.9093323349952698,
"learning_rate": 9.637327038102902e-06,
"loss": 0.2218,
"step": 283
},
{
"epoch": 1.2197092084006462,
"grad_norm": 0.8968831300735474,
"learning_rate": 9.634791171514585e-06,
"loss": 0.2298,
"step": 284
},
{
"epoch": 1.2240172320947766,
"grad_norm": 0.8954552412033081,
"learning_rate": 9.632246806199242e-06,
"loss": 0.2173,
"step": 285
},
{
"epoch": 1.2283252557889068,
"grad_norm": 0.8648595213890076,
"learning_rate": 9.629693946822423e-06,
"loss": 0.2273,
"step": 286
},
{
"epoch": 1.2326332794830372,
"grad_norm": 0.8752464056015015,
"learning_rate": 9.627132598065258e-06,
"loss": 0.2239,
"step": 287
},
{
"epoch": 1.2369413031771674,
"grad_norm": 0.830769419670105,
"learning_rate": 9.624562764624445e-06,
"loss": 0.2295,
"step": 288
},
{
"epoch": 1.2412493268712979,
"grad_norm": 0.8199811577796936,
"learning_rate": 9.621984451212237e-06,
"loss": 0.2256,
"step": 289
},
{
"epoch": 1.245557350565428,
"grad_norm": 0.8059448599815369,
"learning_rate": 9.619397662556434e-06,
"loss": 0.2196,
"step": 290
},
{
"epoch": 1.2498653742595585,
"grad_norm": 0.8064429759979248,
"learning_rate": 9.616802403400384e-06,
"loss": 0.2208,
"step": 291
},
{
"epoch": 1.2541733979536889,
"grad_norm": 0.8114342093467712,
"learning_rate": 9.614198678502965e-06,
"loss": 0.2172,
"step": 292
},
{
"epoch": 1.258481421647819,
"grad_norm": 0.7933779954910278,
"learning_rate": 9.611586492638573e-06,
"loss": 0.2121,
"step": 293
},
{
"epoch": 1.2627894453419493,
"grad_norm": 0.7357353568077087,
"learning_rate": 9.608965850597125e-06,
"loss": 0.2237,
"step": 294
},
{
"epoch": 1.2670974690360797,
"grad_norm": 0.7709231972694397,
"learning_rate": 9.606336757184041e-06,
"loss": 0.22,
"step": 295
},
{
"epoch": 1.27140549273021,
"grad_norm": 0.7920863032341003,
"learning_rate": 9.603699217220239e-06,
"loss": 0.2214,
"step": 296
},
{
"epoch": 1.2757135164243403,
"grad_norm": 0.8977172374725342,
"learning_rate": 9.601053235542124e-06,
"loss": 0.214,
"step": 297
},
{
"epoch": 1.2800215401184707,
"grad_norm": 0.750146210193634,
"learning_rate": 9.598398817001585e-06,
"loss": 0.2127,
"step": 298
},
{
"epoch": 1.284329563812601,
"grad_norm": 0.8150034546852112,
"learning_rate": 9.595735966465973e-06,
"loss": 0.2209,
"step": 299
},
{
"epoch": 1.2886375875067313,
"grad_norm": 0.819951057434082,
"learning_rate": 9.59306468881811e-06,
"loss": 0.2291,
"step": 300
},
{
"epoch": 1.2929456112008615,
"grad_norm": 0.7801006436347961,
"learning_rate": 9.590384988956264e-06,
"loss": 0.215,
"step": 301
},
{
"epoch": 1.297253634894992,
"grad_norm": 0.8777364492416382,
"learning_rate": 9.587696871794148e-06,
"loss": 0.2321,
"step": 302
},
{
"epoch": 1.3015616585891223,
"grad_norm": 0.8642171621322632,
"learning_rate": 9.585000342260914e-06,
"loss": 0.218,
"step": 303
},
{
"epoch": 1.3058696822832525,
"grad_norm": 0.9009878635406494,
"learning_rate": 9.582295405301131e-06,
"loss": 0.2283,
"step": 304
},
{
"epoch": 1.310177705977383,
"grad_norm": 0.9107750058174133,
"learning_rate": 9.579582065874794e-06,
"loss": 0.2297,
"step": 305
},
{
"epoch": 1.3144857296715131,
"grad_norm": 0.8410093784332275,
"learning_rate": 9.576860328957299e-06,
"loss": 0.2275,
"step": 306
},
{
"epoch": 1.3187937533656435,
"grad_norm": 0.8049638271331787,
"learning_rate": 9.574130199539443e-06,
"loss": 0.209,
"step": 307
},
{
"epoch": 1.3231017770597737,
"grad_norm": 0.7018632888793945,
"learning_rate": 9.571391682627413e-06,
"loss": 0.2034,
"step": 308
},
{
"epoch": 1.3274098007539041,
"grad_norm": 0.779147207736969,
"learning_rate": 9.568644783242771e-06,
"loss": 0.225,
"step": 309
},
{
"epoch": 1.3317178244480345,
"grad_norm": 0.725441575050354,
"learning_rate": 9.565889506422457e-06,
"loss": 0.2082,
"step": 310
},
{
"epoch": 1.3360258481421647,
"grad_norm": 0.7808791995048523,
"learning_rate": 9.563125857218766e-06,
"loss": 0.2217,
"step": 311
},
{
"epoch": 1.3403338718362952,
"grad_norm": 0.833756148815155,
"learning_rate": 9.56035384069935e-06,
"loss": 0.2377,
"step": 312
},
{
"epoch": 1.3446418955304253,
"grad_norm": 0.9143878817558289,
"learning_rate": 9.557573461947201e-06,
"loss": 0.2221,
"step": 313
},
{
"epoch": 1.3489499192245558,
"grad_norm": 0.8486077785491943,
"learning_rate": 9.554784726060647e-06,
"loss": 0.2403,
"step": 314
},
{
"epoch": 1.353257942918686,
"grad_norm": 0.7151229381561279,
"learning_rate": 9.551987638153339e-06,
"loss": 0.2175,
"step": 315
},
{
"epoch": 1.3575659666128164,
"grad_norm": 0.8472189903259277,
"learning_rate": 9.549182203354241e-06,
"loss": 0.232,
"step": 316
},
{
"epoch": 1.3618739903069468,
"grad_norm": 0.8481770753860474,
"learning_rate": 9.546368426807628e-06,
"loss": 0.2335,
"step": 317
},
{
"epoch": 1.366182014001077,
"grad_norm": 0.8769595623016357,
"learning_rate": 9.543546313673065e-06,
"loss": 0.226,
"step": 318
},
{
"epoch": 1.3704900376952074,
"grad_norm": 0.7629135251045227,
"learning_rate": 9.540715869125407e-06,
"loss": 0.2204,
"step": 319
},
{
"epoch": 1.3747980613893376,
"grad_norm": 0.7563349008560181,
"learning_rate": 9.537877098354787e-06,
"loss": 0.2123,
"step": 320
},
{
"epoch": 1.379106085083468,
"grad_norm": 0.6988772749900818,
"learning_rate": 9.5350300065666e-06,
"loss": 0.2211,
"step": 321
},
{
"epoch": 1.3834141087775982,
"grad_norm": 0.7229964137077332,
"learning_rate": 9.532174598981507e-06,
"loss": 0.2222,
"step": 322
},
{
"epoch": 1.3877221324717286,
"grad_norm": 0.7834879755973816,
"learning_rate": 9.529310880835414e-06,
"loss": 0.2038,
"step": 323
},
{
"epoch": 1.392030156165859,
"grad_norm": 0.7434644103050232,
"learning_rate": 9.526438857379463e-06,
"loss": 0.231,
"step": 324
},
{
"epoch": 1.3963381798599892,
"grad_norm": 0.9631347060203552,
"learning_rate": 9.52355853388003e-06,
"loss": 0.2262,
"step": 325
},
{
"epoch": 1.4006462035541196,
"grad_norm": 0.6573905348777771,
"learning_rate": 9.520669915618708e-06,
"loss": 0.2094,
"step": 326
},
{
"epoch": 1.4049542272482498,
"grad_norm": 0.8617786765098572,
"learning_rate": 9.5177730078923e-06,
"loss": 0.2167,
"step": 327
},
{
"epoch": 1.4092622509423802,
"grad_norm": 0.7743338942527771,
"learning_rate": 9.514867816012809e-06,
"loss": 0.2188,
"step": 328
},
{
"epoch": 1.4135702746365104,
"grad_norm": 0.8752392530441284,
"learning_rate": 9.511954345307432e-06,
"loss": 0.2245,
"step": 329
},
{
"epoch": 1.4178782983306408,
"grad_norm": 0.8186882734298706,
"learning_rate": 9.509032601118541e-06,
"loss": 0.2185,
"step": 330
},
{
"epoch": 1.4221863220247712,
"grad_norm": 0.781764805316925,
"learning_rate": 9.506102588803683e-06,
"loss": 0.2392,
"step": 331
},
{
"epoch": 1.4264943457189014,
"grad_norm": 0.758326530456543,
"learning_rate": 9.503164313735566e-06,
"loss": 0.2127,
"step": 332
},
{
"epoch": 1.4308023694130318,
"grad_norm": 0.7992167472839355,
"learning_rate": 9.500217781302048e-06,
"loss": 0.2087,
"step": 333
},
{
"epoch": 1.435110393107162,
"grad_norm": 0.764522910118103,
"learning_rate": 9.497262996906126e-06,
"loss": 0.209,
"step": 334
},
{
"epoch": 1.4394184168012925,
"grad_norm": 1.0435947179794312,
"learning_rate": 9.494299965965935e-06,
"loss": 0.2547,
"step": 335
},
{
"epoch": 1.4437264404954226,
"grad_norm": 0.826876163482666,
"learning_rate": 9.491328693914723e-06,
"loss": 0.2252,
"step": 336
},
{
"epoch": 1.448034464189553,
"grad_norm": 0.6937296986579895,
"learning_rate": 9.488349186200858e-06,
"loss": 0.2105,
"step": 337
},
{
"epoch": 1.4523424878836835,
"grad_norm": 0.8464234471321106,
"learning_rate": 9.485361448287804e-06,
"loss": 0.2039,
"step": 338
},
{
"epoch": 1.4566505115778137,
"grad_norm": 0.8548002243041992,
"learning_rate": 9.482365485654118e-06,
"loss": 0.232,
"step": 339
},
{
"epoch": 1.4609585352719439,
"grad_norm": 0.7955244779586792,
"learning_rate": 9.479361303793441e-06,
"loss": 0.217,
"step": 340
},
{
"epoch": 1.4652665589660743,
"grad_norm": 0.7540826797485352,
"learning_rate": 9.476348908214482e-06,
"loss": 0.2287,
"step": 341
},
{
"epoch": 1.4695745826602047,
"grad_norm": 0.8318302631378174,
"learning_rate": 9.47332830444101e-06,
"loss": 0.2193,
"step": 342
},
{
"epoch": 1.4738826063543349,
"grad_norm": 0.8411695957183838,
"learning_rate": 9.470299498011851e-06,
"loss": 0.2377,
"step": 343
},
{
"epoch": 1.4781906300484653,
"grad_norm": 0.837867021560669,
"learning_rate": 9.46726249448087e-06,
"loss": 0.2201,
"step": 344
},
{
"epoch": 1.4824986537425957,
"grad_norm": 0.7983565926551819,
"learning_rate": 9.464217299416956e-06,
"loss": 0.2238,
"step": 345
},
{
"epoch": 1.486806677436726,
"grad_norm": 0.763559103012085,
"learning_rate": 9.46116391840403e-06,
"loss": 0.2173,
"step": 346
},
{
"epoch": 1.491114701130856,
"grad_norm": 0.769709587097168,
"learning_rate": 9.458102357041017e-06,
"loss": 0.2347,
"step": 347
},
{
"epoch": 1.4954227248249865,
"grad_norm": 0.9357376098632812,
"learning_rate": 9.45503262094184e-06,
"loss": 0.2402,
"step": 348
},
{
"epoch": 1.499730748519117,
"grad_norm": 0.729263186454773,
"learning_rate": 9.451954715735416e-06,
"loss": 0.226,
"step": 349
},
{
"epoch": 1.504038772213247,
"grad_norm": 0.84844970703125,
"learning_rate": 9.448868647065644e-06,
"loss": 0.2489,
"step": 350
},
{
"epoch": 1.5083467959073775,
"grad_norm": 0.9604891538619995,
"learning_rate": 9.445774420591382e-06,
"loss": 0.2311,
"step": 351
},
{
"epoch": 1.512654819601508,
"grad_norm": 0.8172181248664856,
"learning_rate": 9.442672041986456e-06,
"loss": 0.2167,
"step": 352
},
{
"epoch": 1.5169628432956381,
"grad_norm": 0.9129267334938049,
"learning_rate": 9.43956151693964e-06,
"loss": 0.2422,
"step": 353
},
{
"epoch": 1.5212708669897683,
"grad_norm": 0.7796788811683655,
"learning_rate": 9.436442851154642e-06,
"loss": 0.2182,
"step": 354
},
{
"epoch": 1.5255788906838987,
"grad_norm": 0.7742928862571716,
"learning_rate": 9.433316050350099e-06,
"loss": 0.2075,
"step": 355
},
{
"epoch": 1.5298869143780292,
"grad_norm": 0.8352739214897156,
"learning_rate": 9.430181120259566e-06,
"loss": 0.2111,
"step": 356
},
{
"epoch": 1.5341949380721593,
"grad_norm": 0.7576451301574707,
"learning_rate": 9.427038066631502e-06,
"loss": 0.2066,
"step": 357
},
{
"epoch": 1.5385029617662898,
"grad_norm": 0.8700200319290161,
"learning_rate": 9.423886895229266e-06,
"loss": 0.2375,
"step": 358
},
{
"epoch": 1.5428109854604202,
"grad_norm": 0.7757928967475891,
"learning_rate": 9.420727611831098e-06,
"loss": 0.2272,
"step": 359
},
{
"epoch": 1.5471190091545504,
"grad_norm": 0.7936514616012573,
"learning_rate": 9.417560222230115e-06,
"loss": 0.2118,
"step": 360
},
{
"epoch": 1.5514270328486806,
"grad_norm": 0.9016469120979309,
"learning_rate": 9.414384732234301e-06,
"loss": 0.229,
"step": 361
},
{
"epoch": 1.555735056542811,
"grad_norm": 0.7308838367462158,
"learning_rate": 9.411201147666486e-06,
"loss": 0.2067,
"step": 362
},
{
"epoch": 1.5600430802369414,
"grad_norm": 0.7090045213699341,
"learning_rate": 9.408009474364353e-06,
"loss": 0.2198,
"step": 363
},
{
"epoch": 1.5643511039310716,
"grad_norm": 0.8225852251052856,
"learning_rate": 9.404809718180408e-06,
"loss": 0.2323,
"step": 364
},
{
"epoch": 1.568659127625202,
"grad_norm": 0.7157912254333496,
"learning_rate": 9.401601884981983e-06,
"loss": 0.2192,
"step": 365
},
{
"epoch": 1.5729671513193324,
"grad_norm": 0.7957601547241211,
"learning_rate": 9.39838598065122e-06,
"loss": 0.2152,
"step": 366
},
{
"epoch": 1.5772751750134626,
"grad_norm": 0.7559109330177307,
"learning_rate": 9.39516201108506e-06,
"loss": 0.242,
"step": 367
},
{
"epoch": 1.5815831987075928,
"grad_norm": 0.8238731026649475,
"learning_rate": 9.391929982195233e-06,
"loss": 0.2238,
"step": 368
},
{
"epoch": 1.5858912224017232,
"grad_norm": 0.8968952298164368,
"learning_rate": 9.38868989990825e-06,
"loss": 0.2228,
"step": 369
},
{
"epoch": 1.5901992460958536,
"grad_norm": 0.724861741065979,
"learning_rate": 9.385441770165385e-06,
"loss": 0.2201,
"step": 370
},
{
"epoch": 1.5945072697899838,
"grad_norm": 0.7094451189041138,
"learning_rate": 9.382185598922674e-06,
"loss": 0.2136,
"step": 371
},
{
"epoch": 1.598815293484114,
"grad_norm": 0.7279621958732605,
"learning_rate": 9.378921392150893e-06,
"loss": 0.2236,
"step": 372
},
{
"epoch": 1.6031233171782446,
"grad_norm": 0.7491332292556763,
"learning_rate": 9.375649155835554e-06,
"loss": 0.197,
"step": 373
},
{
"epoch": 1.6074313408723748,
"grad_norm": 0.796010434627533,
"learning_rate": 9.372368895976896e-06,
"loss": 0.2332,
"step": 374
},
{
"epoch": 1.611739364566505,
"grad_norm": 0.7918517589569092,
"learning_rate": 9.369080618589866e-06,
"loss": 0.2238,
"step": 375
},
{
"epoch": 1.6160473882606354,
"grad_norm": 0.9035218954086304,
"learning_rate": 9.365784329704114e-06,
"loss": 0.2495,
"step": 376
},
{
"epoch": 1.6203554119547658,
"grad_norm": 0.9317935705184937,
"learning_rate": 9.362480035363987e-06,
"loss": 0.2425,
"step": 377
},
{
"epoch": 1.624663435648896,
"grad_norm": 0.7318428158760071,
"learning_rate": 9.3591677416285e-06,
"loss": 0.2084,
"step": 378
},
{
"epoch": 1.6289714593430262,
"grad_norm": 0.798697292804718,
"learning_rate": 9.35584745457134e-06,
"loss": 0.2235,
"step": 379
},
{
"epoch": 1.6332794830371569,
"grad_norm": 0.7385568022727966,
"learning_rate": 9.352519180280862e-06,
"loss": 0.2124,
"step": 380
},
{
"epoch": 1.637587506731287,
"grad_norm": 0.7316162586212158,
"learning_rate": 9.34918292486005e-06,
"loss": 0.2357,
"step": 381
},
{
"epoch": 1.6418955304254172,
"grad_norm": 0.8386530876159668,
"learning_rate": 9.345838694426535e-06,
"loss": 0.2426,
"step": 382
},
{
"epoch": 1.6462035541195477,
"grad_norm": 0.7922302484512329,
"learning_rate": 9.342486495112566e-06,
"loss": 0.234,
"step": 383
},
{
"epoch": 1.650511577813678,
"grad_norm": 0.8359029293060303,
"learning_rate": 9.339126333065008e-06,
"loss": 0.2255,
"step": 384
},
{
"epoch": 1.6548196015078083,
"grad_norm": 0.7093521356582642,
"learning_rate": 9.335758214445323e-06,
"loss": 0.2255,
"step": 385
},
{
"epoch": 1.6591276252019385,
"grad_norm": 0.7251277565956116,
"learning_rate": 9.332382145429568e-06,
"loss": 0.2144,
"step": 386
},
{
"epoch": 1.6634356488960689,
"grad_norm": 0.7854430675506592,
"learning_rate": 9.328998132208373e-06,
"loss": 0.2191,
"step": 387
},
{
"epoch": 1.6677436725901993,
"grad_norm": 0.750257670879364,
"learning_rate": 9.325606180986938e-06,
"loss": 0.216,
"step": 388
},
{
"epoch": 1.6720516962843295,
"grad_norm": 0.7411738038063049,
"learning_rate": 9.32220629798502e-06,
"loss": 0.2179,
"step": 389
},
{
"epoch": 1.67635971997846,
"grad_norm": 0.7300736308097839,
"learning_rate": 9.318798489436917e-06,
"loss": 0.2101,
"step": 390
},
{
"epoch": 1.6806677436725903,
"grad_norm": 0.8182831406593323,
"learning_rate": 9.315382761591463e-06,
"loss": 0.214,
"step": 391
},
{
"epoch": 1.6849757673667205,
"grad_norm": 0.7903637290000916,
"learning_rate": 9.311959120712012e-06,
"loss": 0.2188,
"step": 392
},
{
"epoch": 1.6892837910608507,
"grad_norm": 0.761500358581543,
"learning_rate": 9.308527573076425e-06,
"loss": 0.2278,
"step": 393
},
{
"epoch": 1.693591814754981,
"grad_norm": 0.7488672733306885,
"learning_rate": 9.30508812497707e-06,
"loss": 0.2152,
"step": 394
},
{
"epoch": 1.6978998384491115,
"grad_norm": 0.76832115650177,
"learning_rate": 9.301640782720792e-06,
"loss": 0.2244,
"step": 395
},
{
"epoch": 1.7022078621432417,
"grad_norm": 0.7345932722091675,
"learning_rate": 9.298185552628917e-06,
"loss": 0.2184,
"step": 396
},
{
"epoch": 1.7065158858373721,
"grad_norm": 0.7528600692749023,
"learning_rate": 9.294722441037238e-06,
"loss": 0.2451,
"step": 397
},
{
"epoch": 1.7108239095315025,
"grad_norm": 0.7345549464225769,
"learning_rate": 9.291251454295989e-06,
"loss": 0.2043,
"step": 398
},
{
"epoch": 1.7151319332256327,
"grad_norm": 0.6699172258377075,
"learning_rate": 9.287772598769855e-06,
"loss": 0.2134,
"step": 399
},
{
"epoch": 1.719439956919763,
"grad_norm": 0.7636283040046692,
"learning_rate": 9.284285880837947e-06,
"loss": 0.2148,
"step": 400
},
{
"epoch": 1.7237479806138933,
"grad_norm": 0.9217869639396667,
"learning_rate": 9.28079130689379e-06,
"loss": 0.2317,
"step": 401
},
{
"epoch": 1.7280560043080238,
"grad_norm": 0.8372043967247009,
"learning_rate": 9.277288883345318e-06,
"loss": 0.2244,
"step": 402
},
{
"epoch": 1.732364028002154,
"grad_norm": 0.7365742325782776,
"learning_rate": 9.273778616614857e-06,
"loss": 0.2215,
"step": 403
},
{
"epoch": 1.7366720516962844,
"grad_norm": 0.712045431137085,
"learning_rate": 9.270260513139116e-06,
"loss": 0.2107,
"step": 404
},
{
"epoch": 1.7409800753904148,
"grad_norm": 0.6611554622650146,
"learning_rate": 9.266734579369172e-06,
"loss": 0.222,
"step": 405
},
{
"epoch": 1.745288099084545,
"grad_norm": 0.8067266941070557,
"learning_rate": 9.263200821770462e-06,
"loss": 0.2069,
"step": 406
},
{
"epoch": 1.7495961227786752,
"grad_norm": 0.7734603881835938,
"learning_rate": 9.25965924682277e-06,
"loss": 0.2177,
"step": 407
},
{
"epoch": 1.7539041464728056,
"grad_norm": 0.9521034359931946,
"learning_rate": 9.256109861020213e-06,
"loss": 0.2181,
"step": 408
},
{
"epoch": 1.758212170166936,
"grad_norm": 0.7550384402275085,
"learning_rate": 9.252552670871232e-06,
"loss": 0.2262,
"step": 409
},
{
"epoch": 1.7625201938610662,
"grad_norm": 0.6783749461174011,
"learning_rate": 9.248987682898576e-06,
"loss": 0.2242,
"step": 410
},
{
"epoch": 1.7668282175551966,
"grad_norm": 0.7524251937866211,
"learning_rate": 9.245414903639295e-06,
"loss": 0.2175,
"step": 411
},
{
"epoch": 1.771136241249327,
"grad_norm": 0.7920904755592346,
"learning_rate": 9.241834339644726e-06,
"loss": 0.2431,
"step": 412
},
{
"epoch": 1.7754442649434572,
"grad_norm": 0.7145002484321594,
"learning_rate": 9.23824599748048e-06,
"loss": 0.2123,
"step": 413
},
{
"epoch": 1.7797522886375874,
"grad_norm": 0.8676689863204956,
"learning_rate": 9.234649883726432e-06,
"loss": 0.245,
"step": 414
},
{
"epoch": 1.7840603123317178,
"grad_norm": 0.7974684834480286,
"learning_rate": 9.231046004976704e-06,
"loss": 0.2204,
"step": 415
},
{
"epoch": 1.7883683360258482,
"grad_norm": 0.6935186386108398,
"learning_rate": 9.22743436783966e-06,
"loss": 0.2284,
"step": 416
},
{
"epoch": 1.7926763597199784,
"grad_norm": 0.8282122611999512,
"learning_rate": 9.223814978937888e-06,
"loss": 0.2542,
"step": 417
},
{
"epoch": 1.7969843834141088,
"grad_norm": 0.7227144837379456,
"learning_rate": 9.220187844908194e-06,
"loss": 0.2142,
"step": 418
},
{
"epoch": 1.8012924071082392,
"grad_norm": 0.74013352394104,
"learning_rate": 9.216552972401582e-06,
"loss": 0.2266,
"step": 419
},
{
"epoch": 1.8056004308023694,
"grad_norm": 0.6855618953704834,
"learning_rate": 9.212910368083246e-06,
"loss": 0.2192,
"step": 420
},
{
"epoch": 1.8099084544964996,
"grad_norm": 0.7998465895652771,
"learning_rate": 9.209260038632562e-06,
"loss": 0.2172,
"step": 421
},
{
"epoch": 1.81421647819063,
"grad_norm": 0.798219621181488,
"learning_rate": 9.205601990743068e-06,
"loss": 0.236,
"step": 422
},
{
"epoch": 1.8185245018847604,
"grad_norm": 0.7570368051528931,
"learning_rate": 9.201936231122453e-06,
"loss": 0.2229,
"step": 423
},
{
"epoch": 1.8228325255788906,
"grad_norm": 0.7358865141868591,
"learning_rate": 9.198262766492554e-06,
"loss": 0.2302,
"step": 424
},
{
"epoch": 1.827140549273021,
"grad_norm": 0.7285174131393433,
"learning_rate": 9.194581603589327e-06,
"loss": 0.2312,
"step": 425
},
{
"epoch": 1.8314485729671515,
"grad_norm": 0.6818633675575256,
"learning_rate": 9.190892749162854e-06,
"loss": 0.2128,
"step": 426
},
{
"epoch": 1.8357565966612817,
"grad_norm": 0.7215419411659241,
"learning_rate": 9.187196209977314e-06,
"loss": 0.2142,
"step": 427
},
{
"epoch": 1.8400646203554119,
"grad_norm": 0.7318853139877319,
"learning_rate": 9.18349199281098e-06,
"loss": 0.2146,
"step": 428
},
{
"epoch": 1.8443726440495423,
"grad_norm": 0.7606166005134583,
"learning_rate": 9.179780104456205e-06,
"loss": 0.2099,
"step": 429
},
{
"epoch": 1.8486806677436727,
"grad_norm": 0.7998718619346619,
"learning_rate": 9.176060551719402e-06,
"loss": 0.2169,
"step": 430
},
{
"epoch": 1.8529886914378029,
"grad_norm": 0.788635790348053,
"learning_rate": 9.172333341421046e-06,
"loss": 0.2084,
"step": 431
},
{
"epoch": 1.857296715131933,
"grad_norm": 0.7982397675514221,
"learning_rate": 9.168598480395653e-06,
"loss": 0.2275,
"step": 432
},
{
"epoch": 1.8616047388260637,
"grad_norm": 0.783639132976532,
"learning_rate": 9.16485597549176e-06,
"loss": 0.2255,
"step": 433
},
{
"epoch": 1.865912762520194,
"grad_norm": 0.683904767036438,
"learning_rate": 9.16110583357193e-06,
"loss": 0.2044,
"step": 434
},
{
"epoch": 1.870220786214324,
"grad_norm": 0.7352596521377563,
"learning_rate": 9.157348061512728e-06,
"loss": 0.211,
"step": 435
},
{
"epoch": 1.8745288099084545,
"grad_norm": 0.7592405080795288,
"learning_rate": 9.153582666204702e-06,
"loss": 0.2208,
"step": 436
},
{
"epoch": 1.878836833602585,
"grad_norm": 0.8083028793334961,
"learning_rate": 9.149809654552387e-06,
"loss": 0.2278,
"step": 437
},
{
"epoch": 1.883144857296715,
"grad_norm": 0.918319046497345,
"learning_rate": 9.146029033474284e-06,
"loss": 0.2447,
"step": 438
},
{
"epoch": 1.8874528809908453,
"grad_norm": 0.8534688949584961,
"learning_rate": 9.142240809902841e-06,
"loss": 0.2135,
"step": 439
},
{
"epoch": 1.891760904684976,
"grad_norm": 0.6360190510749817,
"learning_rate": 9.138444990784455e-06,
"loss": 0.2143,
"step": 440
},
{
"epoch": 1.8960689283791061,
"grad_norm": 0.7315630316734314,
"learning_rate": 9.13464158307944e-06,
"loss": 0.2249,
"step": 441
},
{
"epoch": 1.9003769520732363,
"grad_norm": 0.7185893058776855,
"learning_rate": 9.130830593762037e-06,
"loss": 0.2219,
"step": 442
},
{
"epoch": 1.9046849757673667,
"grad_norm": 0.7180191874504089,
"learning_rate": 9.12701202982038e-06,
"loss": 0.2262,
"step": 443
},
{
"epoch": 1.9089929994614971,
"grad_norm": 0.844063937664032,
"learning_rate": 9.123185898256497e-06,
"loss": 0.2342,
"step": 444
},
{
"epoch": 1.9133010231556273,
"grad_norm": 0.7702510356903076,
"learning_rate": 9.119352206086292e-06,
"loss": 0.2277,
"step": 445
},
{
"epoch": 1.9176090468497575,
"grad_norm": 0.7082570791244507,
"learning_rate": 9.115510960339533e-06,
"loss": 0.2195,
"step": 446
},
{
"epoch": 1.921917070543888,
"grad_norm": 0.7621147632598877,
"learning_rate": 9.111662168059836e-06,
"loss": 0.2167,
"step": 447
},
{
"epoch": 1.9262250942380184,
"grad_norm": 0.7724840641021729,
"learning_rate": 9.107805836304658e-06,
"loss": 0.2276,
"step": 448
},
{
"epoch": 1.9305331179321485,
"grad_norm": 0.8185726404190063,
"learning_rate": 9.10394197214528e-06,
"loss": 0.227,
"step": 449
},
{
"epoch": 1.934841141626279,
"grad_norm": 0.8165621757507324,
"learning_rate": 9.100070582666796e-06,
"loss": 0.2159,
"step": 450
},
{
"epoch": 1.9391491653204094,
"grad_norm": 0.6630677580833435,
"learning_rate": 9.096191674968095e-06,
"loss": 0.2025,
"step": 451
},
{
"epoch": 1.9434571890145396,
"grad_norm": 0.8972369432449341,
"learning_rate": 9.09230525616186e-06,
"loss": 0.2288,
"step": 452
},
{
"epoch": 1.9477652127086698,
"grad_norm": 0.6295329332351685,
"learning_rate": 9.088411333374539e-06,
"loss": 0.1988,
"step": 453
},
{
"epoch": 1.9520732364028002,
"grad_norm": 0.8167666792869568,
"learning_rate": 9.084509913746342e-06,
"loss": 0.2316,
"step": 454
},
{
"epoch": 1.9563812600969306,
"grad_norm": 0.9114188551902771,
"learning_rate": 9.08060100443123e-06,
"loss": 0.2467,
"step": 455
},
{
"epoch": 1.9606892837910608,
"grad_norm": 0.6885513663291931,
"learning_rate": 9.076684612596891e-06,
"loss": 0.2033,
"step": 456
},
{
"epoch": 1.9649973074851912,
"grad_norm": 0.8051403760910034,
"learning_rate": 9.07276074542474e-06,
"loss": 0.2264,
"step": 457
},
{
"epoch": 1.9693053311793216,
"grad_norm": 0.7468147277832031,
"learning_rate": 9.068829410109893e-06,
"loss": 0.2166,
"step": 458
},
{
"epoch": 1.9736133548734518,
"grad_norm": 0.7809402346611023,
"learning_rate": 9.064890613861168e-06,
"loss": 0.2222,
"step": 459
},
{
"epoch": 1.977921378567582,
"grad_norm": 0.8526117205619812,
"learning_rate": 9.060944363901057e-06,
"loss": 0.2437,
"step": 460
},
{
"epoch": 1.9822294022617124,
"grad_norm": 0.7447945475578308,
"learning_rate": 9.05699066746572e-06,
"loss": 0.2191,
"step": 461
},
{
"epoch": 1.9865374259558428,
"grad_norm": 0.6959477663040161,
"learning_rate": 9.05302953180498e-06,
"loss": 0.2098,
"step": 462
},
{
"epoch": 1.990845449649973,
"grad_norm": 0.9427213668823242,
"learning_rate": 9.04906096418229e-06,
"loss": 0.2205,
"step": 463
},
{
"epoch": 1.9951534733441034,
"grad_norm": 0.8688002228736877,
"learning_rate": 9.045084971874738e-06,
"loss": 0.2068,
"step": 464
},
{
"epoch": 1.9994614970382338,
"grad_norm": 0.8308051824569702,
"learning_rate": 9.041101562173023e-06,
"loss": 0.2221,
"step": 465
},
{
"epoch": 2.0,
"grad_norm": 0.8308051824569702,
"learning_rate": 9.037110742381445e-06,
"loss": 0.2036,
"step": 466
},
{
"epoch": 2.00430802369413,
"grad_norm": 1.945624828338623,
"learning_rate": 9.033112519817897e-06,
"loss": 0.2128,
"step": 467
},
{
"epoch": 2.008616047388261,
"grad_norm": 0.7928284406661987,
"learning_rate": 9.02910690181384e-06,
"loss": 0.1988,
"step": 468
},
{
"epoch": 2.012924071082391,
"grad_norm": 0.7106955647468567,
"learning_rate": 9.0250938957143e-06,
"loss": 0.2005,
"step": 469
},
{
"epoch": 2.017232094776521,
"grad_norm": 0.7833051085472107,
"learning_rate": 9.021073508877845e-06,
"loss": 0.1958,
"step": 470
},
{
"epoch": 2.0215401184706514,
"grad_norm": 0.7723597288131714,
"learning_rate": 9.017045748676584e-06,
"loss": 0.1849,
"step": 471
},
{
"epoch": 2.025848142164782,
"grad_norm": 0.7105690836906433,
"learning_rate": 9.013010622496145e-06,
"loss": 0.2006,
"step": 472
},
{
"epoch": 2.0301561658589122,
"grad_norm": 0.7699000239372253,
"learning_rate": 9.008968137735655e-06,
"loss": 0.1946,
"step": 473
},
{
"epoch": 2.0344641895530424,
"grad_norm": 0.7681946158409119,
"learning_rate": 9.004918301807746e-06,
"loss": 0.1967,
"step": 474
},
{
"epoch": 2.038772213247173,
"grad_norm": 0.7109311819076538,
"learning_rate": 9.000861122138518e-06,
"loss": 0.1941,
"step": 475
},
{
"epoch": 2.0430802369413033,
"grad_norm": 0.7828920483589172,
"learning_rate": 8.996796606167549e-06,
"loss": 0.1931,
"step": 476
},
{
"epoch": 2.0473882606354334,
"grad_norm": 0.7319113612174988,
"learning_rate": 8.99272476134786e-06,
"loss": 0.2126,
"step": 477
},
{
"epoch": 2.0516962843295636,
"grad_norm": 0.7852508425712585,
"learning_rate": 8.988645595145913e-06,
"loss": 0.2024,
"step": 478
},
{
"epoch": 2.0560043080236943,
"grad_norm": 0.8128734827041626,
"learning_rate": 8.9845591150416e-06,
"loss": 0.2046,
"step": 479
},
{
"epoch": 2.0603123317178245,
"grad_norm": 0.8232097029685974,
"learning_rate": 8.98046532852822e-06,
"loss": 0.2083,
"step": 480
},
{
"epoch": 2.0646203554119547,
"grad_norm": 0.7391591668128967,
"learning_rate": 8.976364243112468e-06,
"loss": 0.1971,
"step": 481
},
{
"epoch": 2.0689283791060853,
"grad_norm": 0.7096617221832275,
"learning_rate": 8.972255866314425e-06,
"loss": 0.1974,
"step": 482
},
{
"epoch": 2.0732364028002155,
"grad_norm": 0.7919265627861023,
"learning_rate": 8.968140205667544e-06,
"loss": 0.1944,
"step": 483
},
{
"epoch": 2.0775444264943457,
"grad_norm": 0.7866784930229187,
"learning_rate": 8.964017268718632e-06,
"loss": 0.205,
"step": 484
},
{
"epoch": 2.081852450188476,
"grad_norm": 0.7036305069923401,
"learning_rate": 8.959887063027837e-06,
"loss": 0.1952,
"step": 485
},
{
"epoch": 2.0861604738826065,
"grad_norm": 0.7657865881919861,
"learning_rate": 8.95574959616864e-06,
"loss": 0.215,
"step": 486
},
{
"epoch": 2.0904684975767367,
"grad_norm": 0.7114456295967102,
"learning_rate": 8.951604875727833e-06,
"loss": 0.1968,
"step": 487
},
{
"epoch": 2.094776521270867,
"grad_norm": 0.7189726829528809,
"learning_rate": 8.94745290930551e-06,
"loss": 0.1929,
"step": 488
},
{
"epoch": 2.099084544964997,
"grad_norm": 0.7715457677841187,
"learning_rate": 8.94329370451505e-06,
"loss": 0.2049,
"step": 489
},
{
"epoch": 2.1033925686591277,
"grad_norm": 0.6534609198570251,
"learning_rate": 8.93912726898311e-06,
"loss": 0.1926,
"step": 490
},
{
"epoch": 2.107700592353258,
"grad_norm": 0.742912769317627,
"learning_rate": 8.934953610349599e-06,
"loss": 0.1961,
"step": 491
},
{
"epoch": 2.112008616047388,
"grad_norm": 0.6816502213478088,
"learning_rate": 8.930772736267675e-06,
"loss": 0.1885,
"step": 492
},
{
"epoch": 2.1163166397415187,
"grad_norm": 0.7203125357627869,
"learning_rate": 8.926584654403725e-06,
"loss": 0.199,
"step": 493
},
{
"epoch": 2.120624663435649,
"grad_norm": 0.714485228061676,
"learning_rate": 8.922389372437357e-06,
"loss": 0.2069,
"step": 494
},
{
"epoch": 2.124932687129779,
"grad_norm": 0.8489207029342651,
"learning_rate": 8.918186898061377e-06,
"loss": 0.1975,
"step": 495
},
{
"epoch": 2.1292407108239093,
"grad_norm": 0.7291126251220703,
"learning_rate": 8.91397723898178e-06,
"loss": 0.1973,
"step": 496
},
{
"epoch": 2.13354873451804,
"grad_norm": 0.7790775895118713,
"learning_rate": 8.909760402917738e-06,
"loss": 0.2002,
"step": 497
},
{
"epoch": 2.13785675821217,
"grad_norm": 0.6789630651473999,
"learning_rate": 8.90553639760158e-06,
"loss": 0.1825,
"step": 498
},
{
"epoch": 2.1421647819063003,
"grad_norm": 0.6676609516143799,
"learning_rate": 8.901305230778783e-06,
"loss": 0.1912,
"step": 499
},
{
"epoch": 2.146472805600431,
"grad_norm": 0.7658909559249878,
"learning_rate": 8.897066910207958e-06,
"loss": 0.1906,
"step": 500
},
{
"epoch": 2.150780829294561,
"grad_norm": 0.7476828694343567,
"learning_rate": 8.892821443660831e-06,
"loss": 0.1889,
"step": 501
},
{
"epoch": 2.1550888529886914,
"grad_norm": 0.7970424294471741,
"learning_rate": 8.888568838922231e-06,
"loss": 0.2097,
"step": 502
},
{
"epoch": 2.1593968766828215,
"grad_norm": 0.775412380695343,
"learning_rate": 8.884309103790078e-06,
"loss": 0.1901,
"step": 503
},
{
"epoch": 2.163704900376952,
"grad_norm": 0.822077214717865,
"learning_rate": 8.880042246075366e-06,
"loss": 0.2022,
"step": 504
},
{
"epoch": 2.1680129240710824,
"grad_norm": 0.7652409672737122,
"learning_rate": 8.875768273602148e-06,
"loss": 0.2038,
"step": 505
},
{
"epoch": 2.1723209477652126,
"grad_norm": 0.6706952452659607,
"learning_rate": 8.871487194207527e-06,
"loss": 0.1898,
"step": 506
},
{
"epoch": 2.176628971459343,
"grad_norm": 0.7346447706222534,
"learning_rate": 8.867199015741632e-06,
"loss": 0.1916,
"step": 507
},
{
"epoch": 2.1809369951534734,
"grad_norm": 0.8773221969604492,
"learning_rate": 8.862903746067619e-06,
"loss": 0.1995,
"step": 508
},
{
"epoch": 2.1852450188476036,
"grad_norm": 0.8744363784790039,
"learning_rate": 8.858601393061634e-06,
"loss": 0.2144,
"step": 509
},
{
"epoch": 2.1895530425417338,
"grad_norm": 0.7618948817253113,
"learning_rate": 8.854291964612824e-06,
"loss": 0.1927,
"step": 510
},
{
"epoch": 2.1938610662358644,
"grad_norm": 0.8064348697662354,
"learning_rate": 8.849975468623302e-06,
"loss": 0.2018,
"step": 511
},
{
"epoch": 2.1981690899299946,
"grad_norm": 0.7798653244972229,
"learning_rate": 8.845651913008145e-06,
"loss": 0.1968,
"step": 512
},
{
"epoch": 2.202477113624125,
"grad_norm": 0.7638295292854309,
"learning_rate": 8.841321305695372e-06,
"loss": 0.2006,
"step": 513
},
{
"epoch": 2.2067851373182554,
"grad_norm": 0.7909740805625916,
"learning_rate": 8.836983654625934e-06,
"loss": 0.2074,
"step": 514
},
{
"epoch": 2.2110931610123856,
"grad_norm": 0.7894330620765686,
"learning_rate": 8.832638967753699e-06,
"loss": 0.1924,
"step": 515
},
{
"epoch": 2.215401184706516,
"grad_norm": 0.7473629117012024,
"learning_rate": 8.828287253045436e-06,
"loss": 0.2004,
"step": 516
},
{
"epoch": 2.219709208400646,
"grad_norm": 0.666713535785675,
"learning_rate": 8.823928518480797e-06,
"loss": 0.1855,
"step": 517
},
{
"epoch": 2.2240172320947766,
"grad_norm": 0.7174916863441467,
"learning_rate": 8.819562772052312e-06,
"loss": 0.1951,
"step": 518
},
{
"epoch": 2.228325255788907,
"grad_norm": 0.8827500343322754,
"learning_rate": 8.815190021765365e-06,
"loss": 0.1914,
"step": 519
},
{
"epoch": 2.232633279483037,
"grad_norm": 0.7642231583595276,
"learning_rate": 8.810810275638183e-06,
"loss": 0.2155,
"step": 520
},
{
"epoch": 2.2369413031771677,
"grad_norm": 0.8573631644248962,
"learning_rate": 8.806423541701824e-06,
"loss": 0.2138,
"step": 521
},
{
"epoch": 2.241249326871298,
"grad_norm": 0.8140626549720764,
"learning_rate": 8.802029828000157e-06,
"loss": 0.2121,
"step": 522
},
{
"epoch": 2.245557350565428,
"grad_norm": 0.8195950388908386,
"learning_rate": 8.797629142589846e-06,
"loss": 0.2035,
"step": 523
},
{
"epoch": 2.2498653742595582,
"grad_norm": 0.8607699871063232,
"learning_rate": 8.793221493540347e-06,
"loss": 0.2148,
"step": 524
},
{
"epoch": 2.254173397953689,
"grad_norm": 0.8024702072143555,
"learning_rate": 8.788806888933881e-06,
"loss": 0.1999,
"step": 525
},
{
"epoch": 2.258481421647819,
"grad_norm": 0.7508124113082886,
"learning_rate": 8.784385336865419e-06,
"loss": 0.2015,
"step": 526
},
{
"epoch": 2.2627894453419493,
"grad_norm": 0.6819286942481995,
"learning_rate": 8.779956845442682e-06,
"loss": 0.2036,
"step": 527
},
{
"epoch": 2.26709746903608,
"grad_norm": 0.7056980729103088,
"learning_rate": 8.775521422786104e-06,
"loss": 0.1968,
"step": 528
},
{
"epoch": 2.27140549273021,
"grad_norm": 0.6826128959655762,
"learning_rate": 8.771079077028836e-06,
"loss": 0.1974,
"step": 529
},
{
"epoch": 2.2757135164243403,
"grad_norm": 0.7513538002967834,
"learning_rate": 8.766629816316722e-06,
"loss": 0.1985,
"step": 530
},
{
"epoch": 2.2800215401184705,
"grad_norm": 0.720900297164917,
"learning_rate": 8.762173648808283e-06,
"loss": 0.2054,
"step": 531
},
{
"epoch": 2.284329563812601,
"grad_norm": 0.90285325050354,
"learning_rate": 8.757710582674708e-06,
"loss": 0.2161,
"step": 532
},
{
"epoch": 2.2886375875067313,
"grad_norm": 0.877210259437561,
"learning_rate": 8.753240626099836e-06,
"loss": 0.2011,
"step": 533
},
{
"epoch": 2.2929456112008615,
"grad_norm": 0.7395567893981934,
"learning_rate": 8.748763787280142e-06,
"loss": 0.2003,
"step": 534
},
{
"epoch": 2.297253634894992,
"grad_norm": 0.7593705058097839,
"learning_rate": 8.744280074424713e-06,
"loss": 0.2002,
"step": 535
},
{
"epoch": 2.3015616585891223,
"grad_norm": 0.7474355697631836,
"learning_rate": 8.739789495755254e-06,
"loss": 0.1963,
"step": 536
},
{
"epoch": 2.3058696822832525,
"grad_norm": 0.6919746398925781,
"learning_rate": 8.735292059506047e-06,
"loss": 0.1955,
"step": 537
},
{
"epoch": 2.3101777059773827,
"grad_norm": 0.7640472650527954,
"learning_rate": 8.730787773923957e-06,
"loss": 0.2001,
"step": 538
},
{
"epoch": 2.3144857296715133,
"grad_norm": 0.6979402899742126,
"learning_rate": 8.726276647268403e-06,
"loss": 0.1985,
"step": 539
},
{
"epoch": 2.3187937533656435,
"grad_norm": 0.7744669914245605,
"learning_rate": 8.721758687811353e-06,
"loss": 0.206,
"step": 540
},
{
"epoch": 2.3231017770597737,
"grad_norm": 0.839337944984436,
"learning_rate": 8.717233903837298e-06,
"loss": 0.2062,
"step": 541
},
{
"epoch": 2.3274098007539044,
"grad_norm": 0.7544538974761963,
"learning_rate": 8.712702303643254e-06,
"loss": 0.2108,
"step": 542
},
{
"epoch": 2.3317178244480345,
"grad_norm": 0.7645125389099121,
"learning_rate": 8.708163895538722e-06,
"loss": 0.1971,
"step": 543
},
{
"epoch": 2.3360258481421647,
"grad_norm": 0.8445392847061157,
"learning_rate": 8.703618687845697e-06,
"loss": 0.1967,
"step": 544
},
{
"epoch": 2.340333871836295,
"grad_norm": 0.8427358865737915,
"learning_rate": 8.699066688898636e-06,
"loss": 0.216,
"step": 545
},
{
"epoch": 2.3446418955304256,
"grad_norm": 0.7722771763801575,
"learning_rate": 8.694507907044454e-06,
"loss": 0.1975,
"step": 546
},
{
"epoch": 2.3489499192245558,
"grad_norm": 0.689138650894165,
"learning_rate": 8.6899423506425e-06,
"loss": 0.2049,
"step": 547
},
{
"epoch": 2.353257942918686,
"grad_norm": 0.7982721328735352,
"learning_rate": 8.685370028064546e-06,
"loss": 0.1983,
"step": 548
},
{
"epoch": 2.3575659666128166,
"grad_norm": 0.8092723488807678,
"learning_rate": 8.680790947694772e-06,
"loss": 0.2092,
"step": 549
},
{
"epoch": 2.361873990306947,
"grad_norm": 0.7706890106201172,
"learning_rate": 8.676205117929752e-06,
"loss": 0.2121,
"step": 550
},
{
"epoch": 2.366182014001077,
"grad_norm": 0.8860268592834473,
"learning_rate": 8.671612547178428e-06,
"loss": 0.2071,
"step": 551
},
{
"epoch": 2.370490037695207,
"grad_norm": 0.8444654941558838,
"learning_rate": 8.667013243862113e-06,
"loss": 0.1887,
"step": 552
},
{
"epoch": 2.374798061389338,
"grad_norm": 0.6994353532791138,
"learning_rate": 8.66240721641446e-06,
"loss": 0.1928,
"step": 553
},
{
"epoch": 2.379106085083468,
"grad_norm": 0.7686805725097656,
"learning_rate": 8.657794473281447e-06,
"loss": 0.1974,
"step": 554
},
{
"epoch": 2.383414108777598,
"grad_norm": 0.6858060956001282,
"learning_rate": 8.65317502292138e-06,
"loss": 0.1914,
"step": 555
},
{
"epoch": 2.387722132471729,
"grad_norm": 0.7314836382865906,
"learning_rate": 8.64854887380485e-06,
"loss": 0.2014,
"step": 556
},
{
"epoch": 2.392030156165859,
"grad_norm": 0.7626047134399414,
"learning_rate": 8.643916034414741e-06,
"loss": 0.1925,
"step": 557
},
{
"epoch": 2.396338179859989,
"grad_norm": 0.6878098845481873,
"learning_rate": 8.639276513246199e-06,
"loss": 0.1925,
"step": 558
},
{
"epoch": 2.4006462035541194,
"grad_norm": 0.8436222076416016,
"learning_rate": 8.634630318806626e-06,
"loss": 0.217,
"step": 559
},
{
"epoch": 2.40495422724825,
"grad_norm": 0.7351590394973755,
"learning_rate": 8.629977459615655e-06,
"loss": 0.2068,
"step": 560
},
{
"epoch": 2.4092622509423802,
"grad_norm": 0.7552523612976074,
"learning_rate": 8.62531794420515e-06,
"loss": 0.2003,
"step": 561
},
{
"epoch": 2.4135702746365104,
"grad_norm": 0.7772282361984253,
"learning_rate": 8.620651781119169e-06,
"loss": 0.1978,
"step": 562
},
{
"epoch": 2.417878298330641,
"grad_norm": 0.8326259255409241,
"learning_rate": 8.615978978913968e-06,
"loss": 0.217,
"step": 563
},
{
"epoch": 2.4221863220247712,
"grad_norm": 0.8465754985809326,
"learning_rate": 8.611299546157973e-06,
"loss": 0.2032,
"step": 564
},
{
"epoch": 2.4264943457189014,
"grad_norm": 0.703048050403595,
"learning_rate": 8.60661349143177e-06,
"loss": 0.1963,
"step": 565
},
{
"epoch": 2.4308023694130316,
"grad_norm": 0.73744797706604,
"learning_rate": 8.601920823328088e-06,
"loss": 0.1893,
"step": 566
},
{
"epoch": 2.4351103931071623,
"grad_norm": 0.8171586990356445,
"learning_rate": 8.59722155045178e-06,
"loss": 0.198,
"step": 567
},
{
"epoch": 2.4394184168012925,
"grad_norm": 0.7568394541740417,
"learning_rate": 8.592515681419812e-06,
"loss": 0.1975,
"step": 568
},
{
"epoch": 2.4437264404954226,
"grad_norm": 0.874150812625885,
"learning_rate": 8.587803224861248e-06,
"loss": 0.2031,
"step": 569
},
{
"epoch": 2.4480344641895533,
"grad_norm": 0.7862420678138733,
"learning_rate": 8.583084189417225e-06,
"loss": 0.2001,
"step": 570
},
{
"epoch": 2.4523424878836835,
"grad_norm": 0.708115816116333,
"learning_rate": 8.578358583740947e-06,
"loss": 0.1989,
"step": 571
},
{
"epoch": 2.4566505115778137,
"grad_norm": 0.7508895397186279,
"learning_rate": 8.573626416497669e-06,
"loss": 0.1982,
"step": 572
},
{
"epoch": 2.460958535271944,
"grad_norm": 0.7255448698997498,
"learning_rate": 8.568887696364673e-06,
"loss": 0.1977,
"step": 573
},
{
"epoch": 2.4652665589660745,
"grad_norm": 0.819625198841095,
"learning_rate": 8.564142432031257e-06,
"loss": 0.1953,
"step": 574
},
{
"epoch": 2.4695745826602047,
"grad_norm": 0.6851021647453308,
"learning_rate": 8.559390632198723e-06,
"loss": 0.1925,
"step": 575
},
{
"epoch": 2.473882606354335,
"grad_norm": 0.8143554329872131,
"learning_rate": 8.554632305580355e-06,
"loss": 0.2123,
"step": 576
},
{
"epoch": 2.4781906300484655,
"grad_norm": 0.7244540452957153,
"learning_rate": 8.549867460901402e-06,
"loss": 0.1924,
"step": 577
},
{
"epoch": 2.4824986537425957,
"grad_norm": 0.7576718330383301,
"learning_rate": 8.545096106899068e-06,
"loss": 0.2173,
"step": 578
},
{
"epoch": 2.486806677436726,
"grad_norm": 0.7695874571800232,
"learning_rate": 8.540318252322493e-06,
"loss": 0.2073,
"step": 579
},
{
"epoch": 2.491114701130856,
"grad_norm": 0.6576676964759827,
"learning_rate": 8.535533905932739e-06,
"loss": 0.1936,
"step": 580
},
{
"epoch": 2.4954227248249867,
"grad_norm": 0.7426583766937256,
"learning_rate": 8.530743076502766e-06,
"loss": 0.1967,
"step": 581
},
{
"epoch": 2.499730748519117,
"grad_norm": 0.7785230278968811,
"learning_rate": 8.525945772817427e-06,
"loss": 0.2053,
"step": 582
},
{
"epoch": 2.504038772213247,
"grad_norm": 0.7582430243492126,
"learning_rate": 8.521142003673447e-06,
"loss": 0.1936,
"step": 583
},
{
"epoch": 2.5083467959073777,
"grad_norm": 0.7534447908401489,
"learning_rate": 8.5163317778794e-06,
"loss": 0.2121,
"step": 584
},
{
"epoch": 2.512654819601508,
"grad_norm": 0.7007425427436829,
"learning_rate": 8.51151510425571e-06,
"loss": 0.2067,
"step": 585
},
{
"epoch": 2.516962843295638,
"grad_norm": 0.7491743564605713,
"learning_rate": 8.506691991634612e-06,
"loss": 0.1993,
"step": 586
},
{
"epoch": 2.5212708669897683,
"grad_norm": 0.7492290139198303,
"learning_rate": 8.501862448860159e-06,
"loss": 0.2011,
"step": 587
},
{
"epoch": 2.5255788906838985,
"grad_norm": 0.7035954594612122,
"learning_rate": 8.497026484788189e-06,
"loss": 0.2085,
"step": 588
},
{
"epoch": 2.529886914378029,
"grad_norm": 0.6668585538864136,
"learning_rate": 8.492184108286316e-06,
"loss": 0.1962,
"step": 589
},
{
"epoch": 2.5341949380721593,
"grad_norm": 0.7750927209854126,
"learning_rate": 8.487335328233912e-06,
"loss": 0.2219,
"step": 590
},
{
"epoch": 2.53850296176629,
"grad_norm": 0.7335095405578613,
"learning_rate": 8.48248015352209e-06,
"loss": 0.1978,
"step": 591
},
{
"epoch": 2.54281098546042,
"grad_norm": 0.6468396782875061,
"learning_rate": 8.477618593053693e-06,
"loss": 0.1945,
"step": 592
},
{
"epoch": 2.5471190091545504,
"grad_norm": 0.7096503376960754,
"learning_rate": 8.47275065574327e-06,
"loss": 0.2109,
"step": 593
},
{
"epoch": 2.5514270328486806,
"grad_norm": 0.771790087223053,
"learning_rate": 8.46787635051706e-06,
"loss": 0.1988,
"step": 594
},
{
"epoch": 2.5557350565428107,
"grad_norm": 0.8515303134918213,
"learning_rate": 8.462995686312985e-06,
"loss": 0.1917,
"step": 595
},
{
"epoch": 2.5600430802369414,
"grad_norm": 0.7744242548942566,
"learning_rate": 8.458108672080624e-06,
"loss": 0.2075,
"step": 596
},
{
"epoch": 2.5643511039310716,
"grad_norm": 0.6713221669197083,
"learning_rate": 8.453215316781205e-06,
"loss": 0.1914,
"step": 597
},
{
"epoch": 2.568659127625202,
"grad_norm": 0.6832510232925415,
"learning_rate": 8.448315629387572e-06,
"loss": 0.2057,
"step": 598
},
{
"epoch": 2.5729671513193324,
"grad_norm": 0.8287082314491272,
"learning_rate": 8.44340961888419e-06,
"loss": 0.2012,
"step": 599
},
{
"epoch": 2.5772751750134626,
"grad_norm": 0.7368465662002563,
"learning_rate": 8.438497294267117e-06,
"loss": 0.2022,
"step": 600
},
{
"epoch": 2.581583198707593,
"grad_norm": 0.7697679996490479,
"learning_rate": 8.433578664543986e-06,
"loss": 0.2098,
"step": 601
},
{
"epoch": 2.585891222401723,
"grad_norm": 0.7776275277137756,
"learning_rate": 8.428653738733996e-06,
"loss": 0.1963,
"step": 602
},
{
"epoch": 2.5901992460958536,
"grad_norm": 0.7150554656982422,
"learning_rate": 8.423722525867883e-06,
"loss": 0.1917,
"step": 603
},
{
"epoch": 2.594507269789984,
"grad_norm": 0.67384934425354,
"learning_rate": 8.418785034987921e-06,
"loss": 0.1931,
"step": 604
},
{
"epoch": 2.598815293484114,
"grad_norm": 0.7748411297798157,
"learning_rate": 8.413841275147893e-06,
"loss": 0.1979,
"step": 605
},
{
"epoch": 2.6031233171782446,
"grad_norm": 0.7929915189743042,
"learning_rate": 8.408891255413072e-06,
"loss": 0.214,
"step": 606
},
{
"epoch": 2.607431340872375,
"grad_norm": 0.7815855145454407,
"learning_rate": 8.403934984860216e-06,
"loss": 0.2106,
"step": 607
},
{
"epoch": 2.611739364566505,
"grad_norm": 0.8815018534660339,
"learning_rate": 8.39897247257754e-06,
"loss": 0.2141,
"step": 608
},
{
"epoch": 2.616047388260635,
"grad_norm": 0.6736514568328857,
"learning_rate": 8.39400372766471e-06,
"loss": 0.204,
"step": 609
},
{
"epoch": 2.620355411954766,
"grad_norm": 0.6993145942687988,
"learning_rate": 8.389028759232816e-06,
"loss": 0.2135,
"step": 610
},
{
"epoch": 2.624663435648896,
"grad_norm": 0.8782504200935364,
"learning_rate": 8.38404757640436e-06,
"loss": 0.2129,
"step": 611
},
{
"epoch": 2.6289714593430262,
"grad_norm": 0.6805502772331238,
"learning_rate": 8.379060188313244e-06,
"loss": 0.2151,
"step": 612
},
{
"epoch": 2.633279483037157,
"grad_norm": 0.8480393886566162,
"learning_rate": 8.374066604104742e-06,
"loss": 0.2181,
"step": 613
},
{
"epoch": 2.637587506731287,
"grad_norm": 0.7601768970489502,
"learning_rate": 8.369066832935498e-06,
"loss": 0.2011,
"step": 614
},
{
"epoch": 2.6418955304254172,
"grad_norm": 0.6898226141929626,
"learning_rate": 8.364060883973488e-06,
"loss": 0.1925,
"step": 615
},
{
"epoch": 2.6462035541195474,
"grad_norm": 0.6958613991737366,
"learning_rate": 8.359048766398032e-06,
"loss": 0.1993,
"step": 616
},
{
"epoch": 2.650511577813678,
"grad_norm": 0.7246137261390686,
"learning_rate": 8.354030489399747e-06,
"loss": 0.1969,
"step": 617
},
{
"epoch": 2.6548196015078083,
"grad_norm": 0.7820032238960266,
"learning_rate": 8.349006062180552e-06,
"loss": 0.1914,
"step": 618
},
{
"epoch": 2.6591276252019385,
"grad_norm": 0.8184999823570251,
"learning_rate": 8.343975493953645e-06,
"loss": 0.1963,
"step": 619
},
{
"epoch": 2.663435648896069,
"grad_norm": 0.7577897906303406,
"learning_rate": 8.338938793943478e-06,
"loss": 0.2011,
"step": 620
},
{
"epoch": 2.6677436725901993,
"grad_norm": 0.7401598691940308,
"learning_rate": 8.333895971385754e-06,
"loss": 0.1919,
"step": 621
},
{
"epoch": 2.6720516962843295,
"grad_norm": 0.7432500720024109,
"learning_rate": 8.328847035527397e-06,
"loss": 0.2049,
"step": 622
},
{
"epoch": 2.6763597199784597,
"grad_norm": 0.7831881046295166,
"learning_rate": 8.323791995626543e-06,
"loss": 0.2062,
"step": 623
},
{
"epoch": 2.6806677436725903,
"grad_norm": 0.8239650130271912,
"learning_rate": 8.318730860952523e-06,
"loss": 0.2153,
"step": 624
},
{
"epoch": 2.6849757673667205,
"grad_norm": 0.7947770357131958,
"learning_rate": 8.313663640785839e-06,
"loss": 0.2047,
"step": 625
},
{
"epoch": 2.6892837910608507,
"grad_norm": 0.8108847141265869,
"learning_rate": 8.308590344418158e-06,
"loss": 0.1959,
"step": 626
},
{
"epoch": 2.6935918147549813,
"grad_norm": 0.7876724004745483,
"learning_rate": 8.303510981152283e-06,
"loss": 0.2138,
"step": 627
},
{
"epoch": 2.6978998384491115,
"grad_norm": 0.8588249683380127,
"learning_rate": 8.298425560302146e-06,
"loss": 0.2119,
"step": 628
},
{
"epoch": 2.7022078621432417,
"grad_norm": 0.7418671250343323,
"learning_rate": 8.293334091192782e-06,
"loss": 0.1929,
"step": 629
},
{
"epoch": 2.706515885837372,
"grad_norm": 0.6973020434379578,
"learning_rate": 8.288236583160322e-06,
"loss": 0.2001,
"step": 630
},
{
"epoch": 2.7108239095315025,
"grad_norm": 0.8735868334770203,
"learning_rate": 8.28313304555197e-06,
"loss": 0.1849,
"step": 631
},
{
"epoch": 2.7151319332256327,
"grad_norm": 0.700727641582489,
"learning_rate": 8.278023487725981e-06,
"loss": 0.1902,
"step": 632
},
{
"epoch": 2.719439956919763,
"grad_norm": 0.7866157293319702,
"learning_rate": 8.272907919051653e-06,
"loss": 0.2,
"step": 633
},
{
"epoch": 2.7237479806138936,
"grad_norm": 0.7389417290687561,
"learning_rate": 8.267786348909306e-06,
"loss": 0.2006,
"step": 634
},
{
"epoch": 2.7280560043080238,
"grad_norm": 0.7321817278862,
"learning_rate": 8.262658786690262e-06,
"loss": 0.2072,
"step": 635
},
{
"epoch": 2.732364028002154,
"grad_norm": 0.6979652047157288,
"learning_rate": 8.257525241796837e-06,
"loss": 0.1992,
"step": 636
},
{
"epoch": 2.736672051696284,
"grad_norm": 0.7699320316314697,
"learning_rate": 8.252385723642312e-06,
"loss": 0.1939,
"step": 637
},
{
"epoch": 2.7409800753904148,
"grad_norm": 0.7373313903808594,
"learning_rate": 8.247240241650918e-06,
"loss": 0.2051,
"step": 638
},
{
"epoch": 2.745288099084545,
"grad_norm": 0.6791808605194092,
"learning_rate": 8.242088805257832e-06,
"loss": 0.1934,
"step": 639
},
{
"epoch": 2.749596122778675,
"grad_norm": 0.6882132291793823,
"learning_rate": 8.23693142390914e-06,
"loss": 0.2032,
"step": 640
},
{
"epoch": 2.753904146472806,
"grad_norm": 0.7233914136886597,
"learning_rate": 8.231768107061831e-06,
"loss": 0.193,
"step": 641
},
{
"epoch": 2.758212170166936,
"grad_norm": 0.7256336212158203,
"learning_rate": 8.226598864183782e-06,
"loss": 0.2061,
"step": 642
},
{
"epoch": 2.762520193861066,
"grad_norm": 0.739739716053009,
"learning_rate": 8.221423704753733e-06,
"loss": 0.214,
"step": 643
},
{
"epoch": 2.7668282175551964,
"grad_norm": 0.6662392020225525,
"learning_rate": 8.216242638261277e-06,
"loss": 0.1992,
"step": 644
},
{
"epoch": 2.771136241249327,
"grad_norm": 0.7961679697036743,
"learning_rate": 8.211055674206828e-06,
"loss": 0.1898,
"step": 645
},
{
"epoch": 2.775444264943457,
"grad_norm": 0.7357142567634583,
"learning_rate": 8.205862822101628e-06,
"loss": 0.1988,
"step": 646
},
{
"epoch": 2.7797522886375874,
"grad_norm": 0.7799228429794312,
"learning_rate": 8.200664091467707e-06,
"loss": 0.2121,
"step": 647
},
{
"epoch": 2.784060312331718,
"grad_norm": 0.8995800018310547,
"learning_rate": 8.195459491837881e-06,
"loss": 0.2165,
"step": 648
},
{
"epoch": 2.788368336025848,
"grad_norm": 0.8075342178344727,
"learning_rate": 8.190249032755717e-06,
"loss": 0.2076,
"step": 649
},
{
"epoch": 2.7926763597199784,
"grad_norm": 0.8986786007881165,
"learning_rate": 8.18503272377554e-06,
"loss": 0.2093,
"step": 650
},
{
"epoch": 2.7969843834141086,
"grad_norm": 0.885172426700592,
"learning_rate": 8.179810574462388e-06,
"loss": 0.206,
"step": 651
},
{
"epoch": 2.8012924071082392,
"grad_norm": 0.9469455480575562,
"learning_rate": 8.17458259439202e-06,
"loss": 0.2077,
"step": 652
},
{
"epoch": 2.8056004308023694,
"grad_norm": 0.8134719729423523,
"learning_rate": 8.169348793150884e-06,
"loss": 0.2039,
"step": 653
},
{
"epoch": 2.8099084544964996,
"grad_norm": 0.7365888357162476,
"learning_rate": 8.164109180336094e-06,
"loss": 0.2014,
"step": 654
},
{
"epoch": 2.8142164781906303,
"grad_norm": 0.7723566293716431,
"learning_rate": 8.15886376555543e-06,
"loss": 0.192,
"step": 655
},
{
"epoch": 2.8185245018847604,
"grad_norm": 0.7469866275787354,
"learning_rate": 8.153612558427311e-06,
"loss": 0.208,
"step": 656
},
{
"epoch": 2.8228325255788906,
"grad_norm": 0.7777121067047119,
"learning_rate": 8.148355568580768e-06,
"loss": 0.1991,
"step": 657
},
{
"epoch": 2.827140549273021,
"grad_norm": 0.7026106119155884,
"learning_rate": 8.143092805655445e-06,
"loss": 0.1996,
"step": 658
},
{
"epoch": 2.8314485729671515,
"grad_norm": 0.778965413570404,
"learning_rate": 8.13782427930157e-06,
"loss": 0.2162,
"step": 659
},
{
"epoch": 2.8357565966612817,
"grad_norm": 0.8295383453369141,
"learning_rate": 8.132549999179934e-06,
"loss": 0.2013,
"step": 660
},
{
"epoch": 2.840064620355412,
"grad_norm": 0.7001696825027466,
"learning_rate": 8.127269974961886e-06,
"loss": 0.209,
"step": 661
},
{
"epoch": 2.8443726440495425,
"grad_norm": 0.6350052356719971,
"learning_rate": 8.121984216329303e-06,
"loss": 0.1946,
"step": 662
},
{
"epoch": 2.8486806677436727,
"grad_norm": 0.7462396621704102,
"learning_rate": 8.116692732974578e-06,
"loss": 0.2067,
"step": 663
},
{
"epoch": 2.852988691437803,
"grad_norm": 0.7569505572319031,
"learning_rate": 8.111395534600604e-06,
"loss": 0.1956,
"step": 664
},
{
"epoch": 2.857296715131933,
"grad_norm": 0.7309179902076721,
"learning_rate": 8.10609263092075e-06,
"loss": 0.2024,
"step": 665
},
{
"epoch": 2.8616047388260637,
"grad_norm": 0.6988533139228821,
"learning_rate": 8.100784031658846e-06,
"loss": 0.2191,
"step": 666
},
{
"epoch": 2.865912762520194,
"grad_norm": 0.7274155616760254,
"learning_rate": 8.095469746549172e-06,
"loss": 0.2057,
"step": 667
},
{
"epoch": 2.870220786214324,
"grad_norm": 0.7919628620147705,
"learning_rate": 8.090149785336426e-06,
"loss": 0.1939,
"step": 668
},
{
"epoch": 2.8745288099084547,
"grad_norm": 0.7325423955917358,
"learning_rate": 8.084824157775719e-06,
"loss": 0.189,
"step": 669
},
{
"epoch": 2.878836833602585,
"grad_norm": 0.8092797994613647,
"learning_rate": 8.079492873632554e-06,
"loss": 0.2127,
"step": 670
},
{
"epoch": 2.883144857296715,
"grad_norm": 0.6685104966163635,
"learning_rate": 8.074155942682803e-06,
"loss": 0.201,
"step": 671
},
{
"epoch": 2.8874528809908453,
"grad_norm": 0.7067998647689819,
"learning_rate": 8.068813374712689e-06,
"loss": 0.2036,
"step": 672
},
{
"epoch": 2.891760904684976,
"grad_norm": 0.7045204043388367,
"learning_rate": 8.06346517951878e-06,
"loss": 0.1982,
"step": 673
},
{
"epoch": 2.896068928379106,
"grad_norm": 0.6677000522613525,
"learning_rate": 8.058111366907957e-06,
"loss": 0.189,
"step": 674
},
{
"epoch": 2.9003769520732363,
"grad_norm": 0.8431633114814758,
"learning_rate": 8.052751946697403e-06,
"loss": 0.2113,
"step": 675
},
{
"epoch": 2.904684975767367,
"grad_norm": 0.781539797782898,
"learning_rate": 8.047386928714583e-06,
"loss": 0.2184,
"step": 676
},
{
"epoch": 2.908992999461497,
"grad_norm": 0.74564129114151,
"learning_rate": 8.042016322797227e-06,
"loss": 0.2001,
"step": 677
},
{
"epoch": 2.9133010231556273,
"grad_norm": 0.6591554880142212,
"learning_rate": 8.03664013879331e-06,
"loss": 0.1947,
"step": 678
},
{
"epoch": 2.9176090468497575,
"grad_norm": 0.7208926677703857,
"learning_rate": 8.031258386561038e-06,
"loss": 0.2105,
"step": 679
},
{
"epoch": 2.9219170705438877,
"grad_norm": 0.7640514373779297,
"learning_rate": 8.025871075968828e-06,
"loss": 0.2109,
"step": 680
},
{
"epoch": 2.9262250942380184,
"grad_norm": 0.7145681977272034,
"learning_rate": 8.020478216895282e-06,
"loss": 0.1874,
"step": 681
},
{
"epoch": 2.9305331179321485,
"grad_norm": 0.7085782885551453,
"learning_rate": 8.015079819229187e-06,
"loss": 0.2117,
"step": 682
},
{
"epoch": 2.934841141626279,
"grad_norm": 0.7184674739837646,
"learning_rate": 8.009675892869478e-06,
"loss": 0.2103,
"step": 683
},
{
"epoch": 2.9391491653204094,
"grad_norm": 0.726003885269165,
"learning_rate": 8.00426644772523e-06,
"loss": 0.2119,
"step": 684
},
{
"epoch": 2.9434571890145396,
"grad_norm": 0.8296722769737244,
"learning_rate": 7.99885149371564e-06,
"loss": 0.2048,
"step": 685
},
{
"epoch": 2.9477652127086698,
"grad_norm": 0.6646244525909424,
"learning_rate": 7.993431040770002e-06,
"loss": 0.2064,
"step": 686
},
{
"epoch": 2.9520732364028,
"grad_norm": 0.8237845301628113,
"learning_rate": 7.988005098827699e-06,
"loss": 0.2195,
"step": 687
},
{
"epoch": 2.9563812600969306,
"grad_norm": 0.7150204181671143,
"learning_rate": 7.982573677838172e-06,
"loss": 0.2053,
"step": 688
},
{
"epoch": 2.9606892837910608,
"grad_norm": 0.6601112484931946,
"learning_rate": 7.977136787760916e-06,
"loss": 0.1835,
"step": 689
},
{
"epoch": 2.9649973074851914,
"grad_norm": 0.7872753739356995,
"learning_rate": 7.97169443856545e-06,
"loss": 0.1974,
"step": 690
},
{
"epoch": 2.9693053311793216,
"grad_norm": 0.7161293625831604,
"learning_rate": 7.966246640231303e-06,
"loss": 0.1963,
"step": 691
},
{
"epoch": 2.973613354873452,
"grad_norm": 0.7552039623260498,
"learning_rate": 7.960793402748001e-06,
"loss": 0.1942,
"step": 692
},
{
"epoch": 2.977921378567582,
"grad_norm": 0.7136242389678955,
"learning_rate": 7.955334736115038e-06,
"loss": 0.2035,
"step": 693
},
{
"epoch": 2.982229402261712,
"grad_norm": 0.781005859375,
"learning_rate": 7.949870650341864e-06,
"loss": 0.194,
"step": 694
},
{
"epoch": 2.986537425955843,
"grad_norm": 0.6988127827644348,
"learning_rate": 7.944401155447872e-06,
"loss": 0.1936,
"step": 695
},
{
"epoch": 2.990845449649973,
"grad_norm": 0.7402964234352112,
"learning_rate": 7.938926261462366e-06,
"loss": 0.2199,
"step": 696
},
{
"epoch": 2.9951534733441036,
"grad_norm": 0.8665761947631836,
"learning_rate": 7.933445978424555e-06,
"loss": 0.2041,
"step": 697
},
{
"epoch": 2.999461497038234,
"grad_norm": 0.7297384142875671,
"learning_rate": 7.927960316383524e-06,
"loss": 0.1965,
"step": 698
},
{
"epoch": 3.0,
"grad_norm": 0.7297384142875671,
"learning_rate": 7.92246928539823e-06,
"loss": 0.2041,
"step": 699
},
{
"epoch": 3.00430802369413,
"grad_norm": 3.1112265586853027,
"learning_rate": 7.916972895537471e-06,
"loss": 0.1905,
"step": 700
},
{
"epoch": 3.008616047388261,
"grad_norm": 0.7145331501960754,
"learning_rate": 7.911471156879866e-06,
"loss": 0.1807,
"step": 701
},
{
"epoch": 3.012924071082391,
"grad_norm": 0.6694017648696899,
"learning_rate": 7.905964079513851e-06,
"loss": 0.1801,
"step": 702
},
{
"epoch": 3.017232094776521,
"grad_norm": 0.6125895380973816,
"learning_rate": 7.900451673537646e-06,
"loss": 0.1847,
"step": 703
},
{
"epoch": 3.0215401184706514,
"grad_norm": 0.8431554436683655,
"learning_rate": 7.894933949059245e-06,
"loss": 0.1821,
"step": 704
},
{
"epoch": 3.025848142164782,
"grad_norm": 0.6408715844154358,
"learning_rate": 7.88941091619639e-06,
"loss": 0.1709,
"step": 705
},
{
"epoch": 3.0301561658589122,
"grad_norm": 0.6939838528633118,
"learning_rate": 7.883882585076558e-06,
"loss": 0.1765,
"step": 706
},
{
"epoch": 3.0344641895530424,
"grad_norm": 0.739832878112793,
"learning_rate": 7.87834896583695e-06,
"loss": 0.179,
"step": 707
},
{
"epoch": 3.038772213247173,
"grad_norm": 0.8355388045310974,
"learning_rate": 7.872810068624452e-06,
"loss": 0.1736,
"step": 708
},
{
"epoch": 3.0430802369413033,
"grad_norm": 0.7089900374412537,
"learning_rate": 7.867265903595632e-06,
"loss": 0.1787,
"step": 709
},
{
"epoch": 3.0473882606354334,
"grad_norm": 0.7276995182037354,
"learning_rate": 7.86171648091672e-06,
"loss": 0.1756,
"step": 710
},
{
"epoch": 3.0516962843295636,
"grad_norm": 0.6631605625152588,
"learning_rate": 7.856161810763584e-06,
"loss": 0.176,
"step": 711
},
{
"epoch": 3.0560043080236943,
"grad_norm": 0.7474228143692017,
"learning_rate": 7.850601903321717e-06,
"loss": 0.1817,
"step": 712
},
{
"epoch": 3.0603123317178245,
"grad_norm": 0.6688998937606812,
"learning_rate": 7.845036768786214e-06,
"loss": 0.1792,
"step": 713
},
{
"epoch": 3.0646203554119547,
"grad_norm": 0.7395163774490356,
"learning_rate": 7.839466417361753e-06,
"loss": 0.1658,
"step": 714
},
{
"epoch": 3.0689283791060853,
"grad_norm": 0.7095337510108948,
"learning_rate": 7.833890859262579e-06,
"loss": 0.1693,
"step": 715
},
{
"epoch": 3.0732364028002155,
"grad_norm": 0.8074468970298767,
"learning_rate": 7.828310104712488e-06,
"loss": 0.1751,
"step": 716
},
{
"epoch": 3.0775444264943457,
"grad_norm": 0.6688709259033203,
"learning_rate": 7.822724163944802e-06,
"loss": 0.1848,
"step": 717
},
{
"epoch": 3.081852450188476,
"grad_norm": 0.7156962752342224,
"learning_rate": 7.81713304720235e-06,
"loss": 0.1664,
"step": 718
},
{
"epoch": 3.0861604738826065,
"grad_norm": 0.7590762376785278,
"learning_rate": 7.811536764737454e-06,
"loss": 0.1786,
"step": 719
},
{
"epoch": 3.0904684975767367,
"grad_norm": 0.7737705111503601,
"learning_rate": 7.805935326811913e-06,
"loss": 0.182,
"step": 720
},
{
"epoch": 3.094776521270867,
"grad_norm": 0.7593074440956116,
"learning_rate": 7.800328743696973e-06,
"loss": 0.1744,
"step": 721
},
{
"epoch": 3.099084544964997,
"grad_norm": 0.8113828301429749,
"learning_rate": 7.794717025673318e-06,
"loss": 0.166,
"step": 722
},
{
"epoch": 3.1033925686591277,
"grad_norm": 0.7906281352043152,
"learning_rate": 7.789100183031045e-06,
"loss": 0.1876,
"step": 723
},
{
"epoch": 3.107700592353258,
"grad_norm": 0.7934765815734863,
"learning_rate": 7.783478226069652e-06,
"loss": 0.1852,
"step": 724
},
{
"epoch": 3.112008616047388,
"grad_norm": 0.814837634563446,
"learning_rate": 7.777851165098012e-06,
"loss": 0.1699,
"step": 725
},
{
"epoch": 3.1163166397415187,
"grad_norm": 0.6663200259208679,
"learning_rate": 7.772219010434359e-06,
"loss": 0.1742,
"step": 726
},
{
"epoch": 3.120624663435649,
"grad_norm": 0.7285510301589966,
"learning_rate": 7.766581772406266e-06,
"loss": 0.1708,
"step": 727
},
{
"epoch": 3.124932687129779,
"grad_norm": 0.6569226384162903,
"learning_rate": 7.760939461350622e-06,
"loss": 0.179,
"step": 728
},
{
"epoch": 3.1292407108239093,
"grad_norm": 0.7777225375175476,
"learning_rate": 7.755292087613635e-06,
"loss": 0.1706,
"step": 729
},
{
"epoch": 3.13354873451804,
"grad_norm": 0.7825784087181091,
"learning_rate": 7.749639661550775e-06,
"loss": 0.1806,
"step": 730
},
{
"epoch": 3.13785675821217,
"grad_norm": 0.8840234875679016,
"learning_rate": 7.743982193526791e-06,
"loss": 0.1839,
"step": 731
},
{
"epoch": 3.1421647819063003,
"grad_norm": 0.6787511110305786,
"learning_rate": 7.738319693915673e-06,
"loss": 0.1761,
"step": 732
},
{
"epoch": 3.146472805600431,
"grad_norm": 0.7972119450569153,
"learning_rate": 7.732652173100634e-06,
"loss": 0.1742,
"step": 733
},
{
"epoch": 3.150780829294561,
"grad_norm": 0.8184621334075928,
"learning_rate": 7.726979641474102e-06,
"loss": 0.1833,
"step": 734
},
{
"epoch": 3.1550888529886914,
"grad_norm": 0.75215083360672,
"learning_rate": 7.721302109437686e-06,
"loss": 0.1819,
"step": 735
},
{
"epoch": 3.1593968766828215,
"grad_norm": 0.8182296752929688,
"learning_rate": 7.715619587402165e-06,
"loss": 0.1819,
"step": 736
},
{
"epoch": 3.163704900376952,
"grad_norm": 0.8826239109039307,
"learning_rate": 7.709932085787473e-06,
"loss": 0.176,
"step": 737
},
{
"epoch": 3.1680129240710824,
"grad_norm": 0.7005189657211304,
"learning_rate": 7.704239615022671e-06,
"loss": 0.1758,
"step": 738
},
{
"epoch": 3.1723209477652126,
"grad_norm": 0.7199857831001282,
"learning_rate": 7.698542185545932e-06,
"loss": 0.1611,
"step": 739
},
{
"epoch": 3.176628971459343,
"grad_norm": 0.7488324046134949,
"learning_rate": 7.692839807804522e-06,
"loss": 0.1767,
"step": 740
},
{
"epoch": 3.1809369951534734,
"grad_norm": 0.7330511808395386,
"learning_rate": 7.687132492254783e-06,
"loss": 0.1746,
"step": 741
},
{
"epoch": 3.1852450188476036,
"grad_norm": 0.7547047138214111,
"learning_rate": 7.681420249362107e-06,
"loss": 0.1662,
"step": 742
},
{
"epoch": 3.1895530425417338,
"grad_norm": 0.6859620809555054,
"learning_rate": 7.675703089600926e-06,
"loss": 0.1768,
"step": 743
},
{
"epoch": 3.1938610662358644,
"grad_norm": 0.8874660730361938,
"learning_rate": 7.669981023454682e-06,
"loss": 0.1961,
"step": 744
},
{
"epoch": 3.1981690899299946,
"grad_norm": 0.9125708341598511,
"learning_rate": 7.664254061415818e-06,
"loss": 0.1797,
"step": 745
},
{
"epoch": 3.202477113624125,
"grad_norm": 0.711672842502594,
"learning_rate": 7.658522213985757e-06,
"loss": 0.1834,
"step": 746
},
{
"epoch": 3.2067851373182554,
"grad_norm": 0.7379571199417114,
"learning_rate": 7.652785491674872e-06,
"loss": 0.1719,
"step": 747
},
{
"epoch": 3.2110931610123856,
"grad_norm": 0.7924142479896545,
"learning_rate": 7.647043905002485e-06,
"loss": 0.1861,
"step": 748
},
{
"epoch": 3.215401184706516,
"grad_norm": 0.8076757192611694,
"learning_rate": 7.641297464496828e-06,
"loss": 0.1801,
"step": 749
},
{
"epoch": 3.219709208400646,
"grad_norm": 0.723019540309906,
"learning_rate": 7.635546180695039e-06,
"loss": 0.1758,
"step": 750
},
{
"epoch": 3.2240172320947766,
"grad_norm": 0.6700244545936584,
"learning_rate": 7.629790064143139e-06,
"loss": 0.1808,
"step": 751
},
{
"epoch": 3.228325255788907,
"grad_norm": 0.7725983262062073,
"learning_rate": 7.624029125396004e-06,
"loss": 0.1795,
"step": 752
},
{
"epoch": 3.232633279483037,
"grad_norm": 0.7336100935935974,
"learning_rate": 7.618263375017358e-06,
"loss": 0.174,
"step": 753
},
{
"epoch": 3.2369413031771677,
"grad_norm": 0.8215693831443787,
"learning_rate": 7.612492823579744e-06,
"loss": 0.1721,
"step": 754
},
{
"epoch": 3.241249326871298,
"grad_norm": 0.7242813110351562,
"learning_rate": 7.606717481664515e-06,
"loss": 0.1934,
"step": 755
},
{
"epoch": 3.245557350565428,
"grad_norm": 0.889627993106842,
"learning_rate": 7.600937359861799e-06,
"loss": 0.1795,
"step": 756
},
{
"epoch": 3.2498653742595582,
"grad_norm": 0.8864472508430481,
"learning_rate": 7.595152468770497e-06,
"loss": 0.1741,
"step": 757
},
{
"epoch": 3.254173397953689,
"grad_norm": 0.7779316306114197,
"learning_rate": 7.589362818998251e-06,
"loss": 0.1874,
"step": 758
},
{
"epoch": 3.258481421647819,
"grad_norm": 0.8183091878890991,
"learning_rate": 7.58356842116143e-06,
"loss": 0.1851,
"step": 759
},
{
"epoch": 3.2627894453419493,
"grad_norm": 0.8073911070823669,
"learning_rate": 7.57776928588511e-06,
"loss": 0.186,
"step": 760
},
{
"epoch": 3.26709746903608,
"grad_norm": 0.8156265020370483,
"learning_rate": 7.571965423803052e-06,
"loss": 0.1735,
"step": 761
},
{
"epoch": 3.27140549273021,
"grad_norm": 0.7891457080841064,
"learning_rate": 7.566156845557684e-06,
"loss": 0.1708,
"step": 762
},
{
"epoch": 3.2757135164243403,
"grad_norm": 0.7164343595504761,
"learning_rate": 7.560343561800087e-06,
"loss": 0.183,
"step": 763
},
{
"epoch": 3.2800215401184705,
"grad_norm": 0.7573089599609375,
"learning_rate": 7.554525583189969e-06,
"loss": 0.1883,
"step": 764
},
{
"epoch": 3.284329563812601,
"grad_norm": 0.7825318574905396,
"learning_rate": 7.548702920395639e-06,
"loss": 0.1787,
"step": 765
},
{
"epoch": 3.2886375875067313,
"grad_norm": 0.7607948184013367,
"learning_rate": 7.542875584094006e-06,
"loss": 0.1934,
"step": 766
},
{
"epoch": 3.2929456112008615,
"grad_norm": 0.8865452408790588,
"learning_rate": 7.537043584970543e-06,
"loss": 0.1713,
"step": 767
},
{
"epoch": 3.297253634894992,
"grad_norm": 0.7793084383010864,
"learning_rate": 7.53120693371927e-06,
"loss": 0.1741,
"step": 768
},
{
"epoch": 3.3015616585891223,
"grad_norm": 0.6900126934051514,
"learning_rate": 7.525365641042749e-06,
"loss": 0.186,
"step": 769
},
{
"epoch": 3.3058696822832525,
"grad_norm": 0.7191606760025024,
"learning_rate": 7.519519717652039e-06,
"loss": 0.1805,
"step": 770
},
{
"epoch": 3.3101777059773827,
"grad_norm": 0.8574715852737427,
"learning_rate": 7.5136691742667e-06,
"loss": 0.1846,
"step": 771
},
{
"epoch": 3.3144857296715133,
"grad_norm": 0.7486770153045654,
"learning_rate": 7.507814021614761e-06,
"loss": 0.1788,
"step": 772
},
{
"epoch": 3.3187937533656435,
"grad_norm": 0.7185223698616028,
"learning_rate": 7.501954270432701e-06,
"loss": 0.1813,
"step": 773
},
{
"epoch": 3.3231017770597737,
"grad_norm": 0.7054628133773804,
"learning_rate": 7.496089931465432e-06,
"loss": 0.192,
"step": 774
},
{
"epoch": 3.3274098007539044,
"grad_norm": 0.8348388671875,
"learning_rate": 7.490221015466279e-06,
"loss": 0.1831,
"step": 775
},
{
"epoch": 3.3317178244480345,
"grad_norm": 0.90952467918396,
"learning_rate": 7.4843475331969614e-06,
"loss": 0.1764,
"step": 776
},
{
"epoch": 3.3360258481421647,
"grad_norm": 0.7084717750549316,
"learning_rate": 7.478469495427569e-06,
"loss": 0.1865,
"step": 777
},
{
"epoch": 3.340333871836295,
"grad_norm": 0.771369457244873,
"learning_rate": 7.4725869129365484e-06,
"loss": 0.1846,
"step": 778
},
{
"epoch": 3.3446418955304256,
"grad_norm": 0.75652676820755,
"learning_rate": 7.4666997965106725e-06,
"loss": 0.1875,
"step": 779
},
{
"epoch": 3.3489499192245558,
"grad_norm": 0.7481139302253723,
"learning_rate": 7.4608081569450365e-06,
"loss": 0.1752,
"step": 780
},
{
"epoch": 3.353257942918686,
"grad_norm": 0.7712534070014954,
"learning_rate": 7.4549120050430265e-06,
"loss": 0.181,
"step": 781
},
{
"epoch": 3.3575659666128166,
"grad_norm": 0.8637974858283997,
"learning_rate": 7.449011351616302e-06,
"loss": 0.1774,
"step": 782
},
{
"epoch": 3.361873990306947,
"grad_norm": 0.712335467338562,
"learning_rate": 7.443106207484776e-06,
"loss": 0.1878,
"step": 783
},
{
"epoch": 3.366182014001077,
"grad_norm": 0.779151201248169,
"learning_rate": 7.437196583476597e-06,
"loss": 0.1728,
"step": 784
},
{
"epoch": 3.370490037695207,
"grad_norm": 0.7468299865722656,
"learning_rate": 7.43128249042813e-06,
"loss": 0.1911,
"step": 785
},
{
"epoch": 3.374798061389338,
"grad_norm": 0.8930696249008179,
"learning_rate": 7.425363939183931e-06,
"loss": 0.1798,
"step": 786
},
{
"epoch": 3.379106085083468,
"grad_norm": 0.9994642734527588,
"learning_rate": 7.419440940596735e-06,
"loss": 0.1881,
"step": 787
},
{
"epoch": 3.383414108777598,
"grad_norm": 0.8835253715515137,
"learning_rate": 7.41351350552743e-06,
"loss": 0.1827,
"step": 788
},
{
"epoch": 3.387722132471729,
"grad_norm": 0.7508931159973145,
"learning_rate": 7.407581644845038e-06,
"loss": 0.1867,
"step": 789
},
{
"epoch": 3.392030156165859,
"grad_norm": 0.8232507109642029,
"learning_rate": 7.401645369426697e-06,
"loss": 0.1807,
"step": 790
},
{
"epoch": 3.396338179859989,
"grad_norm": 0.755211353302002,
"learning_rate": 7.395704690157644e-06,
"loss": 0.1841,
"step": 791
},
{
"epoch": 3.4006462035541194,
"grad_norm": 0.8512404561042786,
"learning_rate": 7.389759617931183e-06,
"loss": 0.1932,
"step": 792
},
{
"epoch": 3.40495422724825,
"grad_norm": 0.7709099054336548,
"learning_rate": 7.383810163648682e-06,
"loss": 0.166,
"step": 793
},
{
"epoch": 3.4092622509423802,
"grad_norm": 0.7232650518417358,
"learning_rate": 7.3778563382195365e-06,
"loss": 0.18,
"step": 794
},
{
"epoch": 3.4135702746365104,
"grad_norm": 0.8082464337348938,
"learning_rate": 7.371898152561166e-06,
"loss": 0.186,
"step": 795
},
{
"epoch": 3.417878298330641,
"grad_norm": 0.7701036334037781,
"learning_rate": 7.365935617598975e-06,
"loss": 0.1739,
"step": 796
},
{
"epoch": 3.4221863220247712,
"grad_norm": 0.7623946666717529,
"learning_rate": 7.359968744266353e-06,
"loss": 0.1933,
"step": 797
},
{
"epoch": 3.4264943457189014,
"grad_norm": 0.81331866979599,
"learning_rate": 7.35399754350464e-06,
"loss": 0.1903,
"step": 798
},
{
"epoch": 3.4308023694130316,
"grad_norm": 0.7704458236694336,
"learning_rate": 7.3480220262631095e-06,
"loss": 0.1756,
"step": 799
},
{
"epoch": 3.4351103931071623,
"grad_norm": 0.7425993084907532,
"learning_rate": 7.342042203498952e-06,
"loss": 0.1786,
"step": 800
},
{
"epoch": 3.4394184168012925,
"grad_norm": 0.8658850789070129,
"learning_rate": 7.336058086177253e-06,
"loss": 0.1834,
"step": 801
},
{
"epoch": 3.4437264404954226,
"grad_norm": 0.8086578249931335,
"learning_rate": 7.330069685270976e-06,
"loss": 0.1808,
"step": 802
},
{
"epoch": 3.4480344641895533,
"grad_norm": 0.7482308745384216,
"learning_rate": 7.3240770117609325e-06,
"loss": 0.1887,
"step": 803
},
{
"epoch": 3.4523424878836835,
"grad_norm": 0.8038567900657654,
"learning_rate": 7.318080076635773e-06,
"loss": 0.1831,
"step": 804
},
{
"epoch": 3.4566505115778137,
"grad_norm": 0.7037034034729004,
"learning_rate": 7.312078890891962e-06,
"loss": 0.1766,
"step": 805
},
{
"epoch": 3.460958535271944,
"grad_norm": 0.7575218081474304,
"learning_rate": 7.306073465533759e-06,
"loss": 0.1753,
"step": 806
},
{
"epoch": 3.4652665589660745,
"grad_norm": 0.831322193145752,
"learning_rate": 7.300063811573194e-06,
"loss": 0.1988,
"step": 807
},
{
"epoch": 3.4695745826602047,
"grad_norm": 0.737129271030426,
"learning_rate": 7.294049940030055e-06,
"loss": 0.1904,
"step": 808
},
{
"epoch": 3.473882606354335,
"grad_norm": 0.7896698713302612,
"learning_rate": 7.2880318619318605e-06,
"loss": 0.1883,
"step": 809
},
{
"epoch": 3.4781906300484655,
"grad_norm": 0.7189210653305054,
"learning_rate": 7.2820095883138456e-06,
"loss": 0.1743,
"step": 810
},
{
"epoch": 3.4824986537425957,
"grad_norm": 0.7875738739967346,
"learning_rate": 7.2759831302189376e-06,
"loss": 0.1917,
"step": 811
},
{
"epoch": 3.486806677436726,
"grad_norm": 0.8502741456031799,
"learning_rate": 7.269952498697734e-06,
"loss": 0.193,
"step": 812
},
{
"epoch": 3.491114701130856,
"grad_norm": 0.7916224002838135,
"learning_rate": 7.2639177048084894e-06,
"loss": 0.1877,
"step": 813
},
{
"epoch": 3.4954227248249867,
"grad_norm": 0.7110468745231628,
"learning_rate": 7.25787875961709e-06,
"loss": 0.1751,
"step": 814
},
{
"epoch": 3.499730748519117,
"grad_norm": 0.7333479523658752,
"learning_rate": 7.2518356741970285e-06,
"loss": 0.1736,
"step": 815
},
{
"epoch": 3.504038772213247,
"grad_norm": 0.7988889813423157,
"learning_rate": 7.245788459629397e-06,
"loss": 0.1756,
"step": 816
},
{
"epoch": 3.5083467959073777,
"grad_norm": 0.7231525778770447,
"learning_rate": 7.239737127002854e-06,
"loss": 0.1748,
"step": 817
},
{
"epoch": 3.512654819601508,
"grad_norm": 0.7418997287750244,
"learning_rate": 7.233681687413614e-06,
"loss": 0.1809,
"step": 818
},
{
"epoch": 3.516962843295638,
"grad_norm": 0.7931709885597229,
"learning_rate": 7.227622151965418e-06,
"loss": 0.19,
"step": 819
},
{
"epoch": 3.5212708669897683,
"grad_norm": 0.8055832386016846,
"learning_rate": 7.221558531769519e-06,
"loss": 0.1841,
"step": 820
},
{
"epoch": 3.5255788906838985,
"grad_norm": 0.8355058431625366,
"learning_rate": 7.21549083794466e-06,
"loss": 0.1834,
"step": 821
},
{
"epoch": 3.529886914378029,
"grad_norm": 0.7648366093635559,
"learning_rate": 7.209419081617055e-06,
"loss": 0.1836,
"step": 822
},
{
"epoch": 3.5341949380721593,
"grad_norm": 0.8649141192436218,
"learning_rate": 7.203343273920365e-06,
"loss": 0.1928,
"step": 823
},
{
"epoch": 3.53850296176629,
"grad_norm": 0.7186275720596313,
"learning_rate": 7.197263425995682e-06,
"loss": 0.1767,
"step": 824
},
{
"epoch": 3.54281098546042,
"grad_norm": 0.7774980068206787,
"learning_rate": 7.191179548991507e-06,
"loss": 0.1797,
"step": 825
},
{
"epoch": 3.5471190091545504,
"grad_norm": 0.7177664041519165,
"learning_rate": 7.185091654063724e-06,
"loss": 0.1845,
"step": 826
},
{
"epoch": 3.5514270328486806,
"grad_norm": 0.7631592750549316,
"learning_rate": 7.1789997523755915e-06,
"loss": 0.1785,
"step": 827
},
{
"epoch": 3.5557350565428107,
"grad_norm": 0.7096567749977112,
"learning_rate": 7.172903855097712e-06,
"loss": 0.1893,
"step": 828
},
{
"epoch": 3.5600430802369414,
"grad_norm": 0.800830066204071,
"learning_rate": 7.166803973408012e-06,
"loss": 0.1864,
"step": 829
},
{
"epoch": 3.5643511039310716,
"grad_norm": 0.7556864619255066,
"learning_rate": 7.160700118491729e-06,
"loss": 0.1876,
"step": 830
},
{
"epoch": 3.568659127625202,
"grad_norm": 0.7628764510154724,
"learning_rate": 7.154592301541383e-06,
"loss": 0.1873,
"step": 831
},
{
"epoch": 3.5729671513193324,
"grad_norm": 0.7047873139381409,
"learning_rate": 7.148480533756759e-06,
"loss": 0.1781,
"step": 832
},
{
"epoch": 3.5772751750134626,
"grad_norm": 0.7988886833190918,
"learning_rate": 7.142364826344891e-06,
"loss": 0.1856,
"step": 833
},
{
"epoch": 3.581583198707593,
"grad_norm": 0.7696588635444641,
"learning_rate": 7.1362451905200285e-06,
"loss": 0.1739,
"step": 834
},
{
"epoch": 3.585891222401723,
"grad_norm": 0.8719116449356079,
"learning_rate": 7.130121637503633e-06,
"loss": 0.1937,
"step": 835
},
{
"epoch": 3.5901992460958536,
"grad_norm": 0.8098923563957214,
"learning_rate": 7.123994178524345e-06,
"loss": 0.1859,
"step": 836
},
{
"epoch": 3.594507269789984,
"grad_norm": 0.8798801898956299,
"learning_rate": 7.117862824817966e-06,
"loss": 0.1802,
"step": 837
},
{
"epoch": 3.598815293484114,
"grad_norm": 0.8650680184364319,
"learning_rate": 7.1117275876274425e-06,
"loss": 0.1806,
"step": 838
},
{
"epoch": 3.6031233171782446,
"grad_norm": 0.8523943424224854,
"learning_rate": 7.105588478202838e-06,
"loss": 0.1752,
"step": 839
},
{
"epoch": 3.607431340872375,
"grad_norm": 0.7416144609451294,
"learning_rate": 7.099445507801324e-06,
"loss": 0.1747,
"step": 840
},
{
"epoch": 3.611739364566505,
"grad_norm": 0.7997528910636902,
"learning_rate": 7.093298687687141e-06,
"loss": 0.1785,
"step": 841
},
{
"epoch": 3.616047388260635,
"grad_norm": 0.8117310404777527,
"learning_rate": 7.0871480291315975e-06,
"loss": 0.179,
"step": 842
},
{
"epoch": 3.620355411954766,
"grad_norm": 0.8026857972145081,
"learning_rate": 7.080993543413035e-06,
"loss": 0.1905,
"step": 843
},
{
"epoch": 3.624663435648896,
"grad_norm": 0.7391301393508911,
"learning_rate": 7.0748352418168174e-06,
"loss": 0.1751,
"step": 844
},
{
"epoch": 3.6289714593430262,
"grad_norm": 0.7324339747428894,
"learning_rate": 7.068673135635302e-06,
"loss": 0.177,
"step": 845
},
{
"epoch": 3.633279483037157,
"grad_norm": 0.752116858959198,
"learning_rate": 7.062507236167826e-06,
"loss": 0.1919,
"step": 846
},
{
"epoch": 3.637587506731287,
"grad_norm": 0.8739787340164185,
"learning_rate": 7.056337554720676e-06,
"loss": 0.18,
"step": 847
},
{
"epoch": 3.6418955304254172,
"grad_norm": 0.9470150470733643,
"learning_rate": 7.050164102607081e-06,
"loss": 0.1889,
"step": 848
},
{
"epoch": 3.6462035541195474,
"grad_norm": 0.8388432264328003,
"learning_rate": 7.043986891147179e-06,
"loss": 0.1911,
"step": 849
},
{
"epoch": 3.650511577813678,
"grad_norm": 0.7548778057098389,
"learning_rate": 7.037805931668006e-06,
"loss": 0.1833,
"step": 850
},
{
"epoch": 3.6548196015078083,
"grad_norm": 0.7865473628044128,
"learning_rate": 7.031621235503464e-06,
"loss": 0.1824,
"step": 851
},
{
"epoch": 3.6591276252019385,
"grad_norm": 0.7438720464706421,
"learning_rate": 7.025432813994315e-06,
"loss": 0.1812,
"step": 852
},
{
"epoch": 3.663435648896069,
"grad_norm": 0.7457842230796814,
"learning_rate": 7.019240678488145e-06,
"loss": 0.1948,
"step": 853
},
{
"epoch": 3.6677436725901993,
"grad_norm": 0.7958986759185791,
"learning_rate": 7.013044840339353e-06,
"loss": 0.1845,
"step": 854
},
{
"epoch": 3.6720516962843295,
"grad_norm": 0.7546960711479187,
"learning_rate": 7.006845310909131e-06,
"loss": 0.1807,
"step": 855
},
{
"epoch": 3.6763597199784597,
"grad_norm": 0.7642547488212585,
"learning_rate": 7.000642101565434e-06,
"loss": 0.175,
"step": 856
},
{
"epoch": 3.6806677436725903,
"grad_norm": 0.6758170127868652,
"learning_rate": 6.994435223682966e-06,
"loss": 0.1836,
"step": 857
},
{
"epoch": 3.6849757673667205,
"grad_norm": 0.7619655728340149,
"learning_rate": 6.9882246886431615e-06,
"loss": 0.1881,
"step": 858
},
{
"epoch": 3.6892837910608507,
"grad_norm": 0.7106654644012451,
"learning_rate": 6.982010507834158e-06,
"loss": 0.182,
"step": 859
},
{
"epoch": 3.6935918147549813,
"grad_norm": 0.7118704319000244,
"learning_rate": 6.975792692650778e-06,
"loss": 0.1894,
"step": 860
},
{
"epoch": 3.6978998384491115,
"grad_norm": 0.8190689086914062,
"learning_rate": 6.969571254494509e-06,
"loss": 0.1806,
"step": 861
},
{
"epoch": 3.7022078621432417,
"grad_norm": 0.8226962089538574,
"learning_rate": 6.963346204773483e-06,
"loss": 0.1784,
"step": 862
},
{
"epoch": 3.706515885837372,
"grad_norm": 0.6322476863861084,
"learning_rate": 6.957117554902452e-06,
"loss": 0.1706,
"step": 863
},
{
"epoch": 3.7108239095315025,
"grad_norm": 0.8125499486923218,
"learning_rate": 6.950885316302773e-06,
"loss": 0.1934,
"step": 864
},
{
"epoch": 3.7151319332256327,
"grad_norm": 0.8503420948982239,
"learning_rate": 6.94464950040238e-06,
"loss": 0.1833,
"step": 865
},
{
"epoch": 3.719439956919763,
"grad_norm": 0.7436283826828003,
"learning_rate": 6.938410118635768e-06,
"loss": 0.1784,
"step": 866
},
{
"epoch": 3.7237479806138936,
"grad_norm": 0.7340200543403625,
"learning_rate": 6.9321671824439715e-06,
"loss": 0.1855,
"step": 867
},
{
"epoch": 3.7280560043080238,
"grad_norm": 0.7100676894187927,
"learning_rate": 6.9259207032745415e-06,
"loss": 0.1699,
"step": 868
},
{
"epoch": 3.732364028002154,
"grad_norm": 0.8664551973342896,
"learning_rate": 6.919670692581526e-06,
"loss": 0.1914,
"step": 869
},
{
"epoch": 3.736672051696284,
"grad_norm": 0.7995779514312744,
"learning_rate": 6.913417161825449e-06,
"loss": 0.1768,
"step": 870
},
{
"epoch": 3.7409800753904148,
"grad_norm": 0.7231603264808655,
"learning_rate": 6.907160122473291e-06,
"loss": 0.1843,
"step": 871
},
{
"epoch": 3.745288099084545,
"grad_norm": 0.7902730107307434,
"learning_rate": 6.90089958599846e-06,
"loss": 0.1811,
"step": 872
},
{
"epoch": 3.749596122778675,
"grad_norm": 0.7823545336723328,
"learning_rate": 6.894635563880785e-06,
"loss": 0.1795,
"step": 873
},
{
"epoch": 3.753904146472806,
"grad_norm": 0.7500615119934082,
"learning_rate": 6.88836806760648e-06,
"loss": 0.1758,
"step": 874
},
{
"epoch": 3.758212170166936,
"grad_norm": 0.7018880844116211,
"learning_rate": 6.882097108668132e-06,
"loss": 0.1733,
"step": 875
},
{
"epoch": 3.762520193861066,
"grad_norm": 0.7450665235519409,
"learning_rate": 6.875822698564678e-06,
"loss": 0.1676,
"step": 876
},
{
"epoch": 3.7668282175551964,
"grad_norm": 0.7359514832496643,
"learning_rate": 6.869544848801383e-06,
"loss": 0.19,
"step": 877
},
{
"epoch": 3.771136241249327,
"grad_norm": 0.7743359208106995,
"learning_rate": 6.863263570889818e-06,
"loss": 0.1688,
"step": 878
},
{
"epoch": 3.775444264943457,
"grad_norm": 0.6148492097854614,
"learning_rate": 6.85697887634784e-06,
"loss": 0.1873,
"step": 879
},
{
"epoch": 3.7797522886375874,
"grad_norm": 0.7440102696418762,
"learning_rate": 6.850690776699574e-06,
"loss": 0.1903,
"step": 880
},
{
"epoch": 3.784060312331718,
"grad_norm": 0.7168632745742798,
"learning_rate": 6.844399283475384e-06,
"loss": 0.191,
"step": 881
},
{
"epoch": 3.788368336025848,
"grad_norm": 0.6857194900512695,
"learning_rate": 6.838104408211862e-06,
"loss": 0.1831,
"step": 882
},
{
"epoch": 3.7926763597199784,
"grad_norm": 0.6851407885551453,
"learning_rate": 6.831806162451799e-06,
"loss": 0.1842,
"step": 883
},
{
"epoch": 3.7969843834141086,
"grad_norm": 0.8678072094917297,
"learning_rate": 6.825504557744167e-06,
"loss": 0.1892,
"step": 884
},
{
"epoch": 3.8012924071082392,
"grad_norm": 0.7824573516845703,
"learning_rate": 6.819199605644093e-06,
"loss": 0.184,
"step": 885
},
{
"epoch": 3.8056004308023694,
"grad_norm": 0.8031753301620483,
"learning_rate": 6.812891317712851e-06,
"loss": 0.1851,
"step": 886
},
{
"epoch": 3.8099084544964996,
"grad_norm": 0.747918963432312,
"learning_rate": 6.806579705517824e-06,
"loss": 0.1791,
"step": 887
},
{
"epoch": 3.8142164781906303,
"grad_norm": 0.6846821904182434,
"learning_rate": 6.800264780632495e-06,
"loss": 0.1867,
"step": 888
},
{
"epoch": 3.8185245018847604,
"grad_norm": 0.6762527227401733,
"learning_rate": 6.793946554636417e-06,
"loss": 0.1946,
"step": 889
},
{
"epoch": 3.8228325255788906,
"grad_norm": 0.8078494071960449,
"learning_rate": 6.7876250391152e-06,
"loss": 0.1811,
"step": 890
},
{
"epoch": 3.827140549273021,
"grad_norm": 0.709995687007904,
"learning_rate": 6.781300245660487e-06,
"loss": 0.177,
"step": 891
},
{
"epoch": 3.8314485729671515,
"grad_norm": 0.7268922328948975,
"learning_rate": 6.774972185869928e-06,
"loss": 0.1909,
"step": 892
},
{
"epoch": 3.8357565966612817,
"grad_norm": 0.7240892052650452,
"learning_rate": 6.768640871347163e-06,
"loss": 0.1906,
"step": 893
},
{
"epoch": 3.840064620355412,
"grad_norm": 0.7050197720527649,
"learning_rate": 6.762306313701803e-06,
"loss": 0.1792,
"step": 894
},
{
"epoch": 3.8443726440495425,
"grad_norm": 0.7163313031196594,
"learning_rate": 6.7559685245494025e-06,
"loss": 0.1844,
"step": 895
},
{
"epoch": 3.8486806677436727,
"grad_norm": 0.7890435457229614,
"learning_rate": 6.749627515511443e-06,
"loss": 0.1886,
"step": 896
},
{
"epoch": 3.852988691437803,
"grad_norm": 0.7992168068885803,
"learning_rate": 6.743283298215312e-06,
"loss": 0.1907,
"step": 897
},
{
"epoch": 3.857296715131933,
"grad_norm": 0.904237687587738,
"learning_rate": 6.736935884294275e-06,
"loss": 0.1774,
"step": 898
},
{
"epoch": 3.8616047388260637,
"grad_norm": 0.7321799397468567,
"learning_rate": 6.730585285387465e-06,
"loss": 0.18,
"step": 899
},
{
"epoch": 3.865912762520194,
"grad_norm": 0.7467935681343079,
"learning_rate": 6.724231513139853e-06,
"loss": 0.1685,
"step": 900
},
{
"epoch": 3.870220786214324,
"grad_norm": 0.7189427614212036,
"learning_rate": 6.717874579202227e-06,
"loss": 0.1807,
"step": 901
},
{
"epoch": 3.8745288099084547,
"grad_norm": 0.810021162033081,
"learning_rate": 6.711514495231173e-06,
"loss": 0.1857,
"step": 902
},
{
"epoch": 3.878836833602585,
"grad_norm": 0.7699490785598755,
"learning_rate": 6.705151272889055e-06,
"loss": 0.1804,
"step": 903
},
{
"epoch": 3.883144857296715,
"grad_norm": 0.723810076713562,
"learning_rate": 6.698784923843993e-06,
"loss": 0.1767,
"step": 904
},
{
"epoch": 3.8874528809908453,
"grad_norm": 0.7115669250488281,
"learning_rate": 6.692415459769835e-06,
"loss": 0.1681,
"step": 905
},
{
"epoch": 3.891760904684976,
"grad_norm": 0.7950160503387451,
"learning_rate": 6.686042892346147e-06,
"loss": 0.1868,
"step": 906
},
{
"epoch": 3.896068928379106,
"grad_norm": 0.7404954433441162,
"learning_rate": 6.679667233258179e-06,
"loss": 0.1793,
"step": 907
},
{
"epoch": 3.9003769520732363,
"grad_norm": 0.7232417464256287,
"learning_rate": 6.673288494196858e-06,
"loss": 0.1698,
"step": 908
},
{
"epoch": 3.904684975767367,
"grad_norm": 0.6922205090522766,
"learning_rate": 6.666906686858753e-06,
"loss": 0.1878,
"step": 909
},
{
"epoch": 3.908992999461497,
"grad_norm": 0.8065887093544006,
"learning_rate": 6.66052182294606e-06,
"loss": 0.172,
"step": 910
},
{
"epoch": 3.9133010231556273,
"grad_norm": 0.6786320805549622,
"learning_rate": 6.654133914166582e-06,
"loss": 0.1818,
"step": 911
},
{
"epoch": 3.9176090468497575,
"grad_norm": 0.9004634022712708,
"learning_rate": 6.647742972233703e-06,
"loss": 0.1814,
"step": 912
},
{
"epoch": 3.9219170705438877,
"grad_norm": 0.928763747215271,
"learning_rate": 6.641349008866369e-06,
"loss": 0.1819,
"step": 913
},
{
"epoch": 3.9262250942380184,
"grad_norm": 0.8380340337753296,
"learning_rate": 6.634952035789069e-06,
"loss": 0.1799,
"step": 914
},
{
"epoch": 3.9305331179321485,
"grad_norm": 0.7506704330444336,
"learning_rate": 6.628552064731807e-06,
"loss": 0.1777,
"step": 915
},
{
"epoch": 3.934841141626279,
"grad_norm": 0.7512242794036865,
"learning_rate": 6.622149107430088e-06,
"loss": 0.1745,
"step": 916
},
{
"epoch": 3.9391491653204094,
"grad_norm": 0.817297101020813,
"learning_rate": 6.6157431756248906e-06,
"loss": 0.1851,
"step": 917
},
{
"epoch": 3.9434571890145396,
"grad_norm": 0.817671537399292,
"learning_rate": 6.609334281062647e-06,
"loss": 0.1791,
"step": 918
},
{
"epoch": 3.9477652127086698,
"grad_norm": 0.8131271004676819,
"learning_rate": 6.602922435495225e-06,
"loss": 0.18,
"step": 919
},
{
"epoch": 3.9520732364028,
"grad_norm": 0.7726566791534424,
"learning_rate": 6.5965076506799e-06,
"loss": 0.1738,
"step": 920
},
{
"epoch": 3.9563812600969306,
"grad_norm": 0.7950050830841064,
"learning_rate": 6.5900899383793415e-06,
"loss": 0.1756,
"step": 921
},
{
"epoch": 3.9606892837910608,
"grad_norm": 0.7116392850875854,
"learning_rate": 6.583669310361583e-06,
"loss": 0.1766,
"step": 922
},
{
"epoch": 3.9649973074851914,
"grad_norm": 0.7608798742294312,
"learning_rate": 6.577245778400006e-06,
"loss": 0.1897,
"step": 923
},
{
"epoch": 3.9693053311793216,
"grad_norm": 0.8389754891395569,
"learning_rate": 6.570819354273317e-06,
"loss": 0.1857,
"step": 924
},
{
"epoch": 3.973613354873452,
"grad_norm": 0.7660545110702515,
"learning_rate": 6.564390049765528e-06,
"loss": 0.1807,
"step": 925
},
{
"epoch": 3.977921378567582,
"grad_norm": 0.736478865146637,
"learning_rate": 6.557957876665926e-06,
"loss": 0.1777,
"step": 926
},
{
"epoch": 3.982229402261712,
"grad_norm": 0.7079669833183289,
"learning_rate": 6.551522846769067e-06,
"loss": 0.1826,
"step": 927
},
{
"epoch": 3.986537425955843,
"grad_norm": 0.7876244187355042,
"learning_rate": 6.545084971874738e-06,
"loss": 0.1795,
"step": 928
},
{
"epoch": 3.990845449649973,
"grad_norm": 0.7773596048355103,
"learning_rate": 6.538644263787948e-06,
"loss": 0.1932,
"step": 929
},
{
"epoch": 3.9951534733441036,
"grad_norm": 0.744175374507904,
"learning_rate": 6.532200734318896e-06,
"loss": 0.1785,
"step": 930
},
{
"epoch": 3.999461497038234,
"grad_norm": 0.6581542491912842,
"learning_rate": 6.525754395282961e-06,
"loss": 0.1776,
"step": 931
},
{
"epoch": 4.0,
"grad_norm": 0.6581542491912842,
"learning_rate": 6.5193052585006666e-06,
"loss": 0.1884,
"step": 932
}
],
"logging_steps": 1,
"max_steps": 2320,
"num_input_tokens_seen": 0,
"num_train_epochs": 10,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 10751167365120.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}