T5La-Large / trainer_state.json
hrezaei's picture
End of training
5354973
Invalid JSON: Unexpected token 'I', ..."ad_norm": Infinity, "... is not valid JSON
{
"best_global_step": null,
"best_metric": 6.391010284423828,
"best_model_checkpoint": null,
"epoch": 1.04632568359375,
"eval_steps": 5000,
"global_step": 524288,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.00095367431640625,
"grad_norm": 0.6624638438224792,
"learning_rate": 4.995241165161133e-05,
"loss": 8.9018,
"step": 500
},
{
"epoch": 0.0019073486328125,
"grad_norm": 0.6431354284286499,
"learning_rate": 4.990472793579102e-05,
"loss": 8.2495,
"step": 1000
},
{
"epoch": 0.00286102294921875,
"grad_norm": 0.5842123031616211,
"learning_rate": 4.98570442199707e-05,
"loss": 7.7924,
"step": 1500
},
{
"epoch": 0.003814697265625,
"grad_norm": 0.5727596282958984,
"learning_rate": 4.9809360504150393e-05,
"loss": 7.5279,
"step": 2000
},
{
"epoch": 0.00476837158203125,
"grad_norm": 0.45176011323928833,
"learning_rate": 4.9761676788330084e-05,
"loss": 7.3262,
"step": 2500
},
{
"epoch": 0.0057220458984375,
"grad_norm": 0.4838062822818756,
"learning_rate": 4.971399307250977e-05,
"loss": 7.219,
"step": 3000
},
{
"epoch": 0.00667572021484375,
"grad_norm": 0.56992506980896,
"learning_rate": 4.966630935668946e-05,
"loss": 7.1168,
"step": 3500
},
{
"epoch": 0.00762939453125,
"grad_norm": 0.47706305980682373,
"learning_rate": 4.961862564086914e-05,
"loss": 7.0533,
"step": 4000
},
{
"epoch": 0.00858306884765625,
"grad_norm": 0.4582304358482361,
"learning_rate": 4.957094192504883e-05,
"loss": 6.9537,
"step": 4500
},
{
"epoch": 0.0095367431640625,
"grad_norm": 0.4827233850955963,
"learning_rate": 4.952325820922852e-05,
"loss": 6.9745,
"step": 5000
},
{
"epoch": 0.0095367431640625,
"eval_accuracy": 0.031426614481409,
"eval_loss": 6.906828880310059,
"eval_runtime": 236.8104,
"eval_samples_per_second": 42.228,
"eval_steps_per_second": 10.557,
"step": 5000
},
{
"epoch": 0.01049041748046875,
"grad_norm": 0.4683932960033417,
"learning_rate": 4.9475574493408205e-05,
"loss": 6.9377,
"step": 5500
},
{
"epoch": 0.011444091796875,
"grad_norm": 0.43654170632362366,
"learning_rate": 4.9427890777587895e-05,
"loss": 6.9642,
"step": 6000
},
{
"epoch": 0.01239776611328125,
"grad_norm": 0.4411381483078003,
"learning_rate": 4.938020706176758e-05,
"loss": 6.9223,
"step": 6500
},
{
"epoch": 0.0133514404296875,
"grad_norm": 0.5460022687911987,
"learning_rate": 4.933252334594727e-05,
"loss": 6.8951,
"step": 7000
},
{
"epoch": 0.01430511474609375,
"grad_norm": 0.48390814661979675,
"learning_rate": 4.928483963012696e-05,
"loss": 6.8717,
"step": 7500
},
{
"epoch": 0.0152587890625,
"grad_norm": 0.5311650037765503,
"learning_rate": 4.923715591430664e-05,
"loss": 6.8812,
"step": 8000
},
{
"epoch": 0.01621246337890625,
"grad_norm": 0.5130866765975952,
"learning_rate": 4.918947219848633e-05,
"loss": 6.8148,
"step": 8500
},
{
"epoch": 0.0171661376953125,
"grad_norm": 0.5294931530952454,
"learning_rate": 4.9141788482666016e-05,
"loss": 6.787,
"step": 9000
},
{
"epoch": 0.01811981201171875,
"grad_norm": 0.544792652130127,
"learning_rate": 4.9094104766845706e-05,
"loss": 6.7906,
"step": 9500
},
{
"epoch": 0.019073486328125,
"grad_norm": 0.568236768245697,
"learning_rate": 4.9046421051025396e-05,
"loss": 6.7778,
"step": 10000
},
{
"epoch": 0.019073486328125,
"eval_accuracy": 0.031314090019569474,
"eval_loss": 6.7409868240356445,
"eval_runtime": 236.3476,
"eval_samples_per_second": 42.311,
"eval_steps_per_second": 10.578,
"step": 10000
},
{
"epoch": 0.02002716064453125,
"grad_norm": 0.7601892352104187,
"learning_rate": 4.899873733520508e-05,
"loss": 6.7391,
"step": 10500
},
{
"epoch": 0.0209808349609375,
"grad_norm": 0.698943555355072,
"learning_rate": 4.895105361938477e-05,
"loss": 6.757,
"step": 11000
},
{
"epoch": 0.02193450927734375,
"grad_norm": 0.6955975294113159,
"learning_rate": 4.890336990356445e-05,
"loss": 6.7612,
"step": 11500
},
{
"epoch": 0.02288818359375,
"grad_norm": 0.625805675983429,
"learning_rate": 4.8855686187744143e-05,
"loss": 6.808,
"step": 12000
},
{
"epoch": 0.02384185791015625,
"grad_norm": 0.8192827701568604,
"learning_rate": 4.8808002471923834e-05,
"loss": 6.7573,
"step": 12500
},
{
"epoch": 0.0247955322265625,
"grad_norm": 0.9873795509338379,
"learning_rate": 4.876031875610352e-05,
"loss": 6.7737,
"step": 13000
},
{
"epoch": 0.02574920654296875,
"grad_norm": 1.016819953918457,
"learning_rate": 4.871263504028321e-05,
"loss": 6.6898,
"step": 13500
},
{
"epoch": 0.026702880859375,
"grad_norm": 1.170738935470581,
"learning_rate": 4.866495132446289e-05,
"loss": 6.7066,
"step": 14000
},
{
"epoch": 0.02765655517578125,
"grad_norm": 2.338644027709961,
"learning_rate": 4.861726760864258e-05,
"loss": 6.7263,
"step": 14500
},
{
"epoch": 0.0286102294921875,
"grad_norm": 1.3942464590072632,
"learning_rate": 4.856958389282227e-05,
"loss": 6.715,
"step": 15000
},
{
"epoch": 0.0286102294921875,
"eval_accuracy": 0.03333365949119374,
"eval_loss": 6.643101692199707,
"eval_runtime": 240.8417,
"eval_samples_per_second": 41.521,
"eval_steps_per_second": 10.38,
"step": 15000
},
{
"epoch": 0.02956390380859375,
"grad_norm": 1.3489205837249756,
"learning_rate": 4.8521900177001955e-05,
"loss": 6.7263,
"step": 15500
},
{
"epoch": 0.030517578125,
"grad_norm": 1.4647587537765503,
"learning_rate": 4.8474216461181645e-05,
"loss": 6.6834,
"step": 16000
},
{
"epoch": 0.03147125244140625,
"grad_norm": 1.9520461559295654,
"learning_rate": 4.842653274536133e-05,
"loss": 6.6735,
"step": 16500
},
{
"epoch": 0.0324249267578125,
"grad_norm": 1.378394365310669,
"learning_rate": 4.837884902954102e-05,
"loss": 6.6402,
"step": 17000
},
{
"epoch": 0.03337860107421875,
"grad_norm": 1.9632004499435425,
"learning_rate": 4.833116531372071e-05,
"loss": 6.6743,
"step": 17500
},
{
"epoch": 0.034332275390625,
"grad_norm": 2.1774470806121826,
"learning_rate": 4.828348159790039e-05,
"loss": 6.71,
"step": 18000
},
{
"epoch": 0.03528594970703125,
"grad_norm": 2.1665990352630615,
"learning_rate": 4.823579788208008e-05,
"loss": 6.6658,
"step": 18500
},
{
"epoch": 0.0362396240234375,
"grad_norm": 1.8897229433059692,
"learning_rate": 4.8188114166259766e-05,
"loss": 6.6589,
"step": 19000
},
{
"epoch": 0.03719329833984375,
"grad_norm": 3.076052188873291,
"learning_rate": 4.8140430450439456e-05,
"loss": 6.6456,
"step": 19500
},
{
"epoch": 0.03814697265625,
"grad_norm": 2.2930281162261963,
"learning_rate": 4.8092746734619146e-05,
"loss": 6.6547,
"step": 20000
},
{
"epoch": 0.03814697265625,
"eval_accuracy": 0.0334894324853229,
"eval_loss": 6.578402996063232,
"eval_runtime": 239.0673,
"eval_samples_per_second": 41.829,
"eval_steps_per_second": 10.457,
"step": 20000
},
{
"epoch": 0.03910064697265625,
"grad_norm": 4.879852771759033,
"learning_rate": 4.804506301879883e-05,
"loss": 6.6624,
"step": 20500
},
{
"epoch": 0.0400543212890625,
"grad_norm": 4.431488037109375,
"learning_rate": 4.799737930297852e-05,
"loss": 6.6653,
"step": 21000
},
{
"epoch": 0.04100799560546875,
"grad_norm": 2.174495220184326,
"learning_rate": 4.79496955871582e-05,
"loss": 6.6332,
"step": 21500
},
{
"epoch": 0.041961669921875,
"grad_norm": 3.470643997192383,
"learning_rate": 4.7902011871337893e-05,
"loss": 6.6515,
"step": 22000
},
{
"epoch": 0.04291534423828125,
"grad_norm": 3.1371328830718994,
"learning_rate": 4.7854328155517584e-05,
"loss": 6.6232,
"step": 22500
},
{
"epoch": 0.0438690185546875,
"grad_norm": 4.542148113250732,
"learning_rate": 4.780664443969727e-05,
"loss": 6.6196,
"step": 23000
},
{
"epoch": 0.04482269287109375,
"grad_norm": 2.4484500885009766,
"learning_rate": 4.775896072387696e-05,
"loss": 6.5961,
"step": 23500
},
{
"epoch": 0.0457763671875,
"grad_norm": 3.348109006881714,
"learning_rate": 4.771127700805664e-05,
"loss": 6.6348,
"step": 24000
},
{
"epoch": 0.04673004150390625,
"grad_norm": 6.815830707550049,
"learning_rate": 4.766359329223633e-05,
"loss": 6.645,
"step": 24500
},
{
"epoch": 0.0476837158203125,
"grad_norm": 4.380649566650391,
"learning_rate": 4.761590957641602e-05,
"loss": 6.5695,
"step": 25000
},
{
"epoch": 0.0476837158203125,
"eval_accuracy": 0.03409236790606654,
"eval_loss": 6.529794692993164,
"eval_runtime": 239.0015,
"eval_samples_per_second": 41.841,
"eval_steps_per_second": 10.46,
"step": 25000
},
{
"epoch": 0.04863739013671875,
"grad_norm": 4.102541446685791,
"learning_rate": 4.7568225860595705e-05,
"loss": 6.6159,
"step": 25500
},
{
"epoch": 0.049591064453125,
"grad_norm": 4.208073139190674,
"learning_rate": 4.7520542144775395e-05,
"loss": 6.5972,
"step": 26000
},
{
"epoch": 0.05054473876953125,
"grad_norm": 4.404722690582275,
"learning_rate": 4.747285842895508e-05,
"loss": 6.5881,
"step": 26500
},
{
"epoch": 0.0514984130859375,
"grad_norm": 4.774337291717529,
"learning_rate": 4.742517471313477e-05,
"loss": 6.6055,
"step": 27000
},
{
"epoch": 0.05245208740234375,
"grad_norm": 4.009048938751221,
"learning_rate": 4.737749099731446e-05,
"loss": 6.5879,
"step": 27500
},
{
"epoch": 0.05340576171875,
"grad_norm": 7.490168571472168,
"learning_rate": 4.732980728149414e-05,
"loss": 6.5723,
"step": 28000
},
{
"epoch": 0.05435943603515625,
"grad_norm": 12.653952598571777,
"learning_rate": 4.728212356567383e-05,
"loss": 6.5784,
"step": 28500
},
{
"epoch": 0.0553131103515625,
"grad_norm": 6.196599960327148,
"learning_rate": 4.7234439849853516e-05,
"loss": 6.5307,
"step": 29000
},
{
"epoch": 0.05626678466796875,
"grad_norm": 8.773276329040527,
"learning_rate": 4.7186756134033206e-05,
"loss": 6.5646,
"step": 29500
},
{
"epoch": 0.057220458984375,
"grad_norm": 6.8058929443359375,
"learning_rate": 4.7139072418212896e-05,
"loss": 6.5283,
"step": 30000
},
{
"epoch": 0.057220458984375,
"eval_accuracy": 0.03399275929549902,
"eval_loss": 6.526458740234375,
"eval_runtime": 241.0876,
"eval_samples_per_second": 41.479,
"eval_steps_per_second": 10.37,
"step": 30000
},
{
"epoch": 0.05817413330078125,
"grad_norm": 9.646665573120117,
"learning_rate": 4.709138870239258e-05,
"loss": 6.5259,
"step": 30500
},
{
"epoch": 0.0591278076171875,
"grad_norm": 11.337198257446289,
"learning_rate": 4.704370498657227e-05,
"loss": 6.5051,
"step": 31000
},
{
"epoch": 0.06008148193359375,
"grad_norm": 17.71413803100586,
"learning_rate": 4.699602127075195e-05,
"loss": 6.6211,
"step": 31500
},
{
"epoch": 0.06103515625,
"grad_norm": 7.909314155578613,
"learning_rate": 4.6948337554931643e-05,
"loss": 6.6081,
"step": 32000
},
{
"epoch": 0.06198883056640625,
"grad_norm": 10.922993659973145,
"learning_rate": 4.6900653839111334e-05,
"loss": 6.5762,
"step": 32500
},
{
"epoch": 0.0629425048828125,
"grad_norm": 17.44997215270996,
"learning_rate": 4.685297012329102e-05,
"loss": 6.5255,
"step": 33000
},
{
"epoch": 0.06389617919921875,
"grad_norm": 9.439545631408691,
"learning_rate": 4.680528640747071e-05,
"loss": 6.6299,
"step": 33500
},
{
"epoch": 0.064849853515625,
"grad_norm": 14.736495018005371,
"learning_rate": 4.675760269165039e-05,
"loss": 6.5564,
"step": 34000
},
{
"epoch": 0.06580352783203125,
"grad_norm": 11.100777626037598,
"learning_rate": 4.670991897583008e-05,
"loss": 6.5923,
"step": 34500
},
{
"epoch": 0.0667572021484375,
"grad_norm": 15.978229522705078,
"learning_rate": 4.666223526000977e-05,
"loss": 6.5826,
"step": 35000
},
{
"epoch": 0.0667572021484375,
"eval_accuracy": 0.03404227005870841,
"eval_loss": 6.490207195281982,
"eval_runtime": 239.3878,
"eval_samples_per_second": 41.773,
"eval_steps_per_second": 10.443,
"step": 35000
},
{
"epoch": 0.06771087646484375,
"grad_norm": 9.80553150177002,
"learning_rate": 4.6614551544189455e-05,
"loss": 6.574,
"step": 35500
},
{
"epoch": 0.06866455078125,
"grad_norm": 9.590557098388672,
"learning_rate": 4.6566867828369145e-05,
"loss": 6.5776,
"step": 36000
},
{
"epoch": 0.06961822509765625,
"grad_norm": 13.170746803283691,
"learning_rate": 4.651918411254883e-05,
"loss": 6.6031,
"step": 36500
},
{
"epoch": 0.0705718994140625,
"grad_norm": 17.351354598999023,
"learning_rate": 4.647150039672852e-05,
"loss": 6.5395,
"step": 37000
},
{
"epoch": 0.07152557373046875,
"grad_norm": 11.295857429504395,
"learning_rate": 4.642381668090821e-05,
"loss": 6.5644,
"step": 37500
},
{
"epoch": 0.072479248046875,
"grad_norm": 13.997299194335938,
"learning_rate": 4.637613296508789e-05,
"loss": 6.57,
"step": 38000
},
{
"epoch": 0.07343292236328125,
"grad_norm": 10.547431945800781,
"learning_rate": 4.632844924926758e-05,
"loss": 6.5339,
"step": 38500
},
{
"epoch": 0.0743865966796875,
"grad_norm": 11.340121269226074,
"learning_rate": 4.6280765533447266e-05,
"loss": 6.5829,
"step": 39000
},
{
"epoch": 0.07534027099609375,
"grad_norm": 11.731833457946777,
"learning_rate": 4.6233081817626956e-05,
"loss": 6.5487,
"step": 39500
},
{
"epoch": 0.0762939453125,
"grad_norm": 21.85376739501953,
"learning_rate": 4.6185398101806646e-05,
"loss": 6.5631,
"step": 40000
},
{
"epoch": 0.0762939453125,
"eval_accuracy": 0.03568610567514677,
"eval_loss": 6.4704718589782715,
"eval_runtime": 253.9333,
"eval_samples_per_second": 39.38,
"eval_steps_per_second": 9.845,
"step": 40000
},
{
"epoch": 0.07724761962890625,
"grad_norm": 25.65736961364746,
"learning_rate": 4.613771438598633e-05,
"loss": 6.5646,
"step": 40500
},
{
"epoch": 0.0782012939453125,
"grad_norm": 13.945578575134277,
"learning_rate": 4.609003067016602e-05,
"loss": 6.5697,
"step": 41000
},
{
"epoch": 0.07915496826171875,
"grad_norm": 19.017955780029297,
"learning_rate": 4.60423469543457e-05,
"loss": 6.5707,
"step": 41500
},
{
"epoch": 0.080108642578125,
"grad_norm": 32.384315490722656,
"learning_rate": 4.5994663238525393e-05,
"loss": 6.571,
"step": 42000
},
{
"epoch": 0.08106231689453125,
"grad_norm": 21.34004783630371,
"learning_rate": 4.5946979522705084e-05,
"loss": 6.5202,
"step": 42500
},
{
"epoch": 0.0820159912109375,
"grad_norm": 14.851140022277832,
"learning_rate": 4.589929580688477e-05,
"loss": 6.5567,
"step": 43000
},
{
"epoch": 0.08296966552734375,
"grad_norm": 39.20492935180664,
"learning_rate": 4.585161209106446e-05,
"loss": 6.531,
"step": 43500
},
{
"epoch": 0.08392333984375,
"grad_norm": 19.704784393310547,
"learning_rate": 4.580392837524414e-05,
"loss": 6.5621,
"step": 44000
},
{
"epoch": 0.08487701416015625,
"grad_norm": 19.00480079650879,
"learning_rate": 4.575624465942383e-05,
"loss": 6.5331,
"step": 44500
},
{
"epoch": 0.0858306884765625,
"grad_norm": 19.51432228088379,
"learning_rate": 4.570856094360352e-05,
"loss": 6.5234,
"step": 45000
},
{
"epoch": 0.0858306884765625,
"eval_accuracy": 0.0339426614481409,
"eval_loss": 6.461673736572266,
"eval_runtime": 240.3668,
"eval_samples_per_second": 41.603,
"eval_steps_per_second": 10.401,
"step": 45000
},
{
"epoch": 0.08678436279296875,
"grad_norm": 21.277442932128906,
"learning_rate": 4.5660877227783205e-05,
"loss": 6.5399,
"step": 45500
},
{
"epoch": 0.087738037109375,
"grad_norm": 13.337013244628906,
"learning_rate": 4.5613193511962895e-05,
"loss": 6.5251,
"step": 46000
},
{
"epoch": 0.08869171142578125,
"grad_norm": 23.25571632385254,
"learning_rate": 4.556550979614258e-05,
"loss": 6.496,
"step": 46500
},
{
"epoch": 0.0896453857421875,
"grad_norm": 29.124269485473633,
"learning_rate": 4.551782608032227e-05,
"loss": 6.5393,
"step": 47000
},
{
"epoch": 0.09059906005859375,
"grad_norm": 26.739830017089844,
"learning_rate": 4.547014236450196e-05,
"loss": 6.4687,
"step": 47500
},
{
"epoch": 0.091552734375,
"grad_norm": 10.364956855773926,
"learning_rate": 4.542245864868164e-05,
"loss": 6.4882,
"step": 48000
},
{
"epoch": 0.09250640869140625,
"grad_norm": 10.050714492797852,
"learning_rate": 4.537477493286133e-05,
"loss": 6.4549,
"step": 48500
},
{
"epoch": 0.0934600830078125,
"grad_norm": 11.470427513122559,
"learning_rate": 4.5327091217041016e-05,
"loss": 6.504,
"step": 49000
},
{
"epoch": 0.09441375732421875,
"grad_norm": 14.701976776123047,
"learning_rate": 4.5279407501220706e-05,
"loss": 6.5087,
"step": 49500
},
{
"epoch": 0.095367431640625,
"grad_norm": 23.34058952331543,
"learning_rate": 4.523172378540039e-05,
"loss": 6.548,
"step": 50000
},
{
"epoch": 0.095367431640625,
"eval_accuracy": 0.03502857142857143,
"eval_loss": 6.44745397567749,
"eval_runtime": 236.7367,
"eval_samples_per_second": 42.241,
"eval_steps_per_second": 10.56,
"step": 50000
},
{
"epoch": 0.09632110595703125,
"grad_norm": 18.535306930541992,
"learning_rate": 4.518404006958008e-05,
"loss": 6.5755,
"step": 50500
},
{
"epoch": 0.0972747802734375,
"grad_norm": 23.872713088989258,
"learning_rate": 4.513635635375977e-05,
"loss": 6.5359,
"step": 51000
},
{
"epoch": 0.09822845458984375,
"grad_norm": 14.181577682495117,
"learning_rate": 4.508867263793945e-05,
"loss": 6.5603,
"step": 51500
},
{
"epoch": 0.09918212890625,
"grad_norm": 29.683629989624023,
"learning_rate": 4.5040988922119143e-05,
"loss": 6.5151,
"step": 52000
},
{
"epoch": 0.10013580322265625,
"grad_norm": 20.120813369750977,
"learning_rate": 4.499330520629883e-05,
"loss": 6.5369,
"step": 52500
},
{
"epoch": 0.1010894775390625,
"grad_norm": 46.01066589355469,
"learning_rate": 4.494562149047852e-05,
"loss": 6.5455,
"step": 53000
},
{
"epoch": 0.10204315185546875,
"grad_norm": 171.9969024658203,
"learning_rate": 4.489793777465821e-05,
"loss": 6.5004,
"step": 53500
},
{
"epoch": 0.102996826171875,
"grad_norm": 17.876941680908203,
"learning_rate": 4.485025405883789e-05,
"loss": 6.5654,
"step": 54000
},
{
"epoch": 0.10395050048828125,
"grad_norm": 55.01047134399414,
"learning_rate": 4.480257034301758e-05,
"loss": 6.4845,
"step": 54500
},
{
"epoch": 0.1049041748046875,
"grad_norm": 41.43305587768555,
"learning_rate": 4.4754886627197264e-05,
"loss": 6.599,
"step": 55000
},
{
"epoch": 0.1049041748046875,
"eval_accuracy": 0.035460665362035224,
"eval_loss": 6.4514336585998535,
"eval_runtime": 239.1754,
"eval_samples_per_second": 41.81,
"eval_steps_per_second": 10.453,
"step": 55000
},
{
"epoch": 0.10585784912109375,
"grad_norm": 65.93293762207031,
"learning_rate": 4.4707202911376955e-05,
"loss": 6.5473,
"step": 55500
},
{
"epoch": 0.1068115234375,
"grad_norm": 52.11703872680664,
"learning_rate": 4.4659519195556645e-05,
"loss": 6.5312,
"step": 56000
},
{
"epoch": 0.10776519775390625,
"grad_norm": 32.834877014160156,
"learning_rate": 4.461183547973633e-05,
"loss": 6.5875,
"step": 56500
},
{
"epoch": 0.1087188720703125,
"grad_norm": 35.849151611328125,
"learning_rate": 4.456415176391602e-05,
"loss": 6.5532,
"step": 57000
},
{
"epoch": 0.10967254638671875,
"grad_norm": 44.165794372558594,
"learning_rate": 4.45164680480957e-05,
"loss": 6.5122,
"step": 57500
},
{
"epoch": 0.110626220703125,
"grad_norm": 22.337329864501953,
"learning_rate": 4.446878433227539e-05,
"loss": 6.3647,
"step": 58000
},
{
"epoch": 0.11157989501953125,
"grad_norm": 21.325393676757812,
"learning_rate": 4.442110061645508e-05,
"loss": 6.5854,
"step": 58500
},
{
"epoch": 0.1125335693359375,
"grad_norm": 49.023765563964844,
"learning_rate": 4.4373416900634766e-05,
"loss": 6.5848,
"step": 59000
},
{
"epoch": 0.11348724365234375,
"grad_norm": 28.078752517700195,
"learning_rate": 4.4325733184814456e-05,
"loss": 6.5471,
"step": 59500
},
{
"epoch": 0.11444091796875,
"grad_norm": 73.4981689453125,
"learning_rate": 4.427804946899414e-05,
"loss": 6.547,
"step": 60000
},
{
"epoch": 0.11444091796875,
"eval_accuracy": 0.03644618395303327,
"eval_loss": 6.456771373748779,
"eval_runtime": 237.6709,
"eval_samples_per_second": 42.075,
"eval_steps_per_second": 10.519,
"step": 60000
},
{
"epoch": 0.11539459228515625,
"grad_norm": 18.256385803222656,
"learning_rate": 4.423036575317383e-05,
"loss": 6.5359,
"step": 60500
},
{
"epoch": 0.1163482666015625,
"grad_norm": 42.84888458251953,
"learning_rate": 4.418268203735352e-05,
"loss": 6.533,
"step": 61000
},
{
"epoch": 0.11730194091796875,
"grad_norm": 23.927953720092773,
"learning_rate": 4.41349983215332e-05,
"loss": 6.5096,
"step": 61500
},
{
"epoch": 0.118255615234375,
"grad_norm": 45.29853057861328,
"learning_rate": 4.4087314605712893e-05,
"loss": 6.5559,
"step": 62000
},
{
"epoch": 0.11920928955078125,
"grad_norm": 26.264976501464844,
"learning_rate": 4.403963088989258e-05,
"loss": 6.5381,
"step": 62500
},
{
"epoch": 0.1201629638671875,
"grad_norm": 44.30765914916992,
"learning_rate": 4.399194717407227e-05,
"loss": 6.5864,
"step": 63000
},
{
"epoch": 0.12111663818359375,
"grad_norm": 21.413312911987305,
"learning_rate": 4.394426345825196e-05,
"loss": 6.5516,
"step": 63500
},
{
"epoch": 0.1220703125,
"grad_norm": 59.862831115722656,
"learning_rate": 4.389657974243164e-05,
"loss": 6.5113,
"step": 64000
},
{
"epoch": 0.12302398681640625,
"grad_norm": 140.16439819335938,
"learning_rate": 4.384889602661133e-05,
"loss": 6.5812,
"step": 64500
},
{
"epoch": 0.1239776611328125,
"grad_norm": 123.18706512451172,
"learning_rate": 4.3801212310791014e-05,
"loss": 6.5265,
"step": 65000
},
{
"epoch": 0.1239776611328125,
"eval_accuracy": 0.038078277886497064,
"eval_loss": 6.470902442932129,
"eval_runtime": 239.0755,
"eval_samples_per_second": 41.828,
"eval_steps_per_second": 10.457,
"step": 65000
},
{
"epoch": 0.12493133544921875,
"grad_norm": 39.1711540222168,
"learning_rate": 4.3753528594970705e-05,
"loss": 6.5254,
"step": 65500
},
{
"epoch": 0.125885009765625,
"grad_norm": 47.702598571777344,
"learning_rate": 4.3705844879150395e-05,
"loss": 6.5341,
"step": 66000
},
{
"epoch": 0.12683868408203125,
"grad_norm": 28.170778274536133,
"learning_rate": 4.365816116333008e-05,
"loss": 6.4917,
"step": 66500
},
{
"epoch": 0.1277923583984375,
"grad_norm": 34.447105407714844,
"learning_rate": 4.361047744750977e-05,
"loss": 6.515,
"step": 67000
},
{
"epoch": 0.12874603271484375,
"grad_norm": 46.904090881347656,
"learning_rate": 4.356279373168945e-05,
"loss": 6.5513,
"step": 67500
},
{
"epoch": 0.12969970703125,
"grad_norm": 27.520286560058594,
"learning_rate": 4.351511001586914e-05,
"loss": 6.6264,
"step": 68000
},
{
"epoch": 0.13065338134765625,
"grad_norm": 77.6899185180664,
"learning_rate": 4.346742630004883e-05,
"loss": 6.5723,
"step": 68500
},
{
"epoch": 0.1316070556640625,
"grad_norm": 35.40989685058594,
"learning_rate": 4.3419742584228516e-05,
"loss": 6.5746,
"step": 69000
},
{
"epoch": 0.13256072998046875,
"grad_norm": 67.315673828125,
"learning_rate": 4.3372058868408206e-05,
"loss": 6.5677,
"step": 69500
},
{
"epoch": 0.133514404296875,
"grad_norm": 94.40605163574219,
"learning_rate": 4.332437515258789e-05,
"loss": 6.5916,
"step": 70000
},
{
"epoch": 0.133514404296875,
"eval_accuracy": 0.03715283757338552,
"eval_loss": 6.496217250823975,
"eval_runtime": 239.7511,
"eval_samples_per_second": 41.71,
"eval_steps_per_second": 10.427,
"step": 70000
},
{
"epoch": 0.13446807861328125,
"grad_norm": 111.58074188232422,
"learning_rate": 4.327669143676758e-05,
"loss": 6.6048,
"step": 70500
},
{
"epoch": 0.1354217529296875,
"grad_norm": 29.336286544799805,
"learning_rate": 4.322900772094727e-05,
"loss": 6.6133,
"step": 71000
},
{
"epoch": 0.13637542724609375,
"grad_norm": 52.75020217895508,
"learning_rate": 4.318132400512695e-05,
"loss": 6.5921,
"step": 71500
},
{
"epoch": 0.1373291015625,
"grad_norm": 72.22640991210938,
"learning_rate": 4.3133640289306643e-05,
"loss": 6.6073,
"step": 72000
},
{
"epoch": 0.13828277587890625,
"grad_norm": 49.1014289855957,
"learning_rate": 4.308595657348633e-05,
"loss": 6.621,
"step": 72500
},
{
"epoch": 0.1392364501953125,
"grad_norm": 59.5632209777832,
"learning_rate": 4.303827285766602e-05,
"loss": 6.5858,
"step": 73000
},
{
"epoch": 0.14019012451171875,
"grad_norm": 87.36494445800781,
"learning_rate": 4.299058914184571e-05,
"loss": 6.6278,
"step": 73500
},
{
"epoch": 0.141143798828125,
"grad_norm": 32.93346405029297,
"learning_rate": 4.294290542602539e-05,
"loss": 6.5914,
"step": 74000
},
{
"epoch": 0.14209747314453125,
"grad_norm": 52.131343841552734,
"learning_rate": 4.289522171020508e-05,
"loss": 6.5675,
"step": 74500
},
{
"epoch": 0.1430511474609375,
"grad_norm": 97.84684753417969,
"learning_rate": 4.2847537994384764e-05,
"loss": 6.5919,
"step": 75000
},
{
"epoch": 0.1430511474609375,
"eval_accuracy": 0.03751819960861057,
"eval_loss": 6.482682704925537,
"eval_runtime": 237.4453,
"eval_samples_per_second": 42.115,
"eval_steps_per_second": 10.529,
"step": 75000
},
{
"epoch": 0.14400482177734375,
"grad_norm": 46.23030090332031,
"learning_rate": 4.2799854278564455e-05,
"loss": 6.6028,
"step": 75500
},
{
"epoch": 0.14495849609375,
"grad_norm": 73.1805191040039,
"learning_rate": 4.2752170562744145e-05,
"loss": 6.5949,
"step": 76000
},
{
"epoch": 0.14591217041015625,
"grad_norm": 103.25483703613281,
"learning_rate": 4.270448684692383e-05,
"loss": 6.6279,
"step": 76500
},
{
"epoch": 0.1468658447265625,
"grad_norm": 39.62518310546875,
"learning_rate": 4.265680313110352e-05,
"loss": 6.6005,
"step": 77000
},
{
"epoch": 0.14781951904296875,
"grad_norm": 117.60237121582031,
"learning_rate": 4.26091194152832e-05,
"loss": 6.5755,
"step": 77500
},
{
"epoch": 0.148773193359375,
"grad_norm": 93.19673156738281,
"learning_rate": 4.256143569946289e-05,
"loss": 6.5823,
"step": 78000
},
{
"epoch": 0.14972686767578125,
"grad_norm": 24.11821174621582,
"learning_rate": 4.251375198364258e-05,
"loss": 6.6,
"step": 78500
},
{
"epoch": 0.1506805419921875,
"grad_norm": 63.973106384277344,
"learning_rate": 4.2466068267822266e-05,
"loss": 6.592,
"step": 79000
},
{
"epoch": 0.15163421630859375,
"grad_norm": 107.64210510253906,
"learning_rate": 4.2418384552001956e-05,
"loss": 6.5633,
"step": 79500
},
{
"epoch": 0.152587890625,
"grad_norm": 59.94293212890625,
"learning_rate": 4.237070083618164e-05,
"loss": 6.4745,
"step": 80000
},
{
"epoch": 0.152587890625,
"eval_accuracy": 0.038450489236790605,
"eval_loss": 6.490172386169434,
"eval_runtime": 236.0018,
"eval_samples_per_second": 42.373,
"eval_steps_per_second": 10.593,
"step": 80000
},
{
"epoch": 0.15354156494140625,
"grad_norm": 46.570556640625,
"learning_rate": 4.232301712036133e-05,
"loss": 6.5729,
"step": 80500
},
{
"epoch": 0.1544952392578125,
"grad_norm": 77.17676544189453,
"learning_rate": 4.227533340454102e-05,
"loss": 6.5433,
"step": 81000
},
{
"epoch": 0.15544891357421875,
"grad_norm": 156.23880004882812,
"learning_rate": 4.22276496887207e-05,
"loss": 6.5623,
"step": 81500
},
{
"epoch": 0.156402587890625,
"grad_norm": 293.4957275390625,
"learning_rate": 4.2179965972900393e-05,
"loss": 6.5936,
"step": 82000
},
{
"epoch": 0.15735626220703125,
"grad_norm": 91.47918701171875,
"learning_rate": 4.213228225708008e-05,
"loss": 6.5374,
"step": 82500
},
{
"epoch": 0.1583099365234375,
"grad_norm": 123.1861801147461,
"learning_rate": 4.208459854125977e-05,
"loss": 6.5576,
"step": 83000
},
{
"epoch": 0.15926361083984375,
"grad_norm": 75.75604248046875,
"learning_rate": 4.203691482543946e-05,
"loss": 6.5649,
"step": 83500
},
{
"epoch": 0.16021728515625,
"grad_norm": 156.376953125,
"learning_rate": 4.198923110961914e-05,
"loss": 6.5516,
"step": 84000
},
{
"epoch": 0.16117095947265625,
"grad_norm": 198.7659149169922,
"learning_rate": 4.194154739379883e-05,
"loss": 6.5659,
"step": 84500
},
{
"epoch": 0.1621246337890625,
"grad_norm": 89.94020080566406,
"learning_rate": 4.1893863677978514e-05,
"loss": 6.6695,
"step": 85000
},
{
"epoch": 0.1621246337890625,
"eval_accuracy": 0.038353033268101765,
"eval_loss": 6.54418420791626,
"eval_runtime": 241.1387,
"eval_samples_per_second": 41.47,
"eval_steps_per_second": 10.367,
"step": 85000
},
{
"epoch": 0.16307830810546875,
"grad_norm": 86.41958618164062,
"learning_rate": 4.1846179962158205e-05,
"loss": 6.6861,
"step": 85500
},
{
"epoch": 0.164031982421875,
"grad_norm": 172.11248779296875,
"learning_rate": 4.1798496246337895e-05,
"loss": 6.6639,
"step": 86000
},
{
"epoch": 0.16498565673828125,
"grad_norm": 115.12784576416016,
"learning_rate": 4.175081253051758e-05,
"loss": 6.6782,
"step": 86500
},
{
"epoch": 0.1659393310546875,
"grad_norm": 248.4824676513672,
"learning_rate": 4.170312881469727e-05,
"loss": 6.702,
"step": 87000
},
{
"epoch": 0.16689300537109375,
"grad_norm": 45.923828125,
"learning_rate": 4.165544509887695e-05,
"loss": 6.6079,
"step": 87500
},
{
"epoch": 0.1678466796875,
"grad_norm": 602.0431518554688,
"learning_rate": 4.160776138305664e-05,
"loss": 6.6195,
"step": 88000
},
{
"epoch": 0.16880035400390625,
"grad_norm": 41.64591598510742,
"learning_rate": 4.156007766723633e-05,
"loss": 6.6371,
"step": 88500
},
{
"epoch": 0.1697540283203125,
"grad_norm": 213.25375366210938,
"learning_rate": 4.1512393951416016e-05,
"loss": 6.5944,
"step": 89000
},
{
"epoch": 0.17070770263671875,
"grad_norm": 87.20841217041016,
"learning_rate": 4.1464710235595706e-05,
"loss": 6.6422,
"step": 89500
},
{
"epoch": 0.171661376953125,
"grad_norm": 62.394229888916016,
"learning_rate": 4.141702651977539e-05,
"loss": 6.5994,
"step": 90000
},
{
"epoch": 0.171661376953125,
"eval_accuracy": 0.03867651663405088,
"eval_loss": 6.503583908081055,
"eval_runtime": 238.2981,
"eval_samples_per_second": 41.964,
"eval_steps_per_second": 10.491,
"step": 90000
},
{
"epoch": 0.17261505126953125,
"grad_norm": 48.49923324584961,
"learning_rate": 4.136934280395508e-05,
"loss": 6.5997,
"step": 90500
},
{
"epoch": 0.1735687255859375,
"grad_norm": 85.41634368896484,
"learning_rate": 4.132165908813477e-05,
"loss": 6.5842,
"step": 91000
},
{
"epoch": 0.17452239990234375,
"grad_norm": 86.25672149658203,
"learning_rate": 4.127397537231445e-05,
"loss": 6.6317,
"step": 91500
},
{
"epoch": 0.17547607421875,
"grad_norm": 57.90113067626953,
"learning_rate": 4.1226291656494143e-05,
"loss": 6.6006,
"step": 92000
},
{
"epoch": 0.17642974853515625,
"grad_norm": 644.6544189453125,
"learning_rate": 4.117860794067383e-05,
"loss": 6.6067,
"step": 92500
},
{
"epoch": 0.1773834228515625,
"grad_norm": 45.653770446777344,
"learning_rate": 4.113092422485352e-05,
"loss": 6.6163,
"step": 93000
},
{
"epoch": 0.17833709716796875,
"grad_norm": 28.696081161499023,
"learning_rate": 4.108324050903321e-05,
"loss": 6.6331,
"step": 93500
},
{
"epoch": 0.179290771484375,
"grad_norm": 87.23275756835938,
"learning_rate": 4.103555679321289e-05,
"loss": 6.6418,
"step": 94000
},
{
"epoch": 0.18024444580078125,
"grad_norm": 57.913978576660156,
"learning_rate": 4.098787307739258e-05,
"loss": 6.5708,
"step": 94500
},
{
"epoch": 0.1811981201171875,
"grad_norm": 65.90794372558594,
"learning_rate": 4.0940189361572264e-05,
"loss": 6.6023,
"step": 95000
},
{
"epoch": 0.1811981201171875,
"eval_accuracy": 0.03850430528375734,
"eval_loss": 6.49788761138916,
"eval_runtime": 242.9451,
"eval_samples_per_second": 41.162,
"eval_steps_per_second": 10.29,
"step": 95000
},
{
"epoch": 0.18215179443359375,
"grad_norm": 174.3852081298828,
"learning_rate": 4.0892505645751955e-05,
"loss": 6.5774,
"step": 95500
},
{
"epoch": 0.18310546875,
"grad_norm": 131.80348205566406,
"learning_rate": 4.0844821929931645e-05,
"loss": 6.5997,
"step": 96000
},
{
"epoch": 0.18405914306640625,
"grad_norm": 80.42842864990234,
"learning_rate": 4.079713821411133e-05,
"loss": 6.5884,
"step": 96500
},
{
"epoch": 0.1850128173828125,
"grad_norm": 102.92145538330078,
"learning_rate": 4.074945449829102e-05,
"loss": 6.6227,
"step": 97000
},
{
"epoch": 0.18596649169921875,
"grad_norm": 83.99175262451172,
"learning_rate": 4.07017707824707e-05,
"loss": 6.6208,
"step": 97500
},
{
"epoch": 0.186920166015625,
"grad_norm": 115.96537017822266,
"learning_rate": 4.065408706665039e-05,
"loss": 6.6298,
"step": 98000
},
{
"epoch": 0.18787384033203125,
"grad_norm": 97.76721954345703,
"learning_rate": 4.060640335083008e-05,
"loss": 6.6488,
"step": 98500
},
{
"epoch": 0.1888275146484375,
"grad_norm": 239.66554260253906,
"learning_rate": 4.0558719635009766e-05,
"loss": 6.6078,
"step": 99000
},
{
"epoch": 0.18978118896484375,
"grad_norm": 143.55519104003906,
"learning_rate": 4.0511035919189456e-05,
"loss": 6.5442,
"step": 99500
},
{
"epoch": 0.19073486328125,
"grad_norm": 489.53350830078125,
"learning_rate": 4.046335220336914e-05,
"loss": 6.5913,
"step": 100000
},
{
"epoch": 0.19073486328125,
"eval_accuracy": 0.03816399217221135,
"eval_loss": 6.582998275756836,
"eval_runtime": 238.7992,
"eval_samples_per_second": 41.876,
"eval_steps_per_second": 10.469,
"step": 100000
},
{
"epoch": 0.19168853759765625,
"grad_norm": 120.52062225341797,
"learning_rate": 4.041566848754883e-05,
"loss": 6.5985,
"step": 100500
},
{
"epoch": 0.1926422119140625,
"grad_norm": 153.3665008544922,
"learning_rate": 4.036798477172852e-05,
"loss": 6.6235,
"step": 101000
},
{
"epoch": 0.19359588623046875,
"grad_norm": 146.99871826171875,
"learning_rate": 4.03203010559082e-05,
"loss": 6.6944,
"step": 101500
},
{
"epoch": 0.194549560546875,
"grad_norm": 128.0095977783203,
"learning_rate": 4.0272617340087893e-05,
"loss": 6.6253,
"step": 102000
},
{
"epoch": 0.19550323486328125,
"grad_norm": 73.54423522949219,
"learning_rate": 4.022493362426758e-05,
"loss": 6.6544,
"step": 102500
},
{
"epoch": 0.1964569091796875,
"grad_norm": 110.41128540039062,
"learning_rate": 4.017724990844727e-05,
"loss": 6.6509,
"step": 103000
},
{
"epoch": 0.19741058349609375,
"grad_norm": 319.73358154296875,
"learning_rate": 4.012956619262696e-05,
"loss": 6.6368,
"step": 103500
},
{
"epoch": 0.1983642578125,
"grad_norm": 207.58070373535156,
"learning_rate": 4.008188247680664e-05,
"loss": 6.6357,
"step": 104000
},
{
"epoch": 0.19931793212890625,
"grad_norm": 55.293548583984375,
"learning_rate": 4.003419876098633e-05,
"loss": 6.6372,
"step": 104500
},
{
"epoch": 0.2002716064453125,
"grad_norm": 284.53204345703125,
"learning_rate": 3.9986515045166014e-05,
"loss": 6.6542,
"step": 105000
},
{
"epoch": 0.2002716064453125,
"eval_accuracy": 0.038425244618395304,
"eval_loss": 6.55495023727417,
"eval_runtime": 240.6255,
"eval_samples_per_second": 41.558,
"eval_steps_per_second": 10.39,
"step": 105000
},
{
"epoch": 0.20122528076171875,
"grad_norm": 150.2811737060547,
"learning_rate": 3.9938831329345705e-05,
"loss": 6.6518,
"step": 105500
},
{
"epoch": 0.202178955078125,
"grad_norm": 80.65524291992188,
"learning_rate": 3.9891147613525395e-05,
"loss": 6.6464,
"step": 106000
},
{
"epoch": 0.20313262939453125,
"grad_norm": 84.09796905517578,
"learning_rate": 3.984346389770508e-05,
"loss": 6.6482,
"step": 106500
},
{
"epoch": 0.2040863037109375,
"grad_norm": 104.57718658447266,
"learning_rate": 3.979578018188477e-05,
"loss": 6.6667,
"step": 107000
},
{
"epoch": 0.20503997802734375,
"grad_norm": 327.62109375,
"learning_rate": 3.974809646606445e-05,
"loss": 6.6351,
"step": 107500
},
{
"epoch": 0.20599365234375,
"grad_norm": 97.85407257080078,
"learning_rate": 3.970041275024414e-05,
"loss": 6.5981,
"step": 108000
},
{
"epoch": 0.20694732666015625,
"grad_norm": 54.872920989990234,
"learning_rate": 3.965272903442383e-05,
"loss": 6.6179,
"step": 108500
},
{
"epoch": 0.2079010009765625,
"grad_norm": 74.83031463623047,
"learning_rate": 3.9605045318603516e-05,
"loss": 6.6678,
"step": 109000
},
{
"epoch": 0.20885467529296875,
"grad_norm": 171.44764709472656,
"learning_rate": 3.9557361602783206e-05,
"loss": 6.6617,
"step": 109500
},
{
"epoch": 0.209808349609375,
"grad_norm": 107.16600799560547,
"learning_rate": 3.950967788696289e-05,
"loss": 6.6602,
"step": 110000
},
{
"epoch": 0.209808349609375,
"eval_accuracy": 0.038764383561643835,
"eval_loss": 6.5538482666015625,
"eval_runtime": 245.2146,
"eval_samples_per_second": 40.781,
"eval_steps_per_second": 10.195,
"step": 110000
},
{
"epoch": 0.21076202392578125,
"grad_norm": 97.79281616210938,
"learning_rate": 3.946199417114258e-05,
"loss": 6.6494,
"step": 110500
},
{
"epoch": 0.2117156982421875,
"grad_norm": 194.7386474609375,
"learning_rate": 3.941431045532227e-05,
"loss": 6.6627,
"step": 111000
},
{
"epoch": 0.21266937255859375,
"grad_norm": 39.52627944946289,
"learning_rate": 3.936662673950195e-05,
"loss": 6.6378,
"step": 111500
},
{
"epoch": 0.213623046875,
"grad_norm": 304.6911926269531,
"learning_rate": 3.9318943023681643e-05,
"loss": 6.6002,
"step": 112000
},
{
"epoch": 0.21457672119140625,
"grad_norm": 318.4564208984375,
"learning_rate": 3.927125930786133e-05,
"loss": 6.5922,
"step": 112500
},
{
"epoch": 0.2155303955078125,
"grad_norm": 72.45951843261719,
"learning_rate": 3.922357559204102e-05,
"loss": 6.6411,
"step": 113000
},
{
"epoch": 0.21648406982421875,
"grad_norm": 145.93966674804688,
"learning_rate": 3.917589187622071e-05,
"loss": 6.618,
"step": 113500
},
{
"epoch": 0.217437744140625,
"grad_norm": 306.8751220703125,
"learning_rate": 3.912820816040039e-05,
"loss": 6.6476,
"step": 114000
},
{
"epoch": 0.21839141845703125,
"grad_norm": 74.63150024414062,
"learning_rate": 3.908052444458008e-05,
"loss": 6.6734,
"step": 114500
},
{
"epoch": 0.2193450927734375,
"grad_norm": 82.34027099609375,
"learning_rate": 3.9032840728759764e-05,
"loss": 6.6113,
"step": 115000
},
{
"epoch": 0.2193450927734375,
"eval_accuracy": 0.039347162426614485,
"eval_loss": 6.550750732421875,
"eval_runtime": 237.2859,
"eval_samples_per_second": 42.143,
"eval_steps_per_second": 10.536,
"step": 115000
},
{
"epoch": 0.22029876708984375,
"grad_norm": 83.30755615234375,
"learning_rate": 3.8985157012939455e-05,
"loss": 6.5598,
"step": 115500
},
{
"epoch": 0.22125244140625,
"grad_norm": 133.1521453857422,
"learning_rate": 3.8937473297119145e-05,
"loss": 6.5783,
"step": 116000
},
{
"epoch": 0.22220611572265625,
"grad_norm": 220.6308135986328,
"learning_rate": 3.888978958129883e-05,
"loss": 6.5847,
"step": 116500
},
{
"epoch": 0.2231597900390625,
"grad_norm": 135.7482147216797,
"learning_rate": 3.884210586547852e-05,
"loss": 6.6044,
"step": 117000
},
{
"epoch": 0.22411346435546875,
"grad_norm": 159.68698120117188,
"learning_rate": 3.87944221496582e-05,
"loss": 6.6709,
"step": 117500
},
{
"epoch": 0.225067138671875,
"grad_norm": 370.2187805175781,
"learning_rate": 3.874673843383789e-05,
"loss": 6.6519,
"step": 118000
},
{
"epoch": 0.22602081298828125,
"grad_norm": 115.01087188720703,
"learning_rate": 3.869905471801758e-05,
"loss": 6.6864,
"step": 118500
},
{
"epoch": 0.2269744873046875,
"grad_norm": 102.77879333496094,
"learning_rate": 3.8651371002197266e-05,
"loss": 6.6534,
"step": 119000
},
{
"epoch": 0.22792816162109375,
"grad_norm": 122.78046417236328,
"learning_rate": 3.8603687286376956e-05,
"loss": 6.6469,
"step": 119500
},
{
"epoch": 0.2288818359375,
"grad_norm": 34.63007354736328,
"learning_rate": 3.855600357055664e-05,
"loss": 6.6568,
"step": 120000
},
{
"epoch": 0.2288818359375,
"eval_accuracy": 0.03878786692759296,
"eval_loss": 6.552664756774902,
"eval_runtime": 241.1332,
"eval_samples_per_second": 41.471,
"eval_steps_per_second": 10.368,
"step": 120000
},
{
"epoch": 0.22983551025390625,
"grad_norm": 104.76294708251953,
"learning_rate": 3.850831985473633e-05,
"loss": 6.6533,
"step": 120500
},
{
"epoch": 0.2307891845703125,
"grad_norm": 224.17337036132812,
"learning_rate": 3.846063613891602e-05,
"loss": 6.6484,
"step": 121000
},
{
"epoch": 0.23174285888671875,
"grad_norm": 73.90188598632812,
"learning_rate": 3.84129524230957e-05,
"loss": 6.6222,
"step": 121500
},
{
"epoch": 0.232696533203125,
"grad_norm": 73.31129455566406,
"learning_rate": 3.8365268707275393e-05,
"loss": 6.6224,
"step": 122000
},
{
"epoch": 0.23365020751953125,
"grad_norm": 64.67500305175781,
"learning_rate": 3.831758499145508e-05,
"loss": 6.6565,
"step": 122500
},
{
"epoch": 0.2346038818359375,
"grad_norm": 175.16627502441406,
"learning_rate": 3.826990127563477e-05,
"loss": 6.6196,
"step": 123000
},
{
"epoch": 0.23555755615234375,
"grad_norm": 449.0741882324219,
"learning_rate": 3.822221755981446e-05,
"loss": 6.6363,
"step": 123500
},
{
"epoch": 0.23651123046875,
"grad_norm": 364.6779479980469,
"learning_rate": 3.817453384399414e-05,
"loss": 6.6512,
"step": 124000
},
{
"epoch": 0.23746490478515625,
"grad_norm": 84.31883239746094,
"learning_rate": 3.812685012817383e-05,
"loss": 6.6305,
"step": 124500
},
{
"epoch": 0.2384185791015625,
"grad_norm": 259.1290588378906,
"learning_rate": 3.8079166412353514e-05,
"loss": 6.6476,
"step": 125000
},
{
"epoch": 0.2384185791015625,
"eval_accuracy": 0.039112133072407046,
"eval_loss": 6.5456719398498535,
"eval_runtime": 238.7402,
"eval_samples_per_second": 41.887,
"eval_steps_per_second": 10.472,
"step": 125000
},
{
"epoch": 0.23937225341796875,
"grad_norm": 68.04620361328125,
"learning_rate": 3.8031482696533205e-05,
"loss": 6.6502,
"step": 125500
},
{
"epoch": 0.240325927734375,
"grad_norm": 46.73642349243164,
"learning_rate": 3.7983798980712895e-05,
"loss": 6.6139,
"step": 126000
},
{
"epoch": 0.24127960205078125,
"grad_norm": 183.39114379882812,
"learning_rate": 3.793611526489258e-05,
"loss": 6.6471,
"step": 126500
},
{
"epoch": 0.2422332763671875,
"grad_norm": 208.6868133544922,
"learning_rate": 3.788843154907227e-05,
"loss": 6.6326,
"step": 127000
},
{
"epoch": 0.24318695068359375,
"grad_norm": 1748.205322265625,
"learning_rate": 3.784074783325195e-05,
"loss": 6.6223,
"step": 127500
},
{
"epoch": 0.244140625,
"grad_norm": 427.6435852050781,
"learning_rate": 3.779306411743164e-05,
"loss": 6.6556,
"step": 128000
},
{
"epoch": 0.24509429931640625,
"grad_norm": 123.13027954101562,
"learning_rate": 3.774538040161133e-05,
"loss": 6.6449,
"step": 128500
},
{
"epoch": 0.2460479736328125,
"grad_norm": 510.61474609375,
"learning_rate": 3.7697696685791016e-05,
"loss": 6.6579,
"step": 129000
},
{
"epoch": 0.24700164794921875,
"grad_norm": 84.33484649658203,
"learning_rate": 3.7650012969970706e-05,
"loss": 6.6373,
"step": 129500
},
{
"epoch": 0.247955322265625,
"grad_norm": 286.0257568359375,
"learning_rate": 3.760232925415039e-05,
"loss": 6.6636,
"step": 130000
},
{
"epoch": 0.247955322265625,
"eval_accuracy": 0.04012641878669276,
"eval_loss": 6.545810699462891,
"eval_runtime": 237.6897,
"eval_samples_per_second": 42.072,
"eval_steps_per_second": 10.518,
"step": 130000
},
{
"epoch": 0.24890899658203125,
"grad_norm": 206.83262634277344,
"learning_rate": 3.755464553833008e-05,
"loss": 6.6333,
"step": 130500
},
{
"epoch": 0.2498626708984375,
"grad_norm": 93.4488525390625,
"learning_rate": 3.750696182250977e-05,
"loss": 6.6089,
"step": 131000
},
{
"epoch": 0.25081634521484375,
"grad_norm": 59.847511291503906,
"learning_rate": 3.745927810668945e-05,
"loss": 6.6127,
"step": 131500
},
{
"epoch": 0.25177001953125,
"grad_norm": 111.11254119873047,
"learning_rate": 3.7411594390869143e-05,
"loss": 6.6167,
"step": 132000
},
{
"epoch": 0.25272369384765625,
"grad_norm": 243.5362548828125,
"learning_rate": 3.736391067504883e-05,
"loss": 6.6063,
"step": 132500
},
{
"epoch": 0.2536773681640625,
"grad_norm": 109.8550796508789,
"learning_rate": 3.731622695922852e-05,
"loss": 6.6519,
"step": 133000
},
{
"epoch": 0.25463104248046875,
"grad_norm": 196.270263671875,
"learning_rate": 3.726854324340821e-05,
"loss": 6.6547,
"step": 133500
},
{
"epoch": 0.255584716796875,
"grad_norm": 102.67740631103516,
"learning_rate": 3.722085952758789e-05,
"loss": 6.6682,
"step": 134000
},
{
"epoch": 0.25653839111328125,
"grad_norm": 75.3838882446289,
"learning_rate": 3.717317581176758e-05,
"loss": 6.6609,
"step": 134500
},
{
"epoch": 0.2574920654296875,
"grad_norm": 247.18411254882812,
"learning_rate": 3.7125492095947264e-05,
"loss": 6.6318,
"step": 135000
},
{
"epoch": 0.2574920654296875,
"eval_accuracy": 0.03993933463796478,
"eval_loss": 6.546828269958496,
"eval_runtime": 238.1837,
"eval_samples_per_second": 41.984,
"eval_steps_per_second": 10.496,
"step": 135000
},
{
"epoch": 0.25844573974609375,
"grad_norm": 186.8214111328125,
"learning_rate": 3.7077808380126955e-05,
"loss": 6.6381,
"step": 135500
},
{
"epoch": 0.2593994140625,
"grad_norm": 123.54873657226562,
"learning_rate": 3.7030124664306645e-05,
"loss": 6.6224,
"step": 136000
},
{
"epoch": 0.26035308837890625,
"grad_norm": 101.96115112304688,
"learning_rate": 3.698244094848633e-05,
"loss": 6.634,
"step": 136500
},
{
"epoch": 0.2613067626953125,
"grad_norm": 238.8611602783203,
"learning_rate": 3.693475723266602e-05,
"loss": 6.6743,
"step": 137000
},
{
"epoch": 0.26226043701171875,
"grad_norm": 238.52011108398438,
"learning_rate": 3.68870735168457e-05,
"loss": 6.6772,
"step": 137500
},
{
"epoch": 0.263214111328125,
"grad_norm": 180.68150329589844,
"learning_rate": 3.683938980102539e-05,
"loss": 6.6486,
"step": 138000
},
{
"epoch": 0.26416778564453125,
"grad_norm": 108.15036010742188,
"learning_rate": 3.679170608520508e-05,
"loss": 6.6651,
"step": 138500
},
{
"epoch": 0.2651214599609375,
"grad_norm": 1145.0054931640625,
"learning_rate": 3.6744022369384766e-05,
"loss": 6.5743,
"step": 139000
},
{
"epoch": 0.26607513427734375,
"grad_norm": 98.10594177246094,
"learning_rate": 3.6696338653564456e-05,
"loss": 6.6723,
"step": 139500
},
{
"epoch": 0.26702880859375,
"grad_norm": 62.280296325683594,
"learning_rate": 3.664865493774414e-05,
"loss": 6.6358,
"step": 140000
},
{
"epoch": 0.26702880859375,
"eval_accuracy": 0.037448923679060664,
"eval_loss": 6.57286262512207,
"eval_runtime": 237.691,
"eval_samples_per_second": 42.071,
"eval_steps_per_second": 10.518,
"step": 140000
},
{
"epoch": 0.26798248291015625,
"grad_norm": 84.15850067138672,
"learning_rate": 3.660097122192383e-05,
"loss": 6.657,
"step": 140500
},
{
"epoch": 0.2689361572265625,
"grad_norm": 103.6788330078125,
"learning_rate": 3.655328750610352e-05,
"loss": 6.6567,
"step": 141000
},
{
"epoch": 0.26988983154296875,
"grad_norm": 101.2646255493164,
"learning_rate": 3.65056037902832e-05,
"loss": 6.6804,
"step": 141500
},
{
"epoch": 0.270843505859375,
"grad_norm": 141.26473999023438,
"learning_rate": 3.6457920074462893e-05,
"loss": 6.6451,
"step": 142000
},
{
"epoch": 0.27179718017578125,
"grad_norm": 179.66439819335938,
"learning_rate": 3.641023635864258e-05,
"loss": 6.6583,
"step": 142500
},
{
"epoch": 0.2727508544921875,
"grad_norm": 114.78199005126953,
"learning_rate": 3.636255264282227e-05,
"loss": 6.6123,
"step": 143000
},
{
"epoch": 0.27370452880859375,
"grad_norm": 47.80651092529297,
"learning_rate": 3.631486892700196e-05,
"loss": 6.6632,
"step": 143500
},
{
"epoch": 0.274658203125,
"grad_norm": 157.00042724609375,
"learning_rate": 3.626718521118164e-05,
"loss": 6.6487,
"step": 144000
},
{
"epoch": 0.27561187744140625,
"grad_norm": 301.52874755859375,
"learning_rate": 3.621950149536133e-05,
"loss": 6.5943,
"step": 144500
},
{
"epoch": 0.2765655517578125,
"grad_norm": 136.7102813720703,
"learning_rate": 3.6171817779541014e-05,
"loss": 6.639,
"step": 145000
},
{
"epoch": 0.2765655517578125,
"eval_accuracy": 0.039391389432485324,
"eval_loss": 6.5531721115112305,
"eval_runtime": 243.7736,
"eval_samples_per_second": 41.022,
"eval_steps_per_second": 10.255,
"step": 145000
},
{
"epoch": 0.27751922607421875,
"grad_norm": 153.75978088378906,
"learning_rate": 3.6124134063720705e-05,
"loss": 6.6015,
"step": 145500
},
{
"epoch": 0.278472900390625,
"grad_norm": 296.2194519042969,
"learning_rate": 3.6076450347900395e-05,
"loss": 6.6052,
"step": 146000
},
{
"epoch": 0.27942657470703125,
"grad_norm": 337.25872802734375,
"learning_rate": 3.602876663208008e-05,
"loss": 6.6046,
"step": 146500
},
{
"epoch": 0.2803802490234375,
"grad_norm": 101.83748626708984,
"learning_rate": 3.598108291625977e-05,
"loss": 6.659,
"step": 147000
},
{
"epoch": 0.28133392333984375,
"grad_norm": 248.71856689453125,
"learning_rate": 3.593339920043945e-05,
"loss": 6.6136,
"step": 147500
},
{
"epoch": 0.28228759765625,
"grad_norm": 209.75669860839844,
"learning_rate": 3.588571548461914e-05,
"loss": 6.5739,
"step": 148000
},
{
"epoch": 0.28324127197265625,
"grad_norm": 86.48981475830078,
"learning_rate": 3.583803176879883e-05,
"loss": 6.5912,
"step": 148500
},
{
"epoch": 0.2841949462890625,
"grad_norm": 64.70508575439453,
"learning_rate": 3.5790348052978516e-05,
"loss": 6.6749,
"step": 149000
},
{
"epoch": 0.28514862060546875,
"grad_norm": 55.55641555786133,
"learning_rate": 3.5742664337158206e-05,
"loss": 6.6692,
"step": 149500
},
{
"epoch": 0.286102294921875,
"grad_norm": 77.39098358154297,
"learning_rate": 3.569498062133789e-05,
"loss": 6.6817,
"step": 150000
},
{
"epoch": 0.286102294921875,
"eval_accuracy": 0.03908512720156556,
"eval_loss": 6.55155611038208,
"eval_runtime": 247.8743,
"eval_samples_per_second": 40.343,
"eval_steps_per_second": 10.086,
"step": 150000
},
{
"epoch": 0.28705596923828125,
"grad_norm": 86.25543975830078,
"learning_rate": 3.564729690551758e-05,
"loss": 6.6456,
"step": 150500
},
{
"epoch": 0.2880096435546875,
"grad_norm": 194.3202362060547,
"learning_rate": 3.559961318969727e-05,
"loss": 6.6447,
"step": 151000
},
{
"epoch": 0.28896331787109375,
"grad_norm": 98.31609344482422,
"learning_rate": 3.555192947387695e-05,
"loss": 6.6226,
"step": 151500
},
{
"epoch": 0.2899169921875,
"grad_norm": 66.98186492919922,
"learning_rate": 3.5504245758056643e-05,
"loss": 6.6559,
"step": 152000
},
{
"epoch": 0.29087066650390625,
"grad_norm": 120.20733642578125,
"learning_rate": 3.545656204223633e-05,
"loss": 6.6021,
"step": 152500
},
{
"epoch": 0.2918243408203125,
"grad_norm": 137.2542724609375,
"learning_rate": 3.540887832641602e-05,
"loss": 6.665,
"step": 153000
},
{
"epoch": 0.29277801513671875,
"grad_norm": 97.15869140625,
"learning_rate": 3.536119461059571e-05,
"loss": 6.6658,
"step": 153500
},
{
"epoch": 0.293731689453125,
"grad_norm": 206.2852020263672,
"learning_rate": 3.531351089477539e-05,
"loss": 6.6735,
"step": 154000
},
{
"epoch": 0.29468536376953125,
"grad_norm": 303.65582275390625,
"learning_rate": 3.526582717895508e-05,
"loss": 6.6759,
"step": 154500
},
{
"epoch": 0.2956390380859375,
"grad_norm": 152.71656799316406,
"learning_rate": 3.5218143463134764e-05,
"loss": 6.6339,
"step": 155000
},
{
"epoch": 0.2956390380859375,
"eval_accuracy": 0.03885968688845401,
"eval_loss": 6.550864219665527,
"eval_runtime": 237.7604,
"eval_samples_per_second": 42.059,
"eval_steps_per_second": 10.515,
"step": 155000
},
{
"epoch": 0.29659271240234375,
"grad_norm": 159.33724975585938,
"learning_rate": 3.5170459747314455e-05,
"loss": 6.6202,
"step": 155500
},
{
"epoch": 0.29754638671875,
"grad_norm": 202.52700805664062,
"learning_rate": 3.5122776031494145e-05,
"loss": 6.6599,
"step": 156000
},
{
"epoch": 0.29850006103515625,
"grad_norm": 1121.404296875,
"learning_rate": 3.507509231567383e-05,
"loss": 6.6501,
"step": 156500
},
{
"epoch": 0.2994537353515625,
"grad_norm": 154.83506774902344,
"learning_rate": 3.502740859985352e-05,
"loss": 6.6193,
"step": 157000
},
{
"epoch": 0.30040740966796875,
"grad_norm": 43.87880325317383,
"learning_rate": 3.49797248840332e-05,
"loss": 6.6633,
"step": 157500
},
{
"epoch": 0.301361083984375,
"grad_norm": 81.32662963867188,
"learning_rate": 3.493204116821289e-05,
"loss": 6.616,
"step": 158000
},
{
"epoch": 0.30231475830078125,
"grad_norm": 62.43561935424805,
"learning_rate": 3.488435745239258e-05,
"loss": 6.6279,
"step": 158500
},
{
"epoch": 0.3032684326171875,
"grad_norm": 67.34500122070312,
"learning_rate": 3.4836673736572266e-05,
"loss": 6.599,
"step": 159000
},
{
"epoch": 0.30422210693359375,
"grad_norm": 123.81517028808594,
"learning_rate": 3.4788990020751956e-05,
"loss": 6.5607,
"step": 159500
},
{
"epoch": 0.30517578125,
"grad_norm": 135.31414794921875,
"learning_rate": 3.474130630493164e-05,
"loss": 6.6132,
"step": 160000
},
{
"epoch": 0.30517578125,
"eval_accuracy": 0.03812700587084149,
"eval_loss": 6.539149284362793,
"eval_runtime": 239.8317,
"eval_samples_per_second": 41.696,
"eval_steps_per_second": 10.424,
"step": 160000
},
{
"epoch": 0.30612945556640625,
"grad_norm": 73.59024047851562,
"learning_rate": 3.469362258911133e-05,
"loss": 6.6094,
"step": 160500
},
{
"epoch": 0.3070831298828125,
"grad_norm": 536.4093017578125,
"learning_rate": 3.464593887329102e-05,
"loss": 6.6274,
"step": 161000
},
{
"epoch": 0.30803680419921875,
"grad_norm": 198.60101318359375,
"learning_rate": 3.45982551574707e-05,
"loss": 6.6282,
"step": 161500
},
{
"epoch": 0.308990478515625,
"grad_norm": 69.17730712890625,
"learning_rate": 3.4550571441650393e-05,
"loss": 6.6177,
"step": 162000
},
{
"epoch": 0.30994415283203125,
"grad_norm": 86.33926391601562,
"learning_rate": 3.450288772583008e-05,
"loss": 6.5894,
"step": 162500
},
{
"epoch": 0.3108978271484375,
"grad_norm": 214.99929809570312,
"learning_rate": 3.445520401000977e-05,
"loss": 6.5999,
"step": 163000
},
{
"epoch": 0.31185150146484375,
"grad_norm": 169.09580993652344,
"learning_rate": 3.440752029418946e-05,
"loss": 6.5878,
"step": 163500
},
{
"epoch": 0.31280517578125,
"grad_norm": 120.34239959716797,
"learning_rate": 3.435983657836914e-05,
"loss": 6.5944,
"step": 164000
},
{
"epoch": 0.31375885009765625,
"grad_norm": 65.94364166259766,
"learning_rate": 3.431215286254883e-05,
"loss": 6.64,
"step": 164500
},
{
"epoch": 0.3147125244140625,
"grad_norm": 75.02478790283203,
"learning_rate": 3.4264469146728514e-05,
"loss": 6.6347,
"step": 165000
},
{
"epoch": 0.3147125244140625,
"eval_accuracy": 0.038639334637964776,
"eval_loss": 6.537603855133057,
"eval_runtime": 248.4175,
"eval_samples_per_second": 40.255,
"eval_steps_per_second": 10.064,
"step": 165000
},
{
"epoch": 0.31566619873046875,
"grad_norm": 55.92184066772461,
"learning_rate": 3.4216785430908205e-05,
"loss": 6.6448,
"step": 165500
},
{
"epoch": 0.316619873046875,
"grad_norm": 58.1660270690918,
"learning_rate": 3.4169101715087895e-05,
"loss": 6.6149,
"step": 166000
},
{
"epoch": 0.31757354736328125,
"grad_norm": 391.7157897949219,
"learning_rate": 3.412141799926758e-05,
"loss": 6.6365,
"step": 166500
},
{
"epoch": 0.3185272216796875,
"grad_norm": 55.8887825012207,
"learning_rate": 3.407373428344727e-05,
"loss": 6.6335,
"step": 167000
},
{
"epoch": 0.31948089599609375,
"grad_norm": 181.20657348632812,
"learning_rate": 3.402605056762695e-05,
"loss": 6.6448,
"step": 167500
},
{
"epoch": 0.3204345703125,
"grad_norm": 94.94180297851562,
"learning_rate": 3.397836685180664e-05,
"loss": 6.6586,
"step": 168000
},
{
"epoch": 0.32138824462890625,
"grad_norm": 184.8909149169922,
"learning_rate": 3.393068313598633e-05,
"loss": 6.6272,
"step": 168500
},
{
"epoch": 0.3223419189453125,
"grad_norm": 67.47586822509766,
"learning_rate": 3.3882999420166016e-05,
"loss": 6.6441,
"step": 169000
},
{
"epoch": 0.32329559326171875,
"grad_norm": 345.04833984375,
"learning_rate": 3.3835315704345706e-05,
"loss": 6.6201,
"step": 169500
},
{
"epoch": 0.324249267578125,
"grad_norm": 114.57809448242188,
"learning_rate": 3.378763198852539e-05,
"loss": 6.6542,
"step": 170000
},
{
"epoch": 0.324249267578125,
"eval_accuracy": 0.03969354207436399,
"eval_loss": 6.537732124328613,
"eval_runtime": 236.6138,
"eval_samples_per_second": 42.263,
"eval_steps_per_second": 10.566,
"step": 170000
},
{
"epoch": 0.32520294189453125,
"grad_norm": 186.77565002441406,
"learning_rate": 3.373994827270508e-05,
"loss": 6.6657,
"step": 170500
},
{
"epoch": 0.3261566162109375,
"grad_norm": 371.9413146972656,
"learning_rate": 3.369226455688477e-05,
"loss": 6.6699,
"step": 171000
},
{
"epoch": 0.32711029052734375,
"grad_norm": 209.8897247314453,
"learning_rate": 3.364458084106445e-05,
"loss": 6.6187,
"step": 171500
},
{
"epoch": 0.32806396484375,
"grad_norm": 206.73399353027344,
"learning_rate": 3.3596897125244143e-05,
"loss": 6.6278,
"step": 172000
},
{
"epoch": 0.32901763916015625,
"grad_norm": 122.4280776977539,
"learning_rate": 3.354921340942383e-05,
"loss": 6.6408,
"step": 172500
},
{
"epoch": 0.3299713134765625,
"grad_norm": 116.70370483398438,
"learning_rate": 3.350152969360352e-05,
"loss": 6.6568,
"step": 173000
},
{
"epoch": 0.33092498779296875,
"grad_norm": 117.74230194091797,
"learning_rate": 3.345384597778321e-05,
"loss": 6.6408,
"step": 173500
},
{
"epoch": 0.331878662109375,
"grad_norm": 258.5967102050781,
"learning_rate": 3.340616226196289e-05,
"loss": 6.6316,
"step": 174000
},
{
"epoch": 0.33283233642578125,
"grad_norm": 142.6446533203125,
"learning_rate": 3.335847854614258e-05,
"loss": 6.6485,
"step": 174500
},
{
"epoch": 0.3337860107421875,
"grad_norm": 119.10045623779297,
"learning_rate": 3.3310794830322264e-05,
"loss": 6.619,
"step": 175000
},
{
"epoch": 0.3337860107421875,
"eval_accuracy": 0.03876301369863014,
"eval_loss": 6.544088840484619,
"eval_runtime": 247.1694,
"eval_samples_per_second": 40.458,
"eval_steps_per_second": 10.115,
"step": 175000
},
{
"epoch": 0.33473968505859375,
"grad_norm": 86.29735565185547,
"learning_rate": 3.3263111114501955e-05,
"loss": 6.6692,
"step": 175500
},
{
"epoch": 0.335693359375,
"grad_norm": 594.6138305664062,
"learning_rate": 3.3215427398681645e-05,
"loss": 6.652,
"step": 176000
},
{
"epoch": 0.33664703369140625,
"grad_norm": 157.3546142578125,
"learning_rate": 3.316774368286133e-05,
"loss": 6.6185,
"step": 176500
},
{
"epoch": 0.3376007080078125,
"grad_norm": 174.54434204101562,
"learning_rate": 3.312005996704102e-05,
"loss": 6.6237,
"step": 177000
},
{
"epoch": 0.33855438232421875,
"grad_norm": 210.47975158691406,
"learning_rate": 3.30723762512207e-05,
"loss": 6.5825,
"step": 177500
},
{
"epoch": 0.339508056640625,
"grad_norm": 146.17979431152344,
"learning_rate": 3.302469253540039e-05,
"loss": 6.5469,
"step": 178000
},
{
"epoch": 0.34046173095703125,
"grad_norm": 114.10115051269531,
"learning_rate": 3.297700881958008e-05,
"loss": 6.5547,
"step": 178500
},
{
"epoch": 0.3414154052734375,
"grad_norm": 75.68565368652344,
"learning_rate": 3.2929325103759766e-05,
"loss": 6.609,
"step": 179000
},
{
"epoch": 0.34236907958984375,
"grad_norm": 79.07707977294922,
"learning_rate": 3.2881641387939456e-05,
"loss": 6.6539,
"step": 179500
},
{
"epoch": 0.34332275390625,
"grad_norm": 584.5313110351562,
"learning_rate": 3.283395767211914e-05,
"loss": 6.6979,
"step": 180000
},
{
"epoch": 0.34332275390625,
"eval_accuracy": 0.03938180039138943,
"eval_loss": 6.555627346038818,
"eval_runtime": 248.0084,
"eval_samples_per_second": 40.321,
"eval_steps_per_second": 10.08,
"step": 180000
},
{
"epoch": 0.34427642822265625,
"grad_norm": 293.7067565917969,
"learning_rate": 3.278627395629883e-05,
"loss": 6.6661,
"step": 180500
},
{
"epoch": 0.3452301025390625,
"grad_norm": 112.20381927490234,
"learning_rate": 3.273859024047852e-05,
"loss": 6.6845,
"step": 181000
},
{
"epoch": 0.34618377685546875,
"grad_norm": 207.96649169921875,
"learning_rate": 3.26909065246582e-05,
"loss": 6.6346,
"step": 181500
},
{
"epoch": 0.347137451171875,
"grad_norm": 114.42713165283203,
"learning_rate": 3.2643222808837893e-05,
"loss": 6.6249,
"step": 182000
},
{
"epoch": 0.34809112548828125,
"grad_norm": 102.50064086914062,
"learning_rate": 3.259553909301758e-05,
"loss": 6.6432,
"step": 182500
},
{
"epoch": 0.3490447998046875,
"grad_norm": 73.7787857055664,
"learning_rate": 3.254785537719727e-05,
"loss": 6.6306,
"step": 183000
},
{
"epoch": 0.34999847412109375,
"grad_norm": 151.01646423339844,
"learning_rate": 3.250017166137696e-05,
"loss": 6.6134,
"step": 183500
},
{
"epoch": 0.3509521484375,
"grad_norm": 58.98166275024414,
"learning_rate": 3.245248794555664e-05,
"loss": 6.649,
"step": 184000
},
{
"epoch": 0.35190582275390625,
"grad_norm": 103.82510375976562,
"learning_rate": 3.240480422973633e-05,
"loss": 6.6382,
"step": 184500
},
{
"epoch": 0.3528594970703125,
"grad_norm": 50.53388977050781,
"learning_rate": 3.2357120513916014e-05,
"loss": 6.6485,
"step": 185000
},
{
"epoch": 0.3528594970703125,
"eval_accuracy": 0.039942270058708415,
"eval_loss": 6.536978244781494,
"eval_runtime": 248.6094,
"eval_samples_per_second": 40.224,
"eval_steps_per_second": 10.056,
"step": 185000
},
{
"epoch": 0.35381317138671875,
"grad_norm": 125.3191909790039,
"learning_rate": 3.2309436798095705e-05,
"loss": 6.6498,
"step": 185500
},
{
"epoch": 0.354766845703125,
"grad_norm": 113.97254943847656,
"learning_rate": 3.2261753082275395e-05,
"loss": 6.5974,
"step": 186000
},
{
"epoch": 0.35572052001953125,
"grad_norm": 134.717529296875,
"learning_rate": 3.221406936645508e-05,
"loss": 6.6141,
"step": 186500
},
{
"epoch": 0.3566741943359375,
"grad_norm": 144.05467224121094,
"learning_rate": 3.216638565063477e-05,
"loss": 6.595,
"step": 187000
},
{
"epoch": 0.35762786865234375,
"grad_norm": 60.121891021728516,
"learning_rate": 3.211870193481445e-05,
"loss": 6.6121,
"step": 187500
},
{
"epoch": 0.35858154296875,
"grad_norm": 44.34535598754883,
"learning_rate": 3.207101821899414e-05,
"loss": 6.6154,
"step": 188000
},
{
"epoch": 0.35953521728515625,
"grad_norm": 227.33172607421875,
"learning_rate": 3.202333450317383e-05,
"loss": 6.594,
"step": 188500
},
{
"epoch": 0.3604888916015625,
"grad_norm": 1378.051025390625,
"learning_rate": 3.1975650787353516e-05,
"loss": 6.6521,
"step": 189000
},
{
"epoch": 0.36144256591796875,
"grad_norm": 46.72209930419922,
"learning_rate": 3.1927967071533206e-05,
"loss": 6.5648,
"step": 189500
},
{
"epoch": 0.362396240234375,
"grad_norm": 131.37269592285156,
"learning_rate": 3.188028335571289e-05,
"loss": 6.6035,
"step": 190000
},
{
"epoch": 0.362396240234375,
"eval_accuracy": 0.03885205479452055,
"eval_loss": 6.5300092697143555,
"eval_runtime": 251.9306,
"eval_samples_per_second": 39.693,
"eval_steps_per_second": 9.923,
"step": 190000
},
{
"epoch": 0.36334991455078125,
"grad_norm": 76.72520446777344,
"learning_rate": 3.183259963989258e-05,
"loss": 6.6213,
"step": 190500
},
{
"epoch": 0.3643035888671875,
"grad_norm": 112.77885437011719,
"learning_rate": 3.178491592407227e-05,
"loss": 6.6474,
"step": 191000
},
{
"epoch": 0.36525726318359375,
"grad_norm": 70.67579650878906,
"learning_rate": 3.173723220825195e-05,
"loss": 6.6288,
"step": 191500
},
{
"epoch": 0.3662109375,
"grad_norm": 59.372642517089844,
"learning_rate": 3.1689548492431643e-05,
"loss": 6.5876,
"step": 192000
},
{
"epoch": 0.36716461181640625,
"grad_norm": 63.93340301513672,
"learning_rate": 3.164186477661133e-05,
"loss": 6.5254,
"step": 192500
},
{
"epoch": 0.3681182861328125,
"grad_norm": 93.77088165283203,
"learning_rate": 3.159418106079102e-05,
"loss": 6.5014,
"step": 193000
},
{
"epoch": 0.36907196044921875,
"grad_norm": 348.0354919433594,
"learning_rate": 3.154649734497071e-05,
"loss": 6.5559,
"step": 193500
},
{
"epoch": 0.370025634765625,
"grad_norm": 108.91474914550781,
"learning_rate": 3.149881362915039e-05,
"loss": 6.5741,
"step": 194000
},
{
"epoch": 0.37097930908203125,
"grad_norm": 121.82709503173828,
"learning_rate": 3.145112991333008e-05,
"loss": 6.6283,
"step": 194500
},
{
"epoch": 0.3719329833984375,
"grad_norm": 123.15145111083984,
"learning_rate": 3.1403446197509764e-05,
"loss": 6.6574,
"step": 195000
},
{
"epoch": 0.3719329833984375,
"eval_accuracy": 0.038532093933463796,
"eval_loss": 6.527206897735596,
"eval_runtime": 254.2011,
"eval_samples_per_second": 39.339,
"eval_steps_per_second": 9.835,
"step": 195000
},
{
"epoch": 0.37288665771484375,
"grad_norm": 97.69001007080078,
"learning_rate": 3.1355762481689455e-05,
"loss": 6.6062,
"step": 195500
},
{
"epoch": 0.37384033203125,
"grad_norm": 160.37985229492188,
"learning_rate": 3.1308078765869145e-05,
"loss": 6.6397,
"step": 196000
},
{
"epoch": 0.37479400634765625,
"grad_norm": 107.5202407836914,
"learning_rate": 3.126039505004883e-05,
"loss": 6.6337,
"step": 196500
},
{
"epoch": 0.3757476806640625,
"grad_norm": 306.0714416503906,
"learning_rate": 3.121271133422852e-05,
"loss": 6.6156,
"step": 197000
},
{
"epoch": 0.37670135498046875,
"grad_norm": 147.80152893066406,
"learning_rate": 3.11650276184082e-05,
"loss": 6.6691,
"step": 197500
},
{
"epoch": 0.377655029296875,
"grad_norm": 69.64010620117188,
"learning_rate": 3.111734390258789e-05,
"loss": 6.6362,
"step": 198000
},
{
"epoch": 0.37860870361328125,
"grad_norm": 61.15127182006836,
"learning_rate": 3.106966018676758e-05,
"loss": 6.6665,
"step": 198500
},
{
"epoch": 0.3795623779296875,
"grad_norm": 524.3397216796875,
"learning_rate": 3.1021976470947266e-05,
"loss": 6.6447,
"step": 199000
},
{
"epoch": 0.38051605224609375,
"grad_norm": 206.949951171875,
"learning_rate": 3.0974292755126956e-05,
"loss": 6.6509,
"step": 199500
},
{
"epoch": 0.3814697265625,
"grad_norm": 109.48726654052734,
"learning_rate": 3.092660903930664e-05,
"loss": 6.6152,
"step": 200000
},
{
"epoch": 0.3814697265625,
"eval_accuracy": 0.03765283757338552,
"eval_loss": 6.532608985900879,
"eval_runtime": 239.0143,
"eval_samples_per_second": 41.839,
"eval_steps_per_second": 10.46,
"step": 200000
},
{
"epoch": 0.38242340087890625,
"grad_norm": 211.54612731933594,
"learning_rate": 3.087892532348633e-05,
"loss": 6.5765,
"step": 200500
},
{
"epoch": 0.3833770751953125,
"grad_norm": 123.34220123291016,
"learning_rate": 3.083124160766602e-05,
"loss": 6.6243,
"step": 201000
},
{
"epoch": 0.38433074951171875,
"grad_norm": 164.30299377441406,
"learning_rate": 3.07835578918457e-05,
"loss": 6.6452,
"step": 201500
},
{
"epoch": 0.385284423828125,
"grad_norm": 156.64134216308594,
"learning_rate": 3.0735874176025393e-05,
"loss": 6.6471,
"step": 202000
},
{
"epoch": 0.38623809814453125,
"grad_norm": 121.67871856689453,
"learning_rate": 3.068819046020508e-05,
"loss": 6.6086,
"step": 202500
},
{
"epoch": 0.3871917724609375,
"grad_norm": 93.36353302001953,
"learning_rate": 3.064050674438477e-05,
"loss": 6.6283,
"step": 203000
},
{
"epoch": 0.38814544677734375,
"grad_norm": 455.4304504394531,
"learning_rate": 3.059282302856446e-05,
"loss": 6.6332,
"step": 203500
},
{
"epoch": 0.38909912109375,
"grad_norm": 118.70697784423828,
"learning_rate": 3.054513931274414e-05,
"loss": 6.6339,
"step": 204000
},
{
"epoch": 0.39005279541015625,
"grad_norm": 59.528892517089844,
"learning_rate": 3.049745559692383e-05,
"loss": 6.5999,
"step": 204500
},
{
"epoch": 0.3910064697265625,
"grad_norm": 169.48011779785156,
"learning_rate": 3.0449771881103518e-05,
"loss": 6.5946,
"step": 205000
},
{
"epoch": 0.3910064697265625,
"eval_accuracy": 0.03863835616438356,
"eval_loss": 6.531431198120117,
"eval_runtime": 247.6939,
"eval_samples_per_second": 40.372,
"eval_steps_per_second": 10.093,
"step": 205000
},
{
"epoch": 0.39196014404296875,
"grad_norm": 188.75572204589844,
"learning_rate": 3.0402088165283205e-05,
"loss": 6.5888,
"step": 205500
},
{
"epoch": 0.392913818359375,
"grad_norm": 67.42155456542969,
"learning_rate": 3.035440444946289e-05,
"loss": 6.586,
"step": 206000
},
{
"epoch": 0.39386749267578125,
"grad_norm": 76.16634368896484,
"learning_rate": 3.0306720733642578e-05,
"loss": 6.5993,
"step": 206500
},
{
"epoch": 0.3948211669921875,
"grad_norm": 259.7588806152344,
"learning_rate": 3.025903701782227e-05,
"loss": 6.6088,
"step": 207000
},
{
"epoch": 0.39577484130859375,
"grad_norm": 75.8446273803711,
"learning_rate": 3.0211353302001955e-05,
"loss": 6.5749,
"step": 207500
},
{
"epoch": 0.396728515625,
"grad_norm": 563.4567260742188,
"learning_rate": 3.0163669586181642e-05,
"loss": 6.5621,
"step": 208000
},
{
"epoch": 0.39768218994140625,
"grad_norm": 181.39927673339844,
"learning_rate": 3.011598587036133e-05,
"loss": 6.5429,
"step": 208500
},
{
"epoch": 0.3986358642578125,
"grad_norm": 78.88523864746094,
"learning_rate": 3.0068302154541016e-05,
"loss": 6.5932,
"step": 209000
},
{
"epoch": 0.39958953857421875,
"grad_norm": 288.15118408203125,
"learning_rate": 3.0020618438720706e-05,
"loss": 6.6502,
"step": 209500
},
{
"epoch": 0.400543212890625,
"grad_norm": 1462.8275146484375,
"learning_rate": 2.9972934722900393e-05,
"loss": 6.6747,
"step": 210000
},
{
"epoch": 0.400543212890625,
"eval_accuracy": 0.039049902152641876,
"eval_loss": 6.518420696258545,
"eval_runtime": 245.3499,
"eval_samples_per_second": 40.758,
"eval_steps_per_second": 10.19,
"step": 210000
},
{
"epoch": 0.40149688720703125,
"grad_norm": 270.0433654785156,
"learning_rate": 2.992525100708008e-05,
"loss": 6.6314,
"step": 210500
},
{
"epoch": 0.4024505615234375,
"grad_norm": 140.15133666992188,
"learning_rate": 2.9877567291259766e-05,
"loss": 6.6099,
"step": 211000
},
{
"epoch": 0.40340423583984375,
"grad_norm": 158.65878295898438,
"learning_rate": 2.9829883575439453e-05,
"loss": 6.6602,
"step": 211500
},
{
"epoch": 0.40435791015625,
"grad_norm": 63.215354919433594,
"learning_rate": 2.9782199859619143e-05,
"loss": 6.5969,
"step": 212000
},
{
"epoch": 0.40531158447265625,
"grad_norm": 434.90972900390625,
"learning_rate": 2.973451614379883e-05,
"loss": 6.5978,
"step": 212500
},
{
"epoch": 0.4062652587890625,
"grad_norm": 123.8073959350586,
"learning_rate": 2.9686832427978517e-05,
"loss": 6.632,
"step": 213000
},
{
"epoch": 0.40721893310546875,
"grad_norm": 131.49026489257812,
"learning_rate": 2.9639148712158204e-05,
"loss": 6.6471,
"step": 213500
},
{
"epoch": 0.408172607421875,
"grad_norm": 275.69775390625,
"learning_rate": 2.959146499633789e-05,
"loss": 6.6522,
"step": 214000
},
{
"epoch": 0.40912628173828125,
"grad_norm": 285.1741638183594,
"learning_rate": 2.954378128051758e-05,
"loss": 6.6183,
"step": 214500
},
{
"epoch": 0.4100799560546875,
"grad_norm": 242.5558624267578,
"learning_rate": 2.9496097564697268e-05,
"loss": 6.618,
"step": 215000
},
{
"epoch": 0.4100799560546875,
"eval_accuracy": 0.03752504892367906,
"eval_loss": 6.5262370109558105,
"eval_runtime": 236.1515,
"eval_samples_per_second": 42.346,
"eval_steps_per_second": 10.586,
"step": 215000
},
{
"epoch": 0.41103363037109375,
"grad_norm": 111.01753234863281,
"learning_rate": 2.9448413848876955e-05,
"loss": 6.5984,
"step": 215500
},
{
"epoch": 0.4119873046875,
"grad_norm": 79.53560638427734,
"learning_rate": 2.940073013305664e-05,
"loss": 6.6158,
"step": 216000
},
{
"epoch": 0.41294097900390625,
"grad_norm": 56.24204635620117,
"learning_rate": 2.9353046417236328e-05,
"loss": 6.6407,
"step": 216500
},
{
"epoch": 0.4138946533203125,
"grad_norm": 407.7383117675781,
"learning_rate": 2.930536270141602e-05,
"loss": 6.5933,
"step": 217000
},
{
"epoch": 0.41484832763671875,
"grad_norm": 111.54969024658203,
"learning_rate": 2.9257678985595705e-05,
"loss": 6.5416,
"step": 217500
},
{
"epoch": 0.415802001953125,
"grad_norm": 573.9189453125,
"learning_rate": 2.9209995269775392e-05,
"loss": 6.627,
"step": 218000
},
{
"epoch": 0.41675567626953125,
"grad_norm": 61.13914489746094,
"learning_rate": 2.916231155395508e-05,
"loss": 6.6041,
"step": 218500
},
{
"epoch": 0.4177093505859375,
"grad_norm": 223.9310302734375,
"learning_rate": 2.9114627838134766e-05,
"loss": 6.6351,
"step": 219000
},
{
"epoch": 0.41866302490234375,
"grad_norm": 222.9752960205078,
"learning_rate": 2.9066944122314456e-05,
"loss": 6.6542,
"step": 219500
},
{
"epoch": 0.41961669921875,
"grad_norm": 93.45650482177734,
"learning_rate": 2.9019260406494143e-05,
"loss": 6.6218,
"step": 220000
},
{
"epoch": 0.41961669921875,
"eval_accuracy": 0.038779843444227004,
"eval_loss": 6.533949375152588,
"eval_runtime": 253.8147,
"eval_samples_per_second": 39.399,
"eval_steps_per_second": 9.85,
"step": 220000
},
{
"epoch": 0.42057037353515625,
"grad_norm": 116.55095672607422,
"learning_rate": 2.897157669067383e-05,
"loss": 6.6154,
"step": 220500
},
{
"epoch": 0.4215240478515625,
"grad_norm": 97.29713439941406,
"learning_rate": 2.8923892974853516e-05,
"loss": 6.5961,
"step": 221000
},
{
"epoch": 0.42247772216796875,
"grad_norm": 54.4910774230957,
"learning_rate": 2.8876209259033203e-05,
"loss": 6.5797,
"step": 221500
},
{
"epoch": 0.423431396484375,
"grad_norm": 365.531494140625,
"learning_rate": 2.8828525543212893e-05,
"loss": 6.5709,
"step": 222000
},
{
"epoch": 0.42438507080078125,
"grad_norm": 59.368221282958984,
"learning_rate": 2.878084182739258e-05,
"loss": 6.5644,
"step": 222500
},
{
"epoch": 0.4253387451171875,
"grad_norm": 70.24651336669922,
"learning_rate": 2.8733158111572267e-05,
"loss": 6.5807,
"step": 223000
},
{
"epoch": 0.42629241943359375,
"grad_norm": 67.14861297607422,
"learning_rate": 2.8685474395751954e-05,
"loss": 6.6404,
"step": 223500
},
{
"epoch": 0.42724609375,
"grad_norm": 115.37001037597656,
"learning_rate": 2.863779067993164e-05,
"loss": 6.6317,
"step": 224000
},
{
"epoch": 0.42819976806640625,
"grad_norm": 134.29444885253906,
"learning_rate": 2.859010696411133e-05,
"loss": 6.653,
"step": 224500
},
{
"epoch": 0.4291534423828125,
"grad_norm": 88.78755187988281,
"learning_rate": 2.8542423248291018e-05,
"loss": 6.6659,
"step": 225000
},
{
"epoch": 0.4291534423828125,
"eval_accuracy": 0.03827005870841487,
"eval_loss": 6.525757312774658,
"eval_runtime": 239.6203,
"eval_samples_per_second": 41.733,
"eval_steps_per_second": 10.433,
"step": 225000
},
{
"epoch": 0.43010711669921875,
"grad_norm": 59.79543685913086,
"learning_rate": 2.8494739532470705e-05,
"loss": 6.6049,
"step": 225500
},
{
"epoch": 0.431060791015625,
"grad_norm": 70.32538604736328,
"learning_rate": 2.844705581665039e-05,
"loss": 6.6035,
"step": 226000
},
{
"epoch": 0.43201446533203125,
"grad_norm": 47.479225158691406,
"learning_rate": 2.8399372100830078e-05,
"loss": 6.665,
"step": 226500
},
{
"epoch": 0.4329681396484375,
"grad_norm": 299.8247375488281,
"learning_rate": 2.835168838500977e-05,
"loss": 6.6103,
"step": 227000
},
{
"epoch": 0.43392181396484375,
"grad_norm": 68.14066314697266,
"learning_rate": 2.8304004669189455e-05,
"loss": 6.6295,
"step": 227500
},
{
"epoch": 0.43487548828125,
"grad_norm": 134.41168212890625,
"learning_rate": 2.8256320953369142e-05,
"loss": 6.5957,
"step": 228000
},
{
"epoch": 0.43582916259765625,
"grad_norm": 121.62809753417969,
"learning_rate": 2.820863723754883e-05,
"loss": 6.646,
"step": 228500
},
{
"epoch": 0.4367828369140625,
"grad_norm": 66.99333190917969,
"learning_rate": 2.8160953521728516e-05,
"loss": 6.6207,
"step": 229000
},
{
"epoch": 0.43773651123046875,
"grad_norm": 105.14512634277344,
"learning_rate": 2.8113269805908206e-05,
"loss": 6.6208,
"step": 229500
},
{
"epoch": 0.438690185546875,
"grad_norm": 80.91681671142578,
"learning_rate": 2.8065586090087893e-05,
"loss": 6.6292,
"step": 230000
},
{
"epoch": 0.438690185546875,
"eval_accuracy": 0.03869960861056752,
"eval_loss": 6.524181365966797,
"eval_runtime": 237.6497,
"eval_samples_per_second": 42.079,
"eval_steps_per_second": 10.52,
"step": 230000
},
{
"epoch": 0.43964385986328125,
"grad_norm": 70.6031494140625,
"learning_rate": 2.801790237426758e-05,
"loss": 6.634,
"step": 230500
},
{
"epoch": 0.4405975341796875,
"grad_norm": 94.75713348388672,
"learning_rate": 2.7970218658447266e-05,
"loss": 6.5882,
"step": 231000
},
{
"epoch": 0.44155120849609375,
"grad_norm": 85.7402572631836,
"learning_rate": 2.7922534942626953e-05,
"loss": 6.6304,
"step": 231500
},
{
"epoch": 0.4425048828125,
"grad_norm": 62.476173400878906,
"learning_rate": 2.7874851226806643e-05,
"loss": 6.602,
"step": 232000
},
{
"epoch": 0.44345855712890625,
"grad_norm": 103.45085144042969,
"learning_rate": 2.782716751098633e-05,
"loss": 6.6056,
"step": 232500
},
{
"epoch": 0.4444122314453125,
"grad_norm": 188.56761169433594,
"learning_rate": 2.7779483795166017e-05,
"loss": 6.611,
"step": 233000
},
{
"epoch": 0.44536590576171875,
"grad_norm": 122.7448959350586,
"learning_rate": 2.7731800079345704e-05,
"loss": 6.5814,
"step": 233500
},
{
"epoch": 0.446319580078125,
"grad_norm": 127.61102294921875,
"learning_rate": 2.768411636352539e-05,
"loss": 6.62,
"step": 234000
},
{
"epoch": 0.44727325439453125,
"grad_norm": 55.343955993652344,
"learning_rate": 2.763643264770508e-05,
"loss": 6.6301,
"step": 234500
},
{
"epoch": 0.4482269287109375,
"grad_norm": 31.93642234802246,
"learning_rate": 2.7588748931884768e-05,
"loss": 6.6608,
"step": 235000
},
{
"epoch": 0.4482269287109375,
"eval_accuracy": 0.035914090019569474,
"eval_loss": 6.558816432952881,
"eval_runtime": 249.3294,
"eval_samples_per_second": 40.108,
"eval_steps_per_second": 10.027,
"step": 235000
},
{
"epoch": 0.44918060302734375,
"grad_norm": 498.0051574707031,
"learning_rate": 2.7541065216064455e-05,
"loss": 6.6317,
"step": 235500
},
{
"epoch": 0.45013427734375,
"grad_norm": 81.84600830078125,
"learning_rate": 2.749338150024414e-05,
"loss": 6.5796,
"step": 236000
},
{
"epoch": 0.45108795166015625,
"grad_norm": 95.89019775390625,
"learning_rate": 2.7445697784423828e-05,
"loss": 6.6014,
"step": 236500
},
{
"epoch": 0.4520416259765625,
"grad_norm": 65.94055938720703,
"learning_rate": 2.739801406860352e-05,
"loss": 6.576,
"step": 237000
},
{
"epoch": 0.45299530029296875,
"grad_norm": 232.42098999023438,
"learning_rate": 2.7350330352783205e-05,
"loss": 6.5618,
"step": 237500
},
{
"epoch": 0.453948974609375,
"grad_norm": 46.52445983886719,
"learning_rate": 2.7302646636962892e-05,
"loss": 6.6523,
"step": 238000
},
{
"epoch": 0.45490264892578125,
"grad_norm": 55.36016082763672,
"learning_rate": 2.725496292114258e-05,
"loss": 6.6198,
"step": 238500
},
{
"epoch": 0.4558563232421875,
"grad_norm": 42.32960891723633,
"learning_rate": 2.7207279205322266e-05,
"loss": 6.6285,
"step": 239000
},
{
"epoch": 0.45680999755859375,
"grad_norm": 65.83206939697266,
"learning_rate": 2.7159595489501956e-05,
"loss": 6.6045,
"step": 239500
},
{
"epoch": 0.457763671875,
"grad_norm": 297.2817077636719,
"learning_rate": 2.7111911773681643e-05,
"loss": 6.5772,
"step": 240000
},
{
"epoch": 0.457763671875,
"eval_accuracy": 0.0389399217221135,
"eval_loss": 6.511717319488525,
"eval_runtime": 239.2997,
"eval_samples_per_second": 41.789,
"eval_steps_per_second": 10.447,
"step": 240000
},
{
"epoch": 0.45871734619140625,
"grad_norm": 284.4686279296875,
"learning_rate": 2.706422805786133e-05,
"loss": 6.6136,
"step": 240500
},
{
"epoch": 0.4596710205078125,
"grad_norm": 57.11293411254883,
"learning_rate": 2.7016544342041016e-05,
"loss": 6.6187,
"step": 241000
},
{
"epoch": 0.46062469482421875,
"grad_norm": 167.8474578857422,
"learning_rate": 2.6968860626220703e-05,
"loss": 6.651,
"step": 241500
},
{
"epoch": 0.461578369140625,
"grad_norm": 47.619842529296875,
"learning_rate": 2.6921176910400393e-05,
"loss": 6.6191,
"step": 242000
},
{
"epoch": 0.46253204345703125,
"grad_norm": 87.00941467285156,
"learning_rate": 2.687349319458008e-05,
"loss": 6.6161,
"step": 242500
},
{
"epoch": 0.4634857177734375,
"grad_norm": 226.7034149169922,
"learning_rate": 2.6825809478759767e-05,
"loss": 6.6149,
"step": 243000
},
{
"epoch": 0.46443939208984375,
"grad_norm": 103.8028335571289,
"learning_rate": 2.6778125762939454e-05,
"loss": 6.6247,
"step": 243500
},
{
"epoch": 0.46539306640625,
"grad_norm": 83.36387634277344,
"learning_rate": 2.673044204711914e-05,
"loss": 6.6019,
"step": 244000
},
{
"epoch": 0.46634674072265625,
"grad_norm": 75.43655395507812,
"learning_rate": 2.668275833129883e-05,
"loss": 6.5738,
"step": 244500
},
{
"epoch": 0.4673004150390625,
"grad_norm": 113.69373321533203,
"learning_rate": 2.6635074615478518e-05,
"loss": 6.5961,
"step": 245000
},
{
"epoch": 0.4673004150390625,
"eval_accuracy": 0.03589549902152642,
"eval_loss": 6.538053035736084,
"eval_runtime": 238.1145,
"eval_samples_per_second": 41.997,
"eval_steps_per_second": 10.499,
"step": 245000
},
{
"epoch": 0.46825408935546875,
"grad_norm": 71.73027801513672,
"learning_rate": 2.6587390899658205e-05,
"loss": 6.6012,
"step": 245500
},
{
"epoch": 0.469207763671875,
"grad_norm": 108.12360382080078,
"learning_rate": 2.653970718383789e-05,
"loss": 6.5581,
"step": 246000
},
{
"epoch": 0.47016143798828125,
"grad_norm": 78.5285873413086,
"learning_rate": 2.6492023468017578e-05,
"loss": 6.5655,
"step": 246500
},
{
"epoch": 0.4711151123046875,
"grad_norm": 42.62543487548828,
"learning_rate": 2.644433975219727e-05,
"loss": 6.595,
"step": 247000
},
{
"epoch": 0.47206878662109375,
"grad_norm": 51.91836929321289,
"learning_rate": 2.6396656036376955e-05,
"loss": 6.5763,
"step": 247500
},
{
"epoch": 0.4730224609375,
"grad_norm": 34.81822204589844,
"learning_rate": 2.6348972320556642e-05,
"loss": 6.6161,
"step": 248000
},
{
"epoch": 0.47397613525390625,
"grad_norm": 33.98869323730469,
"learning_rate": 2.630128860473633e-05,
"loss": 6.5918,
"step": 248500
},
{
"epoch": 0.4749298095703125,
"grad_norm": 36.41916275024414,
"learning_rate": 2.6253604888916016e-05,
"loss": 6.6115,
"step": 249000
},
{
"epoch": 0.47588348388671875,
"grad_norm": 144.3103485107422,
"learning_rate": 2.6205921173095706e-05,
"loss": 6.6107,
"step": 249500
},
{
"epoch": 0.476837158203125,
"grad_norm": 67.23009490966797,
"learning_rate": 2.6158237457275393e-05,
"loss": 6.563,
"step": 250000
},
{
"epoch": 0.476837158203125,
"eval_accuracy": 0.04007847358121331,
"eval_loss": 6.519131183624268,
"eval_runtime": 239.77,
"eval_samples_per_second": 41.707,
"eval_steps_per_second": 10.427,
"step": 250000
},
{
"epoch": 0.47779083251953125,
"grad_norm": 52.45304870605469,
"learning_rate": 2.611055374145508e-05,
"loss": 6.5189,
"step": 250500
},
{
"epoch": 0.4787445068359375,
"grad_norm": 144.16065979003906,
"learning_rate": 2.6062870025634766e-05,
"loss": 6.5234,
"step": 251000
},
{
"epoch": 0.47969818115234375,
"grad_norm": 53.3181037902832,
"learning_rate": 2.6015186309814453e-05,
"loss": 6.5775,
"step": 251500
},
{
"epoch": 0.48065185546875,
"grad_norm": 61.429264068603516,
"learning_rate": 2.5967502593994143e-05,
"loss": 6.5551,
"step": 252000
},
{
"epoch": 0.48160552978515625,
"grad_norm": 77.74535369873047,
"learning_rate": 2.591981887817383e-05,
"loss": 6.6271,
"step": 252500
},
{
"epoch": 0.4825592041015625,
"grad_norm": 62.36492156982422,
"learning_rate": 2.5872135162353517e-05,
"loss": 6.599,
"step": 253000
},
{
"epoch": 0.48351287841796875,
"grad_norm": 48.22749710083008,
"learning_rate": 2.5824451446533204e-05,
"loss": 6.6033,
"step": 253500
},
{
"epoch": 0.484466552734375,
"grad_norm": 66.73870086669922,
"learning_rate": 2.577676773071289e-05,
"loss": 6.5897,
"step": 254000
},
{
"epoch": 0.48542022705078125,
"grad_norm": 114.62262725830078,
"learning_rate": 2.572908401489258e-05,
"loss": 6.4891,
"step": 254500
},
{
"epoch": 0.4863739013671875,
"grad_norm": 47.45481491088867,
"learning_rate": 2.5681400299072268e-05,
"loss": 6.5651,
"step": 255000
},
{
"epoch": 0.4863739013671875,
"eval_accuracy": 0.0385412915851272,
"eval_loss": 6.4979939460754395,
"eval_runtime": 243.7128,
"eval_samples_per_second": 41.032,
"eval_steps_per_second": 10.258,
"step": 255000
},
{
"epoch": 0.48732757568359375,
"grad_norm": 168.27577209472656,
"learning_rate": 2.5633716583251955e-05,
"loss": 6.5793,
"step": 255500
},
{
"epoch": 0.48828125,
"grad_norm": 58.58369064331055,
"learning_rate": 2.558603286743164e-05,
"loss": 6.5713,
"step": 256000
},
{
"epoch": 0.48923492431640625,
"grad_norm": 136.4830780029297,
"learning_rate": 2.5538349151611328e-05,
"loss": 6.5813,
"step": 256500
},
{
"epoch": 0.4901885986328125,
"grad_norm": 141.25311279296875,
"learning_rate": 2.549066543579102e-05,
"loss": 6.5841,
"step": 257000
},
{
"epoch": 0.49114227294921875,
"grad_norm": 62.92488479614258,
"learning_rate": 2.5442981719970705e-05,
"loss": 6.5859,
"step": 257500
},
{
"epoch": 0.492095947265625,
"grad_norm": 322.6897888183594,
"learning_rate": 2.5395298004150392e-05,
"loss": 6.5814,
"step": 258000
},
{
"epoch": 0.49304962158203125,
"grad_norm": 223.2647247314453,
"learning_rate": 2.534761428833008e-05,
"loss": 6.5797,
"step": 258500
},
{
"epoch": 0.4940032958984375,
"grad_norm": 50.54922103881836,
"learning_rate": 2.5299930572509766e-05,
"loss": 6.5791,
"step": 259000
},
{
"epoch": 0.49495697021484375,
"grad_norm": 206.4066619873047,
"learning_rate": 2.5252246856689456e-05,
"loss": 6.6171,
"step": 259500
},
{
"epoch": 0.49591064453125,
"grad_norm": 82.16864776611328,
"learning_rate": 2.5204563140869143e-05,
"loss": 6.5398,
"step": 260000
},
{
"epoch": 0.49591064453125,
"eval_accuracy": 0.03885655577299413,
"eval_loss": 6.49983024597168,
"eval_runtime": 237.2672,
"eval_samples_per_second": 42.147,
"eval_steps_per_second": 10.537,
"step": 260000
},
{
"epoch": 0.49686431884765625,
"grad_norm": 100.80988311767578,
"learning_rate": 2.515687942504883e-05,
"loss": 6.5823,
"step": 260500
},
{
"epoch": 0.4978179931640625,
"grad_norm": 40.91484069824219,
"learning_rate": 2.5109195709228516e-05,
"loss": 6.5807,
"step": 261000
},
{
"epoch": 0.49877166748046875,
"grad_norm": 198.70712280273438,
"learning_rate": 2.5061511993408203e-05,
"loss": 6.5759,
"step": 261500
},
{
"epoch": 0.499725341796875,
"grad_norm": 102.08186340332031,
"learning_rate": 2.5013828277587893e-05,
"loss": 6.6002,
"step": 262000
},
{
"epoch": 0.5006790161132812,
"grad_norm": 114.76848602294922,
"learning_rate": 2.496614456176758e-05,
"loss": 6.5748,
"step": 262500
},
{
"epoch": 0.5016326904296875,
"grad_norm": 475.52337646484375,
"learning_rate": 2.4918460845947267e-05,
"loss": 6.5824,
"step": 263000
},
{
"epoch": 0.5025863647460938,
"grad_norm": 63.79277420043945,
"learning_rate": 2.4870777130126954e-05,
"loss": 6.5644,
"step": 263500
},
{
"epoch": 0.5035400390625,
"grad_norm": 273.6373291015625,
"learning_rate": 2.482309341430664e-05,
"loss": 6.6185,
"step": 264000
},
{
"epoch": 0.5044937133789062,
"grad_norm": 174.2074432373047,
"learning_rate": 2.477540969848633e-05,
"loss": 6.5506,
"step": 264500
},
{
"epoch": 0.5054473876953125,
"grad_norm": 1188.002685546875,
"learning_rate": 2.4727725982666018e-05,
"loss": 6.5368,
"step": 265000
},
{
"epoch": 0.5054473876953125,
"eval_accuracy": 0.03887651663405088,
"eval_loss": 6.4984235763549805,
"eval_runtime": 238.5437,
"eval_samples_per_second": 41.921,
"eval_steps_per_second": 10.48,
"step": 265000
},
{
"epoch": 0.5064010620117188,
"grad_norm": 373.5093078613281,
"learning_rate": 2.4680042266845705e-05,
"loss": 6.5601,
"step": 265500
},
{
"epoch": 0.507354736328125,
"grad_norm": 106.3024673461914,
"learning_rate": 2.463235855102539e-05,
"loss": 6.5036,
"step": 266000
},
{
"epoch": 0.5083084106445312,
"grad_norm": 143.9859619140625,
"learning_rate": 2.4584674835205078e-05,
"loss": 6.5817,
"step": 266500
},
{
"epoch": 0.5092620849609375,
"grad_norm": 62.84480285644531,
"learning_rate": 2.453699111938477e-05,
"loss": 6.5998,
"step": 267000
},
{
"epoch": 0.5102157592773438,
"grad_norm": 235.6577606201172,
"learning_rate": 2.4489307403564455e-05,
"loss": 6.6033,
"step": 267500
},
{
"epoch": 0.51116943359375,
"grad_norm": 140.92276000976562,
"learning_rate": 2.4441623687744142e-05,
"loss": 6.535,
"step": 268000
},
{
"epoch": 0.5121231079101562,
"grad_norm": 153.2882843017578,
"learning_rate": 2.439393997192383e-05,
"loss": 6.5595,
"step": 268500
},
{
"epoch": 0.5130767822265625,
"grad_norm": 66.25297546386719,
"learning_rate": 2.4346256256103516e-05,
"loss": 6.5609,
"step": 269000
},
{
"epoch": 0.5140304565429688,
"grad_norm": 174.41998291015625,
"learning_rate": 2.4298572540283206e-05,
"loss": 6.5772,
"step": 269500
},
{
"epoch": 0.514984130859375,
"grad_norm": 126.60942077636719,
"learning_rate": 2.4250888824462893e-05,
"loss": 6.599,
"step": 270000
},
{
"epoch": 0.514984130859375,
"eval_accuracy": 0.0387761252446184,
"eval_loss": 6.493450164794922,
"eval_runtime": 239.9879,
"eval_samples_per_second": 41.669,
"eval_steps_per_second": 10.417,
"step": 270000
},
{
"epoch": 0.5159378051757812,
"grad_norm": 226.81134033203125,
"learning_rate": 2.420320510864258e-05,
"loss": 6.5803,
"step": 270500
},
{
"epoch": 0.5168914794921875,
"grad_norm": 174.19500732421875,
"learning_rate": 2.4155521392822266e-05,
"loss": 6.5671,
"step": 271000
},
{
"epoch": 0.5178451538085938,
"grad_norm": 124.677001953125,
"learning_rate": 2.4107837677001953e-05,
"loss": 6.5904,
"step": 271500
},
{
"epoch": 0.518798828125,
"grad_norm": 76.14005279541016,
"learning_rate": 2.406015396118164e-05,
"loss": 6.5971,
"step": 272000
},
{
"epoch": 0.5197525024414062,
"grad_norm": 129.07936096191406,
"learning_rate": 2.401247024536133e-05,
"loss": 6.5572,
"step": 272500
},
{
"epoch": 0.5207061767578125,
"grad_norm": 77.52312469482422,
"learning_rate": 2.3964786529541017e-05,
"loss": 6.5508,
"step": 273000
},
{
"epoch": 0.5216598510742188,
"grad_norm": 96.4195556640625,
"learning_rate": 2.3917102813720704e-05,
"loss": 6.5542,
"step": 273500
},
{
"epoch": 0.522613525390625,
"grad_norm": 304.2214050292969,
"learning_rate": 2.386941909790039e-05,
"loss": 6.5653,
"step": 274000
},
{
"epoch": 0.5235671997070312,
"grad_norm": 114.71865844726562,
"learning_rate": 2.3821735382080078e-05,
"loss": 6.5624,
"step": 274500
},
{
"epoch": 0.5245208740234375,
"grad_norm": 87.2974853515625,
"learning_rate": 2.3774051666259768e-05,
"loss": 6.6015,
"step": 275000
},
{
"epoch": 0.5245208740234375,
"eval_accuracy": 0.03879902152641879,
"eval_loss": 6.489101409912109,
"eval_runtime": 237.0997,
"eval_samples_per_second": 42.176,
"eval_steps_per_second": 10.544,
"step": 275000
},
{
"epoch": 0.5254745483398438,
"grad_norm": 89.22550201416016,
"learning_rate": 2.3726367950439455e-05,
"loss": 6.5772,
"step": 275500
},
{
"epoch": 0.52642822265625,
"grad_norm": 239.70816040039062,
"learning_rate": 2.367868423461914e-05,
"loss": 6.5974,
"step": 276000
},
{
"epoch": 0.5273818969726562,
"grad_norm": 100.86592102050781,
"learning_rate": 2.3631000518798828e-05,
"loss": 6.5683,
"step": 276500
},
{
"epoch": 0.5283355712890625,
"grad_norm": 173.64987182617188,
"learning_rate": 2.3583316802978515e-05,
"loss": 6.5515,
"step": 277000
},
{
"epoch": 0.5292892456054688,
"grad_norm": 48.169734954833984,
"learning_rate": 2.3535633087158205e-05,
"loss": 6.5881,
"step": 277500
},
{
"epoch": 0.530242919921875,
"grad_norm": 121.87443542480469,
"learning_rate": 2.3487949371337892e-05,
"loss": 6.5795,
"step": 278000
},
{
"epoch": 0.5311965942382812,
"grad_norm": 149.2621612548828,
"learning_rate": 2.344026565551758e-05,
"loss": 6.5027,
"step": 278500
},
{
"epoch": 0.5321502685546875,
"grad_norm": 98.72933197021484,
"learning_rate": 2.3392581939697266e-05,
"loss": 6.5023,
"step": 279000
},
{
"epoch": 0.5331039428710938,
"grad_norm": 95.72864532470703,
"learning_rate": 2.3344898223876953e-05,
"loss": 6.4943,
"step": 279500
},
{
"epoch": 0.5340576171875,
"grad_norm": 1422.8848876953125,
"learning_rate": 2.3297214508056643e-05,
"loss": 6.5597,
"step": 280000
},
{
"epoch": 0.5340576171875,
"eval_accuracy": 0.03942974559686888,
"eval_loss": 6.477899074554443,
"eval_runtime": 243.6605,
"eval_samples_per_second": 41.041,
"eval_steps_per_second": 10.26,
"step": 280000
},
{
"epoch": 0.5350112915039062,
"grad_norm": 58.64554214477539,
"learning_rate": 2.324953079223633e-05,
"loss": 6.6398,
"step": 280500
},
{
"epoch": 0.5359649658203125,
"grad_norm": 179.9173126220703,
"learning_rate": 2.3201847076416016e-05,
"loss": 6.5941,
"step": 281000
},
{
"epoch": 0.5369186401367188,
"grad_norm": 243.6712188720703,
"learning_rate": 2.3154163360595703e-05,
"loss": 6.5901,
"step": 281500
},
{
"epoch": 0.537872314453125,
"grad_norm": 62.67595291137695,
"learning_rate": 2.310647964477539e-05,
"loss": 6.5734,
"step": 282000
},
{
"epoch": 0.5388259887695312,
"grad_norm": 249.361572265625,
"learning_rate": 2.305879592895508e-05,
"loss": 6.5552,
"step": 282500
},
{
"epoch": 0.5397796630859375,
"grad_norm": 85.4103012084961,
"learning_rate": 2.3011112213134767e-05,
"loss": 6.5943,
"step": 283000
},
{
"epoch": 0.5407333374023438,
"grad_norm": 194.80154418945312,
"learning_rate": 2.2963428497314454e-05,
"loss": 6.5562,
"step": 283500
},
{
"epoch": 0.54168701171875,
"grad_norm": 113.66175842285156,
"learning_rate": 2.291574478149414e-05,
"loss": 6.5685,
"step": 284000
},
{
"epoch": 0.5426406860351562,
"grad_norm": 173.5185546875,
"learning_rate": 2.2868061065673828e-05,
"loss": 6.565,
"step": 284500
},
{
"epoch": 0.5435943603515625,
"grad_norm": 932.0955810546875,
"learning_rate": 2.2820377349853518e-05,
"loss": 6.5695,
"step": 285000
},
{
"epoch": 0.5435943603515625,
"eval_accuracy": 0.039506066536203525,
"eval_loss": 6.4823150634765625,
"eval_runtime": 238.0878,
"eval_samples_per_second": 42.001,
"eval_steps_per_second": 10.5,
"step": 285000
},
{
"epoch": 0.5445480346679688,
"grad_norm": 426.4958190917969,
"learning_rate": 2.2772693634033205e-05,
"loss": 6.5879,
"step": 285500
},
{
"epoch": 0.545501708984375,
"grad_norm": 165.65635681152344,
"learning_rate": 2.272500991821289e-05,
"loss": 6.5646,
"step": 286000
},
{
"epoch": 0.5464553833007812,
"grad_norm": 98.79949951171875,
"learning_rate": 2.2677326202392578e-05,
"loss": 6.5702,
"step": 286500
},
{
"epoch": 0.5474090576171875,
"grad_norm": 86.36907196044922,
"learning_rate": 2.2629642486572265e-05,
"loss": 6.5367,
"step": 287000
},
{
"epoch": 0.5483627319335938,
"grad_norm": 74.77356719970703,
"learning_rate": 2.2581958770751955e-05,
"loss": 6.5822,
"step": 287500
},
{
"epoch": 0.54931640625,
"grad_norm": 353.52392578125,
"learning_rate": 2.2534275054931642e-05,
"loss": 6.5747,
"step": 288000
},
{
"epoch": 0.5502700805664062,
"grad_norm": 166.5744171142578,
"learning_rate": 2.248659133911133e-05,
"loss": 6.5156,
"step": 288500
},
{
"epoch": 0.5512237548828125,
"grad_norm": 63.69950485229492,
"learning_rate": 2.2438907623291016e-05,
"loss": 6.5647,
"step": 289000
},
{
"epoch": 0.5521774291992188,
"grad_norm": 127.87213897705078,
"learning_rate": 2.2391223907470703e-05,
"loss": 6.5736,
"step": 289500
},
{
"epoch": 0.553131103515625,
"grad_norm": 202.01283264160156,
"learning_rate": 2.2343540191650393e-05,
"loss": 6.5809,
"step": 290000
},
{
"epoch": 0.553131103515625,
"eval_accuracy": 0.038223287671232876,
"eval_loss": 6.4925103187561035,
"eval_runtime": 237.5747,
"eval_samples_per_second": 42.092,
"eval_steps_per_second": 10.523,
"step": 290000
},
{
"epoch": 0.5540847778320312,
"grad_norm": 481.0960998535156,
"learning_rate": 2.229585647583008e-05,
"loss": 6.559,
"step": 290500
},
{
"epoch": 0.5550384521484375,
"grad_norm": 279.78582763671875,
"learning_rate": 2.2248172760009766e-05,
"loss": 6.5773,
"step": 291000
},
{
"epoch": 0.5559921264648438,
"grad_norm": 1313.1634521484375,
"learning_rate": 2.2200489044189453e-05,
"loss": 6.5843,
"step": 291500
},
{
"epoch": 0.55694580078125,
"grad_norm": 417.1656494140625,
"learning_rate": 2.215280532836914e-05,
"loss": 6.4884,
"step": 292000
},
{
"epoch": 0.5578994750976562,
"grad_norm": 187.26255798339844,
"learning_rate": 2.210512161254883e-05,
"loss": 6.5431,
"step": 292500
},
{
"epoch": 0.5588531494140625,
"grad_norm": 130.42347717285156,
"learning_rate": 2.2057437896728517e-05,
"loss": 6.5248,
"step": 293000
},
{
"epoch": 0.5598068237304688,
"grad_norm": 266.1620178222656,
"learning_rate": 2.2009754180908204e-05,
"loss": 6.4695,
"step": 293500
},
{
"epoch": 0.560760498046875,
"grad_norm": 62.113685607910156,
"learning_rate": 2.196207046508789e-05,
"loss": 6.6018,
"step": 294000
},
{
"epoch": 0.5617141723632812,
"grad_norm": 159.71209716796875,
"learning_rate": 2.1914386749267578e-05,
"loss": 6.6181,
"step": 294500
},
{
"epoch": 0.5626678466796875,
"grad_norm": 75.88334655761719,
"learning_rate": 2.1866703033447268e-05,
"loss": 6.6522,
"step": 295000
},
{
"epoch": 0.5626678466796875,
"eval_accuracy": 0.03903522504892368,
"eval_loss": 6.489808082580566,
"eval_runtime": 237.5721,
"eval_samples_per_second": 42.092,
"eval_steps_per_second": 10.523,
"step": 295000
},
{
"epoch": 0.5636215209960938,
"grad_norm": 397.1622314453125,
"learning_rate": 2.1819019317626955e-05,
"loss": 6.5981,
"step": 295500
},
{
"epoch": 0.5645751953125,
"grad_norm": 100.62149047851562,
"learning_rate": 2.177133560180664e-05,
"loss": 6.5702,
"step": 296000
},
{
"epoch": 0.5655288696289062,
"grad_norm": 139.2002410888672,
"learning_rate": 2.1723651885986328e-05,
"loss": 6.5627,
"step": 296500
},
{
"epoch": 0.5664825439453125,
"grad_norm": 209.5233917236328,
"learning_rate": 2.1675968170166015e-05,
"loss": 6.5707,
"step": 297000
},
{
"epoch": 0.5674362182617188,
"grad_norm": 252.32769775390625,
"learning_rate": 2.1628284454345705e-05,
"loss": 6.6021,
"step": 297500
},
{
"epoch": 0.568389892578125,
"grad_norm": 120.24504852294922,
"learning_rate": 2.1580600738525392e-05,
"loss": 6.6022,
"step": 298000
},
{
"epoch": 0.5693435668945312,
"grad_norm": 428.54132080078125,
"learning_rate": 2.153291702270508e-05,
"loss": 6.5826,
"step": 298500
},
{
"epoch": 0.5702972412109375,
"grad_norm": 136.40740966796875,
"learning_rate": 2.1485233306884766e-05,
"loss": 6.5511,
"step": 299000
},
{
"epoch": 0.5712509155273438,
"grad_norm": 98.76660919189453,
"learning_rate": 2.1437549591064453e-05,
"loss": 6.5606,
"step": 299500
},
{
"epoch": 0.57220458984375,
"grad_norm": 121.40555572509766,
"learning_rate": 2.1389865875244143e-05,
"loss": 6.5688,
"step": 300000
},
{
"epoch": 0.57220458984375,
"eval_accuracy": 0.03886947162426614,
"eval_loss": 6.491418838500977,
"eval_runtime": 239.8089,
"eval_samples_per_second": 41.7,
"eval_steps_per_second": 10.425,
"step": 300000
},
{
"epoch": 0.5731582641601562,
"grad_norm": 65.11746215820312,
"learning_rate": 2.134218215942383e-05,
"loss": 6.593,
"step": 300500
},
{
"epoch": 0.5741119384765625,
"grad_norm": 131.08642578125,
"learning_rate": 2.1294498443603516e-05,
"loss": 6.5569,
"step": 301000
},
{
"epoch": 0.5750656127929688,
"grad_norm": 213.5546875,
"learning_rate": 2.1246814727783203e-05,
"loss": 6.5626,
"step": 301500
},
{
"epoch": 0.576019287109375,
"grad_norm": 164.9312744140625,
"learning_rate": 2.119913101196289e-05,
"loss": 6.5405,
"step": 302000
},
{
"epoch": 0.5769729614257812,
"grad_norm": 130.9062042236328,
"learning_rate": 2.115144729614258e-05,
"loss": 6.5618,
"step": 302500
},
{
"epoch": 0.5779266357421875,
"grad_norm": 75.20059204101562,
"learning_rate": 2.1103763580322267e-05,
"loss": 6.573,
"step": 303000
},
{
"epoch": 0.5788803100585938,
"grad_norm": 98.30027770996094,
"learning_rate": 2.1056079864501954e-05,
"loss": 6.5728,
"step": 303500
},
{
"epoch": 0.579833984375,
"grad_norm": 57.365936279296875,
"learning_rate": 2.100839614868164e-05,
"loss": 6.5246,
"step": 304000
},
{
"epoch": 0.5807876586914062,
"grad_norm": 118.35264587402344,
"learning_rate": 2.0960712432861328e-05,
"loss": 6.5676,
"step": 304500
},
{
"epoch": 0.5817413330078125,
"grad_norm": 92.39289093017578,
"learning_rate": 2.0913028717041018e-05,
"loss": 6.5445,
"step": 305000
},
{
"epoch": 0.5817413330078125,
"eval_accuracy": 0.038926614481409,
"eval_loss": 6.482814788818359,
"eval_runtime": 237.5203,
"eval_samples_per_second": 42.102,
"eval_steps_per_second": 10.525,
"step": 305000
},
{
"epoch": 0.5826950073242188,
"grad_norm": 139.6799774169922,
"learning_rate": 2.0865345001220705e-05,
"loss": 6.5893,
"step": 305500
},
{
"epoch": 0.583648681640625,
"grad_norm": 220.11317443847656,
"learning_rate": 2.081766128540039e-05,
"loss": 6.5063,
"step": 306000
},
{
"epoch": 0.5846023559570312,
"grad_norm": 123.89678955078125,
"learning_rate": 2.0769977569580078e-05,
"loss": 6.541,
"step": 306500
},
{
"epoch": 0.5855560302734375,
"grad_norm": 150.82089233398438,
"learning_rate": 2.0722293853759765e-05,
"loss": 6.5261,
"step": 307000
},
{
"epoch": 0.5865097045898438,
"grad_norm": 110.43388366699219,
"learning_rate": 2.0674610137939455e-05,
"loss": 6.533,
"step": 307500
},
{
"epoch": 0.58746337890625,
"grad_norm": 158.13780212402344,
"learning_rate": 2.0626926422119142e-05,
"loss": 6.6145,
"step": 308000
},
{
"epoch": 0.5884170532226562,
"grad_norm": 944.8779907226562,
"learning_rate": 2.057924270629883e-05,
"loss": 6.6023,
"step": 308500
},
{
"epoch": 0.5893707275390625,
"grad_norm": 105.20059204101562,
"learning_rate": 2.0531558990478516e-05,
"loss": 6.5803,
"step": 309000
},
{
"epoch": 0.5903244018554688,
"grad_norm": 88.02117919921875,
"learning_rate": 2.0483875274658203e-05,
"loss": 6.5734,
"step": 309500
},
{
"epoch": 0.591278076171875,
"grad_norm": 220.67440795898438,
"learning_rate": 2.0436191558837893e-05,
"loss": 6.5674,
"step": 310000
},
{
"epoch": 0.591278076171875,
"eval_accuracy": 0.03877632093933464,
"eval_loss": 6.493968963623047,
"eval_runtime": 240.9683,
"eval_samples_per_second": 41.499,
"eval_steps_per_second": 10.375,
"step": 310000
},
{
"epoch": 0.5922317504882812,
"grad_norm": 151.35154724121094,
"learning_rate": 2.038850784301758e-05,
"loss": 6.544,
"step": 310500
},
{
"epoch": 0.5931854248046875,
"grad_norm": 280.2740173339844,
"learning_rate": 2.0340824127197266e-05,
"loss": 6.566,
"step": 311000
},
{
"epoch": 0.5941390991210938,
"grad_norm": 130.1758270263672,
"learning_rate": 2.0293140411376953e-05,
"loss": 6.5556,
"step": 311500
},
{
"epoch": 0.5950927734375,
"grad_norm": 1289.0264892578125,
"learning_rate": 2.024545669555664e-05,
"loss": 6.6155,
"step": 312000
},
{
"epoch": 0.5960464477539062,
"grad_norm": 324.85662841796875,
"learning_rate": 2.019777297973633e-05,
"loss": 6.5985,
"step": 312500
},
{
"epoch": 0.5970001220703125,
"grad_norm": 251.70790100097656,
"learning_rate": 2.0150089263916017e-05,
"loss": 6.5858,
"step": 313000
},
{
"epoch": 0.5979537963867188,
"grad_norm": 154.8018341064453,
"learning_rate": 2.0102405548095704e-05,
"loss": 6.5304,
"step": 313500
},
{
"epoch": 0.598907470703125,
"grad_norm": 891.773681640625,
"learning_rate": 2.005472183227539e-05,
"loss": 6.5531,
"step": 314000
},
{
"epoch": 0.5998611450195312,
"grad_norm": 116.1676025390625,
"learning_rate": 2.0007038116455078e-05,
"loss": 6.5902,
"step": 314500
},
{
"epoch": 0.6008148193359375,
"grad_norm": 64.54122924804688,
"learning_rate": 1.9959354400634768e-05,
"loss": 6.5926,
"step": 315000
},
{
"epoch": 0.6008148193359375,
"eval_accuracy": 0.03876692759295499,
"eval_loss": 6.478638172149658,
"eval_runtime": 239.6249,
"eval_samples_per_second": 41.732,
"eval_steps_per_second": 10.433,
"step": 315000
},
{
"epoch": 0.6017684936523438,
"grad_norm": 150.49513244628906,
"learning_rate": 1.9911670684814455e-05,
"loss": 6.5588,
"step": 315500
},
{
"epoch": 0.60272216796875,
"grad_norm": 206.52194213867188,
"learning_rate": 1.986398696899414e-05,
"loss": 6.5864,
"step": 316000
},
{
"epoch": 0.6036758422851562,
"grad_norm": 87.1997299194336,
"learning_rate": 1.9816303253173828e-05,
"loss": 6.5414,
"step": 316500
},
{
"epoch": 0.6046295166015625,
"grad_norm": 737.2801513671875,
"learning_rate": 1.9768619537353515e-05,
"loss": 6.5796,
"step": 317000
},
{
"epoch": 0.6055831909179688,
"grad_norm": 67.21781158447266,
"learning_rate": 1.9720935821533205e-05,
"loss": 6.5526,
"step": 317500
},
{
"epoch": 0.606536865234375,
"grad_norm": 148.92437744140625,
"learning_rate": 1.9673252105712892e-05,
"loss": 6.5496,
"step": 318000
},
{
"epoch": 0.6074905395507812,
"grad_norm": 159.55218505859375,
"learning_rate": 1.962556838989258e-05,
"loss": 6.5567,
"step": 318500
},
{
"epoch": 0.6084442138671875,
"grad_norm": 298.22796630859375,
"learning_rate": 1.9577884674072266e-05,
"loss": 6.5438,
"step": 319000
},
{
"epoch": 0.6093978881835938,
"grad_norm": 119.46234893798828,
"learning_rate": 1.9530200958251953e-05,
"loss": 6.5746,
"step": 319500
},
{
"epoch": 0.6103515625,
"grad_norm": 497.9232482910156,
"learning_rate": 1.9482517242431643e-05,
"loss": 6.4979,
"step": 320000
},
{
"epoch": 0.6103515625,
"eval_accuracy": 0.03913796477495108,
"eval_loss": 6.475390434265137,
"eval_runtime": 245.799,
"eval_samples_per_second": 40.684,
"eval_steps_per_second": 10.171,
"step": 320000
},
{
"epoch": 0.6113052368164062,
"grad_norm": 95.33645629882812,
"learning_rate": 1.943483352661133e-05,
"loss": 6.5177,
"step": 320500
},
{
"epoch": 0.6122589111328125,
"grad_norm": 177.78366088867188,
"learning_rate": 1.9387149810791016e-05,
"loss": 6.5235,
"step": 321000
},
{
"epoch": 0.6132125854492188,
"grad_norm": 450.78155517578125,
"learning_rate": 1.9339466094970703e-05,
"loss": 6.6073,
"step": 321500
},
{
"epoch": 0.614166259765625,
"grad_norm": 176.4562225341797,
"learning_rate": 1.929178237915039e-05,
"loss": 6.6167,
"step": 322000
},
{
"epoch": 0.6151199340820312,
"grad_norm": 158.9343719482422,
"learning_rate": 1.924409866333008e-05,
"loss": 6.6144,
"step": 322500
},
{
"epoch": 0.6160736083984375,
"grad_norm": 130.91563415527344,
"learning_rate": 1.9196414947509767e-05,
"loss": 6.6087,
"step": 323000
},
{
"epoch": 0.6170272827148438,
"grad_norm": 158.33616638183594,
"learning_rate": 1.9148731231689454e-05,
"loss": 6.5553,
"step": 323500
},
{
"epoch": 0.61798095703125,
"grad_norm": 104.18647003173828,
"learning_rate": 1.910104751586914e-05,
"loss": 6.5952,
"step": 324000
},
{
"epoch": 0.6189346313476562,
"grad_norm": 151.77406311035156,
"learning_rate": 1.9053363800048828e-05,
"loss": 6.5814,
"step": 324500
},
{
"epoch": 0.6198883056640625,
"grad_norm": 356.1595153808594,
"learning_rate": 1.9005680084228518e-05,
"loss": 6.5669,
"step": 325000
},
{
"epoch": 0.6198883056640625,
"eval_accuracy": 0.039391780821917806,
"eval_loss": 6.472179889678955,
"eval_runtime": 244.9459,
"eval_samples_per_second": 40.825,
"eval_steps_per_second": 10.206,
"step": 325000
},
{
"epoch": 0.6208419799804688,
"grad_norm": 314.9290466308594,
"learning_rate": 1.8957996368408205e-05,
"loss": 6.5539,
"step": 325500
},
{
"epoch": 0.621795654296875,
"grad_norm": 184.68072509765625,
"learning_rate": 1.891031265258789e-05,
"loss": 6.5505,
"step": 326000
},
{
"epoch": 0.6227493286132812,
"grad_norm": 72.39010620117188,
"learning_rate": 1.8862628936767578e-05,
"loss": 6.6057,
"step": 326500
},
{
"epoch": 0.6237030029296875,
"grad_norm": 98.22920989990234,
"learning_rate": 1.8814945220947265e-05,
"loss": 6.536,
"step": 327000
},
{
"epoch": 0.6246566772460938,
"grad_norm": 92.31875610351562,
"learning_rate": 1.8767261505126955e-05,
"loss": 6.5872,
"step": 327500
},
{
"epoch": 0.6256103515625,
"grad_norm": 105.9183578491211,
"learning_rate": 1.8719577789306642e-05,
"loss": 6.5717,
"step": 328000
},
{
"epoch": 0.6265640258789062,
"grad_norm": 109.58235931396484,
"learning_rate": 1.867189407348633e-05,
"loss": 6.5964,
"step": 328500
},
{
"epoch": 0.6275177001953125,
"grad_norm": 99.37423706054688,
"learning_rate": 1.8624210357666016e-05,
"loss": 6.5457,
"step": 329000
},
{
"epoch": 0.6284713745117188,
"grad_norm": 116.90908813476562,
"learning_rate": 1.8576526641845703e-05,
"loss": 6.5854,
"step": 329500
},
{
"epoch": 0.629425048828125,
"grad_norm": 106.40868377685547,
"learning_rate": 1.8528842926025393e-05,
"loss": 6.5335,
"step": 330000
},
{
"epoch": 0.629425048828125,
"eval_accuracy": 0.03977103718199609,
"eval_loss": 6.473107814788818,
"eval_runtime": 239.2374,
"eval_samples_per_second": 41.799,
"eval_steps_per_second": 10.45,
"step": 330000
},
{
"epoch": 0.6303787231445312,
"grad_norm": 146.23626708984375,
"learning_rate": 1.848115921020508e-05,
"loss": 6.626,
"step": 330500
},
{
"epoch": 0.6313323974609375,
"grad_norm": 85.45753479003906,
"learning_rate": 1.8433475494384766e-05,
"loss": 6.5638,
"step": 331000
},
{
"epoch": 0.6322860717773438,
"grad_norm": 141.6040802001953,
"learning_rate": 1.8385791778564453e-05,
"loss": 6.5574,
"step": 331500
},
{
"epoch": 0.63323974609375,
"grad_norm": 85.09579467773438,
"learning_rate": 1.833810806274414e-05,
"loss": 6.5948,
"step": 332000
},
{
"epoch": 0.6341934204101562,
"grad_norm": 55.52738952636719,
"learning_rate": 1.829042434692383e-05,
"loss": 6.5454,
"step": 332500
},
{
"epoch": 0.6351470947265625,
"grad_norm": 248.0428924560547,
"learning_rate": 1.8242740631103517e-05,
"loss": 6.5177,
"step": 333000
},
{
"epoch": 0.6361007690429688,
"grad_norm": 205.9241943359375,
"learning_rate": 1.8195056915283204e-05,
"loss": 6.5136,
"step": 333500
},
{
"epoch": 0.637054443359375,
"grad_norm": 81.16824340820312,
"learning_rate": 1.814737319946289e-05,
"loss": 6.5282,
"step": 334000
},
{
"epoch": 0.6380081176757812,
"grad_norm": 130.18638610839844,
"learning_rate": 1.8099689483642578e-05,
"loss": 6.537,
"step": 334500
},
{
"epoch": 0.6389617919921875,
"grad_norm": 223.24957275390625,
"learning_rate": 1.8052005767822268e-05,
"loss": 6.5727,
"step": 335000
},
{
"epoch": 0.6389617919921875,
"eval_accuracy": 0.039358708414872795,
"eval_loss": 6.476977825164795,
"eval_runtime": 241.6364,
"eval_samples_per_second": 41.384,
"eval_steps_per_second": 10.346,
"step": 335000
},
{
"epoch": 0.6399154663085938,
"grad_norm": 257.26629638671875,
"learning_rate": 1.8004322052001955e-05,
"loss": 6.6013,
"step": 335500
},
{
"epoch": 0.640869140625,
"grad_norm": 526.728271484375,
"learning_rate": 1.795663833618164e-05,
"loss": 6.5763,
"step": 336000
},
{
"epoch": 0.6418228149414062,
"grad_norm": 127.42972564697266,
"learning_rate": 1.7908954620361328e-05,
"loss": 6.5615,
"step": 336500
},
{
"epoch": 0.6427764892578125,
"grad_norm": 62.457923889160156,
"learning_rate": 1.7861270904541015e-05,
"loss": 6.5083,
"step": 337000
},
{
"epoch": 0.6437301635742188,
"grad_norm": 148.3224334716797,
"learning_rate": 1.7813587188720705e-05,
"loss": 6.5489,
"step": 337500
},
{
"epoch": 0.644683837890625,
"grad_norm": 293.37652587890625,
"learning_rate": 1.7765903472900392e-05,
"loss": 6.5661,
"step": 338000
},
{
"epoch": 0.6456375122070312,
"grad_norm": 233.91455078125,
"learning_rate": 1.771821975708008e-05,
"loss": 6.5598,
"step": 338500
},
{
"epoch": 0.6465911865234375,
"grad_norm": 153.2530059814453,
"learning_rate": 1.7670536041259766e-05,
"loss": 6.5883,
"step": 339000
},
{
"epoch": 0.6475448608398438,
"grad_norm": 193.26214599609375,
"learning_rate": 1.7622852325439453e-05,
"loss": 6.5564,
"step": 339500
},
{
"epoch": 0.64849853515625,
"grad_norm": 347.178955078125,
"learning_rate": 1.7575168609619143e-05,
"loss": 6.5735,
"step": 340000
},
{
"epoch": 0.64849853515625,
"eval_accuracy": 0.03903444227005871,
"eval_loss": 6.481354713439941,
"eval_runtime": 245.5901,
"eval_samples_per_second": 40.718,
"eval_steps_per_second": 10.18,
"step": 340000
},
{
"epoch": 0.6494522094726562,
"grad_norm": 110.26327514648438,
"learning_rate": 1.752748489379883e-05,
"loss": 6.5545,
"step": 340500
},
{
"epoch": 0.6504058837890625,
"grad_norm": 295.2226867675781,
"learning_rate": 1.7479801177978516e-05,
"loss": 6.5221,
"step": 341000
},
{
"epoch": 0.6513595581054688,
"grad_norm": 3240.28369140625,
"learning_rate": 1.7432117462158203e-05,
"loss": 6.5653,
"step": 341500
},
{
"epoch": 0.652313232421875,
"grad_norm": 415.33953857421875,
"learning_rate": 1.738443374633789e-05,
"loss": 6.5583,
"step": 342000
},
{
"epoch": 0.6532669067382812,
"grad_norm": 546.7062377929688,
"learning_rate": 1.733675003051758e-05,
"loss": 6.5814,
"step": 342500
},
{
"epoch": 0.6542205810546875,
"grad_norm": 139.72181701660156,
"learning_rate": 1.7289066314697267e-05,
"loss": 6.5635,
"step": 343000
},
{
"epoch": 0.6551742553710938,
"grad_norm": 192.7628936767578,
"learning_rate": 1.7241382598876954e-05,
"loss": 6.5176,
"step": 343500
},
{
"epoch": 0.6561279296875,
"grad_norm": 296.3890686035156,
"learning_rate": 1.719369888305664e-05,
"loss": 6.5611,
"step": 344000
},
{
"epoch": 0.6570816040039062,
"grad_norm": 371.4485778808594,
"learning_rate": 1.7146015167236328e-05,
"loss": 6.5325,
"step": 344500
},
{
"epoch": 0.6580352783203125,
"grad_norm": 171.9521942138672,
"learning_rate": 1.7098331451416018e-05,
"loss": 6.5672,
"step": 345000
},
{
"epoch": 0.6580352783203125,
"eval_accuracy": 0.03944207436399217,
"eval_loss": 6.461095809936523,
"eval_runtime": 240.9495,
"eval_samples_per_second": 41.502,
"eval_steps_per_second": 10.376,
"step": 345000
},
{
"epoch": 0.6589889526367188,
"grad_norm": 135.15243530273438,
"learning_rate": 1.7050647735595705e-05,
"loss": 6.548,
"step": 345500
},
{
"epoch": 0.659942626953125,
"grad_norm": 263.35400390625,
"learning_rate": 1.700296401977539e-05,
"loss": 6.5543,
"step": 346000
},
{
"epoch": 0.6608963012695312,
"grad_norm": 144.79660034179688,
"learning_rate": 1.6955280303955078e-05,
"loss": 6.5284,
"step": 346500
},
{
"epoch": 0.6618499755859375,
"grad_norm": 112.84705352783203,
"learning_rate": 1.6907596588134765e-05,
"loss": 6.4988,
"step": 347000
},
{
"epoch": 0.6628036499023438,
"grad_norm": 122.29241943359375,
"learning_rate": 1.6859912872314455e-05,
"loss": 6.5073,
"step": 347500
},
{
"epoch": 0.66375732421875,
"grad_norm": 65.16890716552734,
"learning_rate": 1.6812229156494142e-05,
"loss": 6.515,
"step": 348000
},
{
"epoch": 0.6647109985351562,
"grad_norm": 198.87631225585938,
"learning_rate": 1.676454544067383e-05,
"loss": 6.6216,
"step": 348500
},
{
"epoch": 0.6656646728515625,
"grad_norm": 141.19139099121094,
"learning_rate": 1.6716861724853516e-05,
"loss": 6.5619,
"step": 349000
},
{
"epoch": 0.6666183471679688,
"grad_norm": 52.12779235839844,
"learning_rate": 1.6669178009033203e-05,
"loss": 6.5689,
"step": 349500
},
{
"epoch": 0.667572021484375,
"grad_norm": 246.66909790039062,
"learning_rate": 1.6621494293212893e-05,
"loss": 6.5223,
"step": 350000
},
{
"epoch": 0.667572021484375,
"eval_accuracy": 0.03970039138943249,
"eval_loss": 6.462384223937988,
"eval_runtime": 239.8458,
"eval_samples_per_second": 41.693,
"eval_steps_per_second": 10.423,
"step": 350000
},
{
"epoch": 0.6685256958007812,
"grad_norm": 319.72589111328125,
"learning_rate": 1.657381057739258e-05,
"loss": 6.5863,
"step": 350500
},
{
"epoch": 0.6694793701171875,
"grad_norm": 196.7140350341797,
"learning_rate": 1.6526126861572266e-05,
"loss": 6.5362,
"step": 351000
},
{
"epoch": 0.6704330444335938,
"grad_norm": 502.68743896484375,
"learning_rate": 1.6478443145751953e-05,
"loss": 6.5761,
"step": 351500
},
{
"epoch": 0.67138671875,
"grad_norm": 92.79645538330078,
"learning_rate": 1.643075942993164e-05,
"loss": 6.5914,
"step": 352000
},
{
"epoch": 0.6723403930664062,
"grad_norm": 236.26942443847656,
"learning_rate": 1.638307571411133e-05,
"loss": 6.5679,
"step": 352500
},
{
"epoch": 0.6732940673828125,
"grad_norm": 123.47889709472656,
"learning_rate": 1.6335391998291017e-05,
"loss": 6.5484,
"step": 353000
},
{
"epoch": 0.6742477416992188,
"grad_norm": 95.9494857788086,
"learning_rate": 1.6287708282470704e-05,
"loss": 6.5179,
"step": 353500
},
{
"epoch": 0.675201416015625,
"grad_norm": 116.54471588134766,
"learning_rate": 1.624002456665039e-05,
"loss": 6.4845,
"step": 354000
},
{
"epoch": 0.6761550903320312,
"grad_norm": 137.38478088378906,
"learning_rate": 1.6192340850830078e-05,
"loss": 6.5456,
"step": 354500
},
{
"epoch": 0.6771087646484375,
"grad_norm": 102.44841766357422,
"learning_rate": 1.6144657135009768e-05,
"loss": 6.518,
"step": 355000
},
{
"epoch": 0.6771087646484375,
"eval_accuracy": 0.040165949119373774,
"eval_loss": 6.473403453826904,
"eval_runtime": 242.3812,
"eval_samples_per_second": 41.257,
"eval_steps_per_second": 10.314,
"step": 355000
},
{
"epoch": 0.6780624389648438,
"grad_norm": 143.36158752441406,
"learning_rate": 1.6096973419189455e-05,
"loss": 6.5395,
"step": 355500
},
{
"epoch": 0.67901611328125,
"grad_norm": 88.15251922607422,
"learning_rate": 1.604928970336914e-05,
"loss": 6.5379,
"step": 356000
},
{
"epoch": 0.6799697875976562,
"grad_norm": 77.68293762207031,
"learning_rate": 1.6001605987548828e-05,
"loss": 6.5462,
"step": 356500
},
{
"epoch": 0.6809234619140625,
"grad_norm": 269.4987487792969,
"learning_rate": 1.5953922271728515e-05,
"loss": 6.5495,
"step": 357000
},
{
"epoch": 0.6818771362304688,
"grad_norm": 158.91729736328125,
"learning_rate": 1.5906238555908205e-05,
"loss": 6.5422,
"step": 357500
},
{
"epoch": 0.682830810546875,
"grad_norm": 104.279052734375,
"learning_rate": 1.5858554840087892e-05,
"loss": 6.53,
"step": 358000
},
{
"epoch": 0.6837844848632812,
"grad_norm": 232.02540588378906,
"learning_rate": 1.581087112426758e-05,
"loss": 6.5474,
"step": 358500
},
{
"epoch": 0.6847381591796875,
"grad_norm": 145.39788818359375,
"learning_rate": 1.5763187408447266e-05,
"loss": 6.5581,
"step": 359000
},
{
"epoch": 0.6856918334960938,
"grad_norm": 208.22569274902344,
"learning_rate": 1.5715503692626953e-05,
"loss": 6.5472,
"step": 359500
},
{
"epoch": 0.6866455078125,
"grad_norm": 187.49270629882812,
"learning_rate": 1.5667819976806643e-05,
"loss": 6.5466,
"step": 360000
},
{
"epoch": 0.6866455078125,
"eval_accuracy": 0.03987984344422701,
"eval_loss": 6.463972568511963,
"eval_runtime": 242.9807,
"eval_samples_per_second": 41.156,
"eval_steps_per_second": 10.289,
"step": 360000
},
{
"epoch": 0.6875991821289062,
"grad_norm": 115.62727355957031,
"learning_rate": 1.562013626098633e-05,
"loss": 6.5215,
"step": 360500
},
{
"epoch": 0.6885528564453125,
"grad_norm": 179.17079162597656,
"learning_rate": 1.5572452545166016e-05,
"loss": 6.5396,
"step": 361000
},
{
"epoch": 0.6895065307617188,
"grad_norm": 518.125,
"learning_rate": 1.5524768829345703e-05,
"loss": 6.5477,
"step": 361500
},
{
"epoch": 0.690460205078125,
"grad_norm": 259.0045166015625,
"learning_rate": 1.547708511352539e-05,
"loss": 6.5385,
"step": 362000
},
{
"epoch": 0.6914138793945312,
"grad_norm": 146.81991577148438,
"learning_rate": 1.542940139770508e-05,
"loss": 6.531,
"step": 362500
},
{
"epoch": 0.6923675537109375,
"grad_norm": 109.12883758544922,
"learning_rate": 1.5381717681884767e-05,
"loss": 6.4949,
"step": 363000
},
{
"epoch": 0.6933212280273438,
"grad_norm": 148.35174560546875,
"learning_rate": 1.5334033966064454e-05,
"loss": 6.4647,
"step": 363500
},
{
"epoch": 0.69427490234375,
"grad_norm": 137.14820861816406,
"learning_rate": 1.528635025024414e-05,
"loss": 6.4803,
"step": 364000
},
{
"epoch": 0.6952285766601562,
"grad_norm": 194.35723876953125,
"learning_rate": 1.523866653442383e-05,
"loss": 6.4629,
"step": 364500
},
{
"epoch": 0.6961822509765625,
"grad_norm": 88.89530944824219,
"learning_rate": 1.5190982818603516e-05,
"loss": 6.5213,
"step": 365000
},
{
"epoch": 0.6961822509765625,
"eval_accuracy": 0.04012818003913894,
"eval_loss": 6.462370872497559,
"eval_runtime": 243.706,
"eval_samples_per_second": 41.033,
"eval_steps_per_second": 10.258,
"step": 365000
},
{
"epoch": 0.6971359252929688,
"grad_norm": Infinity,
"learning_rate": 1.5143299102783205e-05,
"loss": 6.6055,
"step": 365500
},
{
"epoch": 0.698089599609375,
"grad_norm": 180.16140747070312,
"learning_rate": 1.5095615386962891e-05,
"loss": 6.5766,
"step": 366000
},
{
"epoch": 0.6990432739257812,
"grad_norm": 196.64334106445312,
"learning_rate": 1.5047931671142578e-05,
"loss": 6.6007,
"step": 366500
},
{
"epoch": 0.6999969482421875,
"grad_norm": 246.12445068359375,
"learning_rate": 1.5000247955322267e-05,
"loss": 6.6011,
"step": 367000
},
{
"epoch": 0.7009506225585938,
"grad_norm": 99.25709533691406,
"learning_rate": 1.4952564239501954e-05,
"loss": 6.528,
"step": 367500
},
{
"epoch": 0.701904296875,
"grad_norm": 255.1567840576172,
"learning_rate": 1.4904880523681642e-05,
"loss": 6.534,
"step": 368000
},
{
"epoch": 0.7028579711914062,
"grad_norm": 472.0591125488281,
"learning_rate": 1.4857196807861329e-05,
"loss": 6.5193,
"step": 368500
},
{
"epoch": 0.7038116455078125,
"grad_norm": 386.86834716796875,
"learning_rate": 1.4809513092041016e-05,
"loss": 6.547,
"step": 369000
},
{
"epoch": 0.7047653198242188,
"grad_norm": 147.0093231201172,
"learning_rate": 1.4761829376220704e-05,
"loss": 6.5236,
"step": 369500
},
{
"epoch": 0.705718994140625,
"grad_norm": 447.5209045410156,
"learning_rate": 1.4714145660400391e-05,
"loss": 6.5881,
"step": 370000
},
{
"epoch": 0.705718994140625,
"eval_accuracy": 0.03990039138943249,
"eval_loss": 6.450451374053955,
"eval_runtime": 244.0341,
"eval_samples_per_second": 40.978,
"eval_steps_per_second": 10.244,
"step": 370000
},
{
"epoch": 0.7066726684570312,
"grad_norm": 243.3019256591797,
"learning_rate": 1.466646194458008e-05,
"loss": 6.5152,
"step": 370500
},
{
"epoch": 0.7076263427734375,
"grad_norm": 129.59954833984375,
"learning_rate": 1.4618778228759766e-05,
"loss": 6.5965,
"step": 371000
},
{
"epoch": 0.7085800170898438,
"grad_norm": 88.1178970336914,
"learning_rate": 1.4571094512939453e-05,
"loss": 6.5418,
"step": 371500
},
{
"epoch": 0.70953369140625,
"grad_norm": 170.9890594482422,
"learning_rate": 1.4523410797119142e-05,
"loss": 6.5358,
"step": 372000
},
{
"epoch": 0.7104873657226562,
"grad_norm": 765.4295654296875,
"learning_rate": 1.4475727081298829e-05,
"loss": 6.5584,
"step": 372500
},
{
"epoch": 0.7114410400390625,
"grad_norm": 296.3957824707031,
"learning_rate": 1.4428043365478517e-05,
"loss": 6.537,
"step": 373000
},
{
"epoch": 0.7123947143554688,
"grad_norm": 238.33856201171875,
"learning_rate": 1.4380359649658204e-05,
"loss": 6.5633,
"step": 373500
},
{
"epoch": 0.713348388671875,
"grad_norm": 175.88967895507812,
"learning_rate": 1.433267593383789e-05,
"loss": 6.5366,
"step": 374000
},
{
"epoch": 0.7143020629882812,
"grad_norm": 101.9058609008789,
"learning_rate": 1.428499221801758e-05,
"loss": 6.5549,
"step": 374500
},
{
"epoch": 0.7152557373046875,
"grad_norm": 140.43959045410156,
"learning_rate": 1.4237308502197266e-05,
"loss": 6.5353,
"step": 375000
},
{
"epoch": 0.7152557373046875,
"eval_accuracy": 0.040435616438356164,
"eval_loss": 6.447012901306152,
"eval_runtime": 239.2555,
"eval_samples_per_second": 41.796,
"eval_steps_per_second": 10.449,
"step": 375000
},
{
"epoch": 0.7162094116210938,
"grad_norm": 141.53201293945312,
"learning_rate": 1.4189624786376955e-05,
"loss": 6.4952,
"step": 375500
},
{
"epoch": 0.7171630859375,
"grad_norm": 264.2409973144531,
"learning_rate": 1.4141941070556641e-05,
"loss": 6.5382,
"step": 376000
},
{
"epoch": 0.7181167602539062,
"grad_norm": 667.0358276367188,
"learning_rate": 1.4094257354736328e-05,
"loss": 6.5211,
"step": 376500
},
{
"epoch": 0.7190704345703125,
"grad_norm": 391.56927490234375,
"learning_rate": 1.4046573638916017e-05,
"loss": 6.5214,
"step": 377000
},
{
"epoch": 0.7200241088867188,
"grad_norm": 141.28468322753906,
"learning_rate": 1.3998889923095704e-05,
"loss": 6.5409,
"step": 377500
},
{
"epoch": 0.720977783203125,
"grad_norm": 377.8066711425781,
"learning_rate": 1.3951206207275392e-05,
"loss": 6.5413,
"step": 378000
},
{
"epoch": 0.7219314575195312,
"grad_norm": 79.05961608886719,
"learning_rate": 1.3903522491455079e-05,
"loss": 6.5492,
"step": 378500
},
{
"epoch": 0.7228851318359375,
"grad_norm": 195.48342895507812,
"learning_rate": 1.3855838775634766e-05,
"loss": 6.5255,
"step": 379000
},
{
"epoch": 0.7238388061523438,
"grad_norm": 131.9173583984375,
"learning_rate": 1.3808155059814454e-05,
"loss": 6.5142,
"step": 379500
},
{
"epoch": 0.72479248046875,
"grad_norm": 187.51319885253906,
"learning_rate": 1.3760471343994141e-05,
"loss": 6.5371,
"step": 380000
},
{
"epoch": 0.72479248046875,
"eval_accuracy": 0.040126810176125245,
"eval_loss": 6.445300102233887,
"eval_runtime": 240.0497,
"eval_samples_per_second": 41.658,
"eval_steps_per_second": 10.415,
"step": 380000
},
{
"epoch": 0.7257461547851562,
"grad_norm": 355.3277893066406,
"learning_rate": 1.371278762817383e-05,
"loss": 6.5643,
"step": 380500
},
{
"epoch": 0.7266998291015625,
"grad_norm": 132.3724365234375,
"learning_rate": 1.3665103912353516e-05,
"loss": 6.5518,
"step": 381000
},
{
"epoch": 0.7276535034179688,
"grad_norm": 138.26895141601562,
"learning_rate": 1.3617420196533203e-05,
"loss": 6.5359,
"step": 381500
},
{
"epoch": 0.728607177734375,
"grad_norm": 127.5734634399414,
"learning_rate": 1.3569736480712892e-05,
"loss": 6.5026,
"step": 382000
},
{
"epoch": 0.7295608520507812,
"grad_norm": 194.17762756347656,
"learning_rate": 1.3522052764892579e-05,
"loss": 6.4611,
"step": 382500
},
{
"epoch": 0.7305145263671875,
"grad_norm": 97.86437225341797,
"learning_rate": 1.3474369049072265e-05,
"loss": 6.4943,
"step": 383000
},
{
"epoch": 0.7314682006835938,
"grad_norm": 88.39631652832031,
"learning_rate": 1.3426685333251954e-05,
"loss": 6.4875,
"step": 383500
},
{
"epoch": 0.732421875,
"grad_norm": 220.22547912597656,
"learning_rate": 1.337900161743164e-05,
"loss": 6.4643,
"step": 384000
},
{
"epoch": 0.7333755493164062,
"grad_norm": 130.28128051757812,
"learning_rate": 1.333131790161133e-05,
"loss": 6.4525,
"step": 384500
},
{
"epoch": 0.7343292236328125,
"grad_norm": 151.72984313964844,
"learning_rate": 1.3283634185791016e-05,
"loss": 6.5153,
"step": 385000
},
{
"epoch": 0.7343292236328125,
"eval_accuracy": 0.04008082191780822,
"eval_loss": 6.446129322052002,
"eval_runtime": 241.548,
"eval_samples_per_second": 41.4,
"eval_steps_per_second": 10.35,
"step": 385000
},
{
"epoch": 0.7352828979492188,
"grad_norm": 91.2538070678711,
"learning_rate": 1.3235950469970703e-05,
"loss": 6.558,
"step": 385500
},
{
"epoch": 0.736236572265625,
"grad_norm": 772.9464721679688,
"learning_rate": 1.3188266754150391e-05,
"loss": 6.5602,
"step": 386000
},
{
"epoch": 0.7371902465820312,
"grad_norm": 136.80430603027344,
"learning_rate": 1.3140583038330078e-05,
"loss": 6.5403,
"step": 386500
},
{
"epoch": 0.7381439208984375,
"grad_norm": 85.5173110961914,
"learning_rate": 1.3092899322509767e-05,
"loss": 6.52,
"step": 387000
},
{
"epoch": 0.7390975952148438,
"grad_norm": 575.4596557617188,
"learning_rate": 1.3045215606689454e-05,
"loss": 6.5036,
"step": 387500
},
{
"epoch": 0.74005126953125,
"grad_norm": 265.74041748046875,
"learning_rate": 1.299753189086914e-05,
"loss": 6.563,
"step": 388000
},
{
"epoch": 0.7410049438476562,
"grad_norm": 202.33619689941406,
"learning_rate": 1.2949848175048829e-05,
"loss": 6.5295,
"step": 388500
},
{
"epoch": 0.7419586181640625,
"grad_norm": 158.96939086914062,
"learning_rate": 1.2902164459228516e-05,
"loss": 6.5082,
"step": 389000
},
{
"epoch": 0.7429122924804688,
"grad_norm": 277.22186279296875,
"learning_rate": 1.2854480743408204e-05,
"loss": 6.5653,
"step": 389500
},
{
"epoch": 0.743865966796875,
"grad_norm": 69.9422836303711,
"learning_rate": 1.2806797027587891e-05,
"loss": 6.5488,
"step": 390000
},
{
"epoch": 0.743865966796875,
"eval_accuracy": 0.040149706457925635,
"eval_loss": 6.443368911743164,
"eval_runtime": 240.93,
"eval_samples_per_second": 41.506,
"eval_steps_per_second": 10.376,
"step": 390000
},
{
"epoch": 0.7448196411132812,
"grad_norm": 84.9781265258789,
"learning_rate": 1.2759113311767578e-05,
"loss": 6.5277,
"step": 390500
},
{
"epoch": 0.7457733154296875,
"grad_norm": 211.03887939453125,
"learning_rate": 1.2711429595947266e-05,
"loss": 6.5508,
"step": 391000
},
{
"epoch": 0.7467269897460938,
"grad_norm": 120.260009765625,
"learning_rate": 1.2663745880126953e-05,
"loss": 6.4981,
"step": 391500
},
{
"epoch": 0.7476806640625,
"grad_norm": 111.86019897460938,
"learning_rate": 1.2616062164306642e-05,
"loss": 6.5016,
"step": 392000
},
{
"epoch": 0.7486343383789062,
"grad_norm": 136.80467224121094,
"learning_rate": 1.2568378448486329e-05,
"loss": 6.5204,
"step": 392500
},
{
"epoch": 0.7495880126953125,
"grad_norm": 221.40573120117188,
"learning_rate": 1.2520694732666015e-05,
"loss": 6.5076,
"step": 393000
},
{
"epoch": 0.7505416870117188,
"grad_norm": 174.58226013183594,
"learning_rate": 1.2473011016845704e-05,
"loss": 6.5282,
"step": 393500
},
{
"epoch": 0.751495361328125,
"grad_norm": 108.00682830810547,
"learning_rate": 1.242532730102539e-05,
"loss": 6.5193,
"step": 394000
},
{
"epoch": 0.7524490356445312,
"grad_norm": 110.358154296875,
"learning_rate": 1.237764358520508e-05,
"loss": 6.522,
"step": 394500
},
{
"epoch": 0.7534027099609375,
"grad_norm": 114.46678924560547,
"learning_rate": 1.2329959869384766e-05,
"loss": 6.5121,
"step": 395000
},
{
"epoch": 0.7534027099609375,
"eval_accuracy": 0.03947162426614481,
"eval_loss": 6.446810722351074,
"eval_runtime": 240.4854,
"eval_samples_per_second": 41.583,
"eval_steps_per_second": 10.396,
"step": 395000
},
{
"epoch": 0.7543563842773438,
"grad_norm": 686.0939331054688,
"learning_rate": 1.2282276153564453e-05,
"loss": 6.4873,
"step": 395500
},
{
"epoch": 0.75531005859375,
"grad_norm": 323.3516540527344,
"learning_rate": 1.2234592437744141e-05,
"loss": 6.4371,
"step": 396000
},
{
"epoch": 0.7562637329101562,
"grad_norm": 552.0935668945312,
"learning_rate": 1.2186908721923828e-05,
"loss": 6.5156,
"step": 396500
},
{
"epoch": 0.7572174072265625,
"grad_norm": 145.94496154785156,
"learning_rate": 1.2139225006103517e-05,
"loss": 6.5244,
"step": 397000
},
{
"epoch": 0.7581710815429688,
"grad_norm": 213.95999145507812,
"learning_rate": 1.2091541290283204e-05,
"loss": 6.5308,
"step": 397500
},
{
"epoch": 0.759124755859375,
"grad_norm": 179.3788299560547,
"learning_rate": 1.204385757446289e-05,
"loss": 6.4978,
"step": 398000
},
{
"epoch": 0.7600784301757812,
"grad_norm": 479.97125244140625,
"learning_rate": 1.1996173858642579e-05,
"loss": 6.5211,
"step": 398500
},
{
"epoch": 0.7610321044921875,
"grad_norm": 182.0722198486328,
"learning_rate": 1.1948490142822266e-05,
"loss": 6.5162,
"step": 399000
},
{
"epoch": 0.7619857788085938,
"grad_norm": 134.76768493652344,
"learning_rate": 1.1900806427001954e-05,
"loss": 6.4966,
"step": 399500
},
{
"epoch": 0.762939453125,
"grad_norm": 229.81442260742188,
"learning_rate": 1.1853122711181641e-05,
"loss": 6.5106,
"step": 400000
},
{
"epoch": 0.762939453125,
"eval_accuracy": 0.039886692759295496,
"eval_loss": 6.436385154724121,
"eval_runtime": 242.7965,
"eval_samples_per_second": 41.187,
"eval_steps_per_second": 10.297,
"step": 400000
},
{
"epoch": 0.7638931274414062,
"grad_norm": 249.11834716796875,
"learning_rate": 1.1805438995361328e-05,
"loss": 6.5044,
"step": 400500
},
{
"epoch": 0.7648468017578125,
"grad_norm": 122.83702850341797,
"learning_rate": 1.1757755279541016e-05,
"loss": 6.4648,
"step": 401000
},
{
"epoch": 0.7658004760742188,
"grad_norm": 123.14421844482422,
"learning_rate": 1.1710071563720703e-05,
"loss": 6.4555,
"step": 401500
},
{
"epoch": 0.766754150390625,
"grad_norm": 275.2742004394531,
"learning_rate": 1.1662387847900392e-05,
"loss": 6.4705,
"step": 402000
},
{
"epoch": 0.7677078247070312,
"grad_norm": 132.36386108398438,
"learning_rate": 1.1614704132080079e-05,
"loss": 6.4411,
"step": 402500
},
{
"epoch": 0.7686614990234375,
"grad_norm": 75.54257202148438,
"learning_rate": 1.1567020416259765e-05,
"loss": 6.5292,
"step": 403000
},
{
"epoch": 0.7696151733398438,
"grad_norm": 352.4110412597656,
"learning_rate": 1.1519336700439454e-05,
"loss": 6.47,
"step": 403500
},
{
"epoch": 0.77056884765625,
"grad_norm": 138.96554565429688,
"learning_rate": 1.147165298461914e-05,
"loss": 6.5295,
"step": 404000
},
{
"epoch": 0.7715225219726562,
"grad_norm": 162.00111389160156,
"learning_rate": 1.142396926879883e-05,
"loss": 6.5852,
"step": 404500
},
{
"epoch": 0.7724761962890625,
"grad_norm": 119.4192886352539,
"learning_rate": 1.1376285552978516e-05,
"loss": 6.5572,
"step": 405000
},
{
"epoch": 0.7724761962890625,
"eval_accuracy": 0.0405041095890411,
"eval_loss": 6.444308757781982,
"eval_runtime": 239.539,
"eval_samples_per_second": 41.747,
"eval_steps_per_second": 10.437,
"step": 405000
},
{
"epoch": 0.7734298706054688,
"grad_norm": 112.89866638183594,
"learning_rate": 1.1328601837158203e-05,
"loss": 6.5389,
"step": 405500
},
{
"epoch": 0.774383544921875,
"grad_norm": 404.556640625,
"learning_rate": 1.1280918121337891e-05,
"loss": 6.563,
"step": 406000
},
{
"epoch": 0.7753372192382812,
"grad_norm": 448.70306396484375,
"learning_rate": 1.1233234405517578e-05,
"loss": 6.5147,
"step": 406500
},
{
"epoch": 0.7762908935546875,
"grad_norm": 48.68925857543945,
"learning_rate": 1.1185550689697267e-05,
"loss": 6.5531,
"step": 407000
},
{
"epoch": 0.7772445678710938,
"grad_norm": 332.9776611328125,
"learning_rate": 1.1137866973876954e-05,
"loss": 6.4849,
"step": 407500
},
{
"epoch": 0.7781982421875,
"grad_norm": 134.01174926757812,
"learning_rate": 1.109018325805664e-05,
"loss": 6.4986,
"step": 408000
},
{
"epoch": 0.7791519165039062,
"grad_norm": 172.95147705078125,
"learning_rate": 1.1042499542236329e-05,
"loss": 6.5161,
"step": 408500
},
{
"epoch": 0.7801055908203125,
"grad_norm": 178.749267578125,
"learning_rate": 1.0994815826416016e-05,
"loss": 6.5186,
"step": 409000
},
{
"epoch": 0.7810592651367188,
"grad_norm": 1029.6717529296875,
"learning_rate": 1.0947132110595704e-05,
"loss": 6.5478,
"step": 409500
},
{
"epoch": 0.782012939453125,
"grad_norm": 202.45565795898438,
"learning_rate": 1.0899448394775391e-05,
"loss": 6.494,
"step": 410000
},
{
"epoch": 0.782012939453125,
"eval_accuracy": 0.040156360078277886,
"eval_loss": 6.433761119842529,
"eval_runtime": 241.6954,
"eval_samples_per_second": 41.374,
"eval_steps_per_second": 10.344,
"step": 410000
},
{
"epoch": 0.7829666137695312,
"grad_norm": 999.4024047851562,
"learning_rate": 1.0851764678955078e-05,
"loss": 6.46,
"step": 410500
},
{
"epoch": 0.7839202880859375,
"grad_norm": 179.73165893554688,
"learning_rate": 1.0804080963134766e-05,
"loss": 6.5209,
"step": 411000
},
{
"epoch": 0.7848739624023438,
"grad_norm": 234.8481903076172,
"learning_rate": 1.0756397247314453e-05,
"loss": 6.5324,
"step": 411500
},
{
"epoch": 0.78582763671875,
"grad_norm": 126.8324966430664,
"learning_rate": 1.0708713531494142e-05,
"loss": 6.5223,
"step": 412000
},
{
"epoch": 0.7867813110351562,
"grad_norm": 92.26342010498047,
"learning_rate": 1.0661029815673829e-05,
"loss": 6.5303,
"step": 412500
},
{
"epoch": 0.7877349853515625,
"grad_norm": 309.4622497558594,
"learning_rate": 1.0613346099853515e-05,
"loss": 6.5365,
"step": 413000
},
{
"epoch": 0.7886886596679688,
"grad_norm": 124.4228286743164,
"learning_rate": 1.0565662384033204e-05,
"loss": 6.5303,
"step": 413500
},
{
"epoch": 0.789642333984375,
"grad_norm": 134.06324768066406,
"learning_rate": 1.051797866821289e-05,
"loss": 6.5122,
"step": 414000
},
{
"epoch": 0.7905960083007812,
"grad_norm": 749.5646362304688,
"learning_rate": 1.047029495239258e-05,
"loss": 6.5199,
"step": 414500
},
{
"epoch": 0.7915496826171875,
"grad_norm": 158.85972595214844,
"learning_rate": 1.0422611236572266e-05,
"loss": 6.5233,
"step": 415000
},
{
"epoch": 0.7915496826171875,
"eval_accuracy": 0.04043620352250489,
"eval_loss": 6.431772232055664,
"eval_runtime": 247.2315,
"eval_samples_per_second": 40.448,
"eval_steps_per_second": 10.112,
"step": 415000
},
{
"epoch": 0.7925033569335938,
"grad_norm": 337.47760009765625,
"learning_rate": 1.0374927520751953e-05,
"loss": 6.5102,
"step": 415500
},
{
"epoch": 0.79345703125,
"grad_norm": 134.16856384277344,
"learning_rate": 1.0327243804931641e-05,
"loss": 6.5078,
"step": 416000
},
{
"epoch": 0.7944107055664062,
"grad_norm": 394.0135192871094,
"learning_rate": 1.0279560089111328e-05,
"loss": 6.5206,
"step": 416500
},
{
"epoch": 0.7953643798828125,
"grad_norm": 176.5369415283203,
"learning_rate": 1.0231876373291017e-05,
"loss": 6.5359,
"step": 417000
},
{
"epoch": 0.7963180541992188,
"grad_norm": 180.64622497558594,
"learning_rate": 1.0184192657470704e-05,
"loss": 6.4871,
"step": 417500
},
{
"epoch": 0.797271728515625,
"grad_norm": 299.3851013183594,
"learning_rate": 1.013650894165039e-05,
"loss": 6.5244,
"step": 418000
},
{
"epoch": 0.7982254028320312,
"grad_norm": 844.39990234375,
"learning_rate": 1.0088825225830079e-05,
"loss": 6.4919,
"step": 418500
},
{
"epoch": 0.7991790771484375,
"grad_norm": 264.82891845703125,
"learning_rate": 1.0041141510009766e-05,
"loss": 6.4495,
"step": 419000
},
{
"epoch": 0.8001327514648438,
"grad_norm": 470.61968994140625,
"learning_rate": 9.993457794189454e-06,
"loss": 6.4704,
"step": 419500
},
{
"epoch": 0.80108642578125,
"grad_norm": 1155.0771484375,
"learning_rate": 9.945774078369141e-06,
"loss": 6.4718,
"step": 420000
},
{
"epoch": 0.80108642578125,
"eval_accuracy": 0.041468884540117414,
"eval_loss": 6.425516128540039,
"eval_runtime": 241.3638,
"eval_samples_per_second": 41.431,
"eval_steps_per_second": 10.358,
"step": 420000
},
{
"epoch": 0.8020401000976562,
"grad_norm": 182.08621215820312,
"learning_rate": 9.898090362548828e-06,
"loss": 6.4815,
"step": 420500
},
{
"epoch": 0.8029937744140625,
"grad_norm": 159.288330078125,
"learning_rate": 9.850406646728516e-06,
"loss": 6.5181,
"step": 421000
},
{
"epoch": 0.8039474487304688,
"grad_norm": 503.52655029296875,
"learning_rate": 9.802722930908203e-06,
"loss": 6.5563,
"step": 421500
},
{
"epoch": 0.804901123046875,
"grad_norm": 136.9956817626953,
"learning_rate": 9.755039215087892e-06,
"loss": 6.5364,
"step": 422000
},
{
"epoch": 0.8058547973632812,
"grad_norm": 124.22718048095703,
"learning_rate": 9.707355499267579e-06,
"loss": 6.5407,
"step": 422500
},
{
"epoch": 0.8068084716796875,
"grad_norm": 471.5389404296875,
"learning_rate": 9.659671783447265e-06,
"loss": 6.5318,
"step": 423000
},
{
"epoch": 0.8077621459960938,
"grad_norm": 334.5975341796875,
"learning_rate": 9.611988067626954e-06,
"loss": 6.5438,
"step": 423500
},
{
"epoch": 0.8087158203125,
"grad_norm": 294.684326171875,
"learning_rate": 9.56430435180664e-06,
"loss": 6.492,
"step": 424000
},
{
"epoch": 0.8096694946289062,
"grad_norm": 143.47584533691406,
"learning_rate": 9.51662063598633e-06,
"loss": 6.5305,
"step": 424500
},
{
"epoch": 0.8106231689453125,
"grad_norm": 219.26388549804688,
"learning_rate": 9.468936920166016e-06,
"loss": 6.4915,
"step": 425000
},
{
"epoch": 0.8106231689453125,
"eval_accuracy": 0.040221135029354205,
"eval_loss": 6.426062107086182,
"eval_runtime": 238.9838,
"eval_samples_per_second": 41.844,
"eval_steps_per_second": 10.461,
"step": 425000
},
{
"epoch": 0.8115768432617188,
"grad_norm": 85.5811996459961,
"learning_rate": 9.421253204345703e-06,
"loss": 6.508,
"step": 425500
},
{
"epoch": 0.812530517578125,
"grad_norm": 134.37989807128906,
"learning_rate": 9.373569488525391e-06,
"loss": 6.5178,
"step": 426000
},
{
"epoch": 0.8134841918945312,
"grad_norm": 824.7955932617188,
"learning_rate": 9.325885772705078e-06,
"loss": 6.5363,
"step": 426500
},
{
"epoch": 0.8144378662109375,
"grad_norm": 134.56495666503906,
"learning_rate": 9.278202056884767e-06,
"loss": 6.5228,
"step": 427000
},
{
"epoch": 0.8153915405273438,
"grad_norm": 320.2661437988281,
"learning_rate": 9.230518341064454e-06,
"loss": 6.5003,
"step": 427500
},
{
"epoch": 0.81634521484375,
"grad_norm": 97.85888671875,
"learning_rate": 9.18283462524414e-06,
"loss": 6.512,
"step": 428000
},
{
"epoch": 0.8172988891601562,
"grad_norm": 113.15918731689453,
"learning_rate": 9.135150909423829e-06,
"loss": 6.5378,
"step": 428500
},
{
"epoch": 0.8182525634765625,
"grad_norm": 104.32511138916016,
"learning_rate": 9.087467193603516e-06,
"loss": 6.498,
"step": 429000
},
{
"epoch": 0.8192062377929688,
"grad_norm": 139.578857421875,
"learning_rate": 9.039783477783204e-06,
"loss": 6.5095,
"step": 429500
},
{
"epoch": 0.820159912109375,
"grad_norm": 152.16114807128906,
"learning_rate": 8.992099761962891e-06,
"loss": 6.5007,
"step": 430000
},
{
"epoch": 0.820159912109375,
"eval_accuracy": 0.0407105675146771,
"eval_loss": 6.424362659454346,
"eval_runtime": 238.8953,
"eval_samples_per_second": 41.859,
"eval_steps_per_second": 10.465,
"step": 430000
},
{
"epoch": 0.8211135864257812,
"grad_norm": 178.8160400390625,
"learning_rate": 8.944416046142578e-06,
"loss": 6.4954,
"step": 430500
},
{
"epoch": 0.8220672607421875,
"grad_norm": 467.9737548828125,
"learning_rate": 8.896732330322266e-06,
"loss": 6.522,
"step": 431000
},
{
"epoch": 0.8230209350585938,
"grad_norm": 283.6147766113281,
"learning_rate": 8.849048614501953e-06,
"loss": 6.5218,
"step": 431500
},
{
"epoch": 0.823974609375,
"grad_norm": 129.71145629882812,
"learning_rate": 8.801364898681642e-06,
"loss": 6.4796,
"step": 432000
},
{
"epoch": 0.8249282836914062,
"grad_norm": 83.04669189453125,
"learning_rate": 8.753681182861329e-06,
"loss": 6.5349,
"step": 432500
},
{
"epoch": 0.8258819580078125,
"grad_norm": 258.3847351074219,
"learning_rate": 8.705997467041015e-06,
"loss": 6.4963,
"step": 433000
},
{
"epoch": 0.8268356323242188,
"grad_norm": 394.8684387207031,
"learning_rate": 8.658313751220704e-06,
"loss": 6.522,
"step": 433500
},
{
"epoch": 0.827789306640625,
"grad_norm": 164.72079467773438,
"learning_rate": 8.61063003540039e-06,
"loss": 6.5048,
"step": 434000
},
{
"epoch": 0.8287429809570312,
"grad_norm": 265.7942199707031,
"learning_rate": 8.56294631958008e-06,
"loss": 6.5104,
"step": 434500
},
{
"epoch": 0.8296966552734375,
"grad_norm": 123.39559936523438,
"learning_rate": 8.515262603759766e-06,
"loss": 6.488,
"step": 435000
},
{
"epoch": 0.8296966552734375,
"eval_accuracy": 0.04050900195694716,
"eval_loss": 6.418701171875,
"eval_runtime": 238.8227,
"eval_samples_per_second": 41.872,
"eval_steps_per_second": 10.468,
"step": 435000
},
{
"epoch": 0.8306503295898438,
"grad_norm": 109.87480926513672,
"learning_rate": 8.467578887939453e-06,
"loss": 6.4846,
"step": 435500
},
{
"epoch": 0.83160400390625,
"grad_norm": 240.57347106933594,
"learning_rate": 8.419895172119141e-06,
"loss": 6.4685,
"step": 436000
},
{
"epoch": 0.8325576782226562,
"grad_norm": 178.5532989501953,
"learning_rate": 8.372211456298828e-06,
"loss": 6.4504,
"step": 436500
},
{
"epoch": 0.8335113525390625,
"grad_norm": 548.3411254882812,
"learning_rate": 8.324527740478517e-06,
"loss": 6.4365,
"step": 437000
},
{
"epoch": 0.8344650268554688,
"grad_norm": 261.7312316894531,
"learning_rate": 8.276844024658204e-06,
"loss": 6.4829,
"step": 437500
},
{
"epoch": 0.835418701171875,
"grad_norm": 180.61085510253906,
"learning_rate": 8.22916030883789e-06,
"loss": 6.4908,
"step": 438000
},
{
"epoch": 0.8363723754882812,
"grad_norm": 141.17413330078125,
"learning_rate": 8.181476593017579e-06,
"loss": 6.5328,
"step": 438500
},
{
"epoch": 0.8373260498046875,
"grad_norm": 100.98048400878906,
"learning_rate": 8.133792877197266e-06,
"loss": 6.5346,
"step": 439000
},
{
"epoch": 0.8382797241210938,
"grad_norm": 131.7880401611328,
"learning_rate": 8.086109161376954e-06,
"loss": 6.5343,
"step": 439500
},
{
"epoch": 0.8392333984375,
"grad_norm": 143.68992614746094,
"learning_rate": 8.038425445556641e-06,
"loss": 6.5093,
"step": 440000
},
{
"epoch": 0.8392333984375,
"eval_accuracy": 0.04017945205479452,
"eval_loss": 6.416311740875244,
"eval_runtime": 239.7034,
"eval_samples_per_second": 41.718,
"eval_steps_per_second": 10.43,
"step": 440000
},
{
"epoch": 0.8401870727539062,
"grad_norm": 221.09878540039062,
"learning_rate": 7.990741729736328e-06,
"loss": 6.5131,
"step": 440500
},
{
"epoch": 0.8411407470703125,
"grad_norm": 147.80081176757812,
"learning_rate": 7.943058013916016e-06,
"loss": 6.4949,
"step": 441000
},
{
"epoch": 0.8420944213867188,
"grad_norm": 98.18143463134766,
"learning_rate": 7.895374298095703e-06,
"loss": 6.4659,
"step": 441500
},
{
"epoch": 0.843048095703125,
"grad_norm": 213.72427368164062,
"learning_rate": 7.847690582275392e-06,
"loss": 6.4745,
"step": 442000
},
{
"epoch": 0.8440017700195312,
"grad_norm": 171.76968383789062,
"learning_rate": 7.800006866455079e-06,
"loss": 6.5142,
"step": 442500
},
{
"epoch": 0.8449554443359375,
"grad_norm": 241.19647216796875,
"learning_rate": 7.752323150634765e-06,
"loss": 6.5102,
"step": 443000
},
{
"epoch": 0.8459091186523438,
"grad_norm": 111.6817626953125,
"learning_rate": 7.704639434814454e-06,
"loss": 6.4809,
"step": 443500
},
{
"epoch": 0.84686279296875,
"grad_norm": 168.18035888671875,
"learning_rate": 7.65695571899414e-06,
"loss": 6.5021,
"step": 444000
},
{
"epoch": 0.8478164672851562,
"grad_norm": 603.36572265625,
"learning_rate": 7.6092720031738284e-06,
"loss": 6.4793,
"step": 444500
},
{
"epoch": 0.8487701416015625,
"grad_norm": 228.9732666015625,
"learning_rate": 7.561588287353516e-06,
"loss": 6.5029,
"step": 445000
},
{
"epoch": 0.8487701416015625,
"eval_accuracy": 0.040287671232876715,
"eval_loss": 6.415238380432129,
"eval_runtime": 238.84,
"eval_samples_per_second": 41.869,
"eval_steps_per_second": 10.467,
"step": 445000
},
{
"epoch": 0.8497238159179688,
"grad_norm": 395.8603820800781,
"learning_rate": 7.513904571533204e-06,
"loss": 6.5165,
"step": 445500
},
{
"epoch": 0.850677490234375,
"grad_norm": 324.56048583984375,
"learning_rate": 7.466220855712891e-06,
"loss": 6.5002,
"step": 446000
},
{
"epoch": 0.8516311645507812,
"grad_norm": 150.86160278320312,
"learning_rate": 7.418537139892578e-06,
"loss": 6.5325,
"step": 446500
},
{
"epoch": 0.8525848388671875,
"grad_norm": 416.0628967285156,
"learning_rate": 7.370853424072266e-06,
"loss": 6.5135,
"step": 447000
},
{
"epoch": 0.8535385131835938,
"grad_norm": 210.53988647460938,
"learning_rate": 7.323169708251954e-06,
"loss": 6.5016,
"step": 447500
},
{
"epoch": 0.8544921875,
"grad_norm": 228.83592224121094,
"learning_rate": 7.275485992431641e-06,
"loss": 6.529,
"step": 448000
},
{
"epoch": 0.8554458618164062,
"grad_norm": 73.7596206665039,
"learning_rate": 7.227802276611328e-06,
"loss": 6.4851,
"step": 448500
},
{
"epoch": 0.8563995361328125,
"grad_norm": 113.41374969482422,
"learning_rate": 7.180118560791016e-06,
"loss": 6.4297,
"step": 449000
},
{
"epoch": 0.8573532104492188,
"grad_norm": 164.508056640625,
"learning_rate": 7.1324348449707034e-06,
"loss": 6.5079,
"step": 449500
},
{
"epoch": 0.858306884765625,
"grad_norm": 103.63333892822266,
"learning_rate": 7.084751129150391e-06,
"loss": 6.3943,
"step": 450000
},
{
"epoch": 0.858306884765625,
"eval_accuracy": 0.04016125244618395,
"eval_loss": 6.420052528381348,
"eval_runtime": 238.5095,
"eval_samples_per_second": 41.927,
"eval_steps_per_second": 10.482,
"step": 450000
},
{
"epoch": 0.8592605590820312,
"grad_norm": 207.37977600097656,
"learning_rate": 7.037067413330079e-06,
"loss": 6.4935,
"step": 450500
},
{
"epoch": 0.8602142333984375,
"grad_norm": 143.06759643554688,
"learning_rate": 6.989383697509766e-06,
"loss": 6.4474,
"step": 451000
},
{
"epoch": 0.8611679077148438,
"grad_norm": 237.50033569335938,
"learning_rate": 6.941699981689453e-06,
"loss": 6.5138,
"step": 451500
},
{
"epoch": 0.86212158203125,
"grad_norm": 86.28911590576172,
"learning_rate": 6.894016265869141e-06,
"loss": 6.4986,
"step": 452000
},
{
"epoch": 0.8630752563476562,
"grad_norm": 249.3565673828125,
"learning_rate": 6.846332550048829e-06,
"loss": 6.4488,
"step": 452500
},
{
"epoch": 0.8640289306640625,
"grad_norm": 161.2953643798828,
"learning_rate": 6.798648834228516e-06,
"loss": 6.4201,
"step": 453000
},
{
"epoch": 0.8649826049804688,
"grad_norm": 94.91120147705078,
"learning_rate": 6.750965118408203e-06,
"loss": 6.4487,
"step": 453500
},
{
"epoch": 0.865936279296875,
"grad_norm": 160.8304901123047,
"learning_rate": 6.703281402587891e-06,
"loss": 6.4593,
"step": 454000
},
{
"epoch": 0.8668899536132812,
"grad_norm": 121.05160522460938,
"learning_rate": 6.6555976867675784e-06,
"loss": 6.5208,
"step": 454500
},
{
"epoch": 0.8678436279296875,
"grad_norm": 161.79888916015625,
"learning_rate": 6.607913970947266e-06,
"loss": 6.5358,
"step": 455000
},
{
"epoch": 0.8678436279296875,
"eval_accuracy": 0.04102739726027397,
"eval_loss": 6.410449028015137,
"eval_runtime": 241.4157,
"eval_samples_per_second": 41.422,
"eval_steps_per_second": 10.356,
"step": 455000
},
{
"epoch": 0.8687973022460938,
"grad_norm": 994.7066040039062,
"learning_rate": 6.560230255126954e-06,
"loss": 6.467,
"step": 455500
},
{
"epoch": 0.8697509765625,
"grad_norm": 101.90473937988281,
"learning_rate": 6.512546539306641e-06,
"loss": 6.5229,
"step": 456000
},
{
"epoch": 0.8707046508789062,
"grad_norm": 317.31573486328125,
"learning_rate": 6.464862823486328e-06,
"loss": 6.4841,
"step": 456500
},
{
"epoch": 0.8716583251953125,
"grad_norm": 152.74020385742188,
"learning_rate": 6.417179107666016e-06,
"loss": 6.4979,
"step": 457000
},
{
"epoch": 0.8726119995117188,
"grad_norm": 78.52469635009766,
"learning_rate": 6.369495391845704e-06,
"loss": 6.4786,
"step": 457500
},
{
"epoch": 0.873565673828125,
"grad_norm": 189.6461181640625,
"learning_rate": 6.321811676025391e-06,
"loss": 6.5052,
"step": 458000
},
{
"epoch": 0.8745193481445312,
"grad_norm": 255.77638244628906,
"learning_rate": 6.274127960205078e-06,
"loss": 6.4747,
"step": 458500
},
{
"epoch": 0.8754730224609375,
"grad_norm": 665.3084716796875,
"learning_rate": 6.226444244384766e-06,
"loss": 6.5084,
"step": 459000
},
{
"epoch": 0.8764266967773438,
"grad_norm": 138.03125,
"learning_rate": 6.1787605285644534e-06,
"loss": 6.4459,
"step": 459500
},
{
"epoch": 0.87738037109375,
"grad_norm": 112.24298095703125,
"learning_rate": 6.131076812744141e-06,
"loss": 6.5185,
"step": 460000
},
{
"epoch": 0.87738037109375,
"eval_accuracy": 0.041087084148727984,
"eval_loss": 6.409872055053711,
"eval_runtime": 238.0851,
"eval_samples_per_second": 42.002,
"eval_steps_per_second": 10.5,
"step": 460000
},
{
"epoch": 0.8783340454101562,
"grad_norm": 120.41625213623047,
"learning_rate": 6.083393096923829e-06,
"loss": 6.5047,
"step": 460500
},
{
"epoch": 0.8792877197265625,
"grad_norm": 269.24188232421875,
"learning_rate": 6.035709381103516e-06,
"loss": 6.5034,
"step": 461000
},
{
"epoch": 0.8802413940429688,
"grad_norm": 229.2635040283203,
"learning_rate": 5.988025665283203e-06,
"loss": 6.4601,
"step": 461500
},
{
"epoch": 0.881195068359375,
"grad_norm": 200.8750457763672,
"learning_rate": 5.940341949462891e-06,
"loss": 6.5129,
"step": 462000
},
{
"epoch": 0.8821487426757812,
"grad_norm": 298.3540344238281,
"learning_rate": 5.892658233642579e-06,
"loss": 6.4728,
"step": 462500
},
{
"epoch": 0.8831024169921875,
"grad_norm": 376.58282470703125,
"learning_rate": 5.844974517822266e-06,
"loss": 6.5131,
"step": 463000
},
{
"epoch": 0.8840560913085938,
"grad_norm": 120.4054946899414,
"learning_rate": 5.797290802001953e-06,
"loss": 6.4966,
"step": 463500
},
{
"epoch": 0.885009765625,
"grad_norm": 137.49420166015625,
"learning_rate": 5.749607086181641e-06,
"loss": 6.4923,
"step": 464000
},
{
"epoch": 0.8859634399414062,
"grad_norm": 191.45240783691406,
"learning_rate": 5.7019233703613284e-06,
"loss": 6.4865,
"step": 464500
},
{
"epoch": 0.8869171142578125,
"grad_norm": 105.69901275634766,
"learning_rate": 5.654239654541016e-06,
"loss": 6.4622,
"step": 465000
},
{
"epoch": 0.8869171142578125,
"eval_accuracy": 0.040835812133072406,
"eval_loss": 6.4109673500061035,
"eval_runtime": 240.5013,
"eval_samples_per_second": 41.58,
"eval_steps_per_second": 10.395,
"step": 465000
},
{
"epoch": 0.8878707885742188,
"grad_norm": 155.20513916015625,
"learning_rate": 5.606555938720704e-06,
"loss": 6.4755,
"step": 465500
},
{
"epoch": 0.888824462890625,
"grad_norm": 138.47914123535156,
"learning_rate": 5.558872222900391e-06,
"loss": 6.4916,
"step": 466000
},
{
"epoch": 0.8897781372070312,
"grad_norm": 183.78927612304688,
"learning_rate": 5.511188507080078e-06,
"loss": 6.4476,
"step": 466500
},
{
"epoch": 0.8907318115234375,
"grad_norm": 213.38343811035156,
"learning_rate": 5.463504791259766e-06,
"loss": 6.468,
"step": 467000
},
{
"epoch": 0.8916854858398438,
"grad_norm": 323.8433532714844,
"learning_rate": 5.415821075439454e-06,
"loss": 6.4651,
"step": 467500
},
{
"epoch": 0.89263916015625,
"grad_norm": 478.55859375,
"learning_rate": 5.368137359619141e-06,
"loss": 6.4735,
"step": 468000
},
{
"epoch": 0.8935928344726562,
"grad_norm": 130.722412109375,
"learning_rate": 5.320453643798828e-06,
"loss": 6.4699,
"step": 468500
},
{
"epoch": 0.8945465087890625,
"grad_norm": 394.5163269042969,
"learning_rate": 5.272769927978516e-06,
"loss": 6.4711,
"step": 469000
},
{
"epoch": 0.8955001831054688,
"grad_norm": 146.37515258789062,
"learning_rate": 5.2250862121582034e-06,
"loss": 6.4345,
"step": 469500
},
{
"epoch": 0.896453857421875,
"grad_norm": 201.2168426513672,
"learning_rate": 5.177402496337891e-06,
"loss": 6.4632,
"step": 470000
},
{
"epoch": 0.896453857421875,
"eval_accuracy": 0.040519569471624266,
"eval_loss": 6.408752918243408,
"eval_runtime": 239.9438,
"eval_samples_per_second": 41.676,
"eval_steps_per_second": 10.419,
"step": 470000
},
{
"epoch": 0.8974075317382812,
"grad_norm": 111.96930694580078,
"learning_rate": 5.129718780517579e-06,
"loss": 6.4518,
"step": 470500
},
{
"epoch": 0.8983612060546875,
"grad_norm": 94.77079010009766,
"learning_rate": 5.082035064697266e-06,
"loss": 6.5325,
"step": 471000
},
{
"epoch": 0.8993148803710938,
"grad_norm": 210.18495178222656,
"learning_rate": 5.034351348876953e-06,
"loss": 6.5485,
"step": 471500
},
{
"epoch": 0.9002685546875,
"grad_norm": 206.44248962402344,
"learning_rate": 4.986667633056641e-06,
"loss": 6.515,
"step": 472000
},
{
"epoch": 0.9012222290039062,
"grad_norm": 157.20700073242188,
"learning_rate": 4.938983917236329e-06,
"loss": 6.4947,
"step": 472500
},
{
"epoch": 0.9021759033203125,
"grad_norm": 218.9573974609375,
"learning_rate": 4.891300201416016e-06,
"loss": 6.467,
"step": 473000
},
{
"epoch": 0.9031295776367188,
"grad_norm": 158.2908477783203,
"learning_rate": 4.843616485595703e-06,
"loss": 6.4824,
"step": 473500
},
{
"epoch": 0.904083251953125,
"grad_norm": 219.37716674804688,
"learning_rate": 4.795932769775391e-06,
"loss": 6.5119,
"step": 474000
},
{
"epoch": 0.9050369262695312,
"grad_norm": 200.3717803955078,
"learning_rate": 4.7482490539550784e-06,
"loss": 6.4907,
"step": 474500
},
{
"epoch": 0.9059906005859375,
"grad_norm": 203.56930541992188,
"learning_rate": 4.700565338134766e-06,
"loss": 6.5168,
"step": 475000
},
{
"epoch": 0.9059906005859375,
"eval_accuracy": 0.04081624266144814,
"eval_loss": 6.405585765838623,
"eval_runtime": 237.2887,
"eval_samples_per_second": 42.143,
"eval_steps_per_second": 10.536,
"step": 475000
},
{
"epoch": 0.9069442749023438,
"grad_norm": 318.3104248046875,
"learning_rate": 4.652881622314453e-06,
"loss": 6.5018,
"step": 475500
},
{
"epoch": 0.90789794921875,
"grad_norm": 182.2261962890625,
"learning_rate": 4.605197906494141e-06,
"loss": 6.5036,
"step": 476000
},
{
"epoch": 0.9088516235351562,
"grad_norm": 209.1435089111328,
"learning_rate": 4.557514190673828e-06,
"loss": 6.4778,
"step": 476500
},
{
"epoch": 0.9098052978515625,
"grad_norm": 351.99151611328125,
"learning_rate": 4.509830474853516e-06,
"loss": 6.489,
"step": 477000
},
{
"epoch": 0.9107589721679688,
"grad_norm": 115.95133972167969,
"learning_rate": 4.462146759033204e-06,
"loss": 6.4559,
"step": 477500
},
{
"epoch": 0.911712646484375,
"grad_norm": 107.39529418945312,
"learning_rate": 4.4144630432128904e-06,
"loss": 6.5008,
"step": 478000
},
{
"epoch": 0.9126663208007812,
"grad_norm": 186.85157775878906,
"learning_rate": 4.366779327392578e-06,
"loss": 6.4971,
"step": 478500
},
{
"epoch": 0.9136199951171875,
"grad_norm": 247.09622192382812,
"learning_rate": 4.319095611572266e-06,
"loss": 6.5137,
"step": 479000
},
{
"epoch": 0.9145736694335938,
"grad_norm": 54.6359977722168,
"learning_rate": 4.2714118957519534e-06,
"loss": 6.5132,
"step": 479500
},
{
"epoch": 0.91552734375,
"grad_norm": 187.11416625976562,
"learning_rate": 4.223728179931641e-06,
"loss": 6.4607,
"step": 480000
},
{
"epoch": 0.91552734375,
"eval_accuracy": 0.04039393346379648,
"eval_loss": 6.399998664855957,
"eval_runtime": 238.4674,
"eval_samples_per_second": 41.934,
"eval_steps_per_second": 10.484,
"step": 480000
},
{
"epoch": 0.9164810180664062,
"grad_norm": 233.7747039794922,
"learning_rate": 4.176044464111328e-06,
"loss": 6.4675,
"step": 480500
},
{
"epoch": 0.9174346923828125,
"grad_norm": 244.14247131347656,
"learning_rate": 4.128360748291016e-06,
"loss": 6.4788,
"step": 481000
},
{
"epoch": 0.9183883666992188,
"grad_norm": 151.67935180664062,
"learning_rate": 4.080677032470703e-06,
"loss": 6.4817,
"step": 481500
},
{
"epoch": 0.919342041015625,
"grad_norm": 113.63164520263672,
"learning_rate": 4.032993316650391e-06,
"loss": 6.4787,
"step": 482000
},
{
"epoch": 0.9202957153320312,
"grad_norm": 368.08343505859375,
"learning_rate": 3.985309600830079e-06,
"loss": 6.5057,
"step": 482500
},
{
"epoch": 0.9212493896484375,
"grad_norm": 159.99847412109375,
"learning_rate": 3.9376258850097654e-06,
"loss": 6.4912,
"step": 483000
},
{
"epoch": 0.9222030639648438,
"grad_norm": 249.7469024658203,
"learning_rate": 3.889942169189453e-06,
"loss": 6.4585,
"step": 483500
},
{
"epoch": 0.92315673828125,
"grad_norm": 156.4268798828125,
"learning_rate": 3.842258453369141e-06,
"loss": 6.5154,
"step": 484000
},
{
"epoch": 0.9241104125976562,
"grad_norm": 930.0221557617188,
"learning_rate": 3.7945747375488284e-06,
"loss": 6.4562,
"step": 484500
},
{
"epoch": 0.9250640869140625,
"grad_norm": 184.3603057861328,
"learning_rate": 3.7468910217285157e-06,
"loss": 6.4444,
"step": 485000
},
{
"epoch": 0.9250640869140625,
"eval_accuracy": 0.04043659491193738,
"eval_loss": 6.402373313903809,
"eval_runtime": 239.0466,
"eval_samples_per_second": 41.833,
"eval_steps_per_second": 10.458,
"step": 485000
},
{
"epoch": 0.9260177612304688,
"grad_norm": 206.1715087890625,
"learning_rate": 3.6992073059082034e-06,
"loss": 6.4102,
"step": 485500
},
{
"epoch": 0.926971435546875,
"grad_norm": 146.68963623046875,
"learning_rate": 3.6515235900878906e-06,
"loss": 6.4127,
"step": 486000
},
{
"epoch": 0.9279251098632812,
"grad_norm": 89.89360046386719,
"learning_rate": 3.6038398742675783e-06,
"loss": 6.4257,
"step": 486500
},
{
"epoch": 0.9288787841796875,
"grad_norm": 242.57437133789062,
"learning_rate": 3.556156158447266e-06,
"loss": 6.4693,
"step": 487000
},
{
"epoch": 0.9298324584960938,
"grad_norm": 125.63064575195312,
"learning_rate": 3.508472442626953e-06,
"loss": 6.4649,
"step": 487500
},
{
"epoch": 0.9307861328125,
"grad_norm": 683.224853515625,
"learning_rate": 3.460788726806641e-06,
"loss": 6.5341,
"step": 488000
},
{
"epoch": 0.9317398071289062,
"grad_norm": 167.0951690673828,
"learning_rate": 3.413105010986328e-06,
"loss": 6.5514,
"step": 488500
},
{
"epoch": 0.9326934814453125,
"grad_norm": 298.4920959472656,
"learning_rate": 3.3654212951660158e-06,
"loss": 6.4831,
"step": 489000
},
{
"epoch": 0.9336471557617188,
"grad_norm": 268.6081237792969,
"learning_rate": 3.3177375793457034e-06,
"loss": 6.5185,
"step": 489500
},
{
"epoch": 0.934600830078125,
"grad_norm": 134.19676208496094,
"learning_rate": 3.2700538635253907e-06,
"loss": 6.5035,
"step": 490000
},
{
"epoch": 0.934600830078125,
"eval_accuracy": 0.04039667318982387,
"eval_loss": 6.399056434631348,
"eval_runtime": 247.0746,
"eval_samples_per_second": 40.474,
"eval_steps_per_second": 10.118,
"step": 490000
},
{
"epoch": 0.9355545043945312,
"grad_norm": 496.64947509765625,
"learning_rate": 3.2223701477050784e-06,
"loss": 6.5038,
"step": 490500
},
{
"epoch": 0.9365081787109375,
"grad_norm": 254.47743225097656,
"learning_rate": 3.1746864318847656e-06,
"loss": 6.5193,
"step": 491000
},
{
"epoch": 0.9374618530273438,
"grad_norm": 567.7495727539062,
"learning_rate": 3.1270027160644533e-06,
"loss": 6.5031,
"step": 491500
},
{
"epoch": 0.93841552734375,
"grad_norm": 133.81227111816406,
"learning_rate": 3.079319000244141e-06,
"loss": 6.4954,
"step": 492000
},
{
"epoch": 0.9393692016601562,
"grad_norm": 515.5515747070312,
"learning_rate": 3.031635284423828e-06,
"loss": 6.6057,
"step": 492500
},
{
"epoch": 0.9403228759765625,
"grad_norm": 133.3535919189453,
"learning_rate": 2.983951568603516e-06,
"loss": 6.4989,
"step": 493000
},
{
"epoch": 0.9412765502929688,
"grad_norm": 206.1186065673828,
"learning_rate": 2.936267852783203e-06,
"loss": 6.4909,
"step": 493500
},
{
"epoch": 0.942230224609375,
"grad_norm": 235.87135314941406,
"learning_rate": 2.8885841369628908e-06,
"loss": 6.4921,
"step": 494000
},
{
"epoch": 0.9431838989257812,
"grad_norm": 92.96924591064453,
"learning_rate": 2.8409004211425784e-06,
"loss": 6.5005,
"step": 494500
},
{
"epoch": 0.9441375732421875,
"grad_norm": 170.0372772216797,
"learning_rate": 2.7932167053222657e-06,
"loss": 6.4861,
"step": 495000
},
{
"epoch": 0.9441375732421875,
"eval_accuracy": 0.04069197651663405,
"eval_loss": 6.399594306945801,
"eval_runtime": 236.9558,
"eval_samples_per_second": 42.202,
"eval_steps_per_second": 10.55,
"step": 495000
},
{
"epoch": 0.9450912475585938,
"grad_norm": 138.89389038085938,
"learning_rate": 2.7455329895019534e-06,
"loss": 6.4904,
"step": 495500
},
{
"epoch": 0.946044921875,
"grad_norm": 428.46466064453125,
"learning_rate": 2.6978492736816406e-06,
"loss": 6.4434,
"step": 496000
},
{
"epoch": 0.9469985961914062,
"grad_norm": 94.90145874023438,
"learning_rate": 2.6501655578613283e-06,
"loss": 6.4063,
"step": 496500
},
{
"epoch": 0.9479522705078125,
"grad_norm": 215.63514709472656,
"learning_rate": 2.602481842041016e-06,
"loss": 6.4792,
"step": 497000
},
{
"epoch": 0.9489059448242188,
"grad_norm": 110.41034698486328,
"learning_rate": 2.554798126220703e-06,
"loss": 6.4369,
"step": 497500
},
{
"epoch": 0.949859619140625,
"grad_norm": 89.65414428710938,
"learning_rate": 2.507114410400391e-06,
"loss": 6.4895,
"step": 498000
},
{
"epoch": 0.9508132934570312,
"grad_norm": 531.5847778320312,
"learning_rate": 2.459430694580078e-06,
"loss": 6.4937,
"step": 498500
},
{
"epoch": 0.9517669677734375,
"grad_norm": 341.2304992675781,
"learning_rate": 2.4117469787597658e-06,
"loss": 6.4776,
"step": 499000
},
{
"epoch": 0.9527206420898438,
"grad_norm": 198.3251495361328,
"learning_rate": 2.3640632629394534e-06,
"loss": 6.4901,
"step": 499500
},
{
"epoch": 0.95367431640625,
"grad_norm": 275.1803894042969,
"learning_rate": 2.3163795471191407e-06,
"loss": 6.4772,
"step": 500000
},
{
"epoch": 0.95367431640625,
"eval_accuracy": 0.04034637964774951,
"eval_loss": 6.396467208862305,
"eval_runtime": 236.7401,
"eval_samples_per_second": 42.24,
"eval_steps_per_second": 10.56,
"step": 500000
},
{
"epoch": 1.0009536743164062,
"grad_norm": 660.521484375,
"learning_rate": 2.2686958312988284e-06,
"loss": 6.4742,
"step": 500500
},
{
"epoch": 1.0019073486328125,
"grad_norm": 61.67089080810547,
"learning_rate": 2.2210121154785156e-06,
"loss": 6.4717,
"step": 501000
},
{
"epoch": 1.0028610229492188,
"grad_norm": 443.39886474609375,
"learning_rate": 2.1733283996582033e-06,
"loss": 6.4642,
"step": 501500
},
{
"epoch": 1.003814697265625,
"grad_norm": 176.8584442138672,
"learning_rate": 2.125644683837891e-06,
"loss": 6.4951,
"step": 502000
},
{
"epoch": 1.0047683715820312,
"grad_norm": 161.3272247314453,
"learning_rate": 2.077960968017578e-06,
"loss": 6.4766,
"step": 502500
},
{
"epoch": 1.0057220458984375,
"grad_norm": 335.29766845703125,
"learning_rate": 2.030277252197266e-06,
"loss": 6.499,
"step": 503000
},
{
"epoch": 1.0066757202148438,
"grad_norm": 1061.899169921875,
"learning_rate": 1.982593536376953e-06,
"loss": 6.4728,
"step": 503500
},
{
"epoch": 1.00762939453125,
"grad_norm": 812.8785400390625,
"learning_rate": 1.9349098205566408e-06,
"loss": 6.4709,
"step": 504000
},
{
"epoch": 1.0085830688476562,
"grad_norm": 449.0191345214844,
"learning_rate": 1.8872261047363282e-06,
"loss": 6.4303,
"step": 504500
},
{
"epoch": 1.0095367431640625,
"grad_norm": 87.57962799072266,
"learning_rate": 1.8395423889160157e-06,
"loss": 6.4699,
"step": 505000
},
{
"epoch": 1.0095367431640625,
"eval_accuracy": 0.04081682974559687,
"eval_loss": 6.394299507141113,
"eval_runtime": 237.5756,
"eval_samples_per_second": 42.092,
"eval_steps_per_second": 10.523,
"step": 505000
},
{
"epoch": 1.0104904174804688,
"grad_norm": 650.8753662109375,
"learning_rate": 1.7918586730957031e-06,
"loss": 6.4638,
"step": 505500
},
{
"epoch": 1.011444091796875,
"grad_norm": 344.9217834472656,
"learning_rate": 1.7441749572753908e-06,
"loss": 6.4907,
"step": 506000
},
{
"epoch": 1.0123977661132812,
"grad_norm": 264.13104248046875,
"learning_rate": 1.6964912414550783e-06,
"loss": 6.4645,
"step": 506500
},
{
"epoch": 1.0133514404296875,
"grad_norm": 110.09146881103516,
"learning_rate": 1.6488075256347657e-06,
"loss": 6.4765,
"step": 507000
},
{
"epoch": 1.0143051147460938,
"grad_norm": 451.4542236328125,
"learning_rate": 1.6011238098144532e-06,
"loss": 6.4731,
"step": 507500
},
{
"epoch": 1.0152587890625,
"grad_norm": 105.0625,
"learning_rate": 1.5534400939941406e-06,
"loss": 6.4823,
"step": 508000
},
{
"epoch": 1.0162124633789062,
"grad_norm": 144.6260528564453,
"learning_rate": 1.505756378173828e-06,
"loss": 6.4399,
"step": 508500
},
{
"epoch": 1.0171661376953125,
"grad_norm": 164.07302856445312,
"learning_rate": 1.4580726623535158e-06,
"loss": 6.4199,
"step": 509000
},
{
"epoch": 1.0181198120117188,
"grad_norm": 532.625244140625,
"learning_rate": 1.4103889465332032e-06,
"loss": 6.4347,
"step": 509500
},
{
"epoch": 1.019073486328125,
"grad_norm": 121.664306640625,
"learning_rate": 1.3627052307128907e-06,
"loss": 6.4423,
"step": 510000
},
{
"epoch": 1.019073486328125,
"eval_accuracy": 0.04074931506849315,
"eval_loss": 6.395938873291016,
"eval_runtime": 238.1708,
"eval_samples_per_second": 41.987,
"eval_steps_per_second": 10.497,
"step": 510000
},
{
"epoch": 1.0200271606445312,
"grad_norm": 144.7412567138672,
"learning_rate": 1.3150215148925781e-06,
"loss": 6.4167,
"step": 510500
},
{
"epoch": 1.0209808349609375,
"grad_norm": 163.46463012695312,
"learning_rate": 1.2673377990722656e-06,
"loss": 6.4411,
"step": 511000
},
{
"epoch": 1.0219345092773438,
"grad_norm": 165.60867309570312,
"learning_rate": 1.2196540832519533e-06,
"loss": 6.4644,
"step": 511500
},
{
"epoch": 1.02288818359375,
"grad_norm": 142.94607543945312,
"learning_rate": 1.1719703674316407e-06,
"loss": 6.5151,
"step": 512000
},
{
"epoch": 1.0238418579101562,
"grad_norm": 203.33934020996094,
"learning_rate": 1.1242866516113282e-06,
"loss": 6.4719,
"step": 512500
},
{
"epoch": 1.0247955322265625,
"grad_norm": 142.940185546875,
"learning_rate": 1.0766029357910156e-06,
"loss": 6.5046,
"step": 513000
},
{
"epoch": 1.0257492065429688,
"grad_norm": 69.12759399414062,
"learning_rate": 1.028919219970703e-06,
"loss": 6.4209,
"step": 513500
},
{
"epoch": 1.026702880859375,
"grad_norm": 334.0024719238281,
"learning_rate": 9.812355041503908e-07,
"loss": 6.4513,
"step": 514000
},
{
"epoch": 1.0276565551757812,
"grad_norm": 188.01004028320312,
"learning_rate": 9.335517883300781e-07,
"loss": 6.4816,
"step": 514500
},
{
"epoch": 1.0286102294921875,
"grad_norm": 247.91525268554688,
"learning_rate": 8.858680725097657e-07,
"loss": 6.4724,
"step": 515000
},
{
"epoch": 1.0286102294921875,
"eval_accuracy": 0.04097964774951076,
"eval_loss": 6.3929643630981445,
"eval_runtime": 239.731,
"eval_samples_per_second": 41.713,
"eval_steps_per_second": 10.428,
"step": 515000
},
{
"epoch": 1.0295639038085938,
"grad_norm": 147.8994598388672,
"learning_rate": 8.381843566894531e-07,
"loss": 6.4946,
"step": 515500
},
{
"epoch": 1.030517578125,
"grad_norm": 105.9547119140625,
"learning_rate": 7.905006408691407e-07,
"loss": 6.4546,
"step": 516000
},
{
"epoch": 1.0314712524414062,
"grad_norm": 259.9665832519531,
"learning_rate": 7.428169250488282e-07,
"loss": 6.457,
"step": 516500
},
{
"epoch": 1.0324249267578125,
"grad_norm": 144.90777587890625,
"learning_rate": 6.951332092285156e-07,
"loss": 6.4264,
"step": 517000
},
{
"epoch": 1.0333786010742188,
"grad_norm": 185.05178833007812,
"learning_rate": 6.474494934082032e-07,
"loss": 6.4623,
"step": 517500
},
{
"epoch": 1.034332275390625,
"grad_norm": 211.78102111816406,
"learning_rate": 5.997657775878906e-07,
"loss": 6.5132,
"step": 518000
},
{
"epoch": 1.0352859497070312,
"grad_norm": 111.77483367919922,
"learning_rate": 5.520820617675782e-07,
"loss": 6.4667,
"step": 518500
},
{
"epoch": 1.0362396240234375,
"grad_norm": 118.17922973632812,
"learning_rate": 5.043983459472657e-07,
"loss": 6.4682,
"step": 519000
},
{
"epoch": 1.0371932983398438,
"grad_norm": 141.0331268310547,
"learning_rate": 4.5671463012695317e-07,
"loss": 6.4657,
"step": 519500
},
{
"epoch": 1.03814697265625,
"grad_norm": 388.150390625,
"learning_rate": 4.0903091430664063e-07,
"loss": 6.4827,
"step": 520000
},
{
"epoch": 1.03814697265625,
"eval_accuracy": 0.04092270058708415,
"eval_loss": 6.391010284423828,
"eval_runtime": 236.8403,
"eval_samples_per_second": 42.223,
"eval_steps_per_second": 10.556,
"step": 520000
},
{
"epoch": 1.0391006469726562,
"grad_norm": 198.49420166015625,
"learning_rate": 3.6134719848632814e-07,
"loss": 6.4965,
"step": 520500
},
{
"epoch": 1.0400543212890625,
"grad_norm": 214.90264892578125,
"learning_rate": 3.1366348266601565e-07,
"loss": 6.5029,
"step": 521000
},
{
"epoch": 1.0410079956054688,
"grad_norm": 111.66552734375,
"learning_rate": 2.6597976684570316e-07,
"loss": 6.4834,
"step": 521500
},
{
"epoch": 1.041961669921875,
"grad_norm": 145.68179321289062,
"learning_rate": 2.1829605102539064e-07,
"loss": 6.4821,
"step": 522000
},
{
"epoch": 1.0429153442382812,
"grad_norm": 174.4090576171875,
"learning_rate": 1.7061233520507813e-07,
"loss": 6.4684,
"step": 522500
},
{
"epoch": 1.0438690185546875,
"grad_norm": 138.6009521484375,
"learning_rate": 1.2292861938476564e-07,
"loss": 6.4652,
"step": 523000
},
{
"epoch": 1.0448226928710938,
"grad_norm": 112.79365539550781,
"learning_rate": 7.524490356445312e-08,
"loss": 6.4511,
"step": 523500
},
{
"epoch": 1.0457763671875,
"grad_norm": 297.8994445800781,
"learning_rate": 2.7561187744140627e-08,
"loss": 6.4939,
"step": 524000
},
{
"epoch": 1.04632568359375,
"step": 524288,
"total_flos": 5.045399375119909e+18,
"train_loss": 6.577412648592144,
"train_runtime": 154506.2809,
"train_samples_per_second": 13.573,
"train_steps_per_second": 3.393
}
],
"logging_steps": 500,
"max_steps": 524288,
"num_input_tokens_seen": 0,
"num_train_epochs": 9223372036854775807,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 5.045399375119909e+18,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}