DCFT-KodCode-V1-SFT-R1_100k-etash / trainer_state.json
neginr's picture
Upload folder using huggingface_hub
d2be6a3 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.9952,
"eval_steps": 500,
"global_step": 585,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.00512,
"grad_norm": 7.427512523661066,
"learning_rate": 1.3559322033898307e-06,
"loss": 1.1833,
"step": 1
},
{
"epoch": 0.01024,
"grad_norm": 7.378805913304633,
"learning_rate": 2.7118644067796613e-06,
"loss": 1.1693,
"step": 2
},
{
"epoch": 0.01536,
"grad_norm": 7.387359456401948,
"learning_rate": 4.067796610169492e-06,
"loss": 1.1699,
"step": 3
},
{
"epoch": 0.02048,
"grad_norm": 5.424792613029199,
"learning_rate": 5.423728813559323e-06,
"loss": 1.112,
"step": 4
},
{
"epoch": 0.0256,
"grad_norm": 3.3531117149966567,
"learning_rate": 6.779661016949153e-06,
"loss": 1.0668,
"step": 5
},
{
"epoch": 0.03072,
"grad_norm": 2.9208579813945645,
"learning_rate": 8.135593220338983e-06,
"loss": 1.0554,
"step": 6
},
{
"epoch": 0.03584,
"grad_norm": 5.338271857324549,
"learning_rate": 9.491525423728815e-06,
"loss": 1.0192,
"step": 7
},
{
"epoch": 0.04096,
"grad_norm": 5.570953182353899,
"learning_rate": 1.0847457627118645e-05,
"loss": 1.0134,
"step": 8
},
{
"epoch": 0.04608,
"grad_norm": 5.770745007584102,
"learning_rate": 1.2203389830508477e-05,
"loss": 0.9676,
"step": 9
},
{
"epoch": 0.0512,
"grad_norm": 4.652978095614179,
"learning_rate": 1.3559322033898305e-05,
"loss": 0.9515,
"step": 10
},
{
"epoch": 0.05632,
"grad_norm": 2.8588327705676626,
"learning_rate": 1.4915254237288137e-05,
"loss": 0.9188,
"step": 11
},
{
"epoch": 0.06144,
"grad_norm": 3.6027109410465066,
"learning_rate": 1.6271186440677967e-05,
"loss": 0.8979,
"step": 12
},
{
"epoch": 0.06656,
"grad_norm": 2.9981487608394635,
"learning_rate": 1.76271186440678e-05,
"loss": 0.8804,
"step": 13
},
{
"epoch": 0.07168,
"grad_norm": 1.9815532395129927,
"learning_rate": 1.898305084745763e-05,
"loss": 0.85,
"step": 14
},
{
"epoch": 0.0768,
"grad_norm": 2.0993713191830765,
"learning_rate": 2.033898305084746e-05,
"loss": 0.8359,
"step": 15
},
{
"epoch": 0.08192,
"grad_norm": 1.8644702221939118,
"learning_rate": 2.169491525423729e-05,
"loss": 0.8325,
"step": 16
},
{
"epoch": 0.08704,
"grad_norm": 1.4001838150615264,
"learning_rate": 2.3050847457627122e-05,
"loss": 0.8145,
"step": 17
},
{
"epoch": 0.09216,
"grad_norm": 1.5105243610460197,
"learning_rate": 2.4406779661016954e-05,
"loss": 0.8061,
"step": 18
},
{
"epoch": 0.09728,
"grad_norm": 1.508229346305267,
"learning_rate": 2.576271186440678e-05,
"loss": 0.8019,
"step": 19
},
{
"epoch": 0.1024,
"grad_norm": 1.313209951669769,
"learning_rate": 2.711864406779661e-05,
"loss": 0.8,
"step": 20
},
{
"epoch": 0.10752,
"grad_norm": 1.173580022336048,
"learning_rate": 2.8474576271186442e-05,
"loss": 0.7764,
"step": 21
},
{
"epoch": 0.11264,
"grad_norm": 1.293902010777013,
"learning_rate": 2.9830508474576274e-05,
"loss": 0.7891,
"step": 22
},
{
"epoch": 0.11776,
"grad_norm": 0.9680297788054414,
"learning_rate": 3.1186440677966106e-05,
"loss": 0.7745,
"step": 23
},
{
"epoch": 0.12288,
"grad_norm": 0.9265872685281233,
"learning_rate": 3.2542372881355934e-05,
"loss": 0.7678,
"step": 24
},
{
"epoch": 0.128,
"grad_norm": 0.9305803654580205,
"learning_rate": 3.389830508474576e-05,
"loss": 0.7559,
"step": 25
},
{
"epoch": 0.13312,
"grad_norm": 1.334014536013991,
"learning_rate": 3.52542372881356e-05,
"loss": 0.7697,
"step": 26
},
{
"epoch": 0.13824,
"grad_norm": 1.7668925981671184,
"learning_rate": 3.6610169491525426e-05,
"loss": 0.7506,
"step": 27
},
{
"epoch": 0.14336,
"grad_norm": 1.2041359597289543,
"learning_rate": 3.796610169491526e-05,
"loss": 0.7446,
"step": 28
},
{
"epoch": 0.14848,
"grad_norm": 1.9838038548617862,
"learning_rate": 3.932203389830509e-05,
"loss": 0.7511,
"step": 29
},
{
"epoch": 0.1536,
"grad_norm": 1.2091908618849432,
"learning_rate": 4.067796610169492e-05,
"loss": 0.743,
"step": 30
},
{
"epoch": 0.15872,
"grad_norm": 2.013739044410424,
"learning_rate": 4.203389830508475e-05,
"loss": 0.739,
"step": 31
},
{
"epoch": 0.16384,
"grad_norm": 1.7387033344897997,
"learning_rate": 4.338983050847458e-05,
"loss": 0.737,
"step": 32
},
{
"epoch": 0.16896,
"grad_norm": 1.7866106146689165,
"learning_rate": 4.474576271186441e-05,
"loss": 0.7392,
"step": 33
},
{
"epoch": 0.17408,
"grad_norm": 1.5407095248074152,
"learning_rate": 4.6101694915254244e-05,
"loss": 0.7286,
"step": 34
},
{
"epoch": 0.1792,
"grad_norm": 1.7355594859856742,
"learning_rate": 4.745762711864407e-05,
"loss": 0.7429,
"step": 35
},
{
"epoch": 0.18432,
"grad_norm": 1.4876420903172791,
"learning_rate": 4.881355932203391e-05,
"loss": 0.7229,
"step": 36
},
{
"epoch": 0.18944,
"grad_norm": 2.2843469216868098,
"learning_rate": 5.016949152542373e-05,
"loss": 0.7288,
"step": 37
},
{
"epoch": 0.19456,
"grad_norm": 1.6645490807582124,
"learning_rate": 5.152542372881356e-05,
"loss": 0.7189,
"step": 38
},
{
"epoch": 0.19968,
"grad_norm": 2.3532899761581096,
"learning_rate": 5.288135593220339e-05,
"loss": 0.7163,
"step": 39
},
{
"epoch": 0.2048,
"grad_norm": 2.497396548783959,
"learning_rate": 5.423728813559322e-05,
"loss": 0.7264,
"step": 40
},
{
"epoch": 0.20992,
"grad_norm": 1.4481682870152812,
"learning_rate": 5.5593220338983056e-05,
"loss": 0.7144,
"step": 41
},
{
"epoch": 0.21504,
"grad_norm": 3.1370809945974587,
"learning_rate": 5.6949152542372884e-05,
"loss": 0.7002,
"step": 42
},
{
"epoch": 0.22016,
"grad_norm": 2.6339360870235264,
"learning_rate": 5.830508474576271e-05,
"loss": 0.7403,
"step": 43
},
{
"epoch": 0.22528,
"grad_norm": 2.2084591884302647,
"learning_rate": 5.966101694915255e-05,
"loss": 0.7236,
"step": 44
},
{
"epoch": 0.2304,
"grad_norm": 2.7378970801102107,
"learning_rate": 6.1016949152542376e-05,
"loss": 0.6975,
"step": 45
},
{
"epoch": 0.23552,
"grad_norm": 1.8558926790043844,
"learning_rate": 6.237288135593221e-05,
"loss": 0.7187,
"step": 46
},
{
"epoch": 0.24064,
"grad_norm": 2.2724542031435866,
"learning_rate": 6.372881355932204e-05,
"loss": 0.7264,
"step": 47
},
{
"epoch": 0.24576,
"grad_norm": 2.8585186655333605,
"learning_rate": 6.508474576271187e-05,
"loss": 0.7122,
"step": 48
},
{
"epoch": 0.25088,
"grad_norm": 2.2069560445393246,
"learning_rate": 6.64406779661017e-05,
"loss": 0.727,
"step": 49
},
{
"epoch": 0.256,
"grad_norm": 3.1071789158018173,
"learning_rate": 6.779661016949152e-05,
"loss": 0.6951,
"step": 50
},
{
"epoch": 0.26112,
"grad_norm": 2.0255083166280694,
"learning_rate": 6.915254237288137e-05,
"loss": 0.6928,
"step": 51
},
{
"epoch": 0.26624,
"grad_norm": 3.3982026768893427,
"learning_rate": 7.05084745762712e-05,
"loss": 0.6963,
"step": 52
},
{
"epoch": 0.27136,
"grad_norm": 1.7595270818326945,
"learning_rate": 7.186440677966102e-05,
"loss": 0.7136,
"step": 53
},
{
"epoch": 0.27648,
"grad_norm": 3.522255480730156,
"learning_rate": 7.322033898305085e-05,
"loss": 0.701,
"step": 54
},
{
"epoch": 0.2816,
"grad_norm": 2.0471574026263473,
"learning_rate": 7.457627118644068e-05,
"loss": 0.7015,
"step": 55
},
{
"epoch": 0.28672,
"grad_norm": 4.076903746052479,
"learning_rate": 7.593220338983052e-05,
"loss": 0.7213,
"step": 56
},
{
"epoch": 0.29184,
"grad_norm": 2.9540886546409,
"learning_rate": 7.728813559322035e-05,
"loss": 0.7072,
"step": 57
},
{
"epoch": 0.29696,
"grad_norm": 2.930300987052726,
"learning_rate": 7.864406779661018e-05,
"loss": 0.7075,
"step": 58
},
{
"epoch": 0.30208,
"grad_norm": 2.309566711188816,
"learning_rate": 8e-05,
"loss": 0.6845,
"step": 59
},
{
"epoch": 0.3072,
"grad_norm": 3.0240560677976482,
"learning_rate": 7.999928656081034e-05,
"loss": 0.7071,
"step": 60
},
{
"epoch": 0.31232,
"grad_norm": 2.1478418033816076,
"learning_rate": 7.999714626869112e-05,
"loss": 0.6906,
"step": 61
},
{
"epoch": 0.31744,
"grad_norm": 3.179230462128308,
"learning_rate": 7.999357919999074e-05,
"loss": 0.7038,
"step": 62
},
{
"epoch": 0.32256,
"grad_norm": 2.4903190109477693,
"learning_rate": 7.998858548195353e-05,
"loss": 0.7036,
"step": 63
},
{
"epoch": 0.32768,
"grad_norm": 2.5167791307656717,
"learning_rate": 7.998216529271523e-05,
"loss": 0.7022,
"step": 64
},
{
"epoch": 0.3328,
"grad_norm": 2.1782283151414723,
"learning_rate": 7.997431886129654e-05,
"loss": 0.7056,
"step": 65
},
{
"epoch": 0.33792,
"grad_norm": 2.4061240485846707,
"learning_rate": 7.996504646759507e-05,
"loss": 0.7033,
"step": 66
},
{
"epoch": 0.34304,
"grad_norm": 1.8094645315702094,
"learning_rate": 7.995434844237524e-05,
"loss": 0.703,
"step": 67
},
{
"epoch": 0.34816,
"grad_norm": 2.3445365799928384,
"learning_rate": 7.994222516725659e-05,
"loss": 0.6953,
"step": 68
},
{
"epoch": 0.35328,
"grad_norm": 1.99410664221794,
"learning_rate": 7.99286770747001e-05,
"loss": 0.6919,
"step": 69
},
{
"epoch": 0.3584,
"grad_norm": 2.230153791562052,
"learning_rate": 7.991370464799278e-05,
"loss": 0.6862,
"step": 70
},
{
"epoch": 0.36352,
"grad_norm": 1.9601599085514845,
"learning_rate": 7.989730842123042e-05,
"loss": 0.7128,
"step": 71
},
{
"epoch": 0.36864,
"grad_norm": 1.5761765132400967,
"learning_rate": 7.987948897929855e-05,
"loss": 0.6745,
"step": 72
},
{
"epoch": 0.37376,
"grad_norm": 1.679276038030434,
"learning_rate": 7.986024695785161e-05,
"loss": 0.6796,
"step": 73
},
{
"epoch": 0.37888,
"grad_norm": 1.6649601965609995,
"learning_rate": 7.98395830432902e-05,
"loss": 0.6745,
"step": 74
},
{
"epoch": 0.384,
"grad_norm": 1.4057964596604813,
"learning_rate": 7.981749797273661e-05,
"loss": 0.6979,
"step": 75
},
{
"epoch": 0.38912,
"grad_norm": 2.014531329083667,
"learning_rate": 7.979399253400862e-05,
"loss": 0.7037,
"step": 76
},
{
"epoch": 0.39424,
"grad_norm": 1.7428321466528032,
"learning_rate": 7.976906756559127e-05,
"loss": 0.6832,
"step": 77
},
{
"epoch": 0.39936,
"grad_norm": 2.167027629573797,
"learning_rate": 7.974272395660703e-05,
"loss": 0.6879,
"step": 78
},
{
"epoch": 0.40448,
"grad_norm": 1.6144350155950205,
"learning_rate": 7.971496264678404e-05,
"loss": 0.6877,
"step": 79
},
{
"epoch": 0.4096,
"grad_norm": 1.5708068654881047,
"learning_rate": 7.968578462642265e-05,
"loss": 0.6989,
"step": 80
},
{
"epoch": 0.41472,
"grad_norm": 1.555445454516115,
"learning_rate": 7.965519093636e-05,
"loss": 0.6992,
"step": 81
},
{
"epoch": 0.41984,
"grad_norm": 1.8059552941628938,
"learning_rate": 7.962318266793294e-05,
"loss": 0.6912,
"step": 82
},
{
"epoch": 0.42496,
"grad_norm": 1.0063199261184264,
"learning_rate": 7.958976096293916e-05,
"loss": 0.6929,
"step": 83
},
{
"epoch": 0.43008,
"grad_norm": 1.7307326550028748,
"learning_rate": 7.955492701359636e-05,
"loss": 0.6902,
"step": 84
},
{
"epoch": 0.4352,
"grad_norm": 1.205147537826931,
"learning_rate": 7.951868206249975e-05,
"loss": 0.6943,
"step": 85
},
{
"epoch": 0.44032,
"grad_norm": 1.4160221907669925,
"learning_rate": 7.948102740257776e-05,
"loss": 0.6976,
"step": 86
},
{
"epoch": 0.44544,
"grad_norm": 2.001829250494979,
"learning_rate": 7.944196437704593e-05,
"loss": 0.7038,
"step": 87
},
{
"epoch": 0.45056,
"grad_norm": 1.1653512775535024,
"learning_rate": 7.940149437935887e-05,
"loss": 0.6849,
"step": 88
},
{
"epoch": 0.45568,
"grad_norm": 1.6823929061601692,
"learning_rate": 7.935961885316074e-05,
"loss": 0.6925,
"step": 89
},
{
"epoch": 0.4608,
"grad_norm": 1.533381264168228,
"learning_rate": 7.93163392922336e-05,
"loss": 0.7078,
"step": 90
},
{
"epoch": 0.46592,
"grad_norm": 1.3854602682735884,
"learning_rate": 7.927165724044418e-05,
"loss": 0.6905,
"step": 91
},
{
"epoch": 0.47104,
"grad_norm": 1.6672169257948595,
"learning_rate": 7.922557429168884e-05,
"loss": 0.6868,
"step": 92
},
{
"epoch": 0.47616,
"grad_norm": 1.3374855328007862,
"learning_rate": 7.917809208983667e-05,
"loss": 0.6872,
"step": 93
},
{
"epoch": 0.48128,
"grad_norm": 1.1062013903355805,
"learning_rate": 7.912921232867082e-05,
"loss": 0.6829,
"step": 94
},
{
"epoch": 0.4864,
"grad_norm": 1.2577822175567075,
"learning_rate": 7.907893675182816e-05,
"loss": 0.6851,
"step": 95
},
{
"epoch": 0.49152,
"grad_norm": 1.1436105339986973,
"learning_rate": 7.902726715273704e-05,
"loss": 0.6801,
"step": 96
},
{
"epoch": 0.49664,
"grad_norm": 1.3288774009958406,
"learning_rate": 7.89742053745533e-05,
"loss": 0.7013,
"step": 97
},
{
"epoch": 0.50176,
"grad_norm": 1.2793250945243777,
"learning_rate": 7.891975331009454e-05,
"loss": 0.6742,
"step": 98
},
{
"epoch": 0.50688,
"grad_norm": 1.3915794984714676,
"learning_rate": 7.886391290177259e-05,
"loss": 0.692,
"step": 99
},
{
"epoch": 0.512,
"grad_norm": 1.683470719946437,
"learning_rate": 7.880668614152426e-05,
"loss": 0.6903,
"step": 100
},
{
"epoch": 0.51712,
"grad_norm": 1.1736596909100243,
"learning_rate": 7.874807507074019e-05,
"loss": 0.6769,
"step": 101
},
{
"epoch": 0.52224,
"grad_norm": 1.7187295695085814,
"learning_rate": 7.868808178019215e-05,
"loss": 0.6904,
"step": 102
},
{
"epoch": 0.52736,
"grad_norm": 0.9916974084040032,
"learning_rate": 7.862670840995836e-05,
"loss": 0.6788,
"step": 103
},
{
"epoch": 0.53248,
"grad_norm": 1.7935336303746816,
"learning_rate": 7.856395714934718e-05,
"loss": 0.6952,
"step": 104
},
{
"epoch": 0.5376,
"grad_norm": 1.4946062420453348,
"learning_rate": 7.849983023681905e-05,
"loss": 0.6853,
"step": 105
},
{
"epoch": 0.54272,
"grad_norm": 1.245185249022715,
"learning_rate": 7.84343299599066e-05,
"loss": 0.6729,
"step": 106
},
{
"epoch": 0.54784,
"grad_norm": 2.0094790105419738,
"learning_rate": 7.836745865513304e-05,
"loss": 0.682,
"step": 107
},
{
"epoch": 0.55296,
"grad_norm": 1.065814539722233,
"learning_rate": 7.829921870792886e-05,
"loss": 0.6764,
"step": 108
},
{
"epoch": 0.55808,
"grad_norm": 1.9579572631250968,
"learning_rate": 7.822961255254669e-05,
"loss": 0.6756,
"step": 109
},
{
"epoch": 0.5632,
"grad_norm": 1.52170370675673,
"learning_rate": 7.815864267197448e-05,
"loss": 0.6816,
"step": 110
},
{
"epoch": 0.56832,
"grad_norm": 1.704295073143993,
"learning_rate": 7.808631159784693e-05,
"loss": 0.6807,
"step": 111
},
{
"epoch": 0.57344,
"grad_norm": 1.0971860063652723,
"learning_rate": 7.801262191035518e-05,
"loss": 0.6775,
"step": 112
},
{
"epoch": 0.57856,
"grad_norm": 1.2658364692181785,
"learning_rate": 7.79375762381548e-05,
"loss": 0.6677,
"step": 113
},
{
"epoch": 0.58368,
"grad_norm": 1.1136078326808156,
"learning_rate": 7.786117725827195e-05,
"loss": 0.6706,
"step": 114
},
{
"epoch": 0.5888,
"grad_norm": 1.2432841023033496,
"learning_rate": 7.778342769600796e-05,
"loss": 0.6604,
"step": 115
},
{
"epoch": 0.59392,
"grad_norm": 0.9447874229879114,
"learning_rate": 7.770433032484204e-05,
"loss": 0.6674,
"step": 116
},
{
"epoch": 0.59904,
"grad_norm": 1.1728862828599786,
"learning_rate": 7.762388796633243e-05,
"loss": 0.6743,
"step": 117
},
{
"epoch": 0.60416,
"grad_norm": 1.0379645631823264,
"learning_rate": 7.754210349001568e-05,
"loss": 0.6746,
"step": 118
},
{
"epoch": 0.60928,
"grad_norm": 0.6784319178622393,
"learning_rate": 7.74589798133043e-05,
"loss": 0.6613,
"step": 119
},
{
"epoch": 0.6144,
"grad_norm": 1.2334508195450562,
"learning_rate": 7.737451990138275e-05,
"loss": 0.6832,
"step": 120
},
{
"epoch": 0.61952,
"grad_norm": 1.390023864901241,
"learning_rate": 7.728872676710155e-05,
"loss": 0.667,
"step": 121
},
{
"epoch": 0.62464,
"grad_norm": 1.0288050461695384,
"learning_rate": 7.720160347086994e-05,
"loss": 0.6823,
"step": 122
},
{
"epoch": 0.62976,
"grad_norm": 1.144163750515486,
"learning_rate": 7.711315312054659e-05,
"loss": 0.6692,
"step": 123
},
{
"epoch": 0.63488,
"grad_norm": 0.9460143556927034,
"learning_rate": 7.702337887132883e-05,
"loss": 0.6608,
"step": 124
},
{
"epoch": 0.64,
"grad_norm": 1.617773004869984,
"learning_rate": 7.693228392564003e-05,
"loss": 0.6912,
"step": 125
},
{
"epoch": 0.64512,
"grad_norm": 0.8796030638057838,
"learning_rate": 7.68398715330154e-05,
"loss": 0.6545,
"step": 126
},
{
"epoch": 0.65024,
"grad_norm": 1.4530036523303451,
"learning_rate": 7.674614498998608e-05,
"loss": 0.667,
"step": 127
},
{
"epoch": 0.65536,
"grad_norm": 1.6001652553788648,
"learning_rate": 7.66511076399615e-05,
"loss": 0.686,
"step": 128
},
{
"epoch": 0.66048,
"grad_norm": 0.9295070052125337,
"learning_rate": 7.655476287311017e-05,
"loss": 0.6726,
"step": 129
},
{
"epoch": 0.6656,
"grad_norm": 1.981385007424261,
"learning_rate": 7.645711412623871e-05,
"loss": 0.6955,
"step": 130
},
{
"epoch": 0.67072,
"grad_norm": 1.1198337635879976,
"learning_rate": 7.635816488266925e-05,
"loss": 0.6747,
"step": 131
},
{
"epoch": 0.67584,
"grad_norm": 1.8403136092865524,
"learning_rate": 7.62579186721152e-05,
"loss": 0.6817,
"step": 132
},
{
"epoch": 0.68096,
"grad_norm": 1.211708500347191,
"learning_rate": 7.615637907055533e-05,
"loss": 0.6673,
"step": 133
},
{
"epoch": 0.68608,
"grad_norm": 1.619702658720517,
"learning_rate": 7.60535497001062e-05,
"loss": 0.6659,
"step": 134
},
{
"epoch": 0.6912,
"grad_norm": 1.3880946828869691,
"learning_rate": 7.594943422889293e-05,
"loss": 0.6876,
"step": 135
},
{
"epoch": 0.69632,
"grad_norm": 1.3898444786546138,
"learning_rate": 7.584403637091839e-05,
"loss": 0.6697,
"step": 136
},
{
"epoch": 0.70144,
"grad_norm": 1.1550848443588233,
"learning_rate": 7.57373598859307e-05,
"loss": 0.6708,
"step": 137
},
{
"epoch": 0.70656,
"grad_norm": 1.2208384976092193,
"learning_rate": 7.562940857928914e-05,
"loss": 0.6547,
"step": 138
},
{
"epoch": 0.71168,
"grad_norm": 0.7766104942706619,
"learning_rate": 7.552018630182829e-05,
"loss": 0.6548,
"step": 139
},
{
"epoch": 0.7168,
"grad_norm": 1.2980119993455839,
"learning_rate": 7.540969694972086e-05,
"loss": 0.6642,
"step": 140
},
{
"epoch": 0.72192,
"grad_norm": 1.1032988591515491,
"learning_rate": 7.529794446433853e-05,
"loss": 0.6567,
"step": 141
},
{
"epoch": 0.72704,
"grad_norm": 1.0687712643010066,
"learning_rate": 7.51849328321114e-05,
"loss": 0.659,
"step": 142
},
{
"epoch": 0.73216,
"grad_norm": 0.9993909705876705,
"learning_rate": 7.507066608438587e-05,
"loss": 0.6653,
"step": 143
},
{
"epoch": 0.73728,
"grad_norm": 1.4634026159707303,
"learning_rate": 7.495514829728072e-05,
"loss": 0.6633,
"step": 144
},
{
"epoch": 0.7424,
"grad_norm": 0.9096724763021085,
"learning_rate": 7.483838359154175e-05,
"loss": 0.6735,
"step": 145
},
{
"epoch": 0.74752,
"grad_norm": 1.359529554918389,
"learning_rate": 7.472037613239487e-05,
"loss": 0.6678,
"step": 146
},
{
"epoch": 0.75264,
"grad_norm": 0.989415493913379,
"learning_rate": 7.460113012939733e-05,
"loss": 0.6637,
"step": 147
},
{
"epoch": 0.75776,
"grad_norm": 1.3223608862498835,
"learning_rate": 7.448064983628775e-05,
"loss": 0.6645,
"step": 148
},
{
"epoch": 0.76288,
"grad_norm": 1.1620124411680315,
"learning_rate": 7.435893955083422e-05,
"loss": 0.6513,
"step": 149
},
{
"epoch": 0.768,
"grad_norm": 1.0648774531623415,
"learning_rate": 7.423600361468117e-05,
"loss": 0.662,
"step": 150
},
{
"epoch": 0.77312,
"grad_norm": 1.3007725912379944,
"learning_rate": 7.411184641319429e-05,
"loss": 0.6631,
"step": 151
},
{
"epoch": 0.77824,
"grad_norm": 0.8104819966143042,
"learning_rate": 7.398647237530426e-05,
"loss": 0.64,
"step": 152
},
{
"epoch": 0.78336,
"grad_norm": 1.0085673778184652,
"learning_rate": 7.385988597334867e-05,
"loss": 0.6574,
"step": 153
},
{
"epoch": 0.78848,
"grad_norm": 1.2945419726019807,
"learning_rate": 7.373209172291253e-05,
"loss": 0.6475,
"step": 154
},
{
"epoch": 0.7936,
"grad_norm": 1.0304441022051967,
"learning_rate": 7.360309418266715e-05,
"loss": 0.6517,
"step": 155
},
{
"epoch": 0.79872,
"grad_norm": 0.9586021230391207,
"learning_rate": 7.347289795420759e-05,
"loss": 0.6582,
"step": 156
},
{
"epoch": 0.80384,
"grad_norm": 1.1290408208864031,
"learning_rate": 7.334150768188839e-05,
"loss": 0.6609,
"step": 157
},
{
"epoch": 0.80896,
"grad_norm": 1.1884226665875801,
"learning_rate": 7.320892805265807e-05,
"loss": 0.6699,
"step": 158
},
{
"epoch": 0.81408,
"grad_norm": 0.8502814476612911,
"learning_rate": 7.307516379589177e-05,
"loss": 0.6493,
"step": 159
},
{
"epoch": 0.8192,
"grad_norm": 0.9952888696907723,
"learning_rate": 7.294021968322261e-05,
"loss": 0.6484,
"step": 160
},
{
"epoch": 0.82432,
"grad_norm": 1.2290779564393681,
"learning_rate": 7.280410052837156e-05,
"loss": 0.6582,
"step": 161
},
{
"epoch": 0.82944,
"grad_norm": 0.8035227492015825,
"learning_rate": 7.266681118697555e-05,
"loss": 0.6699,
"step": 162
},
{
"epoch": 0.83456,
"grad_norm": 1.1834030000315219,
"learning_rate": 7.252835655641445e-05,
"loss": 0.6625,
"step": 163
},
{
"epoch": 0.83968,
"grad_norm": 0.8159782777745277,
"learning_rate": 7.23887415756362e-05,
"loss": 0.6474,
"step": 164
},
{
"epoch": 0.8448,
"grad_norm": 1.5925994868655784,
"learning_rate": 7.224797122498074e-05,
"loss": 0.6568,
"step": 165
},
{
"epoch": 0.84992,
"grad_norm": 0.7327801301614796,
"learning_rate": 7.210605052600233e-05,
"loss": 0.6535,
"step": 166
},
{
"epoch": 0.85504,
"grad_norm": 1.3897449798624986,
"learning_rate": 7.196298454129037e-05,
"loss": 0.6586,
"step": 167
},
{
"epoch": 0.86016,
"grad_norm": 0.9296756247599517,
"learning_rate": 7.18187783742889e-05,
"loss": 0.6493,
"step": 168
},
{
"epoch": 0.86528,
"grad_norm": 1.4033864221213181,
"learning_rate": 7.167343716911446e-05,
"loss": 0.6535,
"step": 169
},
{
"epoch": 0.8704,
"grad_norm": 1.2953910254634462,
"learning_rate": 7.15269661103726e-05,
"loss": 0.6522,
"step": 170
},
{
"epoch": 0.87552,
"grad_norm": 0.7630866652770001,
"learning_rate": 7.137937042297303e-05,
"loss": 0.6533,
"step": 171
},
{
"epoch": 0.88064,
"grad_norm": 1.6581576394049042,
"learning_rate": 7.123065537194311e-05,
"loss": 0.6466,
"step": 172
},
{
"epoch": 0.88576,
"grad_norm": 0.894168267427839,
"learning_rate": 7.108082626224014e-05,
"loss": 0.6749,
"step": 173
},
{
"epoch": 0.89088,
"grad_norm": 1.2060587657486854,
"learning_rate": 7.092988843856202e-05,
"loss": 0.6644,
"step": 174
},
{
"epoch": 0.896,
"grad_norm": 0.9959417990440771,
"learning_rate": 7.077784728515668e-05,
"loss": 0.6578,
"step": 175
},
{
"epoch": 0.90112,
"grad_norm": 1.0821074322876068,
"learning_rate": 7.062470822563003e-05,
"loss": 0.6543,
"step": 176
},
{
"epoch": 0.90624,
"grad_norm": 0.7740564812790837,
"learning_rate": 7.047047672275235e-05,
"loss": 0.6595,
"step": 177
},
{
"epoch": 0.91136,
"grad_norm": 0.7885588166965639,
"learning_rate": 7.031515827826356e-05,
"loss": 0.6677,
"step": 178
},
{
"epoch": 0.91648,
"grad_norm": 0.8964594506190875,
"learning_rate": 7.015875843267696e-05,
"loss": 0.668,
"step": 179
},
{
"epoch": 0.9216,
"grad_norm": 0.7922996726063177,
"learning_rate": 7.000128276508146e-05,
"loss": 0.6613,
"step": 180
},
{
"epoch": 0.92672,
"grad_norm": 0.7656175088558786,
"learning_rate": 6.984273689294272e-05,
"loss": 0.6474,
"step": 181
},
{
"epoch": 0.93184,
"grad_norm": 0.9214614344976504,
"learning_rate": 6.968312647190267e-05,
"loss": 0.6326,
"step": 182
},
{
"epoch": 0.93696,
"grad_norm": 1.2164859593883526,
"learning_rate": 6.952245719557777e-05,
"loss": 0.6465,
"step": 183
},
{
"epoch": 0.94208,
"grad_norm": 0.6778217095702366,
"learning_rate": 6.936073479535592e-05,
"loss": 0.657,
"step": 184
},
{
"epoch": 0.9472,
"grad_norm": 0.8101497456748098,
"learning_rate": 6.919796504019208e-05,
"loss": 0.6762,
"step": 185
},
{
"epoch": 0.95232,
"grad_norm": 0.8876944556782685,
"learning_rate": 6.903415373640231e-05,
"loss": 0.6475,
"step": 186
},
{
"epoch": 0.95744,
"grad_norm": 1.0000115870935409,
"learning_rate": 6.886930672745684e-05,
"loss": 0.6561,
"step": 187
},
{
"epoch": 0.96256,
"grad_norm": 1.0242604450962167,
"learning_rate": 6.870342989377146e-05,
"loss": 0.6434,
"step": 188
},
{
"epoch": 0.96768,
"grad_norm": 0.9780023684498255,
"learning_rate": 6.853652915249786e-05,
"loss": 0.6545,
"step": 189
},
{
"epoch": 0.9728,
"grad_norm": 0.9266467661951784,
"learning_rate": 6.836861045731254e-05,
"loss": 0.6501,
"step": 190
},
{
"epoch": 0.97792,
"grad_norm": 0.9727642879091534,
"learning_rate": 6.819967979820441e-05,
"loss": 0.6457,
"step": 191
},
{
"epoch": 0.98304,
"grad_norm": 0.6002892901190532,
"learning_rate": 6.802974320126105e-05,
"loss": 0.6535,
"step": 192
},
{
"epoch": 0.98816,
"grad_norm": 0.7042555907925779,
"learning_rate": 6.785880672845391e-05,
"loss": 0.6497,
"step": 193
},
{
"epoch": 0.99328,
"grad_norm": 1.0659679020014914,
"learning_rate": 6.768687647742188e-05,
"loss": 0.6594,
"step": 194
},
{
"epoch": 0.9984,
"grad_norm": 0.919967416151866,
"learning_rate": 6.751395858125393e-05,
"loss": 0.6454,
"step": 195
},
{
"epoch": 1.00352,
"grad_norm": 0.832989636341993,
"learning_rate": 6.734005920827023e-05,
"loss": 0.6229,
"step": 196
},
{
"epoch": 1.00864,
"grad_norm": 0.9268003420050892,
"learning_rate": 6.716518456180219e-05,
"loss": 0.6067,
"step": 197
},
{
"epoch": 1.01376,
"grad_norm": 0.9264505307476663,
"learning_rate": 6.69893408799711e-05,
"loss": 0.594,
"step": 198
},
{
"epoch": 1.01888,
"grad_norm": 0.992069998033644,
"learning_rate": 6.681253443546566e-05,
"loss": 0.6027,
"step": 199
},
{
"epoch": 1.024,
"grad_norm": 0.9820410181623107,
"learning_rate": 6.663477153531819e-05,
"loss": 0.6091,
"step": 200
},
{
"epoch": 1.02912,
"grad_norm": 0.7956602066339694,
"learning_rate": 6.645605852067966e-05,
"loss": 0.6016,
"step": 201
},
{
"epoch": 1.03424,
"grad_norm": 1.1030274948771646,
"learning_rate": 6.627640176659349e-05,
"loss": 0.6128,
"step": 202
},
{
"epoch": 1.03936,
"grad_norm": 0.8027746768903161,
"learning_rate": 6.609580768176813e-05,
"loss": 0.6202,
"step": 203
},
{
"epoch": 1.04448,
"grad_norm": 0.6182011917758258,
"learning_rate": 6.591428270834848e-05,
"loss": 0.6197,
"step": 204
},
{
"epoch": 1.0496,
"grad_norm": 0.6933362237087946,
"learning_rate": 6.573183332168599e-05,
"loss": 0.5973,
"step": 205
},
{
"epoch": 1.05472,
"grad_norm": 0.7261413202596659,
"learning_rate": 6.554846603010783e-05,
"loss": 0.6033,
"step": 206
},
{
"epoch": 1.05984,
"grad_norm": 0.8175867268516304,
"learning_rate": 6.536418737468458e-05,
"loss": 0.5962,
"step": 207
},
{
"epoch": 1.06496,
"grad_norm": 0.8967324447987707,
"learning_rate": 6.517900392899696e-05,
"loss": 0.6138,
"step": 208
},
{
"epoch": 1.07008,
"grad_norm": 0.8644025442427714,
"learning_rate": 6.499292229890136e-05,
"loss": 0.6143,
"step": 209
},
{
"epoch": 1.0752,
"grad_norm": 0.8340587352046295,
"learning_rate": 6.480594912229415e-05,
"loss": 0.5932,
"step": 210
},
{
"epoch": 1.08032,
"grad_norm": 0.7992015971845043,
"learning_rate": 6.46180910688749e-05,
"loss": 0.5946,
"step": 211
},
{
"epoch": 1.08544,
"grad_norm": 0.7883742168020988,
"learning_rate": 6.442935483990848e-05,
"loss": 0.5912,
"step": 212
},
{
"epoch": 1.09056,
"grad_norm": 0.7062624710725375,
"learning_rate": 6.4239747167986e-05,
"loss": 0.6006,
"step": 213
},
{
"epoch": 1.09568,
"grad_norm": 0.4879826112399788,
"learning_rate": 6.404927481678465e-05,
"loss": 0.5908,
"step": 214
},
{
"epoch": 1.1008,
"grad_norm": 0.46707410887787343,
"learning_rate": 6.385794458082645e-05,
"loss": 0.6016,
"step": 215
},
{
"epoch": 1.10592,
"grad_norm": 0.46215479735805276,
"learning_rate": 6.366576328523579e-05,
"loss": 0.6059,
"step": 216
},
{
"epoch": 1.11104,
"grad_norm": 0.4872892344668909,
"learning_rate": 6.347273778549609e-05,
"loss": 0.5979,
"step": 217
},
{
"epoch": 1.11616,
"grad_norm": 0.4117850505120705,
"learning_rate": 6.327887496720513e-05,
"loss": 0.6042,
"step": 218
},
{
"epoch": 1.12128,
"grad_norm": 0.6094192942803528,
"learning_rate": 6.308418174582952e-05,
"loss": 0.6098,
"step": 219
},
{
"epoch": 1.1264,
"grad_norm": 0.47370404685867984,
"learning_rate": 6.288866506645799e-05,
"loss": 0.5967,
"step": 220
},
{
"epoch": 1.13152,
"grad_norm": 0.5812822920441131,
"learning_rate": 6.269233190355355e-05,
"loss": 0.589,
"step": 221
},
{
"epoch": 1.13664,
"grad_norm": 0.581689390915006,
"learning_rate": 6.249518926070488e-05,
"loss": 0.6033,
"step": 222
},
{
"epoch": 1.14176,
"grad_norm": 0.7708083440150924,
"learning_rate": 6.229724417037632e-05,
"loss": 0.6008,
"step": 223
},
{
"epoch": 1.14688,
"grad_norm": 0.9437745661416408,
"learning_rate": 6.209850369365714e-05,
"loss": 0.5872,
"step": 224
},
{
"epoch": 1.152,
"grad_norm": 1.0285174187897348,
"learning_rate": 6.189897492000953e-05,
"loss": 0.6034,
"step": 225
},
{
"epoch": 1.15712,
"grad_norm": 0.9256911501210775,
"learning_rate": 6.169866496701586e-05,
"loss": 0.6065,
"step": 226
},
{
"epoch": 1.16224,
"grad_norm": 0.8931346889464913,
"learning_rate": 6.149758098012464e-05,
"loss": 0.6162,
"step": 227
},
{
"epoch": 1.16736,
"grad_norm": 0.906249956782627,
"learning_rate": 6.129573013239571e-05,
"loss": 0.609,
"step": 228
},
{
"epoch": 1.17248,
"grad_norm": 0.9031799418470178,
"learning_rate": 6.109311962424433e-05,
"loss": 0.6043,
"step": 229
},
{
"epoch": 1.1776,
"grad_norm": 0.7749991050050256,
"learning_rate": 6.0889756683184324e-05,
"loss": 0.609,
"step": 230
},
{
"epoch": 1.18272,
"grad_norm": 0.6113291239741755,
"learning_rate": 6.068564856357031e-05,
"loss": 0.6041,
"step": 231
},
{
"epoch": 1.18784,
"grad_norm": 0.4610907845132779,
"learning_rate": 6.048080254633885e-05,
"loss": 0.6032,
"step": 232
},
{
"epoch": 1.19296,
"grad_norm": 0.7535273976467541,
"learning_rate": 6.0275225938748764e-05,
"loss": 0.5856,
"step": 233
},
{
"epoch": 1.19808,
"grad_norm": 0.7357293529763054,
"learning_rate": 6.006892607412047e-05,
"loss": 0.5871,
"step": 234
},
{
"epoch": 1.2032,
"grad_norm": 0.7850706730871837,
"learning_rate": 5.98619103115744e-05,
"loss": 0.6022,
"step": 235
},
{
"epoch": 1.20832,
"grad_norm": 0.9520874763788266,
"learning_rate": 5.965418603576842e-05,
"loss": 0.6029,
"step": 236
},
{
"epoch": 1.21344,
"grad_norm": 0.6234215842388057,
"learning_rate": 5.944576065663449e-05,
"loss": 0.5989,
"step": 237
},
{
"epoch": 1.21856,
"grad_norm": 0.5581618761087828,
"learning_rate": 5.9236641609114295e-05,
"loss": 0.5852,
"step": 238
},
{
"epoch": 1.2236799999999999,
"grad_norm": 0.8280934815331484,
"learning_rate": 5.902683635289403e-05,
"loss": 0.6065,
"step": 239
},
{
"epoch": 1.2288000000000001,
"grad_norm": 0.8871893394061044,
"learning_rate": 5.881635237213829e-05,
"loss": 0.5901,
"step": 240
},
{
"epoch": 1.23392,
"grad_norm": 0.6767238651749208,
"learning_rate": 5.8605197175223095e-05,
"loss": 0.6073,
"step": 241
},
{
"epoch": 1.23904,
"grad_norm": 0.688896399179518,
"learning_rate": 5.839337829446809e-05,
"loss": 0.6049,
"step": 242
},
{
"epoch": 1.24416,
"grad_norm": 0.5804470630643502,
"learning_rate": 5.8180903285867804e-05,
"loss": 0.6036,
"step": 243
},
{
"epoch": 1.24928,
"grad_norm": 0.6457016614930509,
"learning_rate": 5.7967779728822134e-05,
"loss": 0.5981,
"step": 244
},
{
"epoch": 1.2544,
"grad_norm": 0.8414381704157707,
"learning_rate": 5.775401522586596e-05,
"loss": 0.6109,
"step": 245
},
{
"epoch": 1.25952,
"grad_norm": 0.8374711125166566,
"learning_rate": 5.753961740239799e-05,
"loss": 0.6005,
"step": 246
},
{
"epoch": 1.26464,
"grad_norm": 0.5542646630500834,
"learning_rate": 5.732459390640868e-05,
"loss": 0.6103,
"step": 247
},
{
"epoch": 1.26976,
"grad_norm": 0.5875945873220082,
"learning_rate": 5.7108952408207485e-05,
"loss": 0.6128,
"step": 248
},
{
"epoch": 1.27488,
"grad_norm": 0.5087534889785712,
"learning_rate": 5.689270060014919e-05,
"loss": 0.5958,
"step": 249
},
{
"epoch": 1.28,
"grad_norm": 0.4031024078366132,
"learning_rate": 5.66758461963595e-05,
"loss": 0.589,
"step": 250
},
{
"epoch": 1.28512,
"grad_norm": 0.4918002116220245,
"learning_rate": 5.645839693245995e-05,
"loss": 0.5982,
"step": 251
},
{
"epoch": 1.29024,
"grad_norm": 0.48513251035233335,
"learning_rate": 5.6240360565291874e-05,
"loss": 0.5981,
"step": 252
},
{
"epoch": 1.29536,
"grad_norm": 0.41091372229331263,
"learning_rate": 5.602174487263971e-05,
"loss": 0.5902,
"step": 253
},
{
"epoch": 1.30048,
"grad_norm": 0.5989492893332392,
"learning_rate": 5.580255765295358e-05,
"loss": 0.6129,
"step": 254
},
{
"epoch": 1.3056,
"grad_norm": 0.5776722852203174,
"learning_rate": 5.5582806725071135e-05,
"loss": 0.606,
"step": 255
},
{
"epoch": 1.3107199999999999,
"grad_norm": 0.4466289358616809,
"learning_rate": 5.536249992793856e-05,
"loss": 0.5992,
"step": 256
},
{
"epoch": 1.3158400000000001,
"grad_norm": 0.4281414323403905,
"learning_rate": 5.514164512033099e-05,
"loss": 0.5865,
"step": 257
},
{
"epoch": 1.32096,
"grad_norm": 0.3767339434891749,
"learning_rate": 5.4920250180572164e-05,
"loss": 0.6049,
"step": 258
},
{
"epoch": 1.32608,
"grad_norm": 0.4541584747628375,
"learning_rate": 5.469832300625343e-05,
"loss": 0.5891,
"step": 259
},
{
"epoch": 1.3312,
"grad_norm": 0.5852329624714224,
"learning_rate": 5.447587151395193e-05,
"loss": 0.6013,
"step": 260
},
{
"epoch": 1.33632,
"grad_norm": 0.5470836134042647,
"learning_rate": 5.4252903638948304e-05,
"loss": 0.5871,
"step": 261
},
{
"epoch": 1.34144,
"grad_norm": 0.3981165895490368,
"learning_rate": 5.402942733494355e-05,
"loss": 0.5988,
"step": 262
},
{
"epoch": 1.34656,
"grad_norm": 0.8090553651844287,
"learning_rate": 5.380545057377533e-05,
"loss": 0.5919,
"step": 263
},
{
"epoch": 1.35168,
"grad_norm": 0.6973495560228751,
"learning_rate": 5.358098134513359e-05,
"loss": 0.6099,
"step": 264
},
{
"epoch": 1.3568,
"grad_norm": 0.4439863252173243,
"learning_rate": 5.335602765627556e-05,
"loss": 0.6112,
"step": 265
},
{
"epoch": 1.36192,
"grad_norm": 0.45728264462579415,
"learning_rate": 5.313059753174011e-05,
"loss": 0.5999,
"step": 266
},
{
"epoch": 1.36704,
"grad_norm": 0.5814907075742051,
"learning_rate": 5.2904699013061525e-05,
"loss": 0.5996,
"step": 267
},
{
"epoch": 1.37216,
"grad_norm": 0.5358642327581291,
"learning_rate": 5.2678340158482605e-05,
"loss": 0.6085,
"step": 268
},
{
"epoch": 1.37728,
"grad_norm": 0.5406775247232574,
"learning_rate": 5.245152904266722e-05,
"loss": 0.6234,
"step": 269
},
{
"epoch": 1.3824,
"grad_norm": 0.7178735700752464,
"learning_rate": 5.2224273756412335e-05,
"loss": 0.6273,
"step": 270
},
{
"epoch": 1.3875199999999999,
"grad_norm": 0.7039970725546744,
"learning_rate": 5.1996582406359286e-05,
"loss": 0.5899,
"step": 271
},
{
"epoch": 1.39264,
"grad_norm": 0.5122004385795155,
"learning_rate": 5.17684631147047e-05,
"loss": 0.5772,
"step": 272
},
{
"epoch": 1.39776,
"grad_norm": 0.5292102951542996,
"learning_rate": 5.1539924018910705e-05,
"loss": 0.5963,
"step": 273
},
{
"epoch": 1.4028800000000001,
"grad_norm": 0.5393146339840262,
"learning_rate": 5.131097327141467e-05,
"loss": 0.6078,
"step": 274
},
{
"epoch": 1.408,
"grad_norm": 0.708939654981038,
"learning_rate": 5.108161903933837e-05,
"loss": 0.6005,
"step": 275
},
{
"epoch": 1.41312,
"grad_norm": 0.7037903511384231,
"learning_rate": 5.085186950419671e-05,
"loss": 0.5942,
"step": 276
},
{
"epoch": 1.41824,
"grad_norm": 0.4165090417742424,
"learning_rate": 5.0621732861605754e-05,
"loss": 0.5906,
"step": 277
},
{
"epoch": 1.42336,
"grad_norm": 0.6642875595643233,
"learning_rate": 5.039121732099052e-05,
"loss": 0.5891,
"step": 278
},
{
"epoch": 1.42848,
"grad_norm": 0.7547032791125439,
"learning_rate": 5.0160331105292036e-05,
"loss": 0.6008,
"step": 279
},
{
"epoch": 1.4336,
"grad_norm": 0.733986554958704,
"learning_rate": 4.992908245067403e-05,
"loss": 0.5915,
"step": 280
},
{
"epoch": 1.43872,
"grad_norm": 0.47614304475036945,
"learning_rate": 4.969747960622912e-05,
"loss": 0.6011,
"step": 281
},
{
"epoch": 1.44384,
"grad_norm": 0.585876239848498,
"learning_rate": 4.9465530833684614e-05,
"loss": 0.5912,
"step": 282
},
{
"epoch": 1.44896,
"grad_norm": 0.6945418578556748,
"learning_rate": 4.923324440710773e-05,
"loss": 0.6106,
"step": 283
},
{
"epoch": 1.45408,
"grad_norm": 0.6261522164548919,
"learning_rate": 4.900062861261044e-05,
"loss": 0.5941,
"step": 284
},
{
"epoch": 1.4592,
"grad_norm": 0.5342894182102067,
"learning_rate": 4.8767691748053955e-05,
"loss": 0.5922,
"step": 285
},
{
"epoch": 1.46432,
"grad_norm": 0.5838783944360918,
"learning_rate": 4.8534442122752674e-05,
"loss": 0.6075,
"step": 286
},
{
"epoch": 1.46944,
"grad_norm": 0.6486938746583781,
"learning_rate": 4.830088805717779e-05,
"loss": 0.5954,
"step": 287
},
{
"epoch": 1.4745599999999999,
"grad_norm": 0.5719218459201613,
"learning_rate": 4.806703788266045e-05,
"loss": 0.6018,
"step": 288
},
{
"epoch": 1.47968,
"grad_norm": 0.5120146948133751,
"learning_rate": 4.783289994109459e-05,
"loss": 0.5912,
"step": 289
},
{
"epoch": 1.4848,
"grad_norm": 0.6804527806275119,
"learning_rate": 4.759848258463942e-05,
"loss": 0.5896,
"step": 290
},
{
"epoch": 1.4899200000000001,
"grad_norm": 0.5487294767485589,
"learning_rate": 4.736379417542134e-05,
"loss": 0.5882,
"step": 291
},
{
"epoch": 1.49504,
"grad_norm": 0.4853965765407641,
"learning_rate": 4.712884308523579e-05,
"loss": 0.5888,
"step": 292
},
{
"epoch": 1.5001600000000002,
"grad_norm": 0.586848411445172,
"learning_rate": 4.689363769524854e-05,
"loss": 0.5959,
"step": 293
},
{
"epoch": 1.50528,
"grad_norm": 0.49095884598520845,
"learning_rate": 4.6658186395696735e-05,
"loss": 0.6054,
"step": 294
},
{
"epoch": 1.5104,
"grad_norm": 0.5564665878511703,
"learning_rate": 4.64224975855896e-05,
"loss": 0.6059,
"step": 295
},
{
"epoch": 1.51552,
"grad_norm": 0.46245666700892735,
"learning_rate": 4.618657967240879e-05,
"loss": 0.5891,
"step": 296
},
{
"epoch": 1.52064,
"grad_norm": 0.533299845495822,
"learning_rate": 4.595044107180857e-05,
"loss": 0.6044,
"step": 297
},
{
"epoch": 1.52576,
"grad_norm": 0.42525913312635544,
"learning_rate": 4.571409020731554e-05,
"loss": 0.5869,
"step": 298
},
{
"epoch": 1.53088,
"grad_norm": 0.5215704284954067,
"learning_rate": 4.547753551002814e-05,
"loss": 0.5923,
"step": 299
},
{
"epoch": 1.536,
"grad_norm": 0.47973640340231194,
"learning_rate": 4.524078541831597e-05,
"loss": 0.5891,
"step": 300
},
{
"epoch": 1.54112,
"grad_norm": 0.4327598342601823,
"learning_rate": 4.500384837751868e-05,
"loss": 0.5958,
"step": 301
},
{
"epoch": 1.54624,
"grad_norm": 0.5990121963555108,
"learning_rate": 4.476673283964481e-05,
"loss": 0.5856,
"step": 302
},
{
"epoch": 1.5513599999999999,
"grad_norm": 0.7131944725290759,
"learning_rate": 4.4529447263070215e-05,
"loss": 0.5942,
"step": 303
},
{
"epoch": 1.55648,
"grad_norm": 0.5329359342241601,
"learning_rate": 4.429200011223638e-05,
"loss": 0.597,
"step": 304
},
{
"epoch": 1.5615999999999999,
"grad_norm": 0.4477137886623582,
"learning_rate": 4.4054399857348426e-05,
"loss": 0.5905,
"step": 305
},
{
"epoch": 1.5667200000000001,
"grad_norm": 0.7235716501912173,
"learning_rate": 4.381665497407304e-05,
"loss": 0.6044,
"step": 306
},
{
"epoch": 1.57184,
"grad_norm": 0.5656431568517662,
"learning_rate": 4.3578773943236066e-05,
"loss": 0.5935,
"step": 307
},
{
"epoch": 1.5769600000000001,
"grad_norm": 0.5443076040573943,
"learning_rate": 4.334076525051998e-05,
"loss": 0.6085,
"step": 308
},
{
"epoch": 1.58208,
"grad_norm": 0.44142667616814085,
"learning_rate": 4.3102637386161214e-05,
"loss": 0.5974,
"step": 309
},
{
"epoch": 1.5872000000000002,
"grad_norm": 0.4881724554496324,
"learning_rate": 4.2864398844647333e-05,
"loss": 0.5982,
"step": 310
},
{
"epoch": 1.59232,
"grad_norm": 0.48224923855420715,
"learning_rate": 4.2626058124413934e-05,
"loss": 0.5957,
"step": 311
},
{
"epoch": 1.59744,
"grad_norm": 0.5087339085522424,
"learning_rate": 4.23876237275415e-05,
"loss": 0.5919,
"step": 312
},
{
"epoch": 1.60256,
"grad_norm": 0.4533960557571388,
"learning_rate": 4.2149104159452194e-05,
"loss": 0.6004,
"step": 313
},
{
"epoch": 1.60768,
"grad_norm": 0.511203081310295,
"learning_rate": 4.191050792860641e-05,
"loss": 0.5994,
"step": 314
},
{
"epoch": 1.6128,
"grad_norm": 0.4223027905990951,
"learning_rate": 4.1671843546199195e-05,
"loss": 0.5887,
"step": 315
},
{
"epoch": 1.61792,
"grad_norm": 0.4494541281616636,
"learning_rate": 4.143311952585673e-05,
"loss": 0.5775,
"step": 316
},
{
"epoch": 1.62304,
"grad_norm": 0.49316734261443734,
"learning_rate": 4.119434438333261e-05,
"loss": 0.5937,
"step": 317
},
{
"epoch": 1.62816,
"grad_norm": 0.36392011844676875,
"learning_rate": 4.095552663620403e-05,
"loss": 0.5962,
"step": 318
},
{
"epoch": 1.63328,
"grad_norm": 0.4968107231897248,
"learning_rate": 4.0716674803568005e-05,
"loss": 0.5919,
"step": 319
},
{
"epoch": 1.6383999999999999,
"grad_norm": 0.37525246925036004,
"learning_rate": 4.0477797405737414e-05,
"loss": 0.5969,
"step": 320
},
{
"epoch": 1.64352,
"grad_norm": 0.5908101591078128,
"learning_rate": 4.023890296393714e-05,
"loss": 0.6099,
"step": 321
},
{
"epoch": 1.6486399999999999,
"grad_norm": 0.3984607947668425,
"learning_rate": 4e-05,
"loss": 0.6004,
"step": 322
},
{
"epoch": 1.6537600000000001,
"grad_norm": 0.5149940417738129,
"learning_rate": 3.976109703606288e-05,
"loss": 0.6006,
"step": 323
},
{
"epoch": 1.65888,
"grad_norm": 0.4005964486620091,
"learning_rate": 3.952220259426259e-05,
"loss": 0.5803,
"step": 324
},
{
"epoch": 1.6640000000000001,
"grad_norm": 0.35086767512207107,
"learning_rate": 3.928332519643201e-05,
"loss": 0.585,
"step": 325
},
{
"epoch": 1.66912,
"grad_norm": 0.42885330848275394,
"learning_rate": 3.9044473363796e-05,
"loss": 0.5958,
"step": 326
},
{
"epoch": 1.67424,
"grad_norm": 0.38751578060305375,
"learning_rate": 3.8805655616667406e-05,
"loss": 0.6014,
"step": 327
},
{
"epoch": 1.67936,
"grad_norm": 0.3603759776041056,
"learning_rate": 3.8566880474143284e-05,
"loss": 0.6001,
"step": 328
},
{
"epoch": 1.68448,
"grad_norm": 0.4283886804732866,
"learning_rate": 3.832815645380083e-05,
"loss": 0.5999,
"step": 329
},
{
"epoch": 1.6896,
"grad_norm": 0.3337168354335494,
"learning_rate": 3.80894920713936e-05,
"loss": 0.5918,
"step": 330
},
{
"epoch": 1.69472,
"grad_norm": 0.383147123590219,
"learning_rate": 3.785089584054781e-05,
"loss": 0.604,
"step": 331
},
{
"epoch": 1.69984,
"grad_norm": 0.39090517742398523,
"learning_rate": 3.761237627245851e-05,
"loss": 0.5784,
"step": 332
},
{
"epoch": 1.70496,
"grad_norm": 0.41160894437025963,
"learning_rate": 3.737394187558608e-05,
"loss": 0.5899,
"step": 333
},
{
"epoch": 1.71008,
"grad_norm": 0.4348994266327126,
"learning_rate": 3.713560115535268e-05,
"loss": 0.584,
"step": 334
},
{
"epoch": 1.7151999999999998,
"grad_norm": 0.44033711538569453,
"learning_rate": 3.689736261383879e-05,
"loss": 0.6019,
"step": 335
},
{
"epoch": 1.72032,
"grad_norm": 0.46185741930953395,
"learning_rate": 3.6659234749480043e-05,
"loss": 0.5979,
"step": 336
},
{
"epoch": 1.7254399999999999,
"grad_norm": 0.4557337264056941,
"learning_rate": 3.642122605676396e-05,
"loss": 0.6102,
"step": 337
},
{
"epoch": 1.73056,
"grad_norm": 0.6447579696011232,
"learning_rate": 3.6183345025926966e-05,
"loss": 0.6094,
"step": 338
},
{
"epoch": 1.73568,
"grad_norm": 0.35483221584671387,
"learning_rate": 3.594560014265159e-05,
"loss": 0.5859,
"step": 339
},
{
"epoch": 1.7408000000000001,
"grad_norm": 0.5619949022675687,
"learning_rate": 3.570799988776364e-05,
"loss": 0.602,
"step": 340
},
{
"epoch": 1.74592,
"grad_norm": 0.4808833851628103,
"learning_rate": 3.5470552736929785e-05,
"loss": 0.59,
"step": 341
},
{
"epoch": 1.7510400000000002,
"grad_norm": 0.4280664310256486,
"learning_rate": 3.5233267160355206e-05,
"loss": 0.5973,
"step": 342
},
{
"epoch": 1.75616,
"grad_norm": 0.34606786385637417,
"learning_rate": 3.4996151622481336e-05,
"loss": 0.5987,
"step": 343
},
{
"epoch": 1.76128,
"grad_norm": 0.47886946752668064,
"learning_rate": 3.475921458168404e-05,
"loss": 0.5974,
"step": 344
},
{
"epoch": 1.7664,
"grad_norm": 0.35895795443414463,
"learning_rate": 3.452246448997187e-05,
"loss": 0.5825,
"step": 345
},
{
"epoch": 1.77152,
"grad_norm": 0.6237524007460633,
"learning_rate": 3.428590979268448e-05,
"loss": 0.593,
"step": 346
},
{
"epoch": 1.77664,
"grad_norm": 0.5291377288153889,
"learning_rate": 3.4049558928191435e-05,
"loss": 0.5885,
"step": 347
},
{
"epoch": 1.78176,
"grad_norm": 0.5075647920106248,
"learning_rate": 3.381342032759122e-05,
"loss": 0.5982,
"step": 348
},
{
"epoch": 1.78688,
"grad_norm": 0.49287865419478344,
"learning_rate": 3.357750241441041e-05,
"loss": 0.5802,
"step": 349
},
{
"epoch": 1.792,
"grad_norm": 0.543103939185933,
"learning_rate": 3.334181360430327e-05,
"loss": 0.5936,
"step": 350
},
{
"epoch": 1.79712,
"grad_norm": 0.3894840623083752,
"learning_rate": 3.310636230475148e-05,
"loss": 0.588,
"step": 351
},
{
"epoch": 1.8022399999999998,
"grad_norm": 0.43995147863147344,
"learning_rate": 3.287115691476422e-05,
"loss": 0.6006,
"step": 352
},
{
"epoch": 1.80736,
"grad_norm": 0.3921709111241505,
"learning_rate": 3.263620582457868e-05,
"loss": 0.5939,
"step": 353
},
{
"epoch": 1.8124799999999999,
"grad_norm": 0.31433310266314246,
"learning_rate": 3.24015174153606e-05,
"loss": 0.5893,
"step": 354
},
{
"epoch": 1.8176,
"grad_norm": 0.43895209082760966,
"learning_rate": 3.2167100058905415e-05,
"loss": 0.5899,
"step": 355
},
{
"epoch": 1.82272,
"grad_norm": 0.35289039516589493,
"learning_rate": 3.1932962117339566e-05,
"loss": 0.6034,
"step": 356
},
{
"epoch": 1.8278400000000001,
"grad_norm": 0.3361411963291542,
"learning_rate": 3.169911194282222e-05,
"loss": 0.5794,
"step": 357
},
{
"epoch": 1.83296,
"grad_norm": 0.35346529955825,
"learning_rate": 3.1465557877247325e-05,
"loss": 0.5872,
"step": 358
},
{
"epoch": 1.8380800000000002,
"grad_norm": 0.3031581900611866,
"learning_rate": 3.123230825194606e-05,
"loss": 0.5771,
"step": 359
},
{
"epoch": 1.8432,
"grad_norm": 0.4329864321858136,
"learning_rate": 3.099937138738958e-05,
"loss": 0.5927,
"step": 360
},
{
"epoch": 1.84832,
"grad_norm": 0.35326273185796286,
"learning_rate": 3.076675559289228e-05,
"loss": 0.5996,
"step": 361
},
{
"epoch": 1.85344,
"grad_norm": 0.35925676205986545,
"learning_rate": 3.05344691663154e-05,
"loss": 0.5781,
"step": 362
},
{
"epoch": 1.85856,
"grad_norm": 0.3536221477128384,
"learning_rate": 3.0302520393770894e-05,
"loss": 0.5934,
"step": 363
},
{
"epoch": 1.86368,
"grad_norm": 0.42419086729757743,
"learning_rate": 3.0070917549325983e-05,
"loss": 0.592,
"step": 364
},
{
"epoch": 1.8688,
"grad_norm": 0.3833286680510395,
"learning_rate": 2.9839668894707974e-05,
"loss": 0.5858,
"step": 365
},
{
"epoch": 1.87392,
"grad_norm": 0.3817012337576983,
"learning_rate": 2.960878267900948e-05,
"loss": 0.5816,
"step": 366
},
{
"epoch": 1.87904,
"grad_norm": 0.400413399573491,
"learning_rate": 2.937826713839426e-05,
"loss": 0.5839,
"step": 367
},
{
"epoch": 1.88416,
"grad_norm": 0.36874490386425984,
"learning_rate": 2.9148130495803307e-05,
"loss": 0.5854,
"step": 368
},
{
"epoch": 1.8892799999999998,
"grad_norm": 0.39202924423336344,
"learning_rate": 2.8918380960661624e-05,
"loss": 0.5941,
"step": 369
},
{
"epoch": 1.8944,
"grad_norm": 0.35104738229454596,
"learning_rate": 2.8689026728585338e-05,
"loss": 0.5845,
"step": 370
},
{
"epoch": 1.8995199999999999,
"grad_norm": 0.42429376258538937,
"learning_rate": 2.8460075981089305e-05,
"loss": 0.5827,
"step": 371
},
{
"epoch": 1.90464,
"grad_norm": 0.36436129019616925,
"learning_rate": 2.82315368852953e-05,
"loss": 0.5754,
"step": 372
},
{
"epoch": 1.90976,
"grad_norm": 0.42556325640476184,
"learning_rate": 2.8003417593640727e-05,
"loss": 0.599,
"step": 373
},
{
"epoch": 1.9148800000000001,
"grad_norm": 0.3851458476612929,
"learning_rate": 2.7775726243587685e-05,
"loss": 0.5931,
"step": 374
},
{
"epoch": 1.92,
"grad_norm": 0.3737224724659946,
"learning_rate": 2.754847095733278e-05,
"loss": 0.6015,
"step": 375
},
{
"epoch": 1.9251200000000002,
"grad_norm": 0.36520508052939255,
"learning_rate": 2.732165984151741e-05,
"loss": 0.5961,
"step": 376
},
{
"epoch": 1.93024,
"grad_norm": 0.3088144098356228,
"learning_rate": 2.709530098693849e-05,
"loss": 0.5898,
"step": 377
},
{
"epoch": 1.93536,
"grad_norm": 0.4046047102015038,
"learning_rate": 2.686940246825989e-05,
"loss": 0.6059,
"step": 378
},
{
"epoch": 1.94048,
"grad_norm": 0.3815617952147954,
"learning_rate": 2.6643972343724453e-05,
"loss": 0.5785,
"step": 379
},
{
"epoch": 1.9456,
"grad_norm": 0.3351980518182547,
"learning_rate": 2.6419018654866415e-05,
"loss": 0.585,
"step": 380
},
{
"epoch": 1.95072,
"grad_norm": 0.4441803690810445,
"learning_rate": 2.6194549426224684e-05,
"loss": 0.5955,
"step": 381
},
{
"epoch": 1.95584,
"grad_norm": 0.2725524980139723,
"learning_rate": 2.5970572665056465e-05,
"loss": 0.5871,
"step": 382
},
{
"epoch": 1.96096,
"grad_norm": 0.4789800834824601,
"learning_rate": 2.57470963610517e-05,
"loss": 0.597,
"step": 383
},
{
"epoch": 1.96608,
"grad_norm": 0.3196138403431241,
"learning_rate": 2.5524128486048073e-05,
"loss": 0.5948,
"step": 384
},
{
"epoch": 1.9712,
"grad_norm": 0.4883702808240898,
"learning_rate": 2.5301676993746592e-05,
"loss": 0.588,
"step": 385
},
{
"epoch": 1.9763199999999999,
"grad_norm": 0.4304710776740572,
"learning_rate": 2.5079749819427842e-05,
"loss": 0.5884,
"step": 386
},
{
"epoch": 1.98144,
"grad_norm": 0.364646635973235,
"learning_rate": 2.4858354879669025e-05,
"loss": 0.5776,
"step": 387
},
{
"epoch": 1.9865599999999999,
"grad_norm": 0.385522115127419,
"learning_rate": 2.463750007206146e-05,
"loss": 0.5842,
"step": 388
},
{
"epoch": 1.9916800000000001,
"grad_norm": 0.3482195171706265,
"learning_rate": 2.441719327492887e-05,
"loss": 0.5886,
"step": 389
},
{
"epoch": 1.9968,
"grad_norm": 0.4205224902290012,
"learning_rate": 2.4197442347046432e-05,
"loss": 0.5969,
"step": 390
},
{
"epoch": 2.00192,
"grad_norm": 0.40631741625288587,
"learning_rate": 2.397825512736032e-05,
"loss": 0.5662,
"step": 391
},
{
"epoch": 2.00704,
"grad_norm": 0.4013123221753668,
"learning_rate": 2.3759639434708142e-05,
"loss": 0.5248,
"step": 392
},
{
"epoch": 2.01216,
"grad_norm": 0.4323654857095571,
"learning_rate": 2.3541603067540063e-05,
"loss": 0.5389,
"step": 393
},
{
"epoch": 2.01728,
"grad_norm": 0.4365874888455653,
"learning_rate": 2.332415380364052e-05,
"loss": 0.5172,
"step": 394
},
{
"epoch": 2.0224,
"grad_norm": 0.38461821215764164,
"learning_rate": 2.3107299399850822e-05,
"loss": 0.5262,
"step": 395
},
{
"epoch": 2.02752,
"grad_norm": 0.47903986310634056,
"learning_rate": 2.2891047591792532e-05,
"loss": 0.5212,
"step": 396
},
{
"epoch": 2.03264,
"grad_norm": 0.35779675128884125,
"learning_rate": 2.2675406093591316e-05,
"loss": 0.5293,
"step": 397
},
{
"epoch": 2.03776,
"grad_norm": 0.4808543092900177,
"learning_rate": 2.2460382597602016e-05,
"loss": 0.5352,
"step": 398
},
{
"epoch": 2.04288,
"grad_norm": 0.3753156744242789,
"learning_rate": 2.2245984774134063e-05,
"loss": 0.5214,
"step": 399
},
{
"epoch": 2.048,
"grad_norm": 0.39281547945021766,
"learning_rate": 2.2032220271177876e-05,
"loss": 0.5287,
"step": 400
},
{
"epoch": 2.05312,
"grad_norm": 0.40725978791424594,
"learning_rate": 2.1819096714132206e-05,
"loss": 0.526,
"step": 401
},
{
"epoch": 2.05824,
"grad_norm": 0.36597234139560436,
"learning_rate": 2.1606621705531934e-05,
"loss": 0.5073,
"step": 402
},
{
"epoch": 2.06336,
"grad_norm": 0.4434308159144889,
"learning_rate": 2.139480282477691e-05,
"loss": 0.5259,
"step": 403
},
{
"epoch": 2.06848,
"grad_norm": 0.36571996113644073,
"learning_rate": 2.1183647627861736e-05,
"loss": 0.5375,
"step": 404
},
{
"epoch": 2.0736,
"grad_norm": 0.5190659641286463,
"learning_rate": 2.097316364710599e-05,
"loss": 0.5204,
"step": 405
},
{
"epoch": 2.07872,
"grad_norm": 0.34838410996336344,
"learning_rate": 2.0763358390885705e-05,
"loss": 0.5183,
"step": 406
},
{
"epoch": 2.08384,
"grad_norm": 0.464065008178319,
"learning_rate": 2.0554239343365524e-05,
"loss": 0.5181,
"step": 407
},
{
"epoch": 2.08896,
"grad_norm": 0.3876057469692389,
"learning_rate": 2.0345813964231604e-05,
"loss": 0.5283,
"step": 408
},
{
"epoch": 2.09408,
"grad_norm": 0.38233945304862516,
"learning_rate": 2.013808968842561e-05,
"loss": 0.5253,
"step": 409
},
{
"epoch": 2.0992,
"grad_norm": 0.38531045185458873,
"learning_rate": 1.9931073925879544e-05,
"loss": 0.5286,
"step": 410
},
{
"epoch": 2.10432,
"grad_norm": 0.3441646148982092,
"learning_rate": 1.9724774061251253e-05,
"loss": 0.5289,
"step": 411
},
{
"epoch": 2.10944,
"grad_norm": 0.3059332693194638,
"learning_rate": 1.9519197453661166e-05,
"loss": 0.5268,
"step": 412
},
{
"epoch": 2.11456,
"grad_norm": 0.34251732288777326,
"learning_rate": 1.9314351436429703e-05,
"loss": 0.535,
"step": 413
},
{
"epoch": 2.11968,
"grad_norm": 0.27043155273405406,
"learning_rate": 1.9110243316815672e-05,
"loss": 0.519,
"step": 414
},
{
"epoch": 2.1248,
"grad_norm": 0.3287810309710835,
"learning_rate": 1.8906880375755686e-05,
"loss": 0.5123,
"step": 415
},
{
"epoch": 2.12992,
"grad_norm": 0.31017397419120324,
"learning_rate": 1.87042698676043e-05,
"loss": 0.5271,
"step": 416
},
{
"epoch": 2.13504,
"grad_norm": 0.27534025617214347,
"learning_rate": 1.8502419019875357e-05,
"loss": 0.5148,
"step": 417
},
{
"epoch": 2.14016,
"grad_norm": 0.34713006182861084,
"learning_rate": 1.8301335032984157e-05,
"loss": 0.5263,
"step": 418
},
{
"epoch": 2.14528,
"grad_norm": 0.2944738685945991,
"learning_rate": 1.8101025079990485e-05,
"loss": 0.5082,
"step": 419
},
{
"epoch": 2.1504,
"grad_norm": 0.2795225009877447,
"learning_rate": 1.7901496306342886e-05,
"loss": 0.5162,
"step": 420
},
{
"epoch": 2.15552,
"grad_norm": 0.2792843641392462,
"learning_rate": 1.7702755829623696e-05,
"loss": 0.516,
"step": 421
},
{
"epoch": 2.16064,
"grad_norm": 0.31863845049393985,
"learning_rate": 1.750481073929514e-05,
"loss": 0.5294,
"step": 422
},
{
"epoch": 2.16576,
"grad_norm": 0.2926787760079869,
"learning_rate": 1.730766809644646e-05,
"loss": 0.5436,
"step": 423
},
{
"epoch": 2.17088,
"grad_norm": 0.26756179726965906,
"learning_rate": 1.711133493354203e-05,
"loss": 0.5391,
"step": 424
},
{
"epoch": 2.176,
"grad_norm": 0.26829916228856365,
"learning_rate": 1.6915818254170485e-05,
"loss": 0.5293,
"step": 425
},
{
"epoch": 2.18112,
"grad_norm": 0.2916809848264421,
"learning_rate": 1.672112503279488e-05,
"loss": 0.5239,
"step": 426
},
{
"epoch": 2.18624,
"grad_norm": 0.2847780874837289,
"learning_rate": 1.6527262214503917e-05,
"loss": 0.5206,
"step": 427
},
{
"epoch": 2.19136,
"grad_norm": 0.31205176032409004,
"learning_rate": 1.6334236714764215e-05,
"loss": 0.5219,
"step": 428
},
{
"epoch": 2.19648,
"grad_norm": 0.28653984101417374,
"learning_rate": 1.6142055419173556e-05,
"loss": 0.5396,
"step": 429
},
{
"epoch": 2.2016,
"grad_norm": 0.29293525576131924,
"learning_rate": 1.595072518321535e-05,
"loss": 0.5186,
"step": 430
},
{
"epoch": 2.20672,
"grad_norm": 0.32036644737076314,
"learning_rate": 1.576025283201401e-05,
"loss": 0.5293,
"step": 431
},
{
"epoch": 2.21184,
"grad_norm": 0.2746440977694007,
"learning_rate": 1.5570645160091534e-05,
"loss": 0.5238,
"step": 432
},
{
"epoch": 2.21696,
"grad_norm": 0.2763614067422842,
"learning_rate": 1.5381908931125112e-05,
"loss": 0.5315,
"step": 433
},
{
"epoch": 2.22208,
"grad_norm": 0.3206600626128733,
"learning_rate": 1.519405087770586e-05,
"loss": 0.5155,
"step": 434
},
{
"epoch": 2.2272,
"grad_norm": 0.28546354885077907,
"learning_rate": 1.5007077701098643e-05,
"loss": 0.5153,
"step": 435
},
{
"epoch": 2.23232,
"grad_norm": 0.32183823115025384,
"learning_rate": 1.4820996071003042e-05,
"loss": 0.5244,
"step": 436
},
{
"epoch": 2.23744,
"grad_norm": 0.2833366198667468,
"learning_rate": 1.4635812625315428e-05,
"loss": 0.5291,
"step": 437
},
{
"epoch": 2.24256,
"grad_norm": 0.35259963599910704,
"learning_rate": 1.445153396989218e-05,
"loss": 0.5334,
"step": 438
},
{
"epoch": 2.24768,
"grad_norm": 0.31168802605945267,
"learning_rate": 1.4268166678314029e-05,
"loss": 0.522,
"step": 439
},
{
"epoch": 2.2528,
"grad_norm": 0.32274340867941653,
"learning_rate": 1.4085717291651534e-05,
"loss": 0.5291,
"step": 440
},
{
"epoch": 2.25792,
"grad_norm": 0.2948948221905903,
"learning_rate": 1.390419231823187e-05,
"loss": 0.5394,
"step": 441
},
{
"epoch": 2.26304,
"grad_norm": 0.2743662929170965,
"learning_rate": 1.3723598233406525e-05,
"loss": 0.5271,
"step": 442
},
{
"epoch": 2.26816,
"grad_norm": 0.2556644467949261,
"learning_rate": 1.3543941479320344e-05,
"loss": 0.5253,
"step": 443
},
{
"epoch": 2.27328,
"grad_norm": 0.2611595534203661,
"learning_rate": 1.3365228464681814e-05,
"loss": 0.5262,
"step": 444
},
{
"epoch": 2.2784,
"grad_norm": 0.26247038877870327,
"learning_rate": 1.3187465564534359e-05,
"loss": 0.5257,
"step": 445
},
{
"epoch": 2.28352,
"grad_norm": 0.2536877324922943,
"learning_rate": 1.30106591200289e-05,
"loss": 0.5174,
"step": 446
},
{
"epoch": 2.28864,
"grad_norm": 0.2691257976797504,
"learning_rate": 1.2834815438197827e-05,
"loss": 0.5421,
"step": 447
},
{
"epoch": 2.29376,
"grad_norm": 0.21048332739110232,
"learning_rate": 1.2659940791729773e-05,
"loss": 0.5183,
"step": 448
},
{
"epoch": 2.29888,
"grad_norm": 0.26799152242066077,
"learning_rate": 1.248604141874608e-05,
"loss": 0.5347,
"step": 449
},
{
"epoch": 2.304,
"grad_norm": 0.23183888170912065,
"learning_rate": 1.2313123522578141e-05,
"loss": 0.528,
"step": 450
},
{
"epoch": 2.30912,
"grad_norm": 0.23153879138735126,
"learning_rate": 1.2141193271546104e-05,
"loss": 0.5402,
"step": 451
},
{
"epoch": 2.31424,
"grad_norm": 0.2365803046413853,
"learning_rate": 1.1970256798738946e-05,
"loss": 0.524,
"step": 452
},
{
"epoch": 2.31936,
"grad_norm": 0.23790371050570536,
"learning_rate": 1.1800320201795609e-05,
"loss": 0.5236,
"step": 453
},
{
"epoch": 2.32448,
"grad_norm": 0.21840868912845707,
"learning_rate": 1.1631389542687455e-05,
"loss": 0.5158,
"step": 454
},
{
"epoch": 2.3296,
"grad_norm": 0.22811000289018,
"learning_rate": 1.1463470847502154e-05,
"loss": 0.5491,
"step": 455
},
{
"epoch": 2.33472,
"grad_norm": 0.2441771477812667,
"learning_rate": 1.1296570106228568e-05,
"loss": 0.5119,
"step": 456
},
{
"epoch": 2.33984,
"grad_norm": 0.2504770922206427,
"learning_rate": 1.1130693272543174e-05,
"loss": 0.5259,
"step": 457
},
{
"epoch": 2.34496,
"grad_norm": 0.22307013304308804,
"learning_rate": 1.0965846263597704e-05,
"loss": 0.5229,
"step": 458
},
{
"epoch": 2.35008,
"grad_norm": 0.20777825026632868,
"learning_rate": 1.0802034959807934e-05,
"loss": 0.5327,
"step": 459
},
{
"epoch": 2.3552,
"grad_norm": 0.2496892379371564,
"learning_rate": 1.063926520464407e-05,
"loss": 0.5173,
"step": 460
},
{
"epoch": 2.3603199999999998,
"grad_norm": 0.2710443329473512,
"learning_rate": 1.047754280442225e-05,
"loss": 0.5093,
"step": 461
},
{
"epoch": 2.36544,
"grad_norm": 0.2221839293934009,
"learning_rate": 1.0316873528097333e-05,
"loss": 0.5339,
"step": 462
},
{
"epoch": 2.3705600000000002,
"grad_norm": 0.21317614214755715,
"learning_rate": 1.0157263107057291e-05,
"loss": 0.5284,
"step": 463
},
{
"epoch": 2.37568,
"grad_norm": 0.22721711775906023,
"learning_rate": 9.998717234918557e-06,
"loss": 0.5351,
"step": 464
},
{
"epoch": 2.3808,
"grad_norm": 0.22893893378395325,
"learning_rate": 9.84124156732305e-06,
"loss": 0.5173,
"step": 465
},
{
"epoch": 2.38592,
"grad_norm": 0.2312282069514675,
"learning_rate": 9.684841721736449e-06,
"loss": 0.5099,
"step": 466
},
{
"epoch": 2.39104,
"grad_norm": 0.23235918033484193,
"learning_rate": 9.529523277247672e-06,
"loss": 0.5353,
"step": 467
},
{
"epoch": 2.39616,
"grad_norm": 0.2122801075559491,
"learning_rate": 9.37529177436998e-06,
"loss": 0.524,
"step": 468
},
{
"epoch": 2.40128,
"grad_norm": 0.2281610275331807,
"learning_rate": 9.222152714843324e-06,
"loss": 0.5198,
"step": 469
},
{
"epoch": 2.4064,
"grad_norm": 0.2276209528821846,
"learning_rate": 9.070111561437994e-06,
"loss": 0.5125,
"step": 470
},
{
"epoch": 2.41152,
"grad_norm": 0.25863996249108634,
"learning_rate": 8.919173737759878e-06,
"loss": 0.5278,
"step": 471
},
{
"epoch": 2.41664,
"grad_norm": 0.24519777588872835,
"learning_rate": 8.769344628056893e-06,
"loss": 0.548,
"step": 472
},
{
"epoch": 2.42176,
"grad_norm": 0.22910426740880335,
"learning_rate": 8.62062957702698e-06,
"loss": 0.5186,
"step": 473
},
{
"epoch": 2.42688,
"grad_norm": 0.2172444686293662,
"learning_rate": 8.473033889627404e-06,
"loss": 0.5316,
"step": 474
},
{
"epoch": 2.432,
"grad_norm": 0.21789541593913744,
"learning_rate": 8.326562830885559e-06,
"loss": 0.4964,
"step": 475
},
{
"epoch": 2.43712,
"grad_norm": 0.26419477579272155,
"learning_rate": 8.181221625711102e-06,
"loss": 0.5285,
"step": 476
},
{
"epoch": 2.44224,
"grad_norm": 0.22975957761205745,
"learning_rate": 8.03701545870963e-06,
"loss": 0.5323,
"step": 477
},
{
"epoch": 2.4473599999999998,
"grad_norm": 0.2192599105025437,
"learning_rate": 7.893949473997682e-06,
"loss": 0.513,
"step": 478
},
{
"epoch": 2.45248,
"grad_norm": 0.21383791397200727,
"learning_rate": 7.752028775019264e-06,
"loss": 0.5287,
"step": 479
},
{
"epoch": 2.4576000000000002,
"grad_norm": 0.23011437791676428,
"learning_rate": 7.61125842436381e-06,
"loss": 0.5158,
"step": 480
},
{
"epoch": 2.46272,
"grad_norm": 0.23089368890405654,
"learning_rate": 7.471643443585561e-06,
"loss": 0.5286,
"step": 481
},
{
"epoch": 2.46784,
"grad_norm": 0.21550660720139894,
"learning_rate": 7.33318881302445e-06,
"loss": 0.5315,
"step": 482
},
{
"epoch": 2.47296,
"grad_norm": 0.21786834523919976,
"learning_rate": 7.1958994716284556e-06,
"loss": 0.5344,
"step": 483
},
{
"epoch": 2.47808,
"grad_norm": 0.23493881832666266,
"learning_rate": 7.059780316777396e-06,
"loss": 0.5298,
"step": 484
},
{
"epoch": 2.4832,
"grad_norm": 0.23179568273154502,
"learning_rate": 6.9248362041082514e-06,
"loss": 0.5395,
"step": 485
},
{
"epoch": 2.48832,
"grad_norm": 0.1997326200659906,
"learning_rate": 6.791071947341939e-06,
"loss": 0.5344,
"step": 486
},
{
"epoch": 2.49344,
"grad_norm": 0.18863063393495233,
"learning_rate": 6.658492318111611e-06,
"loss": 0.528,
"step": 487
},
{
"epoch": 2.49856,
"grad_norm": 0.2042454987848812,
"learning_rate": 6.527102045792424e-06,
"loss": 0.5057,
"step": 488
},
{
"epoch": 2.50368,
"grad_norm": 0.21633481485073822,
"learning_rate": 6.39690581733285e-06,
"loss": 0.5231,
"step": 489
},
{
"epoch": 2.5088,
"grad_norm": 0.20671940715964057,
"learning_rate": 6.267908277087489e-06,
"loss": 0.519,
"step": 490
},
{
"epoch": 2.51392,
"grad_norm": 0.20567854945985983,
"learning_rate": 6.140114026651338e-06,
"loss": 0.5246,
"step": 491
},
{
"epoch": 2.51904,
"grad_norm": 0.21668723148852223,
"learning_rate": 6.01352762469575e-06,
"loss": 0.5234,
"step": 492
},
{
"epoch": 2.52416,
"grad_norm": 0.1955709273332618,
"learning_rate": 5.888153586805723e-06,
"loss": 0.4833,
"step": 493
},
{
"epoch": 2.52928,
"grad_norm": 0.19018764122168655,
"learning_rate": 5.763996385318838e-06,
"loss": 0.5241,
"step": 494
},
{
"epoch": 2.5343999999999998,
"grad_norm": 0.21895623789510293,
"learning_rate": 5.641060449165774e-06,
"loss": 0.5022,
"step": 495
},
{
"epoch": 2.53952,
"grad_norm": 0.23471379305483125,
"learning_rate": 5.5193501637122605e-06,
"loss": 0.527,
"step": 496
},
{
"epoch": 2.5446400000000002,
"grad_norm": 0.2119274689582658,
"learning_rate": 5.39886987060267e-06,
"loss": 0.5376,
"step": 497
},
{
"epoch": 2.54976,
"grad_norm": 0.19216483721783073,
"learning_rate": 5.279623867605144e-06,
"loss": 0.5221,
"step": 498
},
{
"epoch": 2.55488,
"grad_norm": 0.2154169515261133,
"learning_rate": 5.161616408458239e-06,
"loss": 0.5188,
"step": 499
},
{
"epoch": 2.56,
"grad_norm": 0.20441281420775267,
"learning_rate": 5.044851702719289e-06,
"loss": 0.5303,
"step": 500
},
{
"epoch": 2.56512,
"grad_norm": 0.2107906688631621,
"learning_rate": 4.929333915614147e-06,
"loss": 0.511,
"step": 501
},
{
"epoch": 2.57024,
"grad_norm": 0.21514247844377765,
"learning_rate": 4.815067167888603e-06,
"loss": 0.5245,
"step": 502
},
{
"epoch": 2.57536,
"grad_norm": 0.2061450148487434,
"learning_rate": 4.702055535661481e-06,
"loss": 0.5079,
"step": 503
},
{
"epoch": 2.58048,
"grad_norm": 0.19909839216020486,
"learning_rate": 4.590303050279144e-06,
"loss": 0.5192,
"step": 504
},
{
"epoch": 2.5856,
"grad_norm": 0.20208792009677243,
"learning_rate": 4.479813698171702e-06,
"loss": 0.5102,
"step": 505
},
{
"epoch": 2.59072,
"grad_norm": 0.2120494653558063,
"learning_rate": 4.370591420710879e-06,
"loss": 0.5276,
"step": 506
},
{
"epoch": 2.59584,
"grad_norm": 0.19182746478809978,
"learning_rate": 4.262640114069303e-06,
"loss": 0.5297,
"step": 507
},
{
"epoch": 2.60096,
"grad_norm": 0.2049595817249684,
"learning_rate": 4.1559636290816165e-06,
"loss": 0.5089,
"step": 508
},
{
"epoch": 2.60608,
"grad_norm": 0.20334632758136215,
"learning_rate": 4.050565771107086e-06,
"loss": 0.5341,
"step": 509
},
{
"epoch": 2.6112,
"grad_norm": 0.21626882759385432,
"learning_rate": 3.946450299893813e-06,
"loss": 0.5356,
"step": 510
},
{
"epoch": 2.61632,
"grad_norm": 0.22272362239503654,
"learning_rate": 3.843620929444667e-06,
"loss": 0.5264,
"step": 511
},
{
"epoch": 2.6214399999999998,
"grad_norm": 0.20142676114788888,
"learning_rate": 3.74208132788481e-06,
"loss": 0.5277,
"step": 512
},
{
"epoch": 2.62656,
"grad_norm": 0.19564873650299325,
"learning_rate": 3.641835117330761e-06,
"loss": 0.532,
"step": 513
},
{
"epoch": 2.6316800000000002,
"grad_norm": 0.21279829206795126,
"learning_rate": 3.542885873761308e-06,
"loss": 0.5163,
"step": 514
},
{
"epoch": 2.6368,
"grad_norm": 0.20724605046419126,
"learning_rate": 3.4452371268898444e-06,
"loss": 0.5074,
"step": 515
},
{
"epoch": 2.64192,
"grad_norm": 0.2140560895573128,
"learning_rate": 3.3488923600385028e-06,
"loss": 0.5331,
"step": 516
},
{
"epoch": 2.64704,
"grad_norm": 0.19348656997795682,
"learning_rate": 3.2538550100139353e-06,
"loss": 0.5273,
"step": 517
},
{
"epoch": 2.65216,
"grad_norm": 0.1959351210860576,
"learning_rate": 3.160128466984609e-06,
"loss": 0.5189,
"step": 518
},
{
"epoch": 2.65728,
"grad_norm": 0.2043908489285227,
"learning_rate": 3.0677160743599788e-06,
"loss": 0.5349,
"step": 519
},
{
"epoch": 2.6624,
"grad_norm": 0.20288115925706235,
"learning_rate": 2.9766211286711868e-06,
"loss": 0.5248,
"step": 520
},
{
"epoch": 2.66752,
"grad_norm": 0.20731847438641798,
"learning_rate": 2.8868468794534243e-06,
"loss": 0.5181,
"step": 521
},
{
"epoch": 2.67264,
"grad_norm": 0.19303823605826137,
"learning_rate": 2.7983965291300765e-06,
"loss": 0.5309,
"step": 522
},
{
"epoch": 2.67776,
"grad_norm": 0.2049514957880367,
"learning_rate": 2.7112732328984594e-06,
"loss": 0.5301,
"step": 523
},
{
"epoch": 2.68288,
"grad_norm": 0.19908648001763588,
"learning_rate": 2.6254800986172635e-06,
"loss": 0.5278,
"step": 524
},
{
"epoch": 2.6879999999999997,
"grad_norm": 0.20892226869833805,
"learning_rate": 2.5410201866957042e-06,
"loss": 0.5236,
"step": 525
},
{
"epoch": 2.69312,
"grad_norm": 0.18255526452984,
"learning_rate": 2.457896509984332e-06,
"loss": 0.525,
"step": 526
},
{
"epoch": 2.69824,
"grad_norm": 0.19075543504463546,
"learning_rate": 2.376112033667579e-06,
"loss": 0.5139,
"step": 527
},
{
"epoch": 2.70336,
"grad_norm": 0.19978943687563103,
"learning_rate": 2.29566967515797e-06,
"loss": 0.523,
"step": 528
},
{
"epoch": 2.7084799999999998,
"grad_norm": 0.17850021044317485,
"learning_rate": 2.216572303992055e-06,
"loss": 0.5327,
"step": 529
},
{
"epoch": 2.7136,
"grad_norm": 0.1854277104912946,
"learning_rate": 2.1388227417280527e-06,
"loss": 0.5233,
"step": 530
},
{
"epoch": 2.7187200000000002,
"grad_norm": 0.18348286998112376,
"learning_rate": 2.0624237618452047e-06,
"loss": 0.5396,
"step": 531
},
{
"epoch": 2.72384,
"grad_norm": 0.1883434076908113,
"learning_rate": 1.98737808964482e-06,
"loss": 0.5197,
"step": 532
},
{
"epoch": 2.72896,
"grad_norm": 0.1841837424394637,
"learning_rate": 1.913688402153082e-06,
"loss": 0.5297,
"step": 533
},
{
"epoch": 2.73408,
"grad_norm": 0.17622787049259134,
"learning_rate": 1.8413573280255326e-06,
"loss": 0.5285,
"step": 534
},
{
"epoch": 2.7392,
"grad_norm": 0.17806753215393287,
"learning_rate": 1.7703874474533167e-06,
"loss": 0.5278,
"step": 535
},
{
"epoch": 2.74432,
"grad_norm": 0.1958676559056442,
"learning_rate": 1.7007812920711408e-06,
"loss": 0.5044,
"step": 536
},
{
"epoch": 2.74944,
"grad_norm": 0.16605984676027447,
"learning_rate": 1.6325413448669625e-06,
"loss": 0.514,
"step": 537
},
{
"epoch": 2.75456,
"grad_norm": 0.18195390968020672,
"learning_rate": 1.5656700400934121e-06,
"loss": 0.5445,
"step": 538
},
{
"epoch": 2.75968,
"grad_norm": 0.16824479934145423,
"learning_rate": 1.500169763180961e-06,
"loss": 0.5212,
"step": 539
},
{
"epoch": 2.7648,
"grad_norm": 0.1819730086116406,
"learning_rate": 1.4360428506528323e-06,
"loss": 0.5232,
"step": 540
},
{
"epoch": 2.76992,
"grad_norm": 0.17926267467820442,
"learning_rate": 1.373291590041661e-06,
"loss": 0.5273,
"step": 541
},
{
"epoch": 2.7750399999999997,
"grad_norm": 0.21059680680377105,
"learning_rate": 1.3119182198078596e-06,
"loss": 0.5131,
"step": 542
},
{
"epoch": 2.78016,
"grad_norm": 0.18407921454967086,
"learning_rate": 1.2519249292598112e-06,
"loss": 0.4969,
"step": 543
},
{
"epoch": 2.78528,
"grad_norm": 0.18395693454820505,
"learning_rate": 1.1933138584757508e-06,
"loss": 0.5188,
"step": 544
},
{
"epoch": 2.7904,
"grad_norm": 0.19315285688390138,
"learning_rate": 1.1360870982274118e-06,
"loss": 0.5186,
"step": 545
},
{
"epoch": 2.79552,
"grad_norm": 0.1887790134903703,
"learning_rate": 1.080246689905473e-06,
"loss": 0.5271,
"step": 546
},
{
"epoch": 2.80064,
"grad_norm": 0.17959548964956856,
"learning_rate": 1.0257946254467145e-06,
"loss": 0.5243,
"step": 547
},
{
"epoch": 2.8057600000000003,
"grad_norm": 0.1817474574269339,
"learning_rate": 9.727328472629716e-07,
"loss": 0.5287,
"step": 548
},
{
"epoch": 2.81088,
"grad_norm": 0.17642219828487612,
"learning_rate": 9.21063248171854e-07,
"loss": 0.526,
"step": 549
},
{
"epoch": 2.816,
"grad_norm": 0.1904246025654594,
"learning_rate": 8.707876713291941e-07,
"loss": 0.5135,
"step": 550
},
{
"epoch": 2.82112,
"grad_norm": 0.17957762634349514,
"learning_rate": 8.219079101633443e-07,
"loss": 0.5215,
"step": 551
},
{
"epoch": 2.82624,
"grad_norm": 0.16263065954048853,
"learning_rate": 7.744257083111662e-07,
"loss": 0.5193,
"step": 552
},
{
"epoch": 2.83136,
"grad_norm": 0.17484720518211333,
"learning_rate": 7.283427595558224e-07,
"loss": 0.5296,
"step": 553
},
{
"epoch": 2.83648,
"grad_norm": 0.1867914151335944,
"learning_rate": 6.836607077664115e-07,
"loss": 0.5383,
"step": 554
},
{
"epoch": 2.8416,
"grad_norm": 0.18028174989209536,
"learning_rate": 6.403811468392685e-07,
"loss": 0.5453,
"step": 555
},
{
"epoch": 2.84672,
"grad_norm": 0.1730198700904891,
"learning_rate": 5.9850562064113e-07,
"loss": 0.5272,
"step": 556
},
{
"epoch": 2.85184,
"grad_norm": 0.17377782938839437,
"learning_rate": 5.580356229540851e-07,
"loss": 0.5298,
"step": 557
},
{
"epoch": 2.85696,
"grad_norm": 0.16971911427369651,
"learning_rate": 5.189725974222448e-07,
"loss": 0.5239,
"step": 558
},
{
"epoch": 2.8620799999999997,
"grad_norm": 0.1853393873016306,
"learning_rate": 4.81317937500263e-07,
"loss": 0.5358,
"step": 559
},
{
"epoch": 2.8672,
"grad_norm": 0.16321076208370755,
"learning_rate": 4.450729864036607e-07,
"loss": 0.511,
"step": 560
},
{
"epoch": 2.87232,
"grad_norm": 0.17617659485628395,
"learning_rate": 4.1023903706084666e-07,
"loss": 0.5251,
"step": 561
},
{
"epoch": 2.87744,
"grad_norm": 0.17229617450371348,
"learning_rate": 3.768173320670654e-07,
"loss": 0.516,
"step": 562
},
{
"epoch": 2.88256,
"grad_norm": 0.18627902011447772,
"learning_rate": 3.448090636400192e-07,
"loss": 0.527,
"step": 563
},
{
"epoch": 2.88768,
"grad_norm": 0.1692917592955478,
"learning_rate": 3.1421537357735567e-07,
"loss": 0.5203,
"step": 564
},
{
"epoch": 2.8928000000000003,
"grad_norm": 0.1742890382438682,
"learning_rate": 2.850373532159578e-07,
"loss": 0.5099,
"step": 565
},
{
"epoch": 2.89792,
"grad_norm": 0.18401122824627672,
"learning_rate": 2.5727604339297996e-07,
"loss": 0.5303,
"step": 566
},
{
"epoch": 2.90304,
"grad_norm": 0.17215703579697378,
"learning_rate": 2.309324344087349e-07,
"loss": 0.5248,
"step": 567
},
{
"epoch": 2.90816,
"grad_norm": 0.1744660017717841,
"learning_rate": 2.060074659913891e-07,
"loss": 0.5067,
"step": 568
},
{
"epoch": 2.91328,
"grad_norm": 0.1753887233423339,
"learning_rate": 1.8250202726339815e-07,
"loss": 0.5142,
"step": 569
},
{
"epoch": 2.9184,
"grad_norm": 0.17222558777930064,
"learning_rate": 1.6041695670981684e-07,
"loss": 0.5271,
"step": 570
},
{
"epoch": 2.92352,
"grad_norm": 0.1749876224664464,
"learning_rate": 1.397530421483939e-07,
"loss": 0.5215,
"step": 571
},
{
"epoch": 2.92864,
"grad_norm": 0.17369587300943137,
"learning_rate": 1.2051102070144816e-07,
"loss": 0.5248,
"step": 572
},
{
"epoch": 2.93376,
"grad_norm": 0.1720447918581037,
"learning_rate": 1.0269157876959146e-07,
"loss": 0.5198,
"step": 573
},
{
"epoch": 2.93888,
"grad_norm": 0.18006932981535365,
"learning_rate": 8.62953520072285e-08,
"loss": 0.5081,
"step": 574
},
{
"epoch": 2.944,
"grad_norm": 0.16965260368509588,
"learning_rate": 7.132292529990814e-08,
"loss": 0.5227,
"step": 575
},
{
"epoch": 2.9491199999999997,
"grad_norm": 0.17833147445276212,
"learning_rate": 5.7774832743415776e-08,
"loss": 0.5269,
"step": 576
},
{
"epoch": 2.95424,
"grad_norm": 0.17095515499148287,
"learning_rate": 4.565155762477069e-08,
"loss": 0.5416,
"step": 577
},
{
"epoch": 2.95936,
"grad_norm": 0.17471514998398122,
"learning_rate": 3.4953532404942146e-08,
"loss": 0.5026,
"step": 578
},
{
"epoch": 2.96448,
"grad_norm": 0.1721020510585504,
"learning_rate": 2.568113870346167e-08,
"loss": 0.5092,
"step": 579
},
{
"epoch": 2.9696,
"grad_norm": 0.18037480336003978,
"learning_rate": 1.783470728477621e-08,
"loss": 0.514,
"step": 580
},
{
"epoch": 2.97472,
"grad_norm": 0.17100517198442003,
"learning_rate": 1.1414518046470868e-08,
"loss": 0.5405,
"step": 581
},
{
"epoch": 2.9798400000000003,
"grad_norm": 0.1702692183309079,
"learning_rate": 6.4208000092769174e-09,
"loss": 0.5228,
"step": 582
},
{
"epoch": 2.98496,
"grad_norm": 0.17252007004331688,
"learning_rate": 2.8537313088961015e-09,
"loss": 0.5295,
"step": 583
},
{
"epoch": 2.99008,
"grad_norm": 0.18840800940002203,
"learning_rate": 7.134391896679305e-10,
"loss": 0.5159,
"step": 584
},
{
"epoch": 2.9952,
"grad_norm": 0.1870617408251541,
"learning_rate": 0.0,
"loss": 0.4924,
"step": 585
},
{
"epoch": 2.9952,
"step": 585,
"total_flos": 4.2089030343416873e+18,
"train_loss": 0.0,
"train_runtime": 5.9478,
"train_samples_per_second": 50438.911,
"train_steps_per_second": 98.356
}
],
"logging_steps": 1,
"max_steps": 585,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 4.2089030343416873e+18,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}