{ "best_metric": null, "best_model_checkpoint": null, "epoch": 5.0, "eval_steps": 500, "global_step": 215, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.023255813953488372, "grad_norm": 5.434192361984504, "learning_rate": 3.6363636363636366e-06, "loss": 0.9705, "step": 1 }, { "epoch": 0.046511627906976744, "grad_norm": 5.477136338528377, "learning_rate": 7.272727272727273e-06, "loss": 0.9693, "step": 2 }, { "epoch": 0.06976744186046512, "grad_norm": 4.994172293091364, "learning_rate": 1.0909090909090909e-05, "loss": 0.971, "step": 3 }, { "epoch": 0.09302325581395349, "grad_norm": 1.9715768923169905, "learning_rate": 1.4545454545454546e-05, "loss": 0.882, "step": 4 }, { "epoch": 0.11627906976744186, "grad_norm": 4.86558025165536, "learning_rate": 1.8181818181818182e-05, "loss": 0.9089, "step": 5 }, { "epoch": 0.13953488372093023, "grad_norm": 6.162689558609649, "learning_rate": 2.1818181818181818e-05, "loss": 0.9031, "step": 6 }, { "epoch": 0.16279069767441862, "grad_norm": 5.545748731779107, "learning_rate": 2.5454545454545457e-05, "loss": 0.9058, "step": 7 }, { "epoch": 0.18604651162790697, "grad_norm": 3.4785773417590877, "learning_rate": 2.9090909090909093e-05, "loss": 0.8423, "step": 8 }, { "epoch": 0.20930232558139536, "grad_norm": 2.16515163827823, "learning_rate": 3.272727272727273e-05, "loss": 0.8043, "step": 9 }, { "epoch": 0.23255813953488372, "grad_norm": 2.664754416131028, "learning_rate": 3.6363636363636364e-05, "loss": 0.7799, "step": 10 }, { "epoch": 0.2558139534883721, "grad_norm": 1.4902721906661134, "learning_rate": 4e-05, "loss": 0.7546, "step": 11 }, { "epoch": 0.27906976744186046, "grad_norm": 2.015625441498336, "learning_rate": 4.3636363636363636e-05, "loss": 0.7408, "step": 12 }, { "epoch": 0.3023255813953488, "grad_norm": 1.1406136624119514, "learning_rate": 4.727272727272728e-05, "loss": 0.7216, "step": 13 }, { "epoch": 0.32558139534883723, "grad_norm": 1.8195500867443861, "learning_rate": 5.0909090909090914e-05, "loss": 0.724, "step": 14 }, { "epoch": 0.3488372093023256, "grad_norm": 1.595476578575812, "learning_rate": 5.4545454545454546e-05, "loss": 0.7084, "step": 15 }, { "epoch": 0.37209302325581395, "grad_norm": 1.9850762532377662, "learning_rate": 5.8181818181818185e-05, "loss": 0.6992, "step": 16 }, { "epoch": 0.3953488372093023, "grad_norm": 1.3156392269832544, "learning_rate": 6.181818181818182e-05, "loss": 0.6969, "step": 17 }, { "epoch": 0.4186046511627907, "grad_norm": 2.165349958578573, "learning_rate": 6.545454545454546e-05, "loss": 0.6819, "step": 18 }, { "epoch": 0.4418604651162791, "grad_norm": 1.860669140085515, "learning_rate": 6.90909090909091e-05, "loss": 0.6986, "step": 19 }, { "epoch": 0.46511627906976744, "grad_norm": 1.7217756323779048, "learning_rate": 7.272727272727273e-05, "loss": 0.6801, "step": 20 }, { "epoch": 0.4883720930232558, "grad_norm": 1.4025893684782413, "learning_rate": 7.636363636363637e-05, "loss": 0.6751, "step": 21 }, { "epoch": 0.5116279069767442, "grad_norm": 1.7449603840187862, "learning_rate": 8e-05, "loss": 0.673, "step": 22 }, { "epoch": 0.5348837209302325, "grad_norm": 2.482316452405395, "learning_rate": 7.999470085828896e-05, "loss": 0.6629, "step": 23 }, { "epoch": 0.5581395348837209, "grad_norm": 1.2207804383595053, "learning_rate": 7.997880483720098e-05, "loss": 0.6716, "step": 24 }, { "epoch": 0.5813953488372093, "grad_norm": 2.568593544907844, "learning_rate": 7.995231614849947e-05, "loss": 0.671, "step": 25 }, { "epoch": 0.6046511627906976, "grad_norm": 1.502461502116564, "learning_rate": 7.991524181055018e-05, "loss": 0.6608, "step": 26 }, { "epoch": 0.627906976744186, "grad_norm": 2.8212929800172915, "learning_rate": 7.986759164646167e-05, "loss": 0.673, "step": 27 }, { "epoch": 0.6511627906976745, "grad_norm": 2.5422318457006825, "learning_rate": 7.980937828148252e-05, "loss": 0.6916, "step": 28 }, { "epoch": 0.6744186046511628, "grad_norm": 1.6426727549059879, "learning_rate": 7.974061713965625e-05, "loss": 0.6487, "step": 29 }, { "epoch": 0.6976744186046512, "grad_norm": 1.8002811870915256, "learning_rate": 7.966132643973464e-05, "loss": 0.6429, "step": 30 }, { "epoch": 0.7209302325581395, "grad_norm": 1.4654311466330405, "learning_rate": 7.95715271903504e-05, "loss": 0.6398, "step": 31 }, { "epoch": 0.7441860465116279, "grad_norm": 1.329866741625803, "learning_rate": 7.947124318445097e-05, "loss": 0.6398, "step": 32 }, { "epoch": 0.7674418604651163, "grad_norm": 1.4731638464740446, "learning_rate": 7.936050099299425e-05, "loss": 0.6243, "step": 33 }, { "epoch": 0.7906976744186046, "grad_norm": 1.4894523263564694, "learning_rate": 7.923932995790858e-05, "loss": 0.6335, "step": 34 }, { "epoch": 0.813953488372093, "grad_norm": 1.173972016468102, "learning_rate": 7.910776218431821e-05, "loss": 0.6224, "step": 35 }, { "epoch": 0.8372093023255814, "grad_norm": 1.6637513739112912, "learning_rate": 7.896583253203703e-05, "loss": 0.6219, "step": 36 }, { "epoch": 0.8604651162790697, "grad_norm": 1.122406312089788, "learning_rate": 7.881357860633204e-05, "loss": 0.6232, "step": 37 }, { "epoch": 0.8837209302325582, "grad_norm": 1.4388855947402022, "learning_rate": 7.865104074795966e-05, "loss": 0.5977, "step": 38 }, { "epoch": 0.9069767441860465, "grad_norm": 1.4139997348656486, "learning_rate": 7.847826202247715e-05, "loss": 0.6193, "step": 39 }, { "epoch": 0.9302325581395349, "grad_norm": 1.4041482845121371, "learning_rate": 7.829528820883205e-05, "loss": 0.604, "step": 40 }, { "epoch": 0.9534883720930233, "grad_norm": 1.1795408353423182, "learning_rate": 7.810216778723275e-05, "loss": 0.5909, "step": 41 }, { "epoch": 0.9767441860465116, "grad_norm": 1.0271878945240036, "learning_rate": 7.789895192630333e-05, "loss": 0.5927, "step": 42 }, { "epoch": 1.0, "grad_norm": 1.7552637527608157, "learning_rate": 7.768569446952603e-05, "loss": 0.5843, "step": 43 }, { "epoch": 1.0232558139534884, "grad_norm": 0.9212131281614243, "learning_rate": 7.746245192097509e-05, "loss": 0.5722, "step": 44 }, { "epoch": 1.0465116279069768, "grad_norm": 1.6756296752481579, "learning_rate": 7.722928343034551e-05, "loss": 0.5773, "step": 45 }, { "epoch": 1.069767441860465, "grad_norm": 1.3543712007327793, "learning_rate": 7.698625077728103e-05, "loss": 0.5672, "step": 46 }, { "epoch": 1.0930232558139534, "grad_norm": 1.7410081511463789, "learning_rate": 7.67334183550051e-05, "loss": 0.5674, "step": 47 }, { "epoch": 1.1162790697674418, "grad_norm": 1.0733210547826335, "learning_rate": 7.647085315325946e-05, "loss": 0.5671, "step": 48 }, { "epoch": 1.1395348837209303, "grad_norm": 2.252268000471556, "learning_rate": 7.619862474055472e-05, "loss": 0.5638, "step": 49 }, { "epoch": 1.1627906976744187, "grad_norm": 1.6138509974039352, "learning_rate": 7.591680524573773e-05, "loss": 0.5625, "step": 50 }, { "epoch": 1.1860465116279069, "grad_norm": 2.0237798746000193, "learning_rate": 7.56254693388805e-05, "loss": 0.5592, "step": 51 }, { "epoch": 1.2093023255813953, "grad_norm": 1.4811712482469028, "learning_rate": 7.532469421149578e-05, "loss": 0.5599, "step": 52 }, { "epoch": 1.2325581395348837, "grad_norm": 1.5279703879919382, "learning_rate": 7.501455955608479e-05, "loss": 0.5559, "step": 53 }, { "epoch": 1.255813953488372, "grad_norm": 1.2468524365094296, "learning_rate": 7.469514754502191e-05, "loss": 0.5549, "step": 54 }, { "epoch": 1.2790697674418605, "grad_norm": 1.2825998428965273, "learning_rate": 7.436654280878269e-05, "loss": 0.546, "step": 55 }, { "epoch": 1.302325581395349, "grad_norm": 0.9710691709326068, "learning_rate": 7.402883241352037e-05, "loss": 0.5492, "step": 56 }, { "epoch": 1.3255813953488373, "grad_norm": 1.4012298372305019, "learning_rate": 7.368210583799701e-05, "loss": 0.5481, "step": 57 }, { "epoch": 1.3488372093023255, "grad_norm": 1.2045446990194086, "learning_rate": 7.332645494987555e-05, "loss": 0.544, "step": 58 }, { "epoch": 1.372093023255814, "grad_norm": 1.5478512274120089, "learning_rate": 7.296197398137878e-05, "loss": 0.5399, "step": 59 }, { "epoch": 1.3953488372093024, "grad_norm": 1.0482231123455121, "learning_rate": 7.258875950432185e-05, "loss": 0.5457, "step": 60 }, { "epoch": 1.4186046511627908, "grad_norm": 1.350025832651234, "learning_rate": 7.220691040452491e-05, "loss": 0.533, "step": 61 }, { "epoch": 1.441860465116279, "grad_norm": 1.6379778302822026, "learning_rate": 7.181652785561254e-05, "loss": 0.5467, "step": 62 }, { "epoch": 1.4651162790697674, "grad_norm": 0.7697335505845303, "learning_rate": 7.141771529220716e-05, "loss": 0.536, "step": 63 }, { "epoch": 1.4883720930232558, "grad_norm": 1.019518212307104, "learning_rate": 7.101057838252324e-05, "loss": 0.5309, "step": 64 }, { "epoch": 1.5116279069767442, "grad_norm": 1.1970177274801466, "learning_rate": 7.05952250003698e-05, "loss": 0.535, "step": 65 }, { "epoch": 1.5348837209302326, "grad_norm": 1.0203220373942377, "learning_rate": 7.017176519656845e-05, "loss": 0.5378, "step": 66 }, { "epoch": 1.558139534883721, "grad_norm": 1.0958072090841686, "learning_rate": 6.974031116979466e-05, "loss": 0.5407, "step": 67 }, { "epoch": 1.5813953488372094, "grad_norm": 1.3397310047597761, "learning_rate": 6.93009772368499e-05, "loss": 0.5375, "step": 68 }, { "epoch": 1.6046511627906976, "grad_norm": 0.8782131687864965, "learning_rate": 6.885387980237262e-05, "loss": 0.5251, "step": 69 }, { "epoch": 1.627906976744186, "grad_norm": 1.4357177318604477, "learning_rate": 6.839913732799604e-05, "loss": 0.5274, "step": 70 }, { "epoch": 1.6511627906976745, "grad_norm": 1.215047188865305, "learning_rate": 6.793687030096085e-05, "loss": 0.5295, "step": 71 }, { "epoch": 1.6744186046511627, "grad_norm": 1.2710521920417013, "learning_rate": 6.746720120219126e-05, "loss": 0.5421, "step": 72 }, { "epoch": 1.697674418604651, "grad_norm": 0.7813066526911904, "learning_rate": 6.699025447384284e-05, "loss": 0.5312, "step": 73 }, { "epoch": 1.7209302325581395, "grad_norm": 1.2496179982961972, "learning_rate": 6.650615648633075e-05, "loss": 0.5246, "step": 74 }, { "epoch": 1.744186046511628, "grad_norm": 1.311820641764223, "learning_rate": 6.601503550484684e-05, "loss": 0.5258, "step": 75 }, { "epoch": 1.7674418604651163, "grad_norm": 1.0893635673008315, "learning_rate": 6.5517021655375e-05, "loss": 0.5271, "step": 76 }, { "epoch": 1.7906976744186047, "grad_norm": 0.7546199150125754, "learning_rate": 6.501224689021341e-05, "loss": 0.5185, "step": 77 }, { "epoch": 1.8139534883720931, "grad_norm": 1.1638269494753801, "learning_rate": 6.450084495301268e-05, "loss": 0.5169, "step": 78 }, { "epoch": 1.8372093023255816, "grad_norm": 1.1455758960103528, "learning_rate": 6.398295134333962e-05, "loss": 0.5197, "step": 79 }, { "epoch": 1.8604651162790697, "grad_norm": 0.9787313133874568, "learning_rate": 6.345870328077569e-05, "loss": 0.5189, "step": 80 }, { "epoch": 1.8837209302325582, "grad_norm": 1.2146657230893994, "learning_rate": 6.292823966855965e-05, "loss": 0.5184, "step": 81 }, { "epoch": 1.9069767441860463, "grad_norm": 0.8315295395061865, "learning_rate": 6.239170105678419e-05, "loss": 0.5151, "step": 82 }, { "epoch": 1.9302325581395348, "grad_norm": 0.43739239970652544, "learning_rate": 6.184922960515616e-05, "loss": 0.5088, "step": 83 }, { "epoch": 1.9534883720930232, "grad_norm": 0.8529923746757689, "learning_rate": 6.13009690453304e-05, "loss": 0.5212, "step": 84 }, { "epoch": 1.9767441860465116, "grad_norm": 0.7915534358179345, "learning_rate": 6.074706464282695e-05, "loss": 0.5081, "step": 85 }, { "epoch": 2.0, "grad_norm": 0.838942938146, "learning_rate": 6.018766315854196e-05, "loss": 0.4843, "step": 86 }, { "epoch": 2.0232558139534884, "grad_norm": 1.0664271997362778, "learning_rate": 5.9622912809862384e-05, "loss": 0.4787, "step": 87 }, { "epoch": 2.046511627906977, "grad_norm": 1.4645846831681002, "learning_rate": 5.905296323139467e-05, "loss": 0.4779, "step": 88 }, { "epoch": 2.0697674418604652, "grad_norm": 1.0222504468766858, "learning_rate": 5.847796543531801e-05, "loss": 0.4708, "step": 89 }, { "epoch": 2.0930232558139537, "grad_norm": 1.446291157107826, "learning_rate": 5.789807177137269e-05, "loss": 0.4747, "step": 90 }, { "epoch": 2.116279069767442, "grad_norm": 0.894564054089064, "learning_rate": 5.731343588649382e-05, "loss": 0.4707, "step": 91 }, { "epoch": 2.13953488372093, "grad_norm": 1.1812623112632408, "learning_rate": 5.6724212684101567e-05, "loss": 0.4724, "step": 92 }, { "epoch": 2.1627906976744184, "grad_norm": 1.1141414953554294, "learning_rate": 5.613055828305838e-05, "loss": 0.4659, "step": 93 }, { "epoch": 2.186046511627907, "grad_norm": 0.7770582052783432, "learning_rate": 5.553262997630419e-05, "loss": 0.4592, "step": 94 }, { "epoch": 2.2093023255813953, "grad_norm": 1.0201596899389407, "learning_rate": 5.493058618918051e-05, "loss": 0.464, "step": 95 }, { "epoch": 2.2325581395348837, "grad_norm": 0.99732039808284, "learning_rate": 5.432458643745456e-05, "loss": 0.4728, "step": 96 }, { "epoch": 2.255813953488372, "grad_norm": 0.9946765474703833, "learning_rate": 5.37147912850544e-05, "loss": 0.4719, "step": 97 }, { "epoch": 2.2790697674418605, "grad_norm": 0.67376712726853, "learning_rate": 5.31013623015264e-05, "loss": 0.4635, "step": 98 }, { "epoch": 2.302325581395349, "grad_norm": 0.5331380825906256, "learning_rate": 5.2484462019226216e-05, "loss": 0.4636, "step": 99 }, { "epoch": 2.3255813953488373, "grad_norm": 0.8114012150052442, "learning_rate": 5.1864253890254744e-05, "loss": 0.4669, "step": 100 }, { "epoch": 2.3488372093023258, "grad_norm": 0.7817511193460437, "learning_rate": 5.1240902243150245e-05, "loss": 0.4643, "step": 101 }, { "epoch": 2.3720930232558137, "grad_norm": 0.5742798063922762, "learning_rate": 5.0614572239348436e-05, "loss": 0.468, "step": 102 }, { "epoch": 2.395348837209302, "grad_norm": 0.7545379109683643, "learning_rate": 4.998542982942169e-05, "loss": 0.4646, "step": 103 }, { "epoch": 2.4186046511627906, "grad_norm": 0.6867300007616275, "learning_rate": 4.9353641709109366e-05, "loss": 0.466, "step": 104 }, { "epoch": 2.441860465116279, "grad_norm": 0.5994645048448306, "learning_rate": 4.87193752751505e-05, "loss": 0.4605, "step": 105 }, { "epoch": 2.4651162790697674, "grad_norm": 0.5648107826549134, "learning_rate": 4.808279858093088e-05, "loss": 0.4633, "step": 106 }, { "epoch": 2.488372093023256, "grad_norm": 0.41656800662182036, "learning_rate": 4.744408029195616e-05, "loss": 0.458, "step": 107 }, { "epoch": 2.511627906976744, "grad_norm": 0.560471313591134, "learning_rate": 4.680338964116267e-05, "loss": 0.4694, "step": 108 }, { "epoch": 2.5348837209302326, "grad_norm": 0.3737750854348531, "learning_rate": 4.616089638407798e-05, "loss": 0.4572, "step": 109 }, { "epoch": 2.558139534883721, "grad_norm": 0.4130815840573508, "learning_rate": 4.551677075384296e-05, "loss": 0.4642, "step": 110 }, { "epoch": 2.5813953488372094, "grad_norm": 0.45613348607122495, "learning_rate": 4.487118341610736e-05, "loss": 0.4558, "step": 111 }, { "epoch": 2.604651162790698, "grad_norm": 0.3913887000750076, "learning_rate": 4.422430542381064e-05, "loss": 0.4646, "step": 112 }, { "epoch": 2.6279069767441863, "grad_norm": 0.30374911049832554, "learning_rate": 4.3576308171860336e-05, "loss": 0.4624, "step": 113 }, { "epoch": 2.6511627906976747, "grad_norm": 0.3917025778134578, "learning_rate": 4.292736335171978e-05, "loss": 0.4566, "step": 114 }, { "epoch": 2.6744186046511627, "grad_norm": 0.3717043272468876, "learning_rate": 4.2277642905917205e-05, "loss": 0.4566, "step": 115 }, { "epoch": 2.697674418604651, "grad_norm": 0.30603425852066757, "learning_rate": 4.162731898248834e-05, "loss": 0.4599, "step": 116 }, { "epoch": 2.7209302325581395, "grad_norm": 0.3251855900379686, "learning_rate": 4.097656388936462e-05, "loss": 0.4623, "step": 117 }, { "epoch": 2.744186046511628, "grad_norm": 0.33442925355720576, "learning_rate": 4.032555004871892e-05, "loss": 0.4587, "step": 118 }, { "epoch": 2.7674418604651163, "grad_norm": 0.33761110069481104, "learning_rate": 3.9674449951281095e-05, "loss": 0.4576, "step": 119 }, { "epoch": 2.7906976744186047, "grad_norm": 0.26743941116764636, "learning_rate": 3.902343611063539e-05, "loss": 0.4556, "step": 120 }, { "epoch": 2.813953488372093, "grad_norm": 0.2128179360983353, "learning_rate": 3.837268101751166e-05, "loss": 0.4558, "step": 121 }, { "epoch": 2.8372093023255816, "grad_norm": 0.2896189154702043, "learning_rate": 3.7722357094082815e-05, "loss": 0.4618, "step": 122 }, { "epoch": 2.8604651162790695, "grad_norm": 0.21296934932760417, "learning_rate": 3.707263664828023e-05, "loss": 0.4591, "step": 123 }, { "epoch": 2.883720930232558, "grad_norm": 0.2791061519264039, "learning_rate": 3.6423691828139685e-05, "loss": 0.4524, "step": 124 }, { "epoch": 2.9069767441860463, "grad_norm": 0.2631221143960164, "learning_rate": 3.577569457618937e-05, "loss": 0.4606, "step": 125 }, { "epoch": 2.9302325581395348, "grad_norm": 0.20684807194617985, "learning_rate": 3.512881658389265e-05, "loss": 0.4578, "step": 126 }, { "epoch": 2.953488372093023, "grad_norm": 0.2697328366910901, "learning_rate": 3.448322924615703e-05, "loss": 0.4544, "step": 127 }, { "epoch": 2.9767441860465116, "grad_norm": 0.22910217651885814, "learning_rate": 3.383910361592204e-05, "loss": 0.458, "step": 128 }, { "epoch": 3.0, "grad_norm": 0.23244831316179806, "learning_rate": 3.319661035883734e-05, "loss": 0.427, "step": 129 }, { "epoch": 3.0232558139534884, "grad_norm": 0.20986791210333858, "learning_rate": 3.2555919708043856e-05, "loss": 0.4232, "step": 130 }, { "epoch": 3.046511627906977, "grad_norm": 0.2570519418980638, "learning_rate": 3.1917201419069126e-05, "loss": 0.4149, "step": 131 }, { "epoch": 3.0697674418604652, "grad_norm": 0.24795179030963221, "learning_rate": 3.128062472484952e-05, "loss": 0.4193, "step": 132 }, { "epoch": 3.0930232558139537, "grad_norm": 0.2576172200398066, "learning_rate": 3.064635829089064e-05, "loss": 0.4113, "step": 133 }, { "epoch": 3.116279069767442, "grad_norm": 0.22687653576900754, "learning_rate": 3.0014570170578324e-05, "loss": 0.417, "step": 134 }, { "epoch": 3.13953488372093, "grad_norm": 0.26047808565271, "learning_rate": 2.938542776065158e-05, "loss": 0.4137, "step": 135 }, { "epoch": 3.1627906976744184, "grad_norm": 0.22319212273415287, "learning_rate": 2.875909775684977e-05, "loss": 0.4211, "step": 136 }, { "epoch": 3.186046511627907, "grad_norm": 0.22428976031004036, "learning_rate": 2.8135746109745273e-05, "loss": 0.4138, "step": 137 }, { "epoch": 3.2093023255813953, "grad_norm": 0.22247453156579614, "learning_rate": 2.751553798077379e-05, "loss": 0.4153, "step": 138 }, { "epoch": 3.2325581395348837, "grad_norm": 0.20007992804768981, "learning_rate": 2.6898637698473605e-05, "loss": 0.4135, "step": 139 }, { "epoch": 3.255813953488372, "grad_norm": 0.2451508001851349, "learning_rate": 2.6285208714945615e-05, "loss": 0.407, "step": 140 }, { "epoch": 3.2790697674418605, "grad_norm": 0.18902377319628208, "learning_rate": 2.5675413562545453e-05, "loss": 0.4044, "step": 141 }, { "epoch": 3.302325581395349, "grad_norm": 0.24194497338579668, "learning_rate": 2.5069413810819503e-05, "loss": 0.4073, "step": 142 }, { "epoch": 3.3255813953488373, "grad_norm": 0.1989749287423866, "learning_rate": 2.446737002369582e-05, "loss": 0.4169, "step": 143 }, { "epoch": 3.3488372093023258, "grad_norm": 0.2095582088688536, "learning_rate": 2.3869441716941627e-05, "loss": 0.4204, "step": 144 }, { "epoch": 3.3720930232558137, "grad_norm": 0.162353572736833, "learning_rate": 2.327578731589843e-05, "loss": 0.4065, "step": 145 }, { "epoch": 3.395348837209302, "grad_norm": 0.17279337899190159, "learning_rate": 2.2686564113506188e-05, "loss": 0.412, "step": 146 }, { "epoch": 3.4186046511627906, "grad_norm": 0.16979965281268794, "learning_rate": 2.210192822862733e-05, "loss": 0.4097, "step": 147 }, { "epoch": 3.441860465116279, "grad_norm": 0.1394513868982289, "learning_rate": 2.1522034564681997e-05, "loss": 0.4136, "step": 148 }, { "epoch": 3.4651162790697674, "grad_norm": 0.15491336187962532, "learning_rate": 2.0947036768605343e-05, "loss": 0.4138, "step": 149 }, { "epoch": 3.488372093023256, "grad_norm": 0.15331640718626022, "learning_rate": 2.0377087190137615e-05, "loss": 0.4121, "step": 150 }, { "epoch": 3.511627906976744, "grad_norm": 0.14333365303471923, "learning_rate": 1.9812336841458032e-05, "loss": 0.412, "step": 151 }, { "epoch": 3.5348837209302326, "grad_norm": 0.15342863178771485, "learning_rate": 1.925293535717306e-05, "loss": 0.4128, "step": 152 }, { "epoch": 3.558139534883721, "grad_norm": 0.1515728462605629, "learning_rate": 1.869903095466962e-05, "loss": 0.4092, "step": 153 }, { "epoch": 3.5813953488372094, "grad_norm": 0.16216900903578507, "learning_rate": 1.815077039484385e-05, "loss": 0.418, "step": 154 }, { "epoch": 3.604651162790698, "grad_norm": 0.14923319773401145, "learning_rate": 1.760829894321582e-05, "loss": 0.4124, "step": 155 }, { "epoch": 3.6279069767441863, "grad_norm": 0.1388357623921274, "learning_rate": 1.7071760331440352e-05, "loss": 0.4134, "step": 156 }, { "epoch": 3.6511627906976747, "grad_norm": 0.14463436978493738, "learning_rate": 1.6541296719224315e-05, "loss": 0.4152, "step": 157 }, { "epoch": 3.6744186046511627, "grad_norm": 0.13948667157444467, "learning_rate": 1.6017048656660392e-05, "loss": 0.4127, "step": 158 }, { "epoch": 3.697674418604651, "grad_norm": 0.12962305417153203, "learning_rate": 1.5499155046987343e-05, "loss": 0.4102, "step": 159 }, { "epoch": 3.7209302325581395, "grad_norm": 0.13488967836960994, "learning_rate": 1.49877531097866e-05, "loss": 0.412, "step": 160 }, { "epoch": 3.744186046511628, "grad_norm": 0.14161978366263686, "learning_rate": 1.4482978344624999e-05, "loss": 0.4112, "step": 161 }, { "epoch": 3.7674418604651163, "grad_norm": 0.12104463455571417, "learning_rate": 1.3984964495153178e-05, "loss": 0.4107, "step": 162 }, { "epoch": 3.7906976744186047, "grad_norm": 0.14778280010357478, "learning_rate": 1.349384351366926e-05, "loss": 0.4103, "step": 163 }, { "epoch": 3.813953488372093, "grad_norm": 0.13974610415305352, "learning_rate": 1.3009745526157165e-05, "loss": 0.4089, "step": 164 }, { "epoch": 3.8372093023255816, "grad_norm": 0.12567459622985308, "learning_rate": 1.2532798797808767e-05, "loss": 0.4082, "step": 165 }, { "epoch": 3.8604651162790695, "grad_norm": 0.13111313669138286, "learning_rate": 1.2063129699039169e-05, "loss": 0.4072, "step": 166 }, { "epoch": 3.883720930232558, "grad_norm": 0.11859583351508307, "learning_rate": 1.1600862672003964e-05, "loss": 0.408, "step": 167 }, { "epoch": 3.9069767441860463, "grad_norm": 0.12426128460062202, "learning_rate": 1.1146120197627375e-05, "loss": 0.41, "step": 168 }, { "epoch": 3.9302325581395348, "grad_norm": 0.12183542106771499, "learning_rate": 1.0699022763150104e-05, "loss": 0.4115, "step": 169 }, { "epoch": 3.953488372093023, "grad_norm": 0.1104533940594946, "learning_rate": 1.0259688830205348e-05, "loss": 0.4055, "step": 170 }, { "epoch": 3.9767441860465116, "grad_norm": 0.11121249489433677, "learning_rate": 9.828234803431559e-06, "loss": 0.411, "step": 171 }, { "epoch": 4.0, "grad_norm": 0.15736411698947347, "learning_rate": 9.404774999630204e-06, "loss": 0.3931, "step": 172 }, { "epoch": 4.023255813953488, "grad_norm": 0.14076644290125107, "learning_rate": 8.989421617476766e-06, "loss": 0.3903, "step": 173 }, { "epoch": 4.046511627906977, "grad_norm": 0.11712681895575112, "learning_rate": 8.582284707792845e-06, "loss": 0.387, "step": 174 }, { "epoch": 4.069767441860465, "grad_norm": 0.12595405435756382, "learning_rate": 8.183472144387456e-06, "loss": 0.3866, "step": 175 }, { "epoch": 4.093023255813954, "grad_norm": 0.13565427155194576, "learning_rate": 7.793089595475094e-06, "loss": 0.3913, "step": 176 }, { "epoch": 4.116279069767442, "grad_norm": 0.1412644950291927, "learning_rate": 7.411240495678158e-06, "loss": 0.3876, "step": 177 }, { "epoch": 4.1395348837209305, "grad_norm": 0.13949397861852225, "learning_rate": 7.038026018621234e-06, "loss": 0.3839, "step": 178 }, { "epoch": 4.162790697674419, "grad_norm": 0.12339126854002234, "learning_rate": 6.673545050124466e-06, "loss": 0.3894, "step": 179 }, { "epoch": 4.186046511627907, "grad_norm": 0.12932008059607125, "learning_rate": 6.317894162003004e-06, "loss": 0.3863, "step": 180 }, { "epoch": 4.209302325581396, "grad_norm": 0.12611084979552667, "learning_rate": 5.971167586479638e-06, "loss": 0.3878, "step": 181 }, { "epoch": 4.232558139534884, "grad_norm": 0.12145402717998689, "learning_rate": 5.633457191217311e-06, "loss": 0.3806, "step": 182 }, { "epoch": 4.2558139534883725, "grad_norm": 0.12362886923322695, "learning_rate": 5.3048524549781066e-06, "loss": 0.3833, "step": 183 }, { "epoch": 4.27906976744186, "grad_norm": 0.1114983032825903, "learning_rate": 4.985440443915223e-06, "loss": 0.3873, "step": 184 }, { "epoch": 4.3023255813953485, "grad_norm": 0.11722293815194844, "learning_rate": 4.6753057885042186e-06, "loss": 0.3827, "step": 185 }, { "epoch": 4.325581395348837, "grad_norm": 0.1250948440370564, "learning_rate": 4.374530661119512e-06, "loss": 0.3855, "step": 186 }, { "epoch": 4.348837209302325, "grad_norm": 0.1106181593146437, "learning_rate": 4.0831947542622655e-06, "loss": 0.3855, "step": 187 }, { "epoch": 4.372093023255814, "grad_norm": 0.10428284018843, "learning_rate": 3.8013752594452834e-06, "loss": 0.3862, "step": 188 }, { "epoch": 4.395348837209302, "grad_norm": 0.1049647850048849, "learning_rate": 3.529146846740559e-06, "loss": 0.3853, "step": 189 }, { "epoch": 4.4186046511627906, "grad_norm": 0.10140234914167741, "learning_rate": 3.266581644994915e-06, "loss": 0.3832, "step": 190 }, { "epoch": 4.441860465116279, "grad_norm": 0.1084178720824487, "learning_rate": 3.0137492227189803e-06, "loss": 0.3874, "step": 191 }, { "epoch": 4.465116279069767, "grad_norm": 0.1035454061342504, "learning_rate": 2.7707165696545035e-06, "loss": 0.3846, "step": 192 }, { "epoch": 4.488372093023256, "grad_norm": 0.10465143616293164, "learning_rate": 2.537548079024923e-06, "loss": 0.3875, "step": 193 }, { "epoch": 4.511627906976744, "grad_norm": 0.10287343108965315, "learning_rate": 2.3143055304739725e-06, "loss": 0.3878, "step": 194 }, { "epoch": 4.534883720930233, "grad_norm": 0.09899078580945446, "learning_rate": 2.1010480736966788e-06, "loss": 0.3844, "step": 195 }, { "epoch": 4.558139534883721, "grad_norm": 0.0947795487504229, "learning_rate": 1.8978322127672522e-06, "loss": 0.3864, "step": 196 }, { "epoch": 4.5813953488372094, "grad_norm": 0.09351592903837026, "learning_rate": 1.7047117911679567e-06, "loss": 0.3764, "step": 197 }, { "epoch": 4.604651162790698, "grad_norm": 0.09075701229280372, "learning_rate": 1.5217379775228503e-06, "loss": 0.3804, "step": 198 }, { "epoch": 4.627906976744186, "grad_norm": 0.09563361674368318, "learning_rate": 1.3489592520403405e-06, "loss": 0.3852, "step": 199 }, { "epoch": 4.651162790697675, "grad_norm": 0.09593704632301235, "learning_rate": 1.186421393667967e-06, "loss": 0.3802, "step": 200 }, { "epoch": 4.674418604651163, "grad_norm": 0.09257390445253719, "learning_rate": 1.034167467962983e-06, "loss": 0.3867, "step": 201 }, { "epoch": 4.6976744186046515, "grad_norm": 0.09289347063379441, "learning_rate": 8.922378156817957e-07, "loss": 0.3829, "step": 202 }, { "epoch": 4.720930232558139, "grad_norm": 0.09257169253912409, "learning_rate": 7.606700420914381e-07, "loss": 0.3821, "step": 203 }, { "epoch": 4.7441860465116275, "grad_norm": 0.09310627098479367, "learning_rate": 6.39499007005746e-07, "loss": 0.3903, "step": 204 }, { "epoch": 4.767441860465116, "grad_norm": 0.08944939258724487, "learning_rate": 5.287568155490341e-07, "loss": 0.3885, "step": 205 }, { "epoch": 4.790697674418604, "grad_norm": 0.09592002472628403, "learning_rate": 4.284728096496027e-07, "loss": 0.3875, "step": 206 }, { "epoch": 4.813953488372093, "grad_norm": 0.09180161205563094, "learning_rate": 3.3867356026537457e-07, "loss": 0.3855, "step": 207 }, { "epoch": 4.837209302325581, "grad_norm": 0.08528621258416194, "learning_rate": 2.593828603437487e-07, "loss": 0.3846, "step": 208 }, { "epoch": 4.8604651162790695, "grad_norm": 0.08815943836016799, "learning_rate": 1.906217185174919e-07, "loss": 0.3839, "step": 209 }, { "epoch": 4.883720930232558, "grad_norm": 0.08604030139119473, "learning_rate": 1.3240835353834247e-07, "loss": 0.3892, "step": 210 }, { "epoch": 4.906976744186046, "grad_norm": 0.0865890927390752, "learning_rate": 8.475818944982728e-08, "loss": 0.3835, "step": 211 }, { "epoch": 4.930232558139535, "grad_norm": 0.09058754572938608, "learning_rate": 4.7683851500548437e-08, "loss": 0.3788, "step": 212 }, { "epoch": 4.953488372093023, "grad_norm": 0.08787858276738178, "learning_rate": 2.1195162799032554e-08, "loss": 0.3903, "step": 213 }, { "epoch": 4.976744186046512, "grad_norm": 0.08813542305445697, "learning_rate": 5.299141711043732e-09, "loss": 0.3865, "step": 214 }, { "epoch": 5.0, "grad_norm": 0.09709093222975328, "learning_rate": 0.0, "loss": 0.379, "step": 215 }, { "epoch": 5.0, "step": 215, "total_flos": 3606870585507840.0, "train_loss": 0.5023871626964835, "train_runtime": 3261.3525, "train_samples_per_second": 33.201, "train_steps_per_second": 0.066 } ], "logging_steps": 1, "max_steps": 215, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3606870585507840.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }