diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,5222 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 5.0, + "eval_steps": 500, + "global_step": 740, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.006756756756756757, + "grad_norm": 3.1352566768118555, + "learning_rate": 1.0810810810810812e-06, + "loss": 1.0885, + "step": 1 + }, + { + "epoch": 0.013513513513513514, + "grad_norm": 3.0707759310558114, + "learning_rate": 2.1621621621621623e-06, + "loss": 1.081, + "step": 2 + }, + { + "epoch": 0.02027027027027027, + "grad_norm": 3.1041877911160487, + "learning_rate": 3.2432432432432437e-06, + "loss": 1.0779, + "step": 3 + }, + { + "epoch": 0.02702702702702703, + "grad_norm": 2.894341433491889, + "learning_rate": 4.324324324324325e-06, + "loss": 1.066, + "step": 4 + }, + { + "epoch": 0.033783783783783786, + "grad_norm": 2.4474574648074654, + "learning_rate": 5.405405405405406e-06, + "loss": 1.0464, + "step": 5 + }, + { + "epoch": 0.04054054054054054, + "grad_norm": 1.7416847938696791, + "learning_rate": 6.486486486486487e-06, + "loss": 1.0062, + "step": 6 + }, + { + "epoch": 0.0472972972972973, + "grad_norm": 1.737012288070095, + "learning_rate": 7.567567567567569e-06, + "loss": 0.9661, + "step": 7 + }, + { + "epoch": 0.05405405405405406, + "grad_norm": 1.54166088046008, + "learning_rate": 8.64864864864865e-06, + "loss": 0.9625, + "step": 8 + }, + { + "epoch": 0.060810810810810814, + "grad_norm": 1.1305140237848132, + "learning_rate": 9.729729729729732e-06, + "loss": 0.9355, + "step": 9 + }, + { + "epoch": 0.06756756756756757, + "grad_norm": 1.6829884643585495, + "learning_rate": 1.0810810810810812e-05, + "loss": 0.8814, + "step": 10 + }, + { + "epoch": 0.07432432432432433, + "grad_norm": 1.2528889517867892, + "learning_rate": 1.1891891891891894e-05, + "loss": 0.8733, + "step": 11 + }, + { + "epoch": 0.08108108108108109, + "grad_norm": 0.9520381902650208, + "learning_rate": 1.2972972972972975e-05, + "loss": 0.8579, + "step": 12 + }, + { + "epoch": 0.08783783783783784, + "grad_norm": 1.4164792552755203, + "learning_rate": 1.4054054054054055e-05, + "loss": 0.82, + "step": 13 + }, + { + "epoch": 0.0945945945945946, + "grad_norm": 1.1661768167320599, + "learning_rate": 1.5135135135135138e-05, + "loss": 0.823, + "step": 14 + }, + { + "epoch": 0.10135135135135136, + "grad_norm": 0.966545525495581, + "learning_rate": 1.6216216216216218e-05, + "loss": 0.8177, + "step": 15 + }, + { + "epoch": 0.10810810810810811, + "grad_norm": 0.9970915856247566, + "learning_rate": 1.72972972972973e-05, + "loss": 0.7828, + "step": 16 + }, + { + "epoch": 0.11486486486486487, + "grad_norm": 0.7510807795940354, + "learning_rate": 1.8378378378378383e-05, + "loss": 0.7752, + "step": 17 + }, + { + "epoch": 0.12162162162162163, + "grad_norm": 0.6801364883047033, + "learning_rate": 1.9459459459459463e-05, + "loss": 0.7802, + "step": 18 + }, + { + "epoch": 0.12837837837837837, + "grad_norm": 0.8052966224978529, + "learning_rate": 2.054054054054054e-05, + "loss": 0.7803, + "step": 19 + }, + { + "epoch": 0.13513513513513514, + "grad_norm": 0.6731724951990521, + "learning_rate": 2.1621621621621624e-05, + "loss": 0.774, + "step": 20 + }, + { + "epoch": 0.14189189189189189, + "grad_norm": 0.5240064900372621, + "learning_rate": 2.2702702702702705e-05, + "loss": 0.7503, + "step": 21 + }, + { + "epoch": 0.14864864864864866, + "grad_norm": 0.5238422824665931, + "learning_rate": 2.378378378378379e-05, + "loss": 0.7452, + "step": 22 + }, + { + "epoch": 0.1554054054054054, + "grad_norm": 0.5685661395814289, + "learning_rate": 2.4864864864864866e-05, + "loss": 0.7357, + "step": 23 + }, + { + "epoch": 0.16216216216216217, + "grad_norm": 0.5559414871229672, + "learning_rate": 2.594594594594595e-05, + "loss": 0.7529, + "step": 24 + }, + { + "epoch": 0.16891891891891891, + "grad_norm": 0.44354381932505604, + "learning_rate": 2.702702702702703e-05, + "loss": 0.7531, + "step": 25 + }, + { + "epoch": 0.17567567567567569, + "grad_norm": 0.5353496542990213, + "learning_rate": 2.810810810810811e-05, + "loss": 0.7443, + "step": 26 + }, + { + "epoch": 0.18243243243243243, + "grad_norm": 0.5011572381597165, + "learning_rate": 2.918918918918919e-05, + "loss": 0.7309, + "step": 27 + }, + { + "epoch": 0.1891891891891892, + "grad_norm": 0.46554848262029747, + "learning_rate": 3.0270270270270275e-05, + "loss": 0.73, + "step": 28 + }, + { + "epoch": 0.19594594594594594, + "grad_norm": 0.5645942632228904, + "learning_rate": 3.135135135135135e-05, + "loss": 0.7218, + "step": 29 + }, + { + "epoch": 0.20270270270270271, + "grad_norm": 0.48122141248903777, + "learning_rate": 3.2432432432432436e-05, + "loss": 0.7272, + "step": 30 + }, + { + "epoch": 0.20945945945945946, + "grad_norm": 0.4302278552572059, + "learning_rate": 3.351351351351351e-05, + "loss": 0.7138, + "step": 31 + }, + { + "epoch": 0.21621621621621623, + "grad_norm": 0.38359657484495446, + "learning_rate": 3.45945945945946e-05, + "loss": 0.716, + "step": 32 + }, + { + "epoch": 0.22297297297297297, + "grad_norm": 0.37734924853300067, + "learning_rate": 3.567567567567568e-05, + "loss": 0.7258, + "step": 33 + }, + { + "epoch": 0.22972972972972974, + "grad_norm": 0.5860873173612458, + "learning_rate": 3.6756756756756765e-05, + "loss": 0.7146, + "step": 34 + }, + { + "epoch": 0.23648648648648649, + "grad_norm": 0.7521759711776034, + "learning_rate": 3.783783783783784e-05, + "loss": 0.727, + "step": 35 + }, + { + "epoch": 0.24324324324324326, + "grad_norm": 0.46287517636393716, + "learning_rate": 3.8918918918918926e-05, + "loss": 0.6978, + "step": 36 + }, + { + "epoch": 0.25, + "grad_norm": 0.3949073220604405, + "learning_rate": 4e-05, + "loss": 0.7143, + "step": 37 + }, + { + "epoch": 0.25675675675675674, + "grad_norm": 0.5390567372094507, + "learning_rate": 4.108108108108108e-05, + "loss": 0.7147, + "step": 38 + }, + { + "epoch": 0.2635135135135135, + "grad_norm": 0.5260020316144047, + "learning_rate": 4.2162162162162164e-05, + "loss": 0.7042, + "step": 39 + }, + { + "epoch": 0.2702702702702703, + "grad_norm": 0.6245421084330798, + "learning_rate": 4.324324324324325e-05, + "loss": 0.7066, + "step": 40 + }, + { + "epoch": 0.27702702702702703, + "grad_norm": 0.4274410936399283, + "learning_rate": 4.4324324324324325e-05, + "loss": 0.7002, + "step": 41 + }, + { + "epoch": 0.28378378378378377, + "grad_norm": 0.5645116869141522, + "learning_rate": 4.540540540540541e-05, + "loss": 0.7017, + "step": 42 + }, + { + "epoch": 0.2905405405405405, + "grad_norm": 0.8531985478315304, + "learning_rate": 4.6486486486486486e-05, + "loss": 0.7105, + "step": 43 + }, + { + "epoch": 0.2972972972972973, + "grad_norm": 0.8463718591267281, + "learning_rate": 4.756756756756758e-05, + "loss": 0.6926, + "step": 44 + }, + { + "epoch": 0.30405405405405406, + "grad_norm": 0.527363152011757, + "learning_rate": 4.8648648648648654e-05, + "loss": 0.698, + "step": 45 + }, + { + "epoch": 0.3108108108108108, + "grad_norm": 0.8921529578247511, + "learning_rate": 4.972972972972973e-05, + "loss": 0.7113, + "step": 46 + }, + { + "epoch": 0.31756756756756754, + "grad_norm": 0.844585341198514, + "learning_rate": 5.081081081081081e-05, + "loss": 0.6918, + "step": 47 + }, + { + "epoch": 0.32432432432432434, + "grad_norm": 0.6610922858373763, + "learning_rate": 5.18918918918919e-05, + "loss": 0.6874, + "step": 48 + }, + { + "epoch": 0.3310810810810811, + "grad_norm": 0.9289696025390548, + "learning_rate": 5.2972972972972976e-05, + "loss": 0.6965, + "step": 49 + }, + { + "epoch": 0.33783783783783783, + "grad_norm": 0.7104021291109865, + "learning_rate": 5.405405405405406e-05, + "loss": 0.7025, + "step": 50 + }, + { + "epoch": 0.34459459459459457, + "grad_norm": 0.7179610269369078, + "learning_rate": 5.513513513513514e-05, + "loss": 0.6906, + "step": 51 + }, + { + "epoch": 0.35135135135135137, + "grad_norm": 0.6612006329248038, + "learning_rate": 5.621621621621622e-05, + "loss": 0.7026, + "step": 52 + }, + { + "epoch": 0.3581081081081081, + "grad_norm": 0.6920205226216654, + "learning_rate": 5.7297297297297305e-05, + "loss": 0.6849, + "step": 53 + }, + { + "epoch": 0.36486486486486486, + "grad_norm": 0.5619919459693192, + "learning_rate": 5.837837837837838e-05, + "loss": 0.6827, + "step": 54 + }, + { + "epoch": 0.3716216216216216, + "grad_norm": 0.789835519140916, + "learning_rate": 5.945945945945946e-05, + "loss": 0.6753, + "step": 55 + }, + { + "epoch": 0.3783783783783784, + "grad_norm": 0.8468946375056802, + "learning_rate": 6.054054054054055e-05, + "loss": 0.6803, + "step": 56 + }, + { + "epoch": 0.38513513513513514, + "grad_norm": 1.1480634289873959, + "learning_rate": 6.162162162162163e-05, + "loss": 0.6849, + "step": 57 + }, + { + "epoch": 0.3918918918918919, + "grad_norm": 0.8338876354018716, + "learning_rate": 6.27027027027027e-05, + "loss": 0.6822, + "step": 58 + }, + { + "epoch": 0.39864864864864863, + "grad_norm": 0.8619635833656949, + "learning_rate": 6.378378378378379e-05, + "loss": 0.6885, + "step": 59 + }, + { + "epoch": 0.40540540540540543, + "grad_norm": 0.9841339840879071, + "learning_rate": 6.486486486486487e-05, + "loss": 0.6941, + "step": 60 + }, + { + "epoch": 0.41216216216216217, + "grad_norm": 0.7846924650076373, + "learning_rate": 6.594594594594596e-05, + "loss": 0.6734, + "step": 61 + }, + { + "epoch": 0.4189189189189189, + "grad_norm": 0.6084845178399872, + "learning_rate": 6.702702702702703e-05, + "loss": 0.6766, + "step": 62 + }, + { + "epoch": 0.42567567567567566, + "grad_norm": 0.8121605538110969, + "learning_rate": 6.810810810810811e-05, + "loss": 0.6882, + "step": 63 + }, + { + "epoch": 0.43243243243243246, + "grad_norm": 0.9962451486752875, + "learning_rate": 6.91891891891892e-05, + "loss": 0.6837, + "step": 64 + }, + { + "epoch": 0.4391891891891892, + "grad_norm": 0.6516040658377318, + "learning_rate": 7.027027027027028e-05, + "loss": 0.6841, + "step": 65 + }, + { + "epoch": 0.44594594594594594, + "grad_norm": 0.683694787120212, + "learning_rate": 7.135135135135136e-05, + "loss": 0.6952, + "step": 66 + }, + { + "epoch": 0.4527027027027027, + "grad_norm": 0.6582713522549495, + "learning_rate": 7.243243243243243e-05, + "loss": 0.6727, + "step": 67 + }, + { + "epoch": 0.4594594594594595, + "grad_norm": 0.673908455931949, + "learning_rate": 7.351351351351353e-05, + "loss": 0.6841, + "step": 68 + }, + { + "epoch": 0.46621621621621623, + "grad_norm": 0.8017390116132226, + "learning_rate": 7.45945945945946e-05, + "loss": 0.67, + "step": 69 + }, + { + "epoch": 0.47297297297297297, + "grad_norm": 0.9799335757769192, + "learning_rate": 7.567567567567568e-05, + "loss": 0.6914, + "step": 70 + }, + { + "epoch": 0.4797297297297297, + "grad_norm": 1.5088822309237815, + "learning_rate": 7.675675675675675e-05, + "loss": 0.6925, + "step": 71 + }, + { + "epoch": 0.4864864864864865, + "grad_norm": 0.7235799329993131, + "learning_rate": 7.783783783783785e-05, + "loss": 0.689, + "step": 72 + }, + { + "epoch": 0.49324324324324326, + "grad_norm": 1.467550439828714, + "learning_rate": 7.891891891891892e-05, + "loss": 0.6927, + "step": 73 + }, + { + "epoch": 0.5, + "grad_norm": 0.9680632895596434, + "learning_rate": 8e-05, + "loss": 0.6729, + "step": 74 + }, + { + "epoch": 0.5067567567567568, + "grad_norm": 1.4048772257060618, + "learning_rate": 7.999955497902857e-05, + "loss": 0.6949, + "step": 75 + }, + { + "epoch": 0.5135135135135135, + "grad_norm": 1.187179427424338, + "learning_rate": 7.999821992601645e-05, + "loss": 0.693, + "step": 76 + }, + { + "epoch": 0.5202702702702703, + "grad_norm": 0.9529005283405405, + "learning_rate": 7.999599487066996e-05, + "loss": 0.6839, + "step": 77 + }, + { + "epoch": 0.527027027027027, + "grad_norm": 0.9173343088851188, + "learning_rate": 7.999287986249894e-05, + "loss": 0.6766, + "step": 78 + }, + { + "epoch": 0.5337837837837838, + "grad_norm": 0.7318824071301274, + "learning_rate": 7.998887497081555e-05, + "loss": 0.6847, + "step": 79 + }, + { + "epoch": 0.5405405405405406, + "grad_norm": 0.6529765700425837, + "learning_rate": 7.998398028473287e-05, + "loss": 0.678, + "step": 80 + }, + { + "epoch": 0.5472972972972973, + "grad_norm": 0.6928101408737816, + "learning_rate": 7.997819591316278e-05, + "loss": 0.6954, + "step": 81 + }, + { + "epoch": 0.5540540540540541, + "grad_norm": 0.5879827240121825, + "learning_rate": 7.99715219848136e-05, + "loss": 0.6642, + "step": 82 + }, + { + "epoch": 0.5608108108108109, + "grad_norm": 0.680605729950147, + "learning_rate": 7.996395864818727e-05, + "loss": 0.6804, + "step": 83 + }, + { + "epoch": 0.5675675675675675, + "grad_norm": 0.5423663061022331, + "learning_rate": 7.995550607157592e-05, + "loss": 0.6688, + "step": 84 + }, + { + "epoch": 0.5743243243243243, + "grad_norm": 0.606163015240631, + "learning_rate": 7.994616444305826e-05, + "loss": 0.6667, + "step": 85 + }, + { + "epoch": 0.581081081081081, + "grad_norm": 0.5950010094955055, + "learning_rate": 7.993593397049533e-05, + "loss": 0.679, + "step": 86 + }, + { + "epoch": 0.5878378378378378, + "grad_norm": 0.6429149705782599, + "learning_rate": 7.992481488152585e-05, + "loss": 0.6685, + "step": 87 + }, + { + "epoch": 0.5945945945945946, + "grad_norm": 0.46140817883542623, + "learning_rate": 7.991280742356124e-05, + "loss": 0.6603, + "step": 88 + }, + { + "epoch": 0.6013513513513513, + "grad_norm": 0.6795669467936494, + "learning_rate": 7.989991186378e-05, + "loss": 0.6699, + "step": 89 + }, + { + "epoch": 0.6081081081081081, + "grad_norm": 0.5738680266825441, + "learning_rate": 7.988612848912186e-05, + "loss": 0.6819, + "step": 90 + }, + { + "epoch": 0.6148648648648649, + "grad_norm": 0.5789217573047567, + "learning_rate": 7.987145760628138e-05, + "loss": 0.6708, + "step": 91 + }, + { + "epoch": 0.6216216216216216, + "grad_norm": 0.49449419241667797, + "learning_rate": 7.985589954170107e-05, + "loss": 0.6665, + "step": 92 + }, + { + "epoch": 0.6283783783783784, + "grad_norm": 0.4987045947630836, + "learning_rate": 7.983945464156419e-05, + "loss": 0.6597, + "step": 93 + }, + { + "epoch": 0.6351351351351351, + "grad_norm": 0.40082592041703324, + "learning_rate": 7.982212327178699e-05, + "loss": 0.6675, + "step": 94 + }, + { + "epoch": 0.6418918918918919, + "grad_norm": 0.3431463481882859, + "learning_rate": 7.980390581801064e-05, + "loss": 0.6616, + "step": 95 + }, + { + "epoch": 0.6486486486486487, + "grad_norm": 0.44148520818681297, + "learning_rate": 7.97848026855926e-05, + "loss": 0.6761, + "step": 96 + }, + { + "epoch": 0.6554054054054054, + "grad_norm": 0.4038247824809848, + "learning_rate": 7.976481429959758e-05, + "loss": 0.6596, + "step": 97 + }, + { + "epoch": 0.6621621621621622, + "grad_norm": 0.3448527603506924, + "learning_rate": 7.974394110478813e-05, + "loss": 0.6536, + "step": 98 + }, + { + "epoch": 0.668918918918919, + "grad_norm": 0.30698753802117784, + "learning_rate": 7.972218356561471e-05, + "loss": 0.6772, + "step": 99 + }, + { + "epoch": 0.6756756756756757, + "grad_norm": 0.30195068085071514, + "learning_rate": 7.96995421662054e-05, + "loss": 0.6527, + "step": 100 + }, + { + "epoch": 0.6824324324324325, + "grad_norm": 0.3166797371131855, + "learning_rate": 7.967601741035507e-05, + "loss": 0.6555, + "step": 101 + }, + { + "epoch": 0.6891891891891891, + "grad_norm": 0.32820963864069536, + "learning_rate": 7.965160982151422e-05, + "loss": 0.6696, + "step": 102 + }, + { + "epoch": 0.6959459459459459, + "grad_norm": 0.3483528954236439, + "learning_rate": 7.962631994277728e-05, + "loss": 0.663, + "step": 103 + }, + { + "epoch": 0.7027027027027027, + "grad_norm": 0.4983197508582807, + "learning_rate": 7.960014833687055e-05, + "loss": 0.6633, + "step": 104 + }, + { + "epoch": 0.7094594594594594, + "grad_norm": 0.7805957154226273, + "learning_rate": 7.957309558613974e-05, + "loss": 0.6587, + "step": 105 + }, + { + "epoch": 0.7162162162162162, + "grad_norm": 1.106633618810969, + "learning_rate": 7.954516229253691e-05, + "loss": 0.662, + "step": 106 + }, + { + "epoch": 0.722972972972973, + "grad_norm": 0.757015003567712, + "learning_rate": 7.951634907760713e-05, + "loss": 0.6593, + "step": 107 + }, + { + "epoch": 0.7297297297297297, + "grad_norm": 0.8427739970040843, + "learning_rate": 7.948665658247463e-05, + "loss": 0.6601, + "step": 108 + }, + { + "epoch": 0.7364864864864865, + "grad_norm": 0.7431541528418906, + "learning_rate": 7.945608546782858e-05, + "loss": 0.667, + "step": 109 + }, + { + "epoch": 0.7432432432432432, + "grad_norm": 0.6845263081020915, + "learning_rate": 7.942463641390834e-05, + "loss": 0.6569, + "step": 110 + }, + { + "epoch": 0.75, + "grad_norm": 0.8712514657793697, + "learning_rate": 7.939231012048833e-05, + "loss": 0.6632, + "step": 111 + }, + { + "epoch": 0.7567567567567568, + "grad_norm": 0.6774356142156878, + "learning_rate": 7.935910730686246e-05, + "loss": 0.6706, + "step": 112 + }, + { + "epoch": 0.7635135135135135, + "grad_norm": 0.7702128806997793, + "learning_rate": 7.932502871182818e-05, + "loss": 0.6679, + "step": 113 + }, + { + "epoch": 0.7702702702702703, + "grad_norm": 0.6032796273856512, + "learning_rate": 7.929007509366994e-05, + "loss": 0.6612, + "step": 114 + }, + { + "epoch": 0.777027027027027, + "grad_norm": 0.7262342972595842, + "learning_rate": 7.925424723014239e-05, + "loss": 0.6639, + "step": 115 + }, + { + "epoch": 0.7837837837837838, + "grad_norm": 0.6319433116490037, + "learning_rate": 7.921754591845307e-05, + "loss": 0.662, + "step": 116 + }, + { + "epoch": 0.7905405405405406, + "grad_norm": 0.6175612756087657, + "learning_rate": 7.917997197524467e-05, + "loss": 0.6566, + "step": 117 + }, + { + "epoch": 0.7972972972972973, + "grad_norm": 0.7923618524838222, + "learning_rate": 7.914152623657678e-05, + "loss": 0.6469, + "step": 118 + }, + { + "epoch": 0.8040540540540541, + "grad_norm": 0.7707195088131242, + "learning_rate": 7.910220955790746e-05, + "loss": 0.6563, + "step": 119 + }, + { + "epoch": 0.8108108108108109, + "grad_norm": 0.4152288654221399, + "learning_rate": 7.906202281407398e-05, + "loss": 0.6519, + "step": 120 + }, + { + "epoch": 0.8175675675675675, + "grad_norm": 0.5439040195272984, + "learning_rate": 7.902096689927355e-05, + "loss": 0.6727, + "step": 121 + }, + { + "epoch": 0.8243243243243243, + "grad_norm": 0.5159119091886112, + "learning_rate": 7.897904272704333e-05, + "loss": 0.6563, + "step": 122 + }, + { + "epoch": 0.831081081081081, + "grad_norm": 0.49623067194442205, + "learning_rate": 7.893625123024011e-05, + "loss": 0.6523, + "step": 123 + }, + { + "epoch": 0.8378378378378378, + "grad_norm": 0.4176229579696035, + "learning_rate": 7.889259336101957e-05, + "loss": 0.6461, + "step": 124 + }, + { + "epoch": 0.8445945945945946, + "grad_norm": 0.5424146541267013, + "learning_rate": 7.884807009081506e-05, + "loss": 0.6535, + "step": 125 + }, + { + "epoch": 0.8513513513513513, + "grad_norm": 0.53245673115827, + "learning_rate": 7.880268241031604e-05, + "loss": 0.6516, + "step": 126 + }, + { + "epoch": 0.8581081081081081, + "grad_norm": 0.4218647663890849, + "learning_rate": 7.875643132944599e-05, + "loss": 0.6557, + "step": 127 + }, + { + "epoch": 0.8648648648648649, + "grad_norm": 0.31388453196671723, + "learning_rate": 7.870931787733996e-05, + "loss": 0.6504, + "step": 128 + }, + { + "epoch": 0.8716216216216216, + "grad_norm": 0.42120451899053063, + "learning_rate": 7.866134310232167e-05, + "loss": 0.6637, + "step": 129 + }, + { + "epoch": 0.8783783783783784, + "grad_norm": 0.3569049182508329, + "learning_rate": 7.861250807188014e-05, + "loss": 0.6565, + "step": 130 + }, + { + "epoch": 0.8851351351351351, + "grad_norm": 0.46980061573845006, + "learning_rate": 7.856281387264603e-05, + "loss": 0.6643, + "step": 131 + }, + { + "epoch": 0.8918918918918919, + "grad_norm": 0.5602330500767254, + "learning_rate": 7.851226161036739e-05, + "loss": 0.6541, + "step": 132 + }, + { + "epoch": 0.8986486486486487, + "grad_norm": 0.555307896703473, + "learning_rate": 7.846085240988503e-05, + "loss": 0.6498, + "step": 133 + }, + { + "epoch": 0.9054054054054054, + "grad_norm": 0.4906170522953401, + "learning_rate": 7.840858741510758e-05, + "loss": 0.6482, + "step": 134 + }, + { + "epoch": 0.9121621621621622, + "grad_norm": 0.4486181714024936, + "learning_rate": 7.835546778898599e-05, + "loss": 0.6457, + "step": 135 + }, + { + "epoch": 0.918918918918919, + "grad_norm": 0.5704154766929971, + "learning_rate": 7.830149471348763e-05, + "loss": 0.6508, + "step": 136 + }, + { + "epoch": 0.9256756756756757, + "grad_norm": 0.738054877738817, + "learning_rate": 7.824666938957004e-05, + "loss": 0.6531, + "step": 137 + }, + { + "epoch": 0.9324324324324325, + "grad_norm": 0.7131460924145132, + "learning_rate": 7.819099303715414e-05, + "loss": 0.6577, + "step": 138 + }, + { + "epoch": 0.9391891891891891, + "grad_norm": 0.5314442328007468, + "learning_rate": 7.813446689509718e-05, + "loss": 0.6472, + "step": 139 + }, + { + "epoch": 0.9459459459459459, + "grad_norm": 0.4354387291682998, + "learning_rate": 7.807709222116506e-05, + "loss": 0.6566, + "step": 140 + }, + { + "epoch": 0.9527027027027027, + "grad_norm": 0.4367558414891408, + "learning_rate": 7.801887029200448e-05, + "loss": 0.6548, + "step": 141 + }, + { + "epoch": 0.9594594594594594, + "grad_norm": 0.49114789228090255, + "learning_rate": 7.795980240311436e-05, + "loss": 0.6651, + "step": 142 + }, + { + "epoch": 0.9662162162162162, + "grad_norm": 0.5563118895714915, + "learning_rate": 7.789988986881719e-05, + "loss": 0.6534, + "step": 143 + }, + { + "epoch": 0.972972972972973, + "grad_norm": 0.5068498475757595, + "learning_rate": 7.78391340222297e-05, + "loss": 0.6635, + "step": 144 + }, + { + "epoch": 0.9797297297297297, + "grad_norm": 0.40377422580342515, + "learning_rate": 7.777753621523316e-05, + "loss": 0.662, + "step": 145 + }, + { + "epoch": 0.9864864864864865, + "grad_norm": 0.48943576691657636, + "learning_rate": 7.771509781844338e-05, + "loss": 0.6465, + "step": 146 + }, + { + "epoch": 0.9932432432432432, + "grad_norm": 0.45222666736741574, + "learning_rate": 7.765182022118014e-05, + "loss": 0.6576, + "step": 147 + }, + { + "epoch": 1.0, + "grad_norm": 0.3524115634430567, + "learning_rate": 7.758770483143634e-05, + "loss": 0.66, + "step": 148 + }, + { + "epoch": 1.0067567567567568, + "grad_norm": 0.3465470718278503, + "learning_rate": 7.752275307584664e-05, + "loss": 0.6528, + "step": 149 + }, + { + "epoch": 1.0135135135135136, + "grad_norm": 0.4289730330622286, + "learning_rate": 7.745696639965569e-05, + "loss": 0.6284, + "step": 150 + }, + { + "epoch": 1.0202702702702702, + "grad_norm": 0.4078417015334056, + "learning_rate": 7.739034626668605e-05, + "loss": 0.6409, + "step": 151 + }, + { + "epoch": 1.027027027027027, + "grad_norm": 0.4224099719890233, + "learning_rate": 7.732289415930549e-05, + "loss": 0.63, + "step": 152 + }, + { + "epoch": 1.0337837837837838, + "grad_norm": 0.43135712270900994, + "learning_rate": 7.725461157839417e-05, + "loss": 0.6606, + "step": 153 + }, + { + "epoch": 1.0405405405405406, + "grad_norm": 0.4832560652031643, + "learning_rate": 7.71855000433111e-05, + "loss": 0.6439, + "step": 154 + }, + { + "epoch": 1.0472972972972974, + "grad_norm": 0.5148651119907238, + "learning_rate": 7.711556109186039e-05, + "loss": 0.6405, + "step": 155 + }, + { + "epoch": 1.054054054054054, + "grad_norm": 0.49381906164540984, + "learning_rate": 7.704479628025704e-05, + "loss": 0.635, + "step": 156 + }, + { + "epoch": 1.0608108108108107, + "grad_norm": 0.4238088291952583, + "learning_rate": 7.697320718309235e-05, + "loss": 0.6363, + "step": 157 + }, + { + "epoch": 1.0675675675675675, + "grad_norm": 0.34639148621347143, + "learning_rate": 7.690079539329875e-05, + "loss": 0.6434, + "step": 158 + }, + { + "epoch": 1.0743243243243243, + "grad_norm": 0.3670601156648982, + "learning_rate": 7.682756252211453e-05, + "loss": 0.6321, + "step": 159 + }, + { + "epoch": 1.0810810810810811, + "grad_norm": 0.48931637412654205, + "learning_rate": 7.675351019904785e-05, + "loss": 0.6371, + "step": 160 + }, + { + "epoch": 1.087837837837838, + "grad_norm": 0.5514281223158661, + "learning_rate": 7.667864007184054e-05, + "loss": 0.6313, + "step": 161 + }, + { + "epoch": 1.0945945945945945, + "grad_norm": 0.48466826872522856, + "learning_rate": 7.660295380643144e-05, + "loss": 0.6358, + "step": 162 + }, + { + "epoch": 1.1013513513513513, + "grad_norm": 0.46142889218279065, + "learning_rate": 7.652645308691933e-05, + "loss": 0.6418, + "step": 163 + }, + { + "epoch": 1.1081081081081081, + "grad_norm": 0.4371908363892829, + "learning_rate": 7.644913961552544e-05, + "loss": 0.6327, + "step": 164 + }, + { + "epoch": 1.114864864864865, + "grad_norm": 0.4159116225859721, + "learning_rate": 7.637101511255554e-05, + "loss": 0.6332, + "step": 165 + }, + { + "epoch": 1.1216216216216217, + "grad_norm": 0.48524137415448176, + "learning_rate": 7.629208131636179e-05, + "loss": 0.6385, + "step": 166 + }, + { + "epoch": 1.1283783783783783, + "grad_norm": 0.502011807585537, + "learning_rate": 7.621233998330387e-05, + "loss": 0.6534, + "step": 167 + }, + { + "epoch": 1.135135135135135, + "grad_norm": 0.4849012769427631, + "learning_rate": 7.61317928877101e-05, + "loss": 0.6382, + "step": 168 + }, + { + "epoch": 1.1418918918918919, + "grad_norm": 0.40497697606883803, + "learning_rate": 7.605044182183779e-05, + "loss": 0.6335, + "step": 169 + }, + { + "epoch": 1.1486486486486487, + "grad_norm": 0.26814241993209303, + "learning_rate": 7.596828859583347e-05, + "loss": 0.6226, + "step": 170 + }, + { + "epoch": 1.1554054054054055, + "grad_norm": 0.36135092127292284, + "learning_rate": 7.588533503769257e-05, + "loss": 0.6286, + "step": 171 + }, + { + "epoch": 1.1621621621621623, + "grad_norm": 0.4608397427011372, + "learning_rate": 7.580158299321872e-05, + "loss": 0.6438, + "step": 172 + }, + { + "epoch": 1.1689189189189189, + "grad_norm": 0.4171913645594413, + "learning_rate": 7.571703432598275e-05, + "loss": 0.6356, + "step": 173 + }, + { + "epoch": 1.1756756756756757, + "grad_norm": 0.4370902558666279, + "learning_rate": 7.563169091728115e-05, + "loss": 0.6422, + "step": 174 + }, + { + "epoch": 1.1824324324324325, + "grad_norm": 0.4283980244016275, + "learning_rate": 7.554555466609425e-05, + "loss": 0.6407, + "step": 175 + }, + { + "epoch": 1.1891891891891893, + "grad_norm": 0.4381556639196284, + "learning_rate": 7.545862748904394e-05, + "loss": 0.6434, + "step": 176 + }, + { + "epoch": 1.195945945945946, + "grad_norm": 0.44668924620805633, + "learning_rate": 7.537091132035111e-05, + "loss": 0.6219, + "step": 177 + }, + { + "epoch": 1.2027027027027026, + "grad_norm": 0.37602920076180757, + "learning_rate": 7.528240811179245e-05, + "loss": 0.6419, + "step": 178 + }, + { + "epoch": 1.2094594594594594, + "grad_norm": 0.2672376131285308, + "learning_rate": 7.519311983265718e-05, + "loss": 0.6366, + "step": 179 + }, + { + "epoch": 1.2162162162162162, + "grad_norm": 0.29369317421661273, + "learning_rate": 7.510304846970311e-05, + "loss": 0.6341, + "step": 180 + }, + { + "epoch": 1.222972972972973, + "grad_norm": 0.3602685800886801, + "learning_rate": 7.501219602711253e-05, + "loss": 0.6432, + "step": 181 + }, + { + "epoch": 1.2297297297297298, + "grad_norm": 0.3237536586353385, + "learning_rate": 7.492056452644753e-05, + "loss": 0.6415, + "step": 182 + }, + { + "epoch": 1.2364864864864864, + "grad_norm": 0.28545557770169117, + "learning_rate": 7.48281560066051e-05, + "loss": 0.6364, + "step": 183 + }, + { + "epoch": 1.2432432432432432, + "grad_norm": 0.2623342902523158, + "learning_rate": 7.473497252377171e-05, + "loss": 0.6333, + "step": 184 + }, + { + "epoch": 1.25, + "grad_norm": 0.3160766346456171, + "learning_rate": 7.464101615137756e-05, + "loss": 0.6256, + "step": 185 + }, + { + "epoch": 1.2567567567567568, + "grad_norm": 0.4173934303596852, + "learning_rate": 7.454628898005043e-05, + "loss": 0.6521, + "step": 186 + }, + { + "epoch": 1.2635135135135136, + "grad_norm": 0.4424346331081593, + "learning_rate": 7.445079311756924e-05, + "loss": 0.6541, + "step": 187 + }, + { + "epoch": 1.2702702702702702, + "grad_norm": 0.38764273759912543, + "learning_rate": 7.435453068881706e-05, + "loss": 0.6385, + "step": 188 + }, + { + "epoch": 1.277027027027027, + "grad_norm": 0.3811463582022792, + "learning_rate": 7.425750383573384e-05, + "loss": 0.6295, + "step": 189 + }, + { + "epoch": 1.2837837837837838, + "grad_norm": 0.5272316642769805, + "learning_rate": 7.415971471726884e-05, + "loss": 0.6307, + "step": 190 + }, + { + "epoch": 1.2905405405405406, + "grad_norm": 0.7635509141776905, + "learning_rate": 7.406116550933246e-05, + "loss": 0.6313, + "step": 191 + }, + { + "epoch": 1.2972972972972974, + "grad_norm": 0.9673085818387788, + "learning_rate": 7.396185840474792e-05, + "loss": 0.6495, + "step": 192 + }, + { + "epoch": 1.304054054054054, + "grad_norm": 0.8691278543652353, + "learning_rate": 7.386179561320243e-05, + "loss": 0.6285, + "step": 193 + }, + { + "epoch": 1.3108108108108107, + "grad_norm": 0.7233647856746338, + "learning_rate": 7.376097936119803e-05, + "loss": 0.6435, + "step": 194 + }, + { + "epoch": 1.3175675675675675, + "grad_norm": 0.4888365541378487, + "learning_rate": 7.365941189200201e-05, + "loss": 0.6345, + "step": 195 + }, + { + "epoch": 1.3243243243243243, + "grad_norm": 0.41440516733521193, + "learning_rate": 7.35570954655971e-05, + "loss": 0.6522, + "step": 196 + }, + { + "epoch": 1.3310810810810811, + "grad_norm": 0.39733170188465466, + "learning_rate": 7.345403235863105e-05, + "loss": 0.636, + "step": 197 + }, + { + "epoch": 1.3378378378378377, + "grad_norm": 0.37538788641282456, + "learning_rate": 7.335022486436608e-05, + "loss": 0.6405, + "step": 198 + }, + { + "epoch": 1.3445945945945945, + "grad_norm": 0.4912257786205075, + "learning_rate": 7.324567529262775e-05, + "loss": 0.6341, + "step": 199 + }, + { + "epoch": 1.3513513513513513, + "grad_norm": 0.3964231351313485, + "learning_rate": 7.31403859697537e-05, + "loss": 0.6329, + "step": 200 + }, + { + "epoch": 1.3581081081081081, + "grad_norm": 0.23258287711574407, + "learning_rate": 7.303435923854172e-05, + "loss": 0.6346, + "step": 201 + }, + { + "epoch": 1.364864864864865, + "grad_norm": 0.36875878813699686, + "learning_rate": 7.292759745819781e-05, + "loss": 0.6423, + "step": 202 + }, + { + "epoch": 1.3716216216216215, + "grad_norm": 0.3816088768921796, + "learning_rate": 7.282010300428351e-05, + "loss": 0.6247, + "step": 203 + }, + { + "epoch": 1.3783783783783785, + "grad_norm": 0.32935135005176236, + "learning_rate": 7.271187826866312e-05, + "loss": 0.6306, + "step": 204 + }, + { + "epoch": 1.385135135135135, + "grad_norm": 0.24849408204445086, + "learning_rate": 7.260292565945049e-05, + "loss": 0.6425, + "step": 205 + }, + { + "epoch": 1.3918918918918919, + "grad_norm": 0.22496521412221002, + "learning_rate": 7.249324760095544e-05, + "loss": 0.64, + "step": 206 + }, + { + "epoch": 1.3986486486486487, + "grad_norm": 0.30450316180118336, + "learning_rate": 7.238284653362977e-05, + "loss": 0.6497, + "step": 207 + }, + { + "epoch": 1.4054054054054055, + "grad_norm": 0.2922608158162527, + "learning_rate": 7.227172491401299e-05, + "loss": 0.6339, + "step": 208 + }, + { + "epoch": 1.4121621621621623, + "grad_norm": 0.23745259799881574, + "learning_rate": 7.215988521467763e-05, + "loss": 0.629, + "step": 209 + }, + { + "epoch": 1.4189189189189189, + "grad_norm": 0.29205738332283676, + "learning_rate": 7.204732992417431e-05, + "loss": 0.6276, + "step": 210 + }, + { + "epoch": 1.4256756756756757, + "grad_norm": 0.2764709836681376, + "learning_rate": 7.193406154697625e-05, + "loss": 0.6325, + "step": 211 + }, + { + "epoch": 1.4324324324324325, + "grad_norm": 0.30413188245977635, + "learning_rate": 7.18200826034236e-05, + "loss": 0.6347, + "step": 212 + }, + { + "epoch": 1.4391891891891893, + "grad_norm": 0.32848243797717097, + "learning_rate": 7.17053956296674e-05, + "loss": 0.6173, + "step": 213 + }, + { + "epoch": 1.445945945945946, + "grad_norm": 0.3056033484058255, + "learning_rate": 7.159000317761305e-05, + "loss": 0.6339, + "step": 214 + }, + { + "epoch": 1.4527027027027026, + "grad_norm": 0.3085891817121419, + "learning_rate": 7.14739078148636e-05, + "loss": 0.6364, + "step": 215 + }, + { + "epoch": 1.4594594594594594, + "grad_norm": 0.3326838165486728, + "learning_rate": 7.135711212466264e-05, + "loss": 0.6422, + "step": 216 + }, + { + "epoch": 1.4662162162162162, + "grad_norm": 0.3216850978230415, + "learning_rate": 7.123961870583671e-05, + "loss": 0.6342, + "step": 217 + }, + { + "epoch": 1.472972972972973, + "grad_norm": 0.35372809908349234, + "learning_rate": 7.112143017273759e-05, + "loss": 0.637, + "step": 218 + }, + { + "epoch": 1.4797297297297298, + "grad_norm": 0.3689688937687713, + "learning_rate": 7.100254915518408e-05, + "loss": 0.6342, + "step": 219 + }, + { + "epoch": 1.4864864864864864, + "grad_norm": 0.4017045063635094, + "learning_rate": 7.088297829840346e-05, + "loss": 0.6286, + "step": 220 + }, + { + "epoch": 1.4932432432432432, + "grad_norm": 0.4446976891407491, + "learning_rate": 7.076272026297268e-05, + "loss": 0.6342, + "step": 221 + }, + { + "epoch": 1.5, + "grad_norm": 0.432808144889846, + "learning_rate": 7.064177772475912e-05, + "loss": 0.6452, + "step": 222 + }, + { + "epoch": 1.5067567567567568, + "grad_norm": 0.3443564755955631, + "learning_rate": 7.052015337486109e-05, + "loss": 0.6378, + "step": 223 + }, + { + "epoch": 1.5135135135135136, + "grad_norm": 0.30789638902879063, + "learning_rate": 7.03978499195479e-05, + "loss": 0.6284, + "step": 224 + }, + { + "epoch": 1.5202702702702702, + "grad_norm": 0.3444085722973672, + "learning_rate": 7.027487008019969e-05, + "loss": 0.6439, + "step": 225 + }, + { + "epoch": 1.527027027027027, + "grad_norm": 0.3698979548144048, + "learning_rate": 7.015121659324678e-05, + "loss": 0.6328, + "step": 226 + }, + { + "epoch": 1.5337837837837838, + "grad_norm": 0.4166719230003736, + "learning_rate": 7.002689221010897e-05, + "loss": 0.6295, + "step": 227 + }, + { + "epoch": 1.5405405405405406, + "grad_norm": 0.4418395627095002, + "learning_rate": 6.990189969713416e-05, + "loss": 0.6303, + "step": 228 + }, + { + "epoch": 1.5472972972972974, + "grad_norm": 0.4102988370573012, + "learning_rate": 6.977624183553676e-05, + "loss": 0.6431, + "step": 229 + }, + { + "epoch": 1.554054054054054, + "grad_norm": 0.39270575425283927, + "learning_rate": 6.964992142133602e-05, + "loss": 0.6333, + "step": 230 + }, + { + "epoch": 1.560810810810811, + "grad_norm": 0.3396456921939933, + "learning_rate": 6.952294126529356e-05, + "loss": 0.6274, + "step": 231 + }, + { + "epoch": 1.5675675675675675, + "grad_norm": 0.29161162699565907, + "learning_rate": 6.939530419285104e-05, + "loss": 0.6346, + "step": 232 + }, + { + "epoch": 1.5743243243243243, + "grad_norm": 0.3270645478162509, + "learning_rate": 6.926701304406713e-05, + "loss": 0.6307, + "step": 233 + }, + { + "epoch": 1.5810810810810811, + "grad_norm": 0.33343232181767685, + "learning_rate": 6.913807067355445e-05, + "loss": 0.6338, + "step": 234 + }, + { + "epoch": 1.5878378378378377, + "grad_norm": 0.34421561990305105, + "learning_rate": 6.90084799504159e-05, + "loss": 0.6417, + "step": 235 + }, + { + "epoch": 1.5945945945945947, + "grad_norm": 0.31139357648881777, + "learning_rate": 6.887824375818099e-05, + "loss": 0.6399, + "step": 236 + }, + { + "epoch": 1.6013513513513513, + "grad_norm": 0.2657664210184094, + "learning_rate": 6.874736499474154e-05, + "loss": 0.6411, + "step": 237 + }, + { + "epoch": 1.6081081081081081, + "grad_norm": 0.25481549866865094, + "learning_rate": 6.861584657228728e-05, + "loss": 0.6418, + "step": 238 + }, + { + "epoch": 1.614864864864865, + "grad_norm": 0.27638932514157166, + "learning_rate": 6.848369141724104e-05, + "loss": 0.6435, + "step": 239 + }, + { + "epoch": 1.6216216216216215, + "grad_norm": 0.32918191799417873, + "learning_rate": 6.835090247019354e-05, + "loss": 0.6269, + "step": 240 + }, + { + "epoch": 1.6283783783783785, + "grad_norm": 0.334162128338808, + "learning_rate": 6.821748268583813e-05, + "loss": 0.6382, + "step": 241 + }, + { + "epoch": 1.635135135135135, + "grad_norm": 0.3172895628393372, + "learning_rate": 6.808343503290491e-05, + "loss": 0.627, + "step": 242 + }, + { + "epoch": 1.6418918918918919, + "grad_norm": 0.3353087016871893, + "learning_rate": 6.79487624940947e-05, + "loss": 0.6228, + "step": 243 + }, + { + "epoch": 1.6486486486486487, + "grad_norm": 0.3718369962163942, + "learning_rate": 6.781346806601273e-05, + "loss": 0.6305, + "step": 244 + }, + { + "epoch": 1.6554054054054053, + "grad_norm": 0.39658980608696004, + "learning_rate": 6.767755475910185e-05, + "loss": 0.6227, + "step": 245 + }, + { + "epoch": 1.6621621621621623, + "grad_norm": 0.4269305577093925, + "learning_rate": 6.754102559757569e-05, + "loss": 0.6338, + "step": 246 + }, + { + "epoch": 1.6689189189189189, + "grad_norm": 0.48834528749062367, + "learning_rate": 6.740388361935125e-05, + "loss": 0.6285, + "step": 247 + }, + { + "epoch": 1.6756756756756757, + "grad_norm": 0.5054797184183533, + "learning_rate": 6.726613187598132e-05, + "loss": 0.6363, + "step": 248 + }, + { + "epoch": 1.6824324324324325, + "grad_norm": 0.46558382405228205, + "learning_rate": 6.712777343258666e-05, + "loss": 0.6342, + "step": 249 + }, + { + "epoch": 1.689189189189189, + "grad_norm": 0.3349216901360824, + "learning_rate": 6.698881136778771e-05, + "loss": 0.638, + "step": 250 + }, + { + "epoch": 1.695945945945946, + "grad_norm": 0.296031271653257, + "learning_rate": 6.684924877363613e-05, + "loss": 0.6449, + "step": 251 + }, + { + "epoch": 1.7027027027027026, + "grad_norm": 0.34651680386096584, + "learning_rate": 6.670908875554594e-05, + "loss": 0.6373, + "step": 252 + }, + { + "epoch": 1.7094594594594594, + "grad_norm": 0.34431135065378676, + "learning_rate": 6.656833443222458e-05, + "loss": 0.6222, + "step": 253 + }, + { + "epoch": 1.7162162162162162, + "grad_norm": 0.3856861975419152, + "learning_rate": 6.642698893560327e-05, + "loss": 0.6389, + "step": 254 + }, + { + "epoch": 1.722972972972973, + "grad_norm": 0.3419900662624624, + "learning_rate": 6.628505541076755e-05, + "loss": 0.6345, + "step": 255 + }, + { + "epoch": 1.7297297297297298, + "grad_norm": 0.25513443874276925, + "learning_rate": 6.614253701588718e-05, + "loss": 0.6337, + "step": 256 + }, + { + "epoch": 1.7364864864864864, + "grad_norm": 0.40205274571625865, + "learning_rate": 6.599943692214587e-05, + "loss": 0.6329, + "step": 257 + }, + { + "epoch": 1.7432432432432432, + "grad_norm": 0.38134230469861735, + "learning_rate": 6.585575831367078e-05, + "loss": 0.6315, + "step": 258 + }, + { + "epoch": 1.75, + "grad_norm": 0.2108025719904083, + "learning_rate": 6.571150438746157e-05, + "loss": 0.6332, + "step": 259 + }, + { + "epoch": 1.7567567567567568, + "grad_norm": 0.2896853669566718, + "learning_rate": 6.55666783533194e-05, + "loss": 0.6254, + "step": 260 + }, + { + "epoch": 1.7635135135135136, + "grad_norm": 0.3811309603864002, + "learning_rate": 6.542128343377536e-05, + "loss": 0.6355, + "step": 261 + }, + { + "epoch": 1.7702702702702702, + "grad_norm": 0.30773816113154867, + "learning_rate": 6.527532286401889e-05, + "loss": 0.6337, + "step": 262 + }, + { + "epoch": 1.777027027027027, + "grad_norm": 0.22231601239252788, + "learning_rate": 6.51287998918257e-05, + "loss": 0.6204, + "step": 263 + }, + { + "epoch": 1.7837837837837838, + "grad_norm": 0.2581820667256423, + "learning_rate": 6.498171777748557e-05, + "loss": 0.6335, + "step": 264 + }, + { + "epoch": 1.7905405405405406, + "grad_norm": 0.3087437556000368, + "learning_rate": 6.483407979372975e-05, + "loss": 0.6221, + "step": 265 + }, + { + "epoch": 1.7972972972972974, + "grad_norm": 0.229624863667569, + "learning_rate": 6.468588922565822e-05, + "loss": 0.6305, + "step": 266 + }, + { + "epoch": 1.804054054054054, + "grad_norm": 0.15965170109349433, + "learning_rate": 6.453714937066648e-05, + "loss": 0.6277, + "step": 267 + }, + { + "epoch": 1.810810810810811, + "grad_norm": 0.16500327907249202, + "learning_rate": 6.438786353837228e-05, + "loss": 0.6279, + "step": 268 + }, + { + "epoch": 1.8175675675675675, + "grad_norm": 0.19228879276238184, + "learning_rate": 6.423803505054193e-05, + "loss": 0.637, + "step": 269 + }, + { + "epoch": 1.8243243243243243, + "grad_norm": 0.1960334960484057, + "learning_rate": 6.408766724101638e-05, + "loss": 0.6344, + "step": 270 + }, + { + "epoch": 1.8310810810810811, + "grad_norm": 0.22246057335109667, + "learning_rate": 6.393676345563708e-05, + "loss": 0.6315, + "step": 271 + }, + { + "epoch": 1.8378378378378377, + "grad_norm": 0.25500558074002594, + "learning_rate": 6.378532705217148e-05, + "loss": 0.6267, + "step": 272 + }, + { + "epoch": 1.8445945945945947, + "grad_norm": 0.2797665813089635, + "learning_rate": 6.363336140023833e-05, + "loss": 0.6198, + "step": 273 + }, + { + "epoch": 1.8513513513513513, + "grad_norm": 0.32237396509938987, + "learning_rate": 6.348086988123274e-05, + "loss": 0.6302, + "step": 274 + }, + { + "epoch": 1.8581081081081081, + "grad_norm": 0.3837879616068762, + "learning_rate": 6.332785588825094e-05, + "loss": 0.6366, + "step": 275 + }, + { + "epoch": 1.864864864864865, + "grad_norm": 0.3847808366177522, + "learning_rate": 6.317432282601469e-05, + "loss": 0.6405, + "step": 276 + }, + { + "epoch": 1.8716216216216215, + "grad_norm": 0.3485054619073992, + "learning_rate": 6.302027411079562e-05, + "loss": 0.6273, + "step": 277 + }, + { + "epoch": 1.8783783783783785, + "grad_norm": 0.3171200624449193, + "learning_rate": 6.286571317033915e-05, + "loss": 0.6337, + "step": 278 + }, + { + "epoch": 1.885135135135135, + "grad_norm": 0.27373328598715324, + "learning_rate": 6.271064344378832e-05, + "loss": 0.6403, + "step": 279 + }, + { + "epoch": 1.8918918918918919, + "grad_norm": 0.3045549423645127, + "learning_rate": 6.255506838160711e-05, + "loss": 0.6317, + "step": 280 + }, + { + "epoch": 1.8986486486486487, + "grad_norm": 0.23258625831405336, + "learning_rate": 6.239899144550383e-05, + "loss": 0.6195, + "step": 281 + }, + { + "epoch": 1.9054054054054053, + "grad_norm": 0.30422094244714903, + "learning_rate": 6.224241610835391e-05, + "loss": 0.6422, + "step": 282 + }, + { + "epoch": 1.9121621621621623, + "grad_norm": 0.3974220280836707, + "learning_rate": 6.208534585412282e-05, + "loss": 0.6341, + "step": 283 + }, + { + "epoch": 1.9189189189189189, + "grad_norm": 0.3588942161699914, + "learning_rate": 6.19277841777884e-05, + "loss": 0.6277, + "step": 284 + }, + { + "epoch": 1.9256756756756757, + "grad_norm": 0.23348715119029995, + "learning_rate": 6.176973458526317e-05, + "loss": 0.6196, + "step": 285 + }, + { + "epoch": 1.9324324324324325, + "grad_norm": 0.18377971391888995, + "learning_rate": 6.161120059331628e-05, + "loss": 0.6332, + "step": 286 + }, + { + "epoch": 1.939189189189189, + "grad_norm": 0.2385215605440726, + "learning_rate": 6.14521857294953e-05, + "loss": 0.6479, + "step": 287 + }, + { + "epoch": 1.945945945945946, + "grad_norm": 0.2797923846654065, + "learning_rate": 6.129269353204769e-05, + "loss": 0.6354, + "step": 288 + }, + { + "epoch": 1.9527027027027026, + "grad_norm": 0.3046327128914545, + "learning_rate": 6.113272754984206e-05, + "loss": 0.6394, + "step": 289 + }, + { + "epoch": 1.9594594594594594, + "grad_norm": 0.2685120513408917, + "learning_rate": 6.0972291342289274e-05, + "loss": 0.6285, + "step": 290 + }, + { + "epoch": 1.9662162162162162, + "grad_norm": 0.19923838754652695, + "learning_rate": 6.081138847926317e-05, + "loss": 0.6239, + "step": 291 + }, + { + "epoch": 1.972972972972973, + "grad_norm": 0.23688012061160577, + "learning_rate": 6.065002254102116e-05, + "loss": 0.6391, + "step": 292 + }, + { + "epoch": 1.9797297297297298, + "grad_norm": 0.2444120597492392, + "learning_rate": 6.048819711812457e-05, + "loss": 0.6315, + "step": 293 + }, + { + "epoch": 1.9864864864864864, + "grad_norm": 0.2374660611243683, + "learning_rate": 6.032591581135878e-05, + "loss": 0.635, + "step": 294 + }, + { + "epoch": 1.9932432432432432, + "grad_norm": 0.19031195961334121, + "learning_rate": 6.0163182231652985e-05, + "loss": 0.6266, + "step": 295 + }, + { + "epoch": 2.0, + "grad_norm": 0.2396442769161295, + "learning_rate": 6.000000000000001e-05, + "loss": 0.6384, + "step": 296 + }, + { + "epoch": 2.0067567567567566, + "grad_norm": 0.2791243307653228, + "learning_rate": 5.983637274737558e-05, + "loss": 0.6251, + "step": 297 + }, + { + "epoch": 2.0135135135135136, + "grad_norm": 0.3160836645945337, + "learning_rate": 5.967230411465768e-05, + "loss": 0.6261, + "step": 298 + }, + { + "epoch": 2.02027027027027, + "grad_norm": 0.2816061126247031, + "learning_rate": 5.950779775254539e-05, + "loss": 0.6146, + "step": 299 + }, + { + "epoch": 2.027027027027027, + "grad_norm": 0.3070948777190884, + "learning_rate": 5.934285732147778e-05, + "loss": 0.6234, + "step": 300 + }, + { + "epoch": 2.0337837837837838, + "grad_norm": 0.31200072769112625, + "learning_rate": 5.91774864915524e-05, + "loss": 0.6129, + "step": 301 + }, + { + "epoch": 2.0405405405405403, + "grad_norm": 0.28687016611033417, + "learning_rate": 5.90116889424436e-05, + "loss": 0.6144, + "step": 302 + }, + { + "epoch": 2.0472972972972974, + "grad_norm": 0.3063930815670496, + "learning_rate": 5.884546836332072e-05, + "loss": 0.6176, + "step": 303 + }, + { + "epoch": 2.054054054054054, + "grad_norm": 0.24478115745724477, + "learning_rate": 5.867882845276593e-05, + "loss": 0.6164, + "step": 304 + }, + { + "epoch": 2.060810810810811, + "grad_norm": 0.17845425335065362, + "learning_rate": 5.851177291869197e-05, + "loss": 0.6118, + "step": 305 + }, + { + "epoch": 2.0675675675675675, + "grad_norm": 0.20058855750068053, + "learning_rate": 5.834430547825964e-05, + "loss": 0.6146, + "step": 306 + }, + { + "epoch": 2.074324324324324, + "grad_norm": 0.21828538712863907, + "learning_rate": 5.8176429857795104e-05, + "loss": 0.6176, + "step": 307 + }, + { + "epoch": 2.081081081081081, + "grad_norm": 0.2218763340787443, + "learning_rate": 5.8008149792706936e-05, + "loss": 0.6059, + "step": 308 + }, + { + "epoch": 2.0878378378378377, + "grad_norm": 0.26010217837913424, + "learning_rate": 5.783946902740304e-05, + "loss": 0.6172, + "step": 309 + }, + { + "epoch": 2.0945945945945947, + "grad_norm": 0.2968234389164342, + "learning_rate": 5.767039131520733e-05, + "loss": 0.6176, + "step": 310 + }, + { + "epoch": 2.1013513513513513, + "grad_norm": 0.29905738657768327, + "learning_rate": 5.750092041827618e-05, + "loss": 0.5998, + "step": 311 + }, + { + "epoch": 2.108108108108108, + "grad_norm": 0.287828129320029, + "learning_rate": 5.7331060107514754e-05, + "loss": 0.597, + "step": 312 + }, + { + "epoch": 2.114864864864865, + "grad_norm": 0.26831745019142034, + "learning_rate": 5.716081416249307e-05, + "loss": 0.6131, + "step": 313 + }, + { + "epoch": 2.1216216216216215, + "grad_norm": 0.28531963857984527, + "learning_rate": 5.699018637136192e-05, + "loss": 0.6056, + "step": 314 + }, + { + "epoch": 2.1283783783783785, + "grad_norm": 0.18651675715743835, + "learning_rate": 5.681918053076858e-05, + "loss": 0.6167, + "step": 315 + }, + { + "epoch": 2.135135135135135, + "grad_norm": 0.21775207949583675, + "learning_rate": 5.664780044577231e-05, + "loss": 0.6219, + "step": 316 + }, + { + "epoch": 2.141891891891892, + "grad_norm": 0.25323829124729375, + "learning_rate": 5.6476049929759714e-05, + "loss": 0.6196, + "step": 317 + }, + { + "epoch": 2.1486486486486487, + "grad_norm": 0.19732115665472313, + "learning_rate": 5.6303932804359857e-05, + "loss": 0.6156, + "step": 318 + }, + { + "epoch": 2.1554054054054053, + "grad_norm": 0.16849505056239242, + "learning_rate": 5.613145289935926e-05, + "loss": 0.6042, + "step": 319 + }, + { + "epoch": 2.1621621621621623, + "grad_norm": 0.19969762470742003, + "learning_rate": 5.595861405261666e-05, + "loss": 0.6031, + "step": 320 + }, + { + "epoch": 2.168918918918919, + "grad_norm": 0.2036043502047779, + "learning_rate": 5.578542010997764e-05, + "loss": 0.633, + "step": 321 + }, + { + "epoch": 2.175675675675676, + "grad_norm": 0.16807494045865312, + "learning_rate": 5.561187492518903e-05, + "loss": 0.6118, + "step": 322 + }, + { + "epoch": 2.1824324324324325, + "grad_norm": 0.1717048805694881, + "learning_rate": 5.5437982359813156e-05, + "loss": 0.6116, + "step": 323 + }, + { + "epoch": 2.189189189189189, + "grad_norm": 0.17594564977890167, + "learning_rate": 5.526374628314195e-05, + "loss": 0.6162, + "step": 324 + }, + { + "epoch": 2.195945945945946, + "grad_norm": 0.21489547073195236, + "learning_rate": 5.50891705721108e-05, + "loss": 0.5984, + "step": 325 + }, + { + "epoch": 2.2027027027027026, + "grad_norm": 0.21447613988889852, + "learning_rate": 5.4914259111212355e-05, + "loss": 0.6198, + "step": 326 + }, + { + "epoch": 2.2094594594594597, + "grad_norm": 0.2207232176733451, + "learning_rate": 5.473901579241e-05, + "loss": 0.6092, + "step": 327 + }, + { + "epoch": 2.2162162162162162, + "grad_norm": 0.19937702101442842, + "learning_rate": 5.4563444515051354e-05, + "loss": 0.6065, + "step": 328 + }, + { + "epoch": 2.222972972972973, + "grad_norm": 0.19293343910472205, + "learning_rate": 5.438754918578144e-05, + "loss": 0.6063, + "step": 329 + }, + { + "epoch": 2.22972972972973, + "grad_norm": 0.17732267437010826, + "learning_rate": 5.4211333718455756e-05, + "loss": 0.6076, + "step": 330 + }, + { + "epoch": 2.2364864864864864, + "grad_norm": 0.18101481051912677, + "learning_rate": 5.4034802034053223e-05, + "loss": 0.6154, + "step": 331 + }, + { + "epoch": 2.2432432432432434, + "grad_norm": 0.19133561762748252, + "learning_rate": 5.3857958060588955e-05, + "loss": 0.5988, + "step": 332 + }, + { + "epoch": 2.25, + "grad_norm": 0.18341857255842758, + "learning_rate": 5.368080573302676e-05, + "loss": 0.6093, + "step": 333 + }, + { + "epoch": 2.2567567567567566, + "grad_norm": 0.2486112524831487, + "learning_rate": 5.3503348993191706e-05, + "loss": 0.6159, + "step": 334 + }, + { + "epoch": 2.2635135135135136, + "grad_norm": 0.21869922307823292, + "learning_rate": 5.332559178968231e-05, + "loss": 0.6128, + "step": 335 + }, + { + "epoch": 2.27027027027027, + "grad_norm": 0.1808704522354482, + "learning_rate": 5.314753807778276e-05, + "loss": 0.6157, + "step": 336 + }, + { + "epoch": 2.277027027027027, + "grad_norm": 0.19637386821805067, + "learning_rate": 5.296919181937485e-05, + "loss": 0.6206, + "step": 337 + }, + { + "epoch": 2.2837837837837838, + "grad_norm": 0.1867094463718795, + "learning_rate": 5.279055698284982e-05, + "loss": 0.6209, + "step": 338 + }, + { + "epoch": 2.2905405405405403, + "grad_norm": 0.17359407157166826, + "learning_rate": 5.261163754302011e-05, + "loss": 0.6012, + "step": 339 + }, + { + "epoch": 2.2972972972972974, + "grad_norm": 0.18569999759822897, + "learning_rate": 5.2432437481030855e-05, + "loss": 0.6048, + "step": 340 + }, + { + "epoch": 2.304054054054054, + "grad_norm": 0.1773148171307398, + "learning_rate": 5.225296078427135e-05, + "loss": 0.5963, + "step": 341 + }, + { + "epoch": 2.310810810810811, + "grad_norm": 0.16243539870707618, + "learning_rate": 5.207321144628628e-05, + "loss": 0.6178, + "step": 342 + }, + { + "epoch": 2.3175675675675675, + "grad_norm": 0.18693063507419128, + "learning_rate": 5.18931934666869e-05, + "loss": 0.6136, + "step": 343 + }, + { + "epoch": 2.3243243243243246, + "grad_norm": 0.1661326777898068, + "learning_rate": 5.171291085106202e-05, + "loss": 0.6235, + "step": 344 + }, + { + "epoch": 2.331081081081081, + "grad_norm": 0.15331403054989767, + "learning_rate": 5.153236761088888e-05, + "loss": 0.6098, + "step": 345 + }, + { + "epoch": 2.3378378378378377, + "grad_norm": 0.1570395157158325, + "learning_rate": 5.135156776344389e-05, + "loss": 0.6137, + "step": 346 + }, + { + "epoch": 2.3445945945945947, + "grad_norm": 0.17430865513764107, + "learning_rate": 5.117051533171321e-05, + "loss": 0.6177, + "step": 347 + }, + { + "epoch": 2.3513513513513513, + "grad_norm": 0.21717070798995733, + "learning_rate": 5.098921434430333e-05, + "loss": 0.6203, + "step": 348 + }, + { + "epoch": 2.358108108108108, + "grad_norm": 0.20392614118792024, + "learning_rate": 5.080766883535129e-05, + "loss": 0.6136, + "step": 349 + }, + { + "epoch": 2.364864864864865, + "grad_norm": 0.21588477709369264, + "learning_rate": 5.062588284443505e-05, + "loss": 0.6145, + "step": 350 + }, + { + "epoch": 2.3716216216216215, + "grad_norm": 0.21049543340334342, + "learning_rate": 5.0443860416483536e-05, + "loss": 0.6019, + "step": 351 + }, + { + "epoch": 2.3783783783783785, + "grad_norm": 0.1730662628064374, + "learning_rate": 5.026160560168661e-05, + "loss": 0.6058, + "step": 352 + }, + { + "epoch": 2.385135135135135, + "grad_norm": 0.17831683554874725, + "learning_rate": 5.0079122455405014e-05, + "loss": 0.6208, + "step": 353 + }, + { + "epoch": 2.391891891891892, + "grad_norm": 0.17744701669503546, + "learning_rate": 4.989641503808011e-05, + "loss": 0.609, + "step": 354 + }, + { + "epoch": 2.3986486486486487, + "grad_norm": 0.17994097783132235, + "learning_rate": 4.971348741514349e-05, + "loss": 0.6066, + "step": 355 + }, + { + "epoch": 2.4054054054054053, + "grad_norm": 0.17459787928555573, + "learning_rate": 4.95303436569266e-05, + "loss": 0.6146, + "step": 356 + }, + { + "epoch": 2.4121621621621623, + "grad_norm": 0.16284962487352878, + "learning_rate": 4.934698783857011e-05, + "loss": 0.6044, + "step": 357 + }, + { + "epoch": 2.418918918918919, + "grad_norm": 0.17767021224773352, + "learning_rate": 4.91634240399332e-05, + "loss": 0.6211, + "step": 358 + }, + { + "epoch": 2.4256756756756754, + "grad_norm": 0.1496311992502912, + "learning_rate": 4.8979656345502904e-05, + "loss": 0.602, + "step": 359 + }, + { + "epoch": 2.4324324324324325, + "grad_norm": 0.16981969578628817, + "learning_rate": 4.8795688844303114e-05, + "loss": 0.6148, + "step": 360 + }, + { + "epoch": 2.439189189189189, + "grad_norm": 0.2089769520148497, + "learning_rate": 4.861152562980362e-05, + "loss": 0.6123, + "step": 361 + }, + { + "epoch": 2.445945945945946, + "grad_norm": 0.20528877980388496, + "learning_rate": 4.8427170799829055e-05, + "loss": 0.6095, + "step": 362 + }, + { + "epoch": 2.4527027027027026, + "grad_norm": 0.1827881621296737, + "learning_rate": 4.824262845646771e-05, + "loss": 0.6112, + "step": 363 + }, + { + "epoch": 2.4594594594594597, + "grad_norm": 0.19580070665324928, + "learning_rate": 4.805790270598021e-05, + "loss": 0.6209, + "step": 364 + }, + { + "epoch": 2.4662162162162162, + "grad_norm": 0.23276386582983333, + "learning_rate": 4.787299765870822e-05, + "loss": 0.6083, + "step": 365 + }, + { + "epoch": 2.472972972972973, + "grad_norm": 0.20338848706607754, + "learning_rate": 4.768791742898292e-05, + "loss": 0.615, + "step": 366 + }, + { + "epoch": 2.47972972972973, + "grad_norm": 0.1521765603212478, + "learning_rate": 4.7502666135033486e-05, + "loss": 0.6154, + "step": 367 + }, + { + "epoch": 2.4864864864864864, + "grad_norm": 0.1559952589700639, + "learning_rate": 4.731724789889547e-05, + "loss": 0.6026, + "step": 368 + }, + { + "epoch": 2.4932432432432434, + "grad_norm": 0.14084577612463858, + "learning_rate": 4.7131666846319036e-05, + "loss": 0.607, + "step": 369 + }, + { + "epoch": 2.5, + "grad_norm": 0.1450823104838612, + "learning_rate": 4.694592710667723e-05, + "loss": 0.6173, + "step": 370 + }, + { + "epoch": 2.506756756756757, + "grad_norm": 0.14280439260452932, + "learning_rate": 4.676003281287397e-05, + "loss": 0.6035, + "step": 371 + }, + { + "epoch": 2.5135135135135136, + "grad_norm": 0.1731629109803712, + "learning_rate": 4.657398810125225e-05, + "loss": 0.6044, + "step": 372 + }, + { + "epoch": 2.52027027027027, + "grad_norm": 0.14551348661799535, + "learning_rate": 4.638779711150198e-05, + "loss": 0.6223, + "step": 373 + }, + { + "epoch": 2.527027027027027, + "grad_norm": 0.15724571828258985, + "learning_rate": 4.620146398656792e-05, + "loss": 0.6174, + "step": 374 + }, + { + "epoch": 2.5337837837837838, + "grad_norm": 0.1622156186595054, + "learning_rate": 4.601499287255748e-05, + "loss": 0.6159, + "step": 375 + }, + { + "epoch": 2.5405405405405403, + "grad_norm": 0.16326776251404446, + "learning_rate": 4.582838791864846e-05, + "loss": 0.6178, + "step": 376 + }, + { + "epoch": 2.5472972972972974, + "grad_norm": 0.16705938089441114, + "learning_rate": 4.5641653276996774e-05, + "loss": 0.6069, + "step": 377 + }, + { + "epoch": 2.554054054054054, + "grad_norm": 0.1240389952109112, + "learning_rate": 4.5454793102644006e-05, + "loss": 0.6129, + "step": 378 + }, + { + "epoch": 2.560810810810811, + "grad_norm": 0.16500872440232617, + "learning_rate": 4.5267811553424945e-05, + "loss": 0.6213, + "step": 379 + }, + { + "epoch": 2.5675675675675675, + "grad_norm": 0.17252309025914744, + "learning_rate": 4.5080712789875154e-05, + "loss": 0.6172, + "step": 380 + }, + { + "epoch": 2.5743243243243246, + "grad_norm": 0.163801171559839, + "learning_rate": 4.489350097513829e-05, + "loss": 0.6205, + "step": 381 + }, + { + "epoch": 2.581081081081081, + "grad_norm": 0.14139376100405637, + "learning_rate": 4.470618027487354e-05, + "loss": 0.6101, + "step": 382 + }, + { + "epoch": 2.5878378378378377, + "grad_norm": 0.14839912258194626, + "learning_rate": 4.451875485716292e-05, + "loss": 0.6214, + "step": 383 + }, + { + "epoch": 2.5945945945945947, + "grad_norm": 0.13849384656147448, + "learning_rate": 4.4331228892418473e-05, + "loss": 0.6111, + "step": 384 + }, + { + "epoch": 2.6013513513513513, + "grad_norm": 0.16314353286789263, + "learning_rate": 4.414360655328957e-05, + "loss": 0.6201, + "step": 385 + }, + { + "epoch": 2.608108108108108, + "grad_norm": 0.13166529990196998, + "learning_rate": 4.395589201457e-05, + "loss": 0.6167, + "step": 386 + }, + { + "epoch": 2.614864864864865, + "grad_norm": 0.14086485553732253, + "learning_rate": 4.376808945310505e-05, + "loss": 0.6219, + "step": 387 + }, + { + "epoch": 2.6216216216216215, + "grad_norm": 0.15193755776745674, + "learning_rate": 4.358020304769867e-05, + "loss": 0.5994, + "step": 388 + }, + { + "epoch": 2.6283783783783785, + "grad_norm": 0.14946480137653986, + "learning_rate": 4.339223697902037e-05, + "loss": 0.61, + "step": 389 + }, + { + "epoch": 2.635135135135135, + "grad_norm": 0.15531265023719382, + "learning_rate": 4.320419542951228e-05, + "loss": 0.6179, + "step": 390 + }, + { + "epoch": 2.641891891891892, + "grad_norm": 0.14264508419628688, + "learning_rate": 4.3016082583296067e-05, + "loss": 0.6167, + "step": 391 + }, + { + "epoch": 2.6486486486486487, + "grad_norm": 0.12288922241404748, + "learning_rate": 4.2827902626079784e-05, + "loss": 0.6195, + "step": 392 + }, + { + "epoch": 2.6554054054054053, + "grad_norm": 0.19330457974395565, + "learning_rate": 4.263965974506483e-05, + "loss": 0.6207, + "step": 393 + }, + { + "epoch": 2.6621621621621623, + "grad_norm": 0.16883088734435447, + "learning_rate": 4.2451358128852654e-05, + "loss": 0.5989, + "step": 394 + }, + { + "epoch": 2.668918918918919, + "grad_norm": 0.1599604515688246, + "learning_rate": 4.22630019673517e-05, + "loss": 0.6142, + "step": 395 + }, + { + "epoch": 2.6756756756756754, + "grad_norm": 0.15518611864455714, + "learning_rate": 4.207459545168405e-05, + "loss": 0.6102, + "step": 396 + }, + { + "epoch": 2.6824324324324325, + "grad_norm": 0.16742006705568305, + "learning_rate": 4.188614277409224e-05, + "loss": 0.6171, + "step": 397 + }, + { + "epoch": 2.689189189189189, + "grad_norm": 0.15553602245497833, + "learning_rate": 4.169764812784594e-05, + "loss": 0.6143, + "step": 398 + }, + { + "epoch": 2.695945945945946, + "grad_norm": 0.1595818544532621, + "learning_rate": 4.1509115707148695e-05, + "loss": 0.6047, + "step": 399 + }, + { + "epoch": 2.7027027027027026, + "grad_norm": 0.13690560623925258, + "learning_rate": 4.132054970704454e-05, + "loss": 0.6092, + "step": 400 + }, + { + "epoch": 2.7094594594594597, + "grad_norm": 0.1341753352699138, + "learning_rate": 4.1131954323324734e-05, + "loss": 0.6023, + "step": 401 + }, + { + "epoch": 2.7162162162162162, + "grad_norm": 0.14452976105100054, + "learning_rate": 4.094333375243428e-05, + "loss": 0.6121, + "step": 402 + }, + { + "epoch": 2.722972972972973, + "grad_norm": 0.17855418886582433, + "learning_rate": 4.075469219137868e-05, + "loss": 0.6192, + "step": 403 + }, + { + "epoch": 2.72972972972973, + "grad_norm": 0.15337873200421298, + "learning_rate": 4.056603383763049e-05, + "loss": 0.6137, + "step": 404 + }, + { + "epoch": 2.7364864864864864, + "grad_norm": 0.14193840867721655, + "learning_rate": 4.0377362889035875e-05, + "loss": 0.5971, + "step": 405 + }, + { + "epoch": 2.743243243243243, + "grad_norm": 0.14049198222822518, + "learning_rate": 4.0188683543721295e-05, + "loss": 0.61, + "step": 406 + }, + { + "epoch": 2.75, + "grad_norm": 0.1643045132978014, + "learning_rate": 4e-05, + "loss": 0.6224, + "step": 407 + }, + { + "epoch": 2.756756756756757, + "grad_norm": 0.14527602919509278, + "learning_rate": 3.981131645627872e-05, + "loss": 0.6164, + "step": 408 + }, + { + "epoch": 2.7635135135135136, + "grad_norm": 0.1384596570378767, + "learning_rate": 3.9622637110964125e-05, + "loss": 0.6158, + "step": 409 + }, + { + "epoch": 2.77027027027027, + "grad_norm": 0.18210804820042478, + "learning_rate": 3.943396616236953e-05, + "loss": 0.6149, + "step": 410 + }, + { + "epoch": 2.777027027027027, + "grad_norm": 0.16943522003893646, + "learning_rate": 3.9245307808621325e-05, + "loss": 0.6159, + "step": 411 + }, + { + "epoch": 2.7837837837837838, + "grad_norm": 0.16603181052517613, + "learning_rate": 3.905666624756573e-05, + "loss": 0.6186, + "step": 412 + }, + { + "epoch": 2.7905405405405403, + "grad_norm": 0.15368628411505494, + "learning_rate": 3.886804567667528e-05, + "loss": 0.6155, + "step": 413 + }, + { + "epoch": 2.7972972972972974, + "grad_norm": 0.1721774875064551, + "learning_rate": 3.867945029295546e-05, + "loss": 0.6099, + "step": 414 + }, + { + "epoch": 2.804054054054054, + "grad_norm": 0.1893984092920587, + "learning_rate": 3.8490884292851325e-05, + "loss": 0.6146, + "step": 415 + }, + { + "epoch": 2.810810810810811, + "grad_norm": 0.1792007727966493, + "learning_rate": 3.830235187215408e-05, + "loss": 0.6195, + "step": 416 + }, + { + "epoch": 2.8175675675675675, + "grad_norm": 0.1569338853010942, + "learning_rate": 3.8113857225907783e-05, + "loss": 0.6173, + "step": 417 + }, + { + "epoch": 2.8243243243243246, + "grad_norm": 0.17031740028829662, + "learning_rate": 3.792540454831596e-05, + "loss": 0.6163, + "step": 418 + }, + { + "epoch": 2.831081081081081, + "grad_norm": 0.15004651120970428, + "learning_rate": 3.7736998032648305e-05, + "loss": 0.6047, + "step": 419 + }, + { + "epoch": 2.8378378378378377, + "grad_norm": 0.142075802466718, + "learning_rate": 3.754864187114736e-05, + "loss": 0.6253, + "step": 420 + }, + { + "epoch": 2.8445945945945947, + "grad_norm": 0.15599780395707422, + "learning_rate": 3.736034025493519e-05, + "loss": 0.593, + "step": 421 + }, + { + "epoch": 2.8513513513513513, + "grad_norm": 0.1563201754706643, + "learning_rate": 3.717209737392022e-05, + "loss": 0.6128, + "step": 422 + }, + { + "epoch": 2.858108108108108, + "grad_norm": 0.15805265922848155, + "learning_rate": 3.698391741670394e-05, + "loss": 0.6059, + "step": 423 + }, + { + "epoch": 2.864864864864865, + "grad_norm": 0.17933457522155424, + "learning_rate": 3.679580457048772e-05, + "loss": 0.6086, + "step": 424 + }, + { + "epoch": 2.8716216216216215, + "grad_norm": 0.17371124474695815, + "learning_rate": 3.660776302097965e-05, + "loss": 0.6079, + "step": 425 + }, + { + "epoch": 2.8783783783783785, + "grad_norm": 0.1315584411453832, + "learning_rate": 3.641979695230135e-05, + "loss": 0.6039, + "step": 426 + }, + { + "epoch": 2.885135135135135, + "grad_norm": 0.148272614562961, + "learning_rate": 3.6231910546894956e-05, + "loss": 0.6161, + "step": 427 + }, + { + "epoch": 2.891891891891892, + "grad_norm": 0.12013666775227018, + "learning_rate": 3.6044107985430015e-05, + "loss": 0.6113, + "step": 428 + }, + { + "epoch": 2.8986486486486487, + "grad_norm": 0.15454206210883253, + "learning_rate": 3.585639344671043e-05, + "loss": 0.6113, + "step": 429 + }, + { + "epoch": 2.9054054054054053, + "grad_norm": 0.130251935264837, + "learning_rate": 3.5668771107581526e-05, + "loss": 0.613, + "step": 430 + }, + { + "epoch": 2.9121621621621623, + "grad_norm": 0.1370783380801396, + "learning_rate": 3.5481245142837095e-05, + "loss": 0.6168, + "step": 431 + }, + { + "epoch": 2.918918918918919, + "grad_norm": 0.13116479132749334, + "learning_rate": 3.5293819725126464e-05, + "loss": 0.6136, + "step": 432 + }, + { + "epoch": 2.9256756756756754, + "grad_norm": 0.14031976128589324, + "learning_rate": 3.5106499024861715e-05, + "loss": 0.6175, + "step": 433 + }, + { + "epoch": 2.9324324324324325, + "grad_norm": 0.12054376627925889, + "learning_rate": 3.491928721012485e-05, + "loss": 0.6134, + "step": 434 + }, + { + "epoch": 2.939189189189189, + "grad_norm": 0.14272909957039542, + "learning_rate": 3.4732188446575055e-05, + "loss": 0.6096, + "step": 435 + }, + { + "epoch": 2.945945945945946, + "grad_norm": 0.1343117824779506, + "learning_rate": 3.454520689735602e-05, + "loss": 0.6203, + "step": 436 + }, + { + "epoch": 2.9527027027027026, + "grad_norm": 0.11376422264610185, + "learning_rate": 3.435834672300324e-05, + "loss": 0.6057, + "step": 437 + }, + { + "epoch": 2.9594594594594597, + "grad_norm": 0.14208820005659473, + "learning_rate": 3.417161208135155e-05, + "loss": 0.6193, + "step": 438 + }, + { + "epoch": 2.9662162162162162, + "grad_norm": 0.11247340517088436, + "learning_rate": 3.398500712744254e-05, + "loss": 0.6218, + "step": 439 + }, + { + "epoch": 2.972972972972973, + "grad_norm": 0.12141011100880744, + "learning_rate": 3.379853601343209e-05, + "loss": 0.6124, + "step": 440 + }, + { + "epoch": 2.97972972972973, + "grad_norm": 0.11441162073496848, + "learning_rate": 3.361220288849804e-05, + "loss": 0.6146, + "step": 441 + }, + { + "epoch": 2.9864864864864864, + "grad_norm": 0.11128267132483809, + "learning_rate": 3.342601189874777e-05, + "loss": 0.6141, + "step": 442 + }, + { + "epoch": 2.993243243243243, + "grad_norm": 0.13308500191665273, + "learning_rate": 3.323996718712605e-05, + "loss": 0.6154, + "step": 443 + }, + { + "epoch": 3.0, + "grad_norm": 0.13634885719680506, + "learning_rate": 3.305407289332279e-05, + "loss": 0.6063, + "step": 444 + }, + { + "epoch": 3.0067567567567566, + "grad_norm": 0.14117643344569666, + "learning_rate": 3.2868333153680964e-05, + "loss": 0.5952, + "step": 445 + }, + { + "epoch": 3.0135135135135136, + "grad_norm": 0.1441355574419299, + "learning_rate": 3.2682752101104536e-05, + "loss": 0.5807, + "step": 446 + }, + { + "epoch": 3.02027027027027, + "grad_norm": 0.13317791462913253, + "learning_rate": 3.249733386496653e-05, + "loss": 0.6011, + "step": 447 + }, + { + "epoch": 3.027027027027027, + "grad_norm": 0.1488227563399244, + "learning_rate": 3.231208257101709e-05, + "loss": 0.6055, + "step": 448 + }, + { + "epoch": 3.0337837837837838, + "grad_norm": 0.13698170944523078, + "learning_rate": 3.212700234129179e-05, + "loss": 0.6012, + "step": 449 + }, + { + "epoch": 3.0405405405405403, + "grad_norm": 0.13426258482203746, + "learning_rate": 3.194209729401979e-05, + "loss": 0.5955, + "step": 450 + }, + { + "epoch": 3.0472972972972974, + "grad_norm": 0.12763528667729904, + "learning_rate": 3.175737154353231e-05, + "loss": 0.5829, + "step": 451 + }, + { + "epoch": 3.054054054054054, + "grad_norm": 0.12302649236306719, + "learning_rate": 3.157282920017096e-05, + "loss": 0.6031, + "step": 452 + }, + { + "epoch": 3.060810810810811, + "grad_norm": 0.12836088172431226, + "learning_rate": 3.1388474370196395e-05, + "loss": 0.6029, + "step": 453 + }, + { + "epoch": 3.0675675675675675, + "grad_norm": 0.12713682511389784, + "learning_rate": 3.12043111556969e-05, + "loss": 0.5898, + "step": 454 + }, + { + "epoch": 3.074324324324324, + "grad_norm": 0.13954610740683865, + "learning_rate": 3.1020343654497096e-05, + "loss": 0.5882, + "step": 455 + }, + { + "epoch": 3.081081081081081, + "grad_norm": 0.14394153328876694, + "learning_rate": 3.083657596006681e-05, + "loss": 0.5878, + "step": 456 + }, + { + "epoch": 3.0878378378378377, + "grad_norm": 0.11738094672449853, + "learning_rate": 3.065301216142991e-05, + "loss": 0.5916, + "step": 457 + }, + { + "epoch": 3.0945945945945947, + "grad_norm": 0.1473396296692266, + "learning_rate": 3.046965634307341e-05, + "loss": 0.6003, + "step": 458 + }, + { + "epoch": 3.1013513513513513, + "grad_norm": 0.12128542181194722, + "learning_rate": 3.028651258485652e-05, + "loss": 0.5807, + "step": 459 + }, + { + "epoch": 3.108108108108108, + "grad_norm": 0.12076772408917641, + "learning_rate": 3.010358496191991e-05, + "loss": 0.5968, + "step": 460 + }, + { + "epoch": 3.114864864864865, + "grad_norm": 0.12207943508497456, + "learning_rate": 2.9920877544595002e-05, + "loss": 0.5911, + "step": 461 + }, + { + "epoch": 3.1216216216216215, + "grad_norm": 0.1193967138970814, + "learning_rate": 2.9738394398313405e-05, + "loss": 0.5867, + "step": 462 + }, + { + "epoch": 3.1283783783783785, + "grad_norm": 0.11942196780407818, + "learning_rate": 2.955613958351647e-05, + "loss": 0.5927, + "step": 463 + }, + { + "epoch": 3.135135135135135, + "grad_norm": 0.12137977118056532, + "learning_rate": 2.9374117155564957e-05, + "loss": 0.5923, + "step": 464 + }, + { + "epoch": 3.141891891891892, + "grad_norm": 0.1204663577707489, + "learning_rate": 2.919233116464872e-05, + "loss": 0.5807, + "step": 465 + }, + { + "epoch": 3.1486486486486487, + "grad_norm": 0.1277645131662919, + "learning_rate": 2.9010785655696698e-05, + "loss": 0.602, + "step": 466 + }, + { + "epoch": 3.1554054054054053, + "grad_norm": 0.14373007650684674, + "learning_rate": 2.88294846682868e-05, + "loss": 0.5952, + "step": 467 + }, + { + "epoch": 3.1621621621621623, + "grad_norm": 0.12531967044629988, + "learning_rate": 2.864843223655613e-05, + "loss": 0.6065, + "step": 468 + }, + { + "epoch": 3.168918918918919, + "grad_norm": 0.1344060345828822, + "learning_rate": 2.8467632389111126e-05, + "loss": 0.5879, + "step": 469 + }, + { + "epoch": 3.175675675675676, + "grad_norm": 0.11307986878514827, + "learning_rate": 2.828708914893799e-05, + "loss": 0.5923, + "step": 470 + }, + { + "epoch": 3.1824324324324325, + "grad_norm": 0.1205066746149822, + "learning_rate": 2.8106806533313106e-05, + "loss": 0.6042, + "step": 471 + }, + { + "epoch": 3.189189189189189, + "grad_norm": 0.11358248639127554, + "learning_rate": 2.7926788553713734e-05, + "loss": 0.6017, + "step": 472 + }, + { + "epoch": 3.195945945945946, + "grad_norm": 0.11273620191303046, + "learning_rate": 2.7747039215728667e-05, + "loss": 0.5909, + "step": 473 + }, + { + "epoch": 3.2027027027027026, + "grad_norm": 0.1527832963288818, + "learning_rate": 2.7567562518969155e-05, + "loss": 0.5884, + "step": 474 + }, + { + "epoch": 3.2094594594594597, + "grad_norm": 0.1251866201941187, + "learning_rate": 2.7388362456979906e-05, + "loss": 0.5942, + "step": 475 + }, + { + "epoch": 3.2162162162162162, + "grad_norm": 0.131897788334417, + "learning_rate": 2.7209443017150193e-05, + "loss": 0.6028, + "step": 476 + }, + { + "epoch": 3.222972972972973, + "grad_norm": 0.10949010701632424, + "learning_rate": 2.703080818062517e-05, + "loss": 0.5899, + "step": 477 + }, + { + "epoch": 3.22972972972973, + "grad_norm": 0.1085593433907961, + "learning_rate": 2.6852461922217253e-05, + "loss": 0.5915, + "step": 478 + }, + { + "epoch": 3.2364864864864864, + "grad_norm": 0.10851474030600673, + "learning_rate": 2.66744082103177e-05, + "loss": 0.5956, + "step": 479 + }, + { + "epoch": 3.2432432432432434, + "grad_norm": 0.10556652124147405, + "learning_rate": 2.6496651006808308e-05, + "loss": 0.5923, + "step": 480 + }, + { + "epoch": 3.25, + "grad_norm": 0.11511281568946971, + "learning_rate": 2.6319194266973256e-05, + "loss": 0.5982, + "step": 481 + }, + { + "epoch": 3.2567567567567566, + "grad_norm": 0.10661745620059551, + "learning_rate": 2.614204193941107e-05, + "loss": 0.6029, + "step": 482 + }, + { + "epoch": 3.2635135135135136, + "grad_norm": 0.12822931878201987, + "learning_rate": 2.5965197965946783e-05, + "loss": 0.5935, + "step": 483 + }, + { + "epoch": 3.27027027027027, + "grad_norm": 0.1156142465379123, + "learning_rate": 2.5788666281544258e-05, + "loss": 0.6142, + "step": 484 + }, + { + "epoch": 3.277027027027027, + "grad_norm": 0.13148336347712752, + "learning_rate": 2.561245081421857e-05, + "loss": 0.5942, + "step": 485 + }, + { + "epoch": 3.2837837837837838, + "grad_norm": 0.10123146621891457, + "learning_rate": 2.5436555484948643e-05, + "loss": 0.5989, + "step": 486 + }, + { + "epoch": 3.2905405405405403, + "grad_norm": 0.1375355982170109, + "learning_rate": 2.5260984207590015e-05, + "loss": 0.5913, + "step": 487 + }, + { + "epoch": 3.2972972972972974, + "grad_norm": 0.1071253679793487, + "learning_rate": 2.5085740888787662e-05, + "loss": 0.6034, + "step": 488 + }, + { + "epoch": 3.304054054054054, + "grad_norm": 0.13183672834367724, + "learning_rate": 2.4910829427889205e-05, + "loss": 0.6016, + "step": 489 + }, + { + "epoch": 3.310810810810811, + "grad_norm": 0.11417300702406326, + "learning_rate": 2.473625371685806e-05, + "loss": 0.5964, + "step": 490 + }, + { + "epoch": 3.3175675675675675, + "grad_norm": 0.11783207513787236, + "learning_rate": 2.4562017640186847e-05, + "loss": 0.5983, + "step": 491 + }, + { + "epoch": 3.3243243243243246, + "grad_norm": 0.12341779828696402, + "learning_rate": 2.4388125074810986e-05, + "loss": 0.5925, + "step": 492 + }, + { + "epoch": 3.331081081081081, + "grad_norm": 0.11637168120668157, + "learning_rate": 2.4214579890022373e-05, + "loss": 0.5771, + "step": 493 + }, + { + "epoch": 3.3378378378378377, + "grad_norm": 0.10923183772815541, + "learning_rate": 2.404138594738335e-05, + "loss": 0.5919, + "step": 494 + }, + { + "epoch": 3.3445945945945947, + "grad_norm": 0.10526844464204445, + "learning_rate": 2.386854710064075e-05, + "loss": 0.607, + "step": 495 + }, + { + "epoch": 3.3513513513513513, + "grad_norm": 0.10474734642926019, + "learning_rate": 2.369606719564015e-05, + "loss": 0.5907, + "step": 496 + }, + { + "epoch": 3.358108108108108, + "grad_norm": 0.11107800655708087, + "learning_rate": 2.35239500702403e-05, + "loss": 0.5981, + "step": 497 + }, + { + "epoch": 3.364864864864865, + "grad_norm": 0.10187850179282668, + "learning_rate": 2.3352199554227698e-05, + "loss": 0.5938, + "step": 498 + }, + { + "epoch": 3.3716216216216215, + "grad_norm": 0.12632250115529653, + "learning_rate": 2.318081946923144e-05, + "loss": 0.6117, + "step": 499 + }, + { + "epoch": 3.3783783783783785, + "grad_norm": 0.10055327704116511, + "learning_rate": 2.3009813628638085e-05, + "loss": 0.5964, + "step": 500 + }, + { + "epoch": 3.385135135135135, + "grad_norm": 0.11916682604562856, + "learning_rate": 2.283918583750695e-05, + "loss": 0.5871, + "step": 501 + }, + { + "epoch": 3.391891891891892, + "grad_norm": 0.1010245925435501, + "learning_rate": 2.266893989248527e-05, + "loss": 0.6004, + "step": 502 + }, + { + "epoch": 3.3986486486486487, + "grad_norm": 0.10446456567129823, + "learning_rate": 2.2499079581723846e-05, + "loss": 0.5942, + "step": 503 + }, + { + "epoch": 3.4054054054054053, + "grad_norm": 0.10099772014538502, + "learning_rate": 2.2329608684792676e-05, + "loss": 0.5956, + "step": 504 + }, + { + "epoch": 3.4121621621621623, + "grad_norm": 0.1005267977448251, + "learning_rate": 2.216053097259697e-05, + "loss": 0.5898, + "step": 505 + }, + { + "epoch": 3.418918918918919, + "grad_norm": 0.10513240417556334, + "learning_rate": 2.1991850207293064e-05, + "loss": 0.5997, + "step": 506 + }, + { + "epoch": 3.4256756756756754, + "grad_norm": 0.09929924802833451, + "learning_rate": 2.1823570142204902e-05, + "loss": 0.5932, + "step": 507 + }, + { + "epoch": 3.4324324324324325, + "grad_norm": 0.10494976279274904, + "learning_rate": 2.1655694521740376e-05, + "loss": 0.5935, + "step": 508 + }, + { + "epoch": 3.439189189189189, + "grad_norm": 0.10732590635113842, + "learning_rate": 2.1488227081308054e-05, + "loss": 0.589, + "step": 509 + }, + { + "epoch": 3.445945945945946, + "grad_norm": 0.1154273309217794, + "learning_rate": 2.132117154723408e-05, + "loss": 0.5927, + "step": 510 + }, + { + "epoch": 3.4527027027027026, + "grad_norm": 0.10250291685305624, + "learning_rate": 2.115453163667929e-05, + "loss": 0.6047, + "step": 511 + }, + { + "epoch": 3.4594594594594597, + "grad_norm": 0.1053185431025155, + "learning_rate": 2.0988311057556397e-05, + "loss": 0.5841, + "step": 512 + }, + { + "epoch": 3.4662162162162162, + "grad_norm": 0.09414135194355852, + "learning_rate": 2.0822513508447608e-05, + "loss": 0.5863, + "step": 513 + }, + { + "epoch": 3.472972972972973, + "grad_norm": 0.10071155506936441, + "learning_rate": 2.065714267852223e-05, + "loss": 0.5923, + "step": 514 + }, + { + "epoch": 3.47972972972973, + "grad_norm": 0.10121850558468308, + "learning_rate": 2.049220224745463e-05, + "loss": 0.5962, + "step": 515 + }, + { + "epoch": 3.4864864864864864, + "grad_norm": 0.10161225481650848, + "learning_rate": 2.032769588534233e-05, + "loss": 0.5925, + "step": 516 + }, + { + "epoch": 3.4932432432432434, + "grad_norm": 0.09239639095614875, + "learning_rate": 2.0163627252624427e-05, + "loss": 0.5944, + "step": 517 + }, + { + "epoch": 3.5, + "grad_norm": 0.11360448222663692, + "learning_rate": 2.0000000000000012e-05, + "loss": 0.6116, + "step": 518 + }, + { + "epoch": 3.506756756756757, + "grad_norm": 0.09631002455489597, + "learning_rate": 1.9836817768347015e-05, + "loss": 0.5947, + "step": 519 + }, + { + "epoch": 3.5135135135135136, + "grad_norm": 0.1167172974336559, + "learning_rate": 1.9674084188641235e-05, + "loss": 0.5843, + "step": 520 + }, + { + "epoch": 3.52027027027027, + "grad_norm": 0.10349743444704453, + "learning_rate": 1.9511802881875438e-05, + "loss": 0.6026, + "step": 521 + }, + { + "epoch": 3.527027027027027, + "grad_norm": 0.11753598782061575, + "learning_rate": 1.9349977458978846e-05, + "loss": 0.5892, + "step": 522 + }, + { + "epoch": 3.5337837837837838, + "grad_norm": 0.09146602359405204, + "learning_rate": 1.9188611520736846e-05, + "loss": 0.5916, + "step": 523 + }, + { + "epoch": 3.5405405405405403, + "grad_norm": 0.10892529356211986, + "learning_rate": 1.902770865771074e-05, + "loss": 0.5854, + "step": 524 + }, + { + "epoch": 3.5472972972972974, + "grad_norm": 0.09418036132529382, + "learning_rate": 1.886727245015794e-05, + "loss": 0.5993, + "step": 525 + }, + { + "epoch": 3.554054054054054, + "grad_norm": 0.1025611805390442, + "learning_rate": 1.8707306467952323e-05, + "loss": 0.5916, + "step": 526 + }, + { + "epoch": 3.560810810810811, + "grad_norm": 0.08852437916251449, + "learning_rate": 1.8547814270504705e-05, + "loss": 0.5946, + "step": 527 + }, + { + "epoch": 3.5675675675675675, + "grad_norm": 0.0999923376365888, + "learning_rate": 1.838879940668373e-05, + "loss": 0.6, + "step": 528 + }, + { + "epoch": 3.5743243243243246, + "grad_norm": 0.10873141849605454, + "learning_rate": 1.823026541473684e-05, + "loss": 0.6017, + "step": 529 + }, + { + "epoch": 3.581081081081081, + "grad_norm": 0.10599260470783008, + "learning_rate": 1.8072215822211613e-05, + "loss": 0.5968, + "step": 530 + }, + { + "epoch": 3.5878378378378377, + "grad_norm": 0.11188333764585037, + "learning_rate": 1.7914654145877187e-05, + "loss": 0.6064, + "step": 531 + }, + { + "epoch": 3.5945945945945947, + "grad_norm": 0.09583819123841314, + "learning_rate": 1.77575838916461e-05, + "loss": 0.5876, + "step": 532 + }, + { + "epoch": 3.6013513513513513, + "grad_norm": 0.10206469144841168, + "learning_rate": 1.760100855449619e-05, + "loss": 0.5948, + "step": 533 + }, + { + "epoch": 3.608108108108108, + "grad_norm": 0.10036289538603702, + "learning_rate": 1.7444931618392894e-05, + "loss": 0.6107, + "step": 534 + }, + { + "epoch": 3.614864864864865, + "grad_norm": 0.08314263844698841, + "learning_rate": 1.7289356556211687e-05, + "loss": 0.6059, + "step": 535 + }, + { + "epoch": 3.6216216216216215, + "grad_norm": 0.10484951196819213, + "learning_rate": 1.7134286829660855e-05, + "loss": 0.5961, + "step": 536 + }, + { + "epoch": 3.6283783783783785, + "grad_norm": 0.07757732001833932, + "learning_rate": 1.697972588920439e-05, + "loss": 0.603, + "step": 537 + }, + { + "epoch": 3.635135135135135, + "grad_norm": 0.09636592692846442, + "learning_rate": 1.6825677173985332e-05, + "loss": 0.597, + "step": 538 + }, + { + "epoch": 3.641891891891892, + "grad_norm": 0.0934961698224681, + "learning_rate": 1.6672144111749066e-05, + "loss": 0.6065, + "step": 539 + }, + { + "epoch": 3.6486486486486487, + "grad_norm": 0.0956796061229755, + "learning_rate": 1.6519130118767258e-05, + "loss": 0.5978, + "step": 540 + }, + { + "epoch": 3.6554054054054053, + "grad_norm": 0.1161980558323059, + "learning_rate": 1.6366638599761676e-05, + "loss": 0.5914, + "step": 541 + }, + { + "epoch": 3.6621621621621623, + "grad_norm": 0.08312306306871219, + "learning_rate": 1.621467294782854e-05, + "loss": 0.5855, + "step": 542 + }, + { + "epoch": 3.668918918918919, + "grad_norm": 0.11701156024970683, + "learning_rate": 1.606323654436293e-05, + "loss": 0.5904, + "step": 543 + }, + { + "epoch": 3.6756756756756754, + "grad_norm": 0.09466465341207426, + "learning_rate": 1.591233275898363e-05, + "loss": 0.5915, + "step": 544 + }, + { + "epoch": 3.6824324324324325, + "grad_norm": 0.10284506462525071, + "learning_rate": 1.5761964949458076e-05, + "loss": 0.5847, + "step": 545 + }, + { + "epoch": 3.689189189189189, + "grad_norm": 0.10382209320021844, + "learning_rate": 1.5612136461627726e-05, + "loss": 0.596, + "step": 546 + }, + { + "epoch": 3.695945945945946, + "grad_norm": 0.09709377064685741, + "learning_rate": 1.546285062933352e-05, + "loss": 0.5909, + "step": 547 + }, + { + "epoch": 3.7027027027027026, + "grad_norm": 0.11272439075208067, + "learning_rate": 1.5314110774341803e-05, + "loss": 0.5953, + "step": 548 + }, + { + "epoch": 3.7094594594594597, + "grad_norm": 0.11904326038877579, + "learning_rate": 1.5165920206270257e-05, + "loss": 0.5929, + "step": 549 + }, + { + "epoch": 3.7162162162162162, + "grad_norm": 0.1039606133991536, + "learning_rate": 1.5018282222514451e-05, + "loss": 0.5877, + "step": 550 + }, + { + "epoch": 3.722972972972973, + "grad_norm": 0.11190632151336002, + "learning_rate": 1.4871200108174306e-05, + "loss": 0.5917, + "step": 551 + }, + { + "epoch": 3.72972972972973, + "grad_norm": 0.09272507731002902, + "learning_rate": 1.4724677135981118e-05, + "loss": 0.5959, + "step": 552 + }, + { + "epoch": 3.7364864864864864, + "grad_norm": 0.09431866458492948, + "learning_rate": 1.457871656622463e-05, + "loss": 0.5887, + "step": 553 + }, + { + "epoch": 3.743243243243243, + "grad_norm": 0.08680813466014699, + "learning_rate": 1.4433321646680614e-05, + "loss": 0.6032, + "step": 554 + }, + { + "epoch": 3.75, + "grad_norm": 0.0844092691423774, + "learning_rate": 1.4288495612538427e-05, + "loss": 0.6062, + "step": 555 + }, + { + "epoch": 3.756756756756757, + "grad_norm": 0.0874893569172525, + "learning_rate": 1.4144241686329236e-05, + "loss": 0.5833, + "step": 556 + }, + { + "epoch": 3.7635135135135136, + "grad_norm": 0.10957613038759607, + "learning_rate": 1.400056307785413e-05, + "loss": 0.6056, + "step": 557 + }, + { + "epoch": 3.77027027027027, + "grad_norm": 0.07882052024046089, + "learning_rate": 1.3857462984112831e-05, + "loss": 0.5912, + "step": 558 + }, + { + "epoch": 3.777027027027027, + "grad_norm": 0.08565900595306047, + "learning_rate": 1.371494458923246e-05, + "loss": 0.5919, + "step": 559 + }, + { + "epoch": 3.7837837837837838, + "grad_norm": 0.09674275812299338, + "learning_rate": 1.3573011064396751e-05, + "loss": 0.5985, + "step": 560 + }, + { + "epoch": 3.7905405405405403, + "grad_norm": 0.08500752883113108, + "learning_rate": 1.3431665567775439e-05, + "loss": 0.5983, + "step": 561 + }, + { + "epoch": 3.7972972972972974, + "grad_norm": 0.11583654081493171, + "learning_rate": 1.3290911244454066e-05, + "loss": 0.5903, + "step": 562 + }, + { + "epoch": 3.804054054054054, + "grad_norm": 0.07940239999432974, + "learning_rate": 1.3150751226363886e-05, + "loss": 0.5931, + "step": 563 + }, + { + "epoch": 3.810810810810811, + "grad_norm": 0.10093820387387048, + "learning_rate": 1.3011188632212307e-05, + "loss": 0.6033, + "step": 564 + }, + { + "epoch": 3.8175675675675675, + "grad_norm": 0.0890166406955722, + "learning_rate": 1.2872226567413346e-05, + "loss": 0.5924, + "step": 565 + }, + { + "epoch": 3.8243243243243246, + "grad_norm": 0.07708649898799794, + "learning_rate": 1.2733868124018694e-05, + "loss": 0.5891, + "step": 566 + }, + { + "epoch": 3.831081081081081, + "grad_norm": 0.11011562257155211, + "learning_rate": 1.2596116380648761e-05, + "loss": 0.5898, + "step": 567 + }, + { + "epoch": 3.8378378378378377, + "grad_norm": 0.07996964493884025, + "learning_rate": 1.2458974402424312e-05, + "loss": 0.5937, + "step": 568 + }, + { + "epoch": 3.8445945945945947, + "grad_norm": 0.09078629614798739, + "learning_rate": 1.2322445240898158e-05, + "loss": 0.5908, + "step": 569 + }, + { + "epoch": 3.8513513513513513, + "grad_norm": 0.08453528616636899, + "learning_rate": 1.2186531933987294e-05, + "loss": 0.5895, + "step": 570 + }, + { + "epoch": 3.858108108108108, + "grad_norm": 0.08207256114306587, + "learning_rate": 1.2051237505905302e-05, + "loss": 0.6, + "step": 571 + }, + { + "epoch": 3.864864864864865, + "grad_norm": 0.08328736066936966, + "learning_rate": 1.19165649670951e-05, + "loss": 0.5901, + "step": 572 + }, + { + "epoch": 3.8716216216216215, + "grad_norm": 0.0796730356081974, + "learning_rate": 1.1782517314161872e-05, + "loss": 0.5825, + "step": 573 + }, + { + "epoch": 3.8783783783783785, + "grad_norm": 0.09614075098109777, + "learning_rate": 1.164909752980648e-05, + "loss": 0.5949, + "step": 574 + }, + { + "epoch": 3.885135135135135, + "grad_norm": 0.08310475221153353, + "learning_rate": 1.1516308582758983e-05, + "loss": 0.604, + "step": 575 + }, + { + "epoch": 3.891891891891892, + "grad_norm": 0.08282668808548053, + "learning_rate": 1.1384153427712729e-05, + "loss": 0.6002, + "step": 576 + }, + { + "epoch": 3.8986486486486487, + "grad_norm": 0.0986705467432139, + "learning_rate": 1.1252635005258466e-05, + "loss": 0.5853, + "step": 577 + }, + { + "epoch": 3.9054054054054053, + "grad_norm": 0.08062037787963591, + "learning_rate": 1.1121756241819023e-05, + "loss": 0.6047, + "step": 578 + }, + { + "epoch": 3.9121621621621623, + "grad_norm": 0.07523397720678549, + "learning_rate": 1.0991520049584112e-05, + "loss": 0.5842, + "step": 579 + }, + { + "epoch": 3.918918918918919, + "grad_norm": 0.08533444706267539, + "learning_rate": 1.0861929326445572e-05, + "loss": 0.5952, + "step": 580 + }, + { + "epoch": 3.9256756756756754, + "grad_norm": 0.0867340533978853, + "learning_rate": 1.0732986955932869e-05, + "loss": 0.5978, + "step": 581 + }, + { + "epoch": 3.9324324324324325, + "grad_norm": 0.06754168833386048, + "learning_rate": 1.0604695807148971e-05, + "loss": 0.5987, + "step": 582 + }, + { + "epoch": 3.939189189189189, + "grad_norm": 0.0796662865647217, + "learning_rate": 1.0477058734706436e-05, + "loss": 0.588, + "step": 583 + }, + { + "epoch": 3.945945945945946, + "grad_norm": 0.0813837559182795, + "learning_rate": 1.0350078578664005e-05, + "loss": 0.6009, + "step": 584 + }, + { + "epoch": 3.9527027027027026, + "grad_norm": 0.06912060191449508, + "learning_rate": 1.0223758164463246e-05, + "loss": 0.6055, + "step": 585 + }, + { + "epoch": 3.9594594594594597, + "grad_norm": 0.06784642823506538, + "learning_rate": 1.0098100302865865e-05, + "loss": 0.5882, + "step": 586 + }, + { + "epoch": 3.9662162162162162, + "grad_norm": 0.0760135339338161, + "learning_rate": 9.973107789891024e-06, + "loss": 0.596, + "step": 587 + }, + { + "epoch": 3.972972972972973, + "grad_norm": 0.07657252850751217, + "learning_rate": 9.848783406753224e-06, + "loss": 0.5962, + "step": 588 + }, + { + "epoch": 3.97972972972973, + "grad_norm": 0.08050666474861323, + "learning_rate": 9.725129919800339e-06, + "loss": 0.5957, + "step": 589 + }, + { + "epoch": 3.9864864864864864, + "grad_norm": 0.0778632825822388, + "learning_rate": 9.60215008045211e-06, + "loss": 0.5952, + "step": 590 + }, + { + "epoch": 3.993243243243243, + "grad_norm": 0.07984922923288139, + "learning_rate": 9.479846625138909e-06, + "loss": 0.6023, + "step": 591 + }, + { + "epoch": 4.0, + "grad_norm": 0.0916488655332951, + "learning_rate": 9.358222275240884e-06, + "loss": 0.5842, + "step": 592 + }, + { + "epoch": 4.006756756756757, + "grad_norm": 0.08766439343096606, + "learning_rate": 9.237279737027326e-06, + "loss": 0.5851, + "step": 593 + }, + { + "epoch": 4.013513513513513, + "grad_norm": 0.07964740730111751, + "learning_rate": 9.117021701596567e-06, + "loss": 0.5776, + "step": 594 + }, + { + "epoch": 4.02027027027027, + "grad_norm": 0.08933138991359177, + "learning_rate": 8.99745084481594e-06, + "loss": 0.5953, + "step": 595 + }, + { + "epoch": 4.027027027027027, + "grad_norm": 0.07914528740585995, + "learning_rate": 8.87856982726243e-06, + "loss": 0.5892, + "step": 596 + }, + { + "epoch": 4.033783783783784, + "grad_norm": 0.07297723048835614, + "learning_rate": 8.7603812941633e-06, + "loss": 0.5823, + "step": 597 + }, + { + "epoch": 4.04054054054054, + "grad_norm": 0.10229836912930287, + "learning_rate": 8.642887875337376e-06, + "loss": 0.5767, + "step": 598 + }, + { + "epoch": 4.047297297297297, + "grad_norm": 0.0781799178126612, + "learning_rate": 8.526092185136394e-06, + "loss": 0.5784, + "step": 599 + }, + { + "epoch": 4.054054054054054, + "grad_norm": 0.09402345888004014, + "learning_rate": 8.409996822386972e-06, + "loss": 0.5828, + "step": 600 + }, + { + "epoch": 4.0608108108108105, + "grad_norm": 0.07679009840525959, + "learning_rate": 8.294604370332613e-06, + "loss": 0.5754, + "step": 601 + }, + { + "epoch": 4.0675675675675675, + "grad_norm": 0.07923822298946714, + "learning_rate": 8.17991739657641e-06, + "loss": 0.573, + "step": 602 + }, + { + "epoch": 4.074324324324325, + "grad_norm": 0.08082893460744046, + "learning_rate": 8.06593845302376e-06, + "loss": 0.5916, + "step": 603 + }, + { + "epoch": 4.081081081081081, + "grad_norm": 0.07578230123409459, + "learning_rate": 7.952670075825702e-06, + "loss": 0.5851, + "step": 604 + }, + { + "epoch": 4.087837837837838, + "grad_norm": 0.07539180401042113, + "learning_rate": 7.840114785322384e-06, + "loss": 0.5939, + "step": 605 + }, + { + "epoch": 4.094594594594595, + "grad_norm": 0.07932706141071251, + "learning_rate": 7.728275085987041e-06, + "loss": 0.5908, + "step": 606 + }, + { + "epoch": 4.101351351351352, + "grad_norm": 0.07344646624670495, + "learning_rate": 7.6171534663702416e-06, + "loss": 0.5915, + "step": 607 + }, + { + "epoch": 4.108108108108108, + "grad_norm": 0.07204192586718106, + "learning_rate": 7.50675239904457e-06, + "loss": 0.5855, + "step": 608 + }, + { + "epoch": 4.114864864864865, + "grad_norm": 0.06717242761020562, + "learning_rate": 7.397074340549508e-06, + "loss": 0.5919, + "step": 609 + }, + { + "epoch": 4.121621621621622, + "grad_norm": 0.07615198090948691, + "learning_rate": 7.288121731336901e-06, + "loss": 0.5883, + "step": 610 + }, + { + "epoch": 4.128378378378378, + "grad_norm": 0.07186386521547546, + "learning_rate": 7.1798969957165025e-06, + "loss": 0.5666, + "step": 611 + }, + { + "epoch": 4.135135135135135, + "grad_norm": 0.07771245688676058, + "learning_rate": 7.072402541802197e-06, + "loss": 0.572, + "step": 612 + }, + { + "epoch": 4.141891891891892, + "grad_norm": 0.06560942291485741, + "learning_rate": 6.965640761458274e-06, + "loss": 0.5853, + "step": 613 + }, + { + "epoch": 4.148648648648648, + "grad_norm": 0.07554536141277478, + "learning_rate": 6.859614030246318e-06, + "loss": 0.5746, + "step": 614 + }, + { + "epoch": 4.155405405405405, + "grad_norm": 0.0766866599574075, + "learning_rate": 6.754324707372264e-06, + "loss": 0.5827, + "step": 615 + }, + { + "epoch": 4.162162162162162, + "grad_norm": 0.0685589990578116, + "learning_rate": 6.649775135633944e-06, + "loss": 0.5708, + "step": 616 + }, + { + "epoch": 4.168918918918919, + "grad_norm": 0.06830496975914782, + "learning_rate": 6.545967641368958e-06, + "loss": 0.5772, + "step": 617 + }, + { + "epoch": 4.175675675675675, + "grad_norm": 0.06792930742316021, + "learning_rate": 6.4429045344029136e-06, + "loss": 0.5801, + "step": 618 + }, + { + "epoch": 4.1824324324324325, + "grad_norm": 0.0728029236733438, + "learning_rate": 6.340588107997994e-06, + "loss": 0.5666, + "step": 619 + }, + { + "epoch": 4.1891891891891895, + "grad_norm": 0.07156462363762543, + "learning_rate": 6.239020638801987e-06, + "loss": 0.5848, + "step": 620 + }, + { + "epoch": 4.195945945945946, + "grad_norm": 0.06757664374848818, + "learning_rate": 6.1382043867975836e-06, + "loss": 0.5679, + "step": 621 + }, + { + "epoch": 4.202702702702703, + "grad_norm": 0.07016074613879056, + "learning_rate": 6.038141595252094e-06, + "loss": 0.5839, + "step": 622 + }, + { + "epoch": 4.20945945945946, + "grad_norm": 0.06809385798805276, + "learning_rate": 5.9388344906675485e-06, + "loss": 0.575, + "step": 623 + }, + { + "epoch": 4.216216216216216, + "grad_norm": 0.061129888925191694, + "learning_rate": 5.840285282731173e-06, + "loss": 0.5914, + "step": 624 + }, + { + "epoch": 4.222972972972973, + "grad_norm": 0.0728841561672064, + "learning_rate": 5.742496164266174e-06, + "loss": 0.578, + "step": 625 + }, + { + "epoch": 4.22972972972973, + "grad_norm": 0.07404139765077528, + "learning_rate": 5.645469311182958e-06, + "loss": 0.5788, + "step": 626 + }, + { + "epoch": 4.236486486486487, + "grad_norm": 0.06599738371244251, + "learning_rate": 5.549206882430773e-06, + "loss": 0.5938, + "step": 627 + }, + { + "epoch": 4.243243243243243, + "grad_norm": 0.07571258452222435, + "learning_rate": 5.453711019949581e-06, + "loss": 0.5761, + "step": 628 + }, + { + "epoch": 4.25, + "grad_norm": 0.06873039122654918, + "learning_rate": 5.358983848622452e-06, + "loss": 0.5865, + "step": 629 + }, + { + "epoch": 4.256756756756757, + "grad_norm": 0.05980111880094078, + "learning_rate": 5.265027476228297e-06, + "loss": 0.58, + "step": 630 + }, + { + "epoch": 4.263513513513513, + "grad_norm": 0.07210626853825096, + "learning_rate": 5.171843993394903e-06, + "loss": 0.5756, + "step": 631 + }, + { + "epoch": 4.27027027027027, + "grad_norm": 0.07396976700448737, + "learning_rate": 5.079435473552474e-06, + "loss": 0.5811, + "step": 632 + }, + { + "epoch": 4.277027027027027, + "grad_norm": 0.09171081358728068, + "learning_rate": 4.987803972887482e-06, + "loss": 0.5871, + "step": 633 + }, + { + "epoch": 4.283783783783784, + "grad_norm": 0.06701824447243687, + "learning_rate": 4.896951530296896e-06, + "loss": 0.579, + "step": 634 + }, + { + "epoch": 4.29054054054054, + "grad_norm": 0.07518459749761504, + "learning_rate": 4.806880167342831e-06, + "loss": 0.5822, + "step": 635 + }, + { + "epoch": 4.297297297297297, + "grad_norm": 0.07098354334790964, + "learning_rate": 4.7175918882075465e-06, + "loss": 0.5861, + "step": 636 + }, + { + "epoch": 4.304054054054054, + "grad_norm": 0.07479138842797883, + "learning_rate": 4.6290886796488946e-06, + "loss": 0.5863, + "step": 637 + }, + { + "epoch": 4.3108108108108105, + "grad_norm": 0.06686921193950773, + "learning_rate": 4.54137251095605e-06, + "loss": 0.5916, + "step": 638 + }, + { + "epoch": 4.3175675675675675, + "grad_norm": 0.07897960962457094, + "learning_rate": 4.454445333905768e-06, + "loss": 0.5948, + "step": 639 + }, + { + "epoch": 4.324324324324325, + "grad_norm": 0.07496951321971122, + "learning_rate": 4.3683090827188666e-06, + "loss": 0.5813, + "step": 640 + }, + { + "epoch": 4.331081081081081, + "grad_norm": 0.07536718259174548, + "learning_rate": 4.282965674017265e-06, + "loss": 0.5872, + "step": 641 + }, + { + "epoch": 4.337837837837838, + "grad_norm": 0.06854859712850378, + "learning_rate": 4.198417006781283e-06, + "loss": 0.5841, + "step": 642 + }, + { + "epoch": 4.344594594594595, + "grad_norm": 0.07618804161631497, + "learning_rate": 4.114664962307439e-06, + "loss": 0.5817, + "step": 643 + }, + { + "epoch": 4.351351351351352, + "grad_norm": 0.07601070844404975, + "learning_rate": 4.031711404166525e-06, + "loss": 0.584, + "step": 644 + }, + { + "epoch": 4.358108108108108, + "grad_norm": 0.06914624176338743, + "learning_rate": 3.949558178162209e-06, + "loss": 0.5888, + "step": 645 + }, + { + "epoch": 4.364864864864865, + "grad_norm": 0.07027655538494314, + "learning_rate": 3.86820711228991e-06, + "loss": 0.5867, + "step": 646 + }, + { + "epoch": 4.371621621621622, + "grad_norm": 0.0760270326599887, + "learning_rate": 3.7876600166961353e-06, + "loss": 0.5961, + "step": 647 + }, + { + "epoch": 4.378378378378378, + "grad_norm": 0.06782313812781374, + "learning_rate": 3.707918683638223e-06, + "loss": 0.5944, + "step": 648 + }, + { + "epoch": 4.385135135135135, + "grad_norm": 0.06192386606423058, + "learning_rate": 3.628984887444462e-06, + "loss": 0.5969, + "step": 649 + }, + { + "epoch": 4.391891891891892, + "grad_norm": 0.062017594322198465, + "learning_rate": 3.550860384474568e-06, + "loss": 0.5863, + "step": 650 + }, + { + "epoch": 4.398648648648648, + "grad_norm": 0.058946604304473224, + "learning_rate": 3.473546913080674e-06, + "loss": 0.5847, + "step": 651 + }, + { + "epoch": 4.405405405405405, + "grad_norm": 0.08803265906901768, + "learning_rate": 3.397046193568558e-06, + "loss": 0.5707, + "step": 652 + }, + { + "epoch": 4.412162162162162, + "grad_norm": 0.06583037908170926, + "learning_rate": 3.3213599281594688e-06, + "loss": 0.582, + "step": 653 + }, + { + "epoch": 4.418918918918919, + "grad_norm": 0.06086938296974132, + "learning_rate": 3.246489800952155e-06, + "loss": 0.5953, + "step": 654 + }, + { + "epoch": 4.425675675675675, + "grad_norm": 0.05932401857518893, + "learning_rate": 3.172437477885475e-06, + "loss": 0.5819, + "step": 655 + }, + { + "epoch": 4.4324324324324325, + "grad_norm": 0.057752799006265246, + "learning_rate": 3.099204606701256e-06, + "loss": 0.5766, + "step": 656 + }, + { + "epoch": 4.4391891891891895, + "grad_norm": 0.06397066535508446, + "learning_rate": 3.026792816907671e-06, + "loss": 0.588, + "step": 657 + }, + { + "epoch": 4.445945945945946, + "grad_norm": 0.06029045959058417, + "learning_rate": 2.955203719742965e-06, + "loss": 0.591, + "step": 658 + }, + { + "epoch": 4.452702702702703, + "grad_norm": 0.05838249627908107, + "learning_rate": 2.884438908139626e-06, + "loss": 0.5929, + "step": 659 + }, + { + "epoch": 4.45945945945946, + "grad_norm": 0.06075718886251464, + "learning_rate": 2.814499956688912e-06, + "loss": 0.5813, + "step": 660 + }, + { + "epoch": 4.466216216216216, + "grad_norm": 0.0719277704750579, + "learning_rate": 2.7453884216058368e-06, + "loss": 0.5795, + "step": 661 + }, + { + "epoch": 4.472972972972973, + "grad_norm": 0.06075719822797747, + "learning_rate": 2.677105840694507e-06, + "loss": 0.5674, + "step": 662 + }, + { + "epoch": 4.47972972972973, + "grad_norm": 0.05975207139447743, + "learning_rate": 2.6096537333139616e-06, + "loss": 0.5805, + "step": 663 + }, + { + "epoch": 4.486486486486487, + "grad_norm": 0.06213586333263577, + "learning_rate": 2.5430336003443045e-06, + "loss": 0.5906, + "step": 664 + }, + { + "epoch": 4.493243243243243, + "grad_norm": 0.05895116376323496, + "learning_rate": 2.4772469241533648e-06, + "loss": 0.5844, + "step": 665 + }, + { + "epoch": 4.5, + "grad_norm": 0.06765046020471, + "learning_rate": 2.4122951685636674e-06, + "loss": 0.5835, + "step": 666 + }, + { + "epoch": 4.506756756756757, + "grad_norm": 0.06650006977809148, + "learning_rate": 2.3481797788198745e-06, + "loss": 0.5789, + "step": 667 + }, + { + "epoch": 4.513513513513513, + "grad_norm": 0.05846137750044436, + "learning_rate": 2.284902181556632e-06, + "loss": 0.5745, + "step": 668 + }, + { + "epoch": 4.52027027027027, + "grad_norm": 0.05989337419115531, + "learning_rate": 2.2224637847668484e-06, + "loss": 0.5733, + "step": 669 + }, + { + "epoch": 4.527027027027027, + "grad_norm": 0.06939319651650724, + "learning_rate": 2.1608659777703033e-06, + "loss": 0.581, + "step": 670 + }, + { + "epoch": 4.533783783783784, + "grad_norm": 0.054679826375991857, + "learning_rate": 2.100110131182813e-06, + "loss": 0.5931, + "step": 671 + }, + { + "epoch": 4.54054054054054, + "grad_norm": 0.05548818045052255, + "learning_rate": 2.0401975968856514e-06, + "loss": 0.5786, + "step": 672 + }, + { + "epoch": 4.547297297297297, + "grad_norm": 0.059389572689147435, + "learning_rate": 1.981129707995542e-06, + "loss": 0.5923, + "step": 673 + }, + { + "epoch": 4.554054054054054, + "grad_norm": 0.06975826873383643, + "learning_rate": 1.9229077788349393e-06, + "loss": 0.577, + "step": 674 + }, + { + "epoch": 4.5608108108108105, + "grad_norm": 0.07220930223960455, + "learning_rate": 1.865533104902828e-06, + "loss": 0.5916, + "step": 675 + }, + { + "epoch": 4.5675675675675675, + "grad_norm": 0.053811020975530956, + "learning_rate": 1.8090069628458583e-06, + "loss": 0.5704, + "step": 676 + }, + { + "epoch": 4.574324324324325, + "grad_norm": 0.05538660298263909, + "learning_rate": 1.7533306104299663e-06, + "loss": 0.5856, + "step": 677 + }, + { + "epoch": 4.581081081081081, + "grad_norm": 0.05829509686058199, + "learning_rate": 1.6985052865123641e-06, + "loss": 0.5754, + "step": 678 + }, + { + "epoch": 4.587837837837838, + "grad_norm": 0.06323909786498662, + "learning_rate": 1.6445322110140116e-06, + "loss": 0.5814, + "step": 679 + }, + { + "epoch": 4.594594594594595, + "grad_norm": 0.056972721072935675, + "learning_rate": 1.59141258489242e-06, + "loss": 0.5756, + "step": 680 + }, + { + "epoch": 4.601351351351351, + "grad_norm": 0.05816167859998497, + "learning_rate": 1.53914759011498e-06, + "loss": 0.5889, + "step": 681 + }, + { + "epoch": 4.608108108108108, + "grad_norm": 0.059044794864098894, + "learning_rate": 1.4877383896326269e-06, + "loss": 0.5781, + "step": 682 + }, + { + "epoch": 4.614864864864865, + "grad_norm": 0.06237941099681927, + "learning_rate": 1.4371861273539778e-06, + "loss": 0.576, + "step": 683 + }, + { + "epoch": 4.621621621621622, + "grad_norm": 0.06531112971211704, + "learning_rate": 1.3874919281198662e-06, + "loss": 0.5799, + "step": 684 + }, + { + "epoch": 4.628378378378378, + "grad_norm": 0.05265860386748638, + "learning_rate": 1.3386568976783453e-06, + "loss": 0.5891, + "step": 685 + }, + { + "epoch": 4.635135135135135, + "grad_norm": 0.05785779557411594, + "learning_rate": 1.2906821226600453e-06, + "loss": 0.5768, + "step": 686 + }, + { + "epoch": 4.641891891891892, + "grad_norm": 0.05639941023635892, + "learning_rate": 1.2435686705540228e-06, + "loss": 0.5842, + "step": 687 + }, + { + "epoch": 4.648648648648649, + "grad_norm": 0.06338285927327728, + "learning_rate": 1.1973175896839684e-06, + "loss": 0.573, + "step": 688 + }, + { + "epoch": 4.655405405405405, + "grad_norm": 0.054671739680034734, + "learning_rate": 1.1519299091849523e-06, + "loss": 0.5774, + "step": 689 + }, + { + "epoch": 4.662162162162162, + "grad_norm": 0.057698821845826635, + "learning_rate": 1.1074066389804395e-06, + "loss": 0.5932, + "step": 690 + }, + { + "epoch": 4.668918918918919, + "grad_norm": 0.05697783615575523, + "learning_rate": 1.0637487697598937e-06, + "loss": 0.5823, + "step": 691 + }, + { + "epoch": 4.675675675675675, + "grad_norm": 0.054051563368746565, + "learning_rate": 1.0209572729566708e-06, + "loss": 0.5787, + "step": 692 + }, + { + "epoch": 4.6824324324324325, + "grad_norm": 0.05344229805841218, + "learning_rate": 9.790331007264543e-07, + "loss": 0.5844, + "step": 693 + }, + { + "epoch": 4.6891891891891895, + "grad_norm": 0.06959576323504683, + "learning_rate": 9.379771859260267e-07, + "loss": 0.5817, + "step": 694 + }, + { + "epoch": 4.695945945945946, + "grad_norm": 0.06318026825084985, + "learning_rate": 8.977904420925543e-07, + "loss": 0.5892, + "step": 695 + }, + { + "epoch": 4.702702702702703, + "grad_norm": 0.056039993579189235, + "learning_rate": 8.584737634232154e-07, + "loss": 0.5835, + "step": 696 + }, + { + "epoch": 4.70945945945946, + "grad_norm": 0.05269309541455996, + "learning_rate": 8.200280247553461e-07, + "loss": 0.5803, + "step": 697 + }, + { + "epoch": 4.716216216216216, + "grad_norm": 0.05510521164357771, + "learning_rate": 7.824540815469306e-07, + "loss": 0.5709, + "step": 698 + }, + { + "epoch": 4.722972972972973, + "grad_norm": 0.05770652073739158, + "learning_rate": 7.457527698576217e-07, + "loss": 0.5855, + "step": 699 + }, + { + "epoch": 4.72972972972973, + "grad_norm": 0.055593016439349656, + "learning_rate": 7.099249063300751e-07, + "loss": 0.5847, + "step": 700 + }, + { + "epoch": 4.736486486486487, + "grad_norm": 0.05730248759901012, + "learning_rate": 6.749712881718306e-07, + "loss": 0.5639, + "step": 701 + }, + { + "epoch": 4.743243243243243, + "grad_norm": 0.05973974788088902, + "learning_rate": 6.408926931375403e-07, + "loss": 0.5808, + "step": 702 + }, + { + "epoch": 4.75, + "grad_norm": 0.05557100230846073, + "learning_rate": 6.076898795116792e-07, + "loss": 0.5848, + "step": 703 + }, + { + "epoch": 4.756756756756757, + "grad_norm": 0.0563081652270169, + "learning_rate": 5.753635860916617e-07, + "loss": 0.5888, + "step": 704 + }, + { + "epoch": 4.763513513513513, + "grad_norm": 0.0536729552491486, + "learning_rate": 5.43914532171419e-07, + "loss": 0.5873, + "step": 705 + }, + { + "epoch": 4.77027027027027, + "grad_norm": 0.049916139412812977, + "learning_rate": 5.133434175253715e-07, + "loss": 0.5792, + "step": 706 + }, + { + "epoch": 4.777027027027027, + "grad_norm": 0.05386272187182479, + "learning_rate": 4.83650922392882e-07, + "loss": 0.5829, + "step": 707 + }, + { + "epoch": 4.783783783783784, + "grad_norm": 0.05503179443312982, + "learning_rate": 4.5483770746309383e-07, + "loss": 0.5873, + "step": 708 + }, + { + "epoch": 4.79054054054054, + "grad_norm": 0.05614293253442416, + "learning_rate": 4.269044138602585e-07, + "loss": 0.5879, + "step": 709 + }, + { + "epoch": 4.797297297297297, + "grad_norm": 0.053484276503556256, + "learning_rate": 3.998516631294491e-07, + "loss": 0.5794, + "step": 710 + }, + { + "epoch": 4.804054054054054, + "grad_norm": 0.055818505320336, + "learning_rate": 3.736800572227317e-07, + "loss": 0.5905, + "step": 711 + }, + { + "epoch": 4.8108108108108105, + "grad_norm": 0.05644903653041531, + "learning_rate": 3.483901784857846e-07, + "loss": 0.5721, + "step": 712 + }, + { + "epoch": 4.8175675675675675, + "grad_norm": 0.04844049723060667, + "learning_rate": 3.239825896449267e-07, + "loss": 0.5826, + "step": 713 + }, + { + "epoch": 4.824324324324325, + "grad_norm": 0.053088243917807275, + "learning_rate": 3.004578337945985e-07, + "loss": 0.5811, + "step": 714 + }, + { + "epoch": 4.831081081081081, + "grad_norm": 0.05140464370774195, + "learning_rate": 2.778164343852918e-07, + "loss": 0.5791, + "step": 715 + }, + { + "epoch": 4.837837837837838, + "grad_norm": 0.05679451911188569, + "learning_rate": 2.5605889521188364e-07, + "loss": 0.5787, + "step": 716 + }, + { + "epoch": 4.844594594594595, + "grad_norm": 0.06082278783885325, + "learning_rate": 2.351857004024316e-07, + "loss": 0.5868, + "step": 717 + }, + { + "epoch": 4.851351351351351, + "grad_norm": 0.05433042972998646, + "learning_rate": 2.1519731440740487e-07, + "loss": 0.5812, + "step": 718 + }, + { + "epoch": 4.858108108108108, + "grad_norm": 0.055224715989184006, + "learning_rate": 1.9609418198935916e-07, + "loss": 0.5881, + "step": 719 + }, + { + "epoch": 4.864864864864865, + "grad_norm": 0.05348313369078817, + "learning_rate": 1.778767282130156e-07, + "loss": 0.5945, + "step": 720 + }, + { + "epoch": 4.871621621621622, + "grad_norm": 0.05163946442780112, + "learning_rate": 1.6054535843582854e-07, + "loss": 0.5891, + "step": 721 + }, + { + "epoch": 4.878378378378378, + "grad_norm": 0.055008800596469655, + "learning_rate": 1.4410045829893915e-07, + "loss": 0.5817, + "step": 722 + }, + { + "epoch": 4.885135135135135, + "grad_norm": 0.05385755694710143, + "learning_rate": 1.2854239371863142e-07, + "loss": 0.5803, + "step": 723 + }, + { + "epoch": 4.891891891891892, + "grad_norm": 0.050256418798137074, + "learning_rate": 1.1387151087814297e-07, + "loss": 0.5776, + "step": 724 + }, + { + "epoch": 4.898648648648649, + "grad_norm": 0.06422788348062423, + "learning_rate": 1.0008813622001345e-07, + "loss": 0.5865, + "step": 725 + }, + { + "epoch": 4.905405405405405, + "grad_norm": 0.058912661052812065, + "learning_rate": 8.719257643877044e-08, + "loss": 0.5904, + "step": 726 + }, + { + "epoch": 4.912162162162162, + "grad_norm": 0.05385241384320565, + "learning_rate": 7.51851184741481e-08, + "loss": 0.5944, + "step": 727 + }, + { + "epoch": 4.918918918918919, + "grad_norm": 0.05443985642845785, + "learning_rate": 6.40660295046791e-08, + "loss": 0.5857, + "step": 728 + }, + { + "epoch": 4.925675675675675, + "grad_norm": 0.05281227387100857, + "learning_rate": 5.3835556941743695e-08, + "loss": 0.5742, + "step": 729 + }, + { + "epoch": 4.9324324324324325, + "grad_norm": 0.05531460046068238, + "learning_rate": 4.449392842408529e-08, + "loss": 0.5743, + "step": 730 + }, + { + "epoch": 4.9391891891891895, + "grad_norm": 0.05538814823209324, + "learning_rate": 3.6041351812743374e-08, + "loss": 0.5877, + "step": 731 + }, + { + "epoch": 4.945945945945946, + "grad_norm": 0.055186199974581227, + "learning_rate": 2.8478015186399477e-08, + "loss": 0.5842, + "step": 732 + }, + { + "epoch": 4.952702702702703, + "grad_norm": 0.055151710910165075, + "learning_rate": 2.1804086837229344e-08, + "loss": 0.5846, + "step": 733 + }, + { + "epoch": 4.95945945945946, + "grad_norm": 0.05558334617444456, + "learning_rate": 1.601971526713708e-08, + "loss": 0.5766, + "step": 734 + }, + { + "epoch": 4.966216216216216, + "grad_norm": 0.049437407136724947, + "learning_rate": 1.112502918445113e-08, + "loss": 0.586, + "step": 735 + }, + { + "epoch": 4.972972972972973, + "grad_norm": 0.05838258812809817, + "learning_rate": 7.12013750107321e-09, + "loss": 0.5918, + "step": 736 + }, + { + "epoch": 4.97972972972973, + "grad_norm": 0.05252990242194222, + "learning_rate": 4.00512933004471e-09, + "loss": 0.5853, + "step": 737 + }, + { + "epoch": 4.986486486486487, + "grad_norm": 0.05539911735758133, + "learning_rate": 1.7800739835616143e-09, + "loss": 0.5731, + "step": 738 + }, + { + "epoch": 4.993243243243243, + "grad_norm": 0.059325115613032664, + "learning_rate": 4.450209714379483e-10, + "loss": 0.5814, + "step": 739 + }, + { + "epoch": 5.0, + "grad_norm": 0.059385070432276595, + "learning_rate": 0.0, + "loss": 0.577, + "step": 740 + }, + { + "epoch": 5.0, + "step": 740, + "total_flos": 4.168626538020864e+16, + "train_loss": 0.6272868978010642, + "train_runtime": 42334.4318, + "train_samples_per_second": 8.913, + "train_steps_per_second": 0.017 + } + ], + "logging_steps": 1, + "max_steps": 740, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 4.168626538020864e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}