| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 2.8094117647058825, |
| "eval_steps": 96, |
| "global_step": 300, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.009411764705882352, |
| "grad_norm": 0.8642500638961792, |
| "learning_rate": 4e-05, |
| "loss": 0.9841, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.018823529411764704, |
| "grad_norm": 0.8003044724464417, |
| "learning_rate": 8e-05, |
| "loss": 0.9727, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.02823529411764706, |
| "grad_norm": 0.7982213497161865, |
| "learning_rate": 0.00012, |
| "loss": 1.0076, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.03764705882352941, |
| "grad_norm": 0.4778502285480499, |
| "learning_rate": 0.00016, |
| "loss": 0.9179, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.047058823529411764, |
| "grad_norm": 0.1798139065504074, |
| "learning_rate": 0.0002, |
| "loss": 0.8544, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.05647058823529412, |
| "grad_norm": 0.10934074968099594, |
| "learning_rate": 0.00019936102236421725, |
| "loss": 0.8259, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.06588235294117648, |
| "grad_norm": 0.451732873916626, |
| "learning_rate": 0.00019872204472843452, |
| "loss": 0.8176, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.07529411764705882, |
| "grad_norm": 0.39763253927230835, |
| "learning_rate": 0.00019808306709265177, |
| "loss": 0.8091, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.08470588235294117, |
| "grad_norm": 0.08421512693166733, |
| "learning_rate": 0.000197444089456869, |
| "loss": 0.8066, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.09411764705882353, |
| "grad_norm": 0.0752546563744545, |
| "learning_rate": 0.00019680511182108628, |
| "loss": 0.788, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.10352941176470588, |
| "grad_norm": 0.07105962187051773, |
| "learning_rate": 0.00019616613418530353, |
| "loss": 0.8142, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.11294117647058824, |
| "grad_norm": 0.07268067449331284, |
| "learning_rate": 0.0001955271565495208, |
| "loss": 0.7427, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.1223529411764706, |
| "grad_norm": 0.071601502597332, |
| "learning_rate": 0.00019488817891373804, |
| "loss": 0.7673, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.13176470588235295, |
| "grad_norm": 0.06190289184451103, |
| "learning_rate": 0.00019424920127795528, |
| "loss": 0.7897, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.1411764705882353, |
| "grad_norm": 0.062015291303396225, |
| "learning_rate": 0.00019361022364217253, |
| "loss": 0.7552, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.15058823529411763, |
| "grad_norm": 0.06343454122543335, |
| "learning_rate": 0.00019297124600638977, |
| "loss": 0.7983, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.16, |
| "grad_norm": 0.0880446657538414, |
| "learning_rate": 0.00019233226837060702, |
| "loss": 0.7331, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.16941176470588235, |
| "grad_norm": 0.06399823725223541, |
| "learning_rate": 0.00019169329073482429, |
| "loss": 0.7017, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.17882352941176471, |
| "grad_norm": 0.06614473462104797, |
| "learning_rate": 0.00019105431309904153, |
| "loss": 0.7346, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.18823529411764706, |
| "grad_norm": 0.05928972363471985, |
| "learning_rate": 0.0001904153354632588, |
| "loss": 0.7088, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.1976470588235294, |
| "grad_norm": 0.05864707753062248, |
| "learning_rate": 0.00018977635782747604, |
| "loss": 0.7017, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.20705882352941177, |
| "grad_norm": 0.0725838840007782, |
| "learning_rate": 0.0001891373801916933, |
| "loss": 0.7255, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.2164705882352941, |
| "grad_norm": 0.07014375925064087, |
| "learning_rate": 0.00018849840255591056, |
| "loss": 0.6699, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.22588235294117648, |
| "grad_norm": 0.08222078531980515, |
| "learning_rate": 0.0001878594249201278, |
| "loss": 0.6802, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.23529411764705882, |
| "grad_norm": 0.07020223140716553, |
| "learning_rate": 0.00018722044728434505, |
| "loss": 0.7013, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.2447058823529412, |
| "grad_norm": 0.05753432214260101, |
| "learning_rate": 0.00018658146964856232, |
| "loss": 0.6547, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.2541176470588235, |
| "grad_norm": 0.06647774577140808, |
| "learning_rate": 0.00018594249201277956, |
| "loss": 0.7132, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.2635294117647059, |
| "grad_norm": 0.061453238129615784, |
| "learning_rate": 0.00018530351437699683, |
| "loss": 0.6891, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.27294117647058824, |
| "grad_norm": 0.055535465478897095, |
| "learning_rate": 0.00018466453674121408, |
| "loss": 0.6542, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.2823529411764706, |
| "grad_norm": 0.05627722293138504, |
| "learning_rate": 0.00018402555910543132, |
| "loss": 0.677, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.2917647058823529, |
| "grad_norm": 0.05428534746170044, |
| "learning_rate": 0.0001833865814696486, |
| "loss": 0.6592, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.30117647058823527, |
| "grad_norm": 0.06012285500764847, |
| "learning_rate": 0.00018274760383386583, |
| "loss": 0.6662, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.31058823529411766, |
| "grad_norm": 0.08650115132331848, |
| "learning_rate": 0.00018210862619808308, |
| "loss": 0.6546, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.32, |
| "grad_norm": 0.058147966861724854, |
| "learning_rate": 0.00018146964856230032, |
| "loss": 0.6505, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.32941176470588235, |
| "grad_norm": 0.054181504994630814, |
| "learning_rate": 0.00018083067092651756, |
| "loss": 0.6428, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.3388235294117647, |
| "grad_norm": 0.05629313737154007, |
| "learning_rate": 0.00018019169329073484, |
| "loss": 0.6916, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.34823529411764703, |
| "grad_norm": 0.05400744453072548, |
| "learning_rate": 0.00017955271565495208, |
| "loss": 0.6646, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.35764705882352943, |
| "grad_norm": 0.053580883890390396, |
| "learning_rate": 0.00017891373801916932, |
| "loss": 0.6297, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.36705882352941177, |
| "grad_norm": 0.05560213327407837, |
| "learning_rate": 0.0001782747603833866, |
| "loss": 0.6284, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.3764705882352941, |
| "grad_norm": 0.05281270295381546, |
| "learning_rate": 0.00017763578274760384, |
| "loss": 0.6392, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.38588235294117645, |
| "grad_norm": 0.06260088086128235, |
| "learning_rate": 0.00017699680511182108, |
| "loss": 0.7056, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.3952941176470588, |
| "grad_norm": 0.058404579758644104, |
| "learning_rate": 0.00017635782747603835, |
| "loss": 0.6138, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.4047058823529412, |
| "grad_norm": 0.050437066704034805, |
| "learning_rate": 0.0001757188498402556, |
| "loss": 0.6569, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.41411764705882353, |
| "grad_norm": 0.05336596816778183, |
| "learning_rate": 0.00017507987220447287, |
| "loss": 0.612, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.4235294117647059, |
| "grad_norm": 0.05253816023468971, |
| "learning_rate": 0.0001744408945686901, |
| "loss": 0.6195, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.4329411764705882, |
| "grad_norm": 0.05323391407728195, |
| "learning_rate": 0.00017380191693290735, |
| "loss": 0.624, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.4423529411764706, |
| "grad_norm": 0.05454220622777939, |
| "learning_rate": 0.00017316293929712462, |
| "loss": 0.6654, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.45176470588235296, |
| "grad_norm": 0.05840761214494705, |
| "learning_rate": 0.00017252396166134187, |
| "loss": 0.6791, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.4611764705882353, |
| "grad_norm": 0.05933906137943268, |
| "learning_rate": 0.0001718849840255591, |
| "loss": 0.6501, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.47058823529411764, |
| "grad_norm": 0.06150069832801819, |
| "learning_rate": 0.00017124600638977638, |
| "loss": 0.6332, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.48, |
| "grad_norm": 0.054139360785484314, |
| "learning_rate": 0.00017060702875399363, |
| "loss": 0.6268, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.4894117647058824, |
| "grad_norm": 0.05578823760151863, |
| "learning_rate": 0.00016996805111821087, |
| "loss": 0.6259, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.4988235294117647, |
| "grad_norm": 0.06352369487285614, |
| "learning_rate": 0.00016932907348242811, |
| "loss": 0.6271, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.508235294117647, |
| "grad_norm": 0.06086834892630577, |
| "learning_rate": 0.00016869009584664536, |
| "loss": 0.631, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.5176470588235295, |
| "grad_norm": 0.05695689842104912, |
| "learning_rate": 0.00016805111821086263, |
| "loss": 0.64, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.5270588235294118, |
| "grad_norm": 0.056874532252550125, |
| "learning_rate": 0.00016741214057507987, |
| "loss": 0.6296, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.5364705882352941, |
| "grad_norm": 0.061433881521224976, |
| "learning_rate": 0.00016677316293929712, |
| "loss": 0.683, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.5458823529411765, |
| "grad_norm": 0.06481802463531494, |
| "learning_rate": 0.0001661341853035144, |
| "loss": 0.599, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.5552941176470588, |
| "grad_norm": 0.06558655947446823, |
| "learning_rate": 0.00016549520766773163, |
| "loss": 0.6362, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.5647058823529412, |
| "grad_norm": 0.057915814220905304, |
| "learning_rate": 0.0001648562300319489, |
| "loss": 0.6396, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.5741176470588235, |
| "grad_norm": 0.06019767373800278, |
| "learning_rate": 0.00016421725239616614, |
| "loss": 0.6444, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.5835294117647059, |
| "grad_norm": 0.06425336748361588, |
| "learning_rate": 0.0001635782747603834, |
| "loss": 0.6175, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.5929411764705882, |
| "grad_norm": 0.062431298196315765, |
| "learning_rate": 0.00016293929712460066, |
| "loss": 0.6458, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.6023529411764705, |
| "grad_norm": 0.057641636580228806, |
| "learning_rate": 0.0001623003194888179, |
| "loss": 0.6262, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.611764705882353, |
| "grad_norm": 0.05894685909152031, |
| "learning_rate": 0.00016166134185303515, |
| "loss": 0.6036, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.6211764705882353, |
| "grad_norm": 0.06207429617643356, |
| "learning_rate": 0.00016102236421725242, |
| "loss": 0.6367, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.6305882352941177, |
| "grad_norm": 0.06500203162431717, |
| "learning_rate": 0.00016038338658146966, |
| "loss": 0.6443, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.64, |
| "grad_norm": 0.06465457379817963, |
| "learning_rate": 0.0001597444089456869, |
| "loss": 0.6105, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.6494117647058824, |
| "grad_norm": 0.060090478509664536, |
| "learning_rate": 0.00015910543130990418, |
| "loss": 0.6447, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.6588235294117647, |
| "grad_norm": 0.06724690645933151, |
| "learning_rate": 0.00015846645367412142, |
| "loss": 0.6062, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.668235294117647, |
| "grad_norm": 0.06291332095861435, |
| "learning_rate": 0.00015782747603833866, |
| "loss": 0.6005, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.6776470588235294, |
| "grad_norm": 0.06592954695224762, |
| "learning_rate": 0.0001571884984025559, |
| "loss": 0.6036, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.6870588235294117, |
| "grad_norm": 0.06050164997577667, |
| "learning_rate": 0.00015654952076677315, |
| "loss": 0.5993, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.6964705882352941, |
| "grad_norm": 0.061909306794404984, |
| "learning_rate": 0.00015591054313099042, |
| "loss": 0.6159, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.7058823529411765, |
| "grad_norm": 0.06927211582660675, |
| "learning_rate": 0.00015527156549520767, |
| "loss": 0.6554, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.7152941176470589, |
| "grad_norm": 0.06098336726427078, |
| "learning_rate": 0.00015463258785942494, |
| "loss": 0.5998, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.7247058823529412, |
| "grad_norm": 0.06612212210893631, |
| "learning_rate": 0.00015399361022364218, |
| "loss": 0.6057, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.7341176470588235, |
| "grad_norm": 0.06455227732658386, |
| "learning_rate": 0.00015335463258785942, |
| "loss": 0.64, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.7435294117647059, |
| "grad_norm": 0.40494394302368164, |
| "learning_rate": 0.0001527156549520767, |
| "loss": 0.6397, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.7529411764705882, |
| "grad_norm": 0.05910542979836464, |
| "learning_rate": 0.00015207667731629394, |
| "loss": 0.6087, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.7623529411764706, |
| "grad_norm": 0.06125890091061592, |
| "learning_rate": 0.00015143769968051118, |
| "loss": 0.6133, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.7717647058823529, |
| "grad_norm": 0.0716337338089943, |
| "learning_rate": 0.00015079872204472845, |
| "loss": 0.6136, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.7811764705882352, |
| "grad_norm": 0.07003322243690491, |
| "learning_rate": 0.0001501597444089457, |
| "loss": 0.6236, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.7905882352941176, |
| "grad_norm": 0.06324715912342072, |
| "learning_rate": 0.00014952076677316297, |
| "loss": 0.6282, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.8, |
| "grad_norm": 0.07135389745235443, |
| "learning_rate": 0.0001488817891373802, |
| "loss": 0.6443, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.8094117647058824, |
| "grad_norm": 0.06127123162150383, |
| "learning_rate": 0.00014824281150159745, |
| "loss": 0.6396, |
| "step": 86 |
| }, |
| { |
| "epoch": 0.8188235294117647, |
| "grad_norm": 0.06172896921634674, |
| "learning_rate": 0.0001476038338658147, |
| "loss": 0.6377, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.8282352941176471, |
| "grad_norm": 0.3543515205383301, |
| "learning_rate": 0.00014696485623003194, |
| "loss": 0.6454, |
| "step": 88 |
| }, |
| { |
| "epoch": 0.8376470588235294, |
| "grad_norm": 0.06527980417013168, |
| "learning_rate": 0.0001463258785942492, |
| "loss": 0.615, |
| "step": 89 |
| }, |
| { |
| "epoch": 0.8470588235294118, |
| "grad_norm": 0.06327975541353226, |
| "learning_rate": 0.00014568690095846646, |
| "loss": 0.6447, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.8564705882352941, |
| "grad_norm": 0.06739926338195801, |
| "learning_rate": 0.0001450479233226837, |
| "loss": 0.6087, |
| "step": 91 |
| }, |
| { |
| "epoch": 0.8658823529411764, |
| "grad_norm": 0.06617800146341324, |
| "learning_rate": 0.00014440894568690097, |
| "loss": 0.6259, |
| "step": 92 |
| }, |
| { |
| "epoch": 0.8752941176470588, |
| "grad_norm": 0.06242835149168968, |
| "learning_rate": 0.00014376996805111821, |
| "loss": 0.6049, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.8847058823529412, |
| "grad_norm": 0.06359254568815231, |
| "learning_rate": 0.00014313099041533546, |
| "loss": 0.6274, |
| "step": 94 |
| }, |
| { |
| "epoch": 0.8941176470588236, |
| "grad_norm": 0.06559597700834274, |
| "learning_rate": 0.00014249201277955273, |
| "loss": 0.5935, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.9035294117647059, |
| "grad_norm": 0.07493000477552414, |
| "learning_rate": 0.00014185303514376997, |
| "loss": 0.5881, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.9035294117647059, |
| "eval_loss": 0.6278424859046936, |
| "eval_runtime": 5.469, |
| "eval_samples_per_second": 8.045, |
| "eval_steps_per_second": 1.097, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.9129411764705883, |
| "grad_norm": 0.06714215129613876, |
| "learning_rate": 0.00014121405750798722, |
| "loss": 0.6438, |
| "step": 97 |
| }, |
| { |
| "epoch": 0.9223529411764706, |
| "grad_norm": 0.06775335222482681, |
| "learning_rate": 0.0001405750798722045, |
| "loss": 0.6256, |
| "step": 98 |
| }, |
| { |
| "epoch": 0.9317647058823529, |
| "grad_norm": 0.07045792043209076, |
| "learning_rate": 0.00013993610223642173, |
| "loss": 0.6712, |
| "step": 99 |
| }, |
| { |
| "epoch": 0.9411764705882353, |
| "grad_norm": 0.06958217918872833, |
| "learning_rate": 0.000139297124600639, |
| "loss": 0.6278, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.9505882352941176, |
| "grad_norm": 0.06508302688598633, |
| "learning_rate": 0.00013865814696485625, |
| "loss": 0.6415, |
| "step": 101 |
| }, |
| { |
| "epoch": 0.96, |
| "grad_norm": 0.06876156479120255, |
| "learning_rate": 0.0001380191693290735, |
| "loss": 0.6211, |
| "step": 102 |
| }, |
| { |
| "epoch": 0.9694117647058823, |
| "grad_norm": 0.07631471753120422, |
| "learning_rate": 0.00013738019169329076, |
| "loss": 0.6276, |
| "step": 103 |
| }, |
| { |
| "epoch": 0.9788235294117648, |
| "grad_norm": 0.07122895866632462, |
| "learning_rate": 0.000136741214057508, |
| "loss": 0.5997, |
| "step": 104 |
| }, |
| { |
| "epoch": 0.9882352941176471, |
| "grad_norm": 0.06862416863441467, |
| "learning_rate": 0.00013610223642172525, |
| "loss": 0.6563, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.9976470588235294, |
| "grad_norm": 0.06701315939426422, |
| "learning_rate": 0.0001354632587859425, |
| "loss": 0.61, |
| "step": 106 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 0.14547984302043915, |
| "learning_rate": 0.00013482428115015973, |
| "loss": 0.6382, |
| "step": 107 |
| }, |
| { |
| "epoch": 1.0094117647058825, |
| "grad_norm": 0.0676453560590744, |
| "learning_rate": 0.000134185303514377, |
| "loss": 0.5766, |
| "step": 108 |
| }, |
| { |
| "epoch": 1.0188235294117647, |
| "grad_norm": 0.06471403688192368, |
| "learning_rate": 0.00013354632587859425, |
| "loss": 0.5769, |
| "step": 109 |
| }, |
| { |
| "epoch": 1.0282352941176471, |
| "grad_norm": 0.07038411498069763, |
| "learning_rate": 0.0001329073482428115, |
| "loss": 0.5828, |
| "step": 110 |
| }, |
| { |
| "epoch": 1.0376470588235294, |
| "grad_norm": 0.0679129809141159, |
| "learning_rate": 0.00013226837060702876, |
| "loss": 0.5755, |
| "step": 111 |
| }, |
| { |
| "epoch": 1.0470588235294118, |
| "grad_norm": 0.07297012954950333, |
| "learning_rate": 0.000131629392971246, |
| "loss": 0.6114, |
| "step": 112 |
| }, |
| { |
| "epoch": 1.056470588235294, |
| "grad_norm": 0.07036986202001572, |
| "learning_rate": 0.00013099041533546325, |
| "loss": 0.6254, |
| "step": 113 |
| }, |
| { |
| "epoch": 1.0658823529411765, |
| "grad_norm": 0.08385265618562698, |
| "learning_rate": 0.00013035143769968052, |
| "loss": 0.6284, |
| "step": 114 |
| }, |
| { |
| "epoch": 1.0752941176470587, |
| "grad_norm": 0.08367498964071274, |
| "learning_rate": 0.00012971246006389777, |
| "loss": 0.6128, |
| "step": 115 |
| }, |
| { |
| "epoch": 1.0847058823529412, |
| "grad_norm": 0.082453154027462, |
| "learning_rate": 0.000129073482428115, |
| "loss": 0.6012, |
| "step": 116 |
| }, |
| { |
| "epoch": 1.0941176470588236, |
| "grad_norm": 0.0714535042643547, |
| "learning_rate": 0.00012843450479233228, |
| "loss": 0.599, |
| "step": 117 |
| }, |
| { |
| "epoch": 1.1035294117647059, |
| "grad_norm": 0.07391790300607681, |
| "learning_rate": 0.00012779552715654952, |
| "loss": 0.5897, |
| "step": 118 |
| }, |
| { |
| "epoch": 1.1129411764705883, |
| "grad_norm": 0.08383256196975708, |
| "learning_rate": 0.0001271565495207668, |
| "loss": 0.634, |
| "step": 119 |
| }, |
| { |
| "epoch": 1.1223529411764706, |
| "grad_norm": 0.08821647614240646, |
| "learning_rate": 0.00012651757188498404, |
| "loss": 0.5869, |
| "step": 120 |
| }, |
| { |
| "epoch": 1.131764705882353, |
| "grad_norm": 0.07556191831827164, |
| "learning_rate": 0.00012587859424920128, |
| "loss": 0.6275, |
| "step": 121 |
| }, |
| { |
| "epoch": 1.1411764705882352, |
| "grad_norm": 0.08794881403446198, |
| "learning_rate": 0.00012523961661341855, |
| "loss": 0.6115, |
| "step": 122 |
| }, |
| { |
| "epoch": 1.1505882352941177, |
| "grad_norm": 0.0785142183303833, |
| "learning_rate": 0.0001246006389776358, |
| "loss": 0.6156, |
| "step": 123 |
| }, |
| { |
| "epoch": 1.16, |
| "grad_norm": 0.07684168219566345, |
| "learning_rate": 0.00012396166134185304, |
| "loss": 0.6107, |
| "step": 124 |
| }, |
| { |
| "epoch": 1.1694117647058824, |
| "grad_norm": 0.07763410359621048, |
| "learning_rate": 0.00012332268370607028, |
| "loss": 0.5888, |
| "step": 125 |
| }, |
| { |
| "epoch": 1.1788235294117646, |
| "grad_norm": 0.09084579348564148, |
| "learning_rate": 0.00012268370607028753, |
| "loss": 0.5912, |
| "step": 126 |
| }, |
| { |
| "epoch": 1.188235294117647, |
| "grad_norm": 0.0838058665394783, |
| "learning_rate": 0.00012204472843450481, |
| "loss": 0.6152, |
| "step": 127 |
| }, |
| { |
| "epoch": 1.1976470588235295, |
| "grad_norm": 0.08908473700284958, |
| "learning_rate": 0.00012140575079872206, |
| "loss": 0.6049, |
| "step": 128 |
| }, |
| { |
| "epoch": 1.2070588235294117, |
| "grad_norm": 0.08239720016717911, |
| "learning_rate": 0.0001207667731629393, |
| "loss": 0.5852, |
| "step": 129 |
| }, |
| { |
| "epoch": 1.2164705882352942, |
| "grad_norm": 0.08228597790002823, |
| "learning_rate": 0.00012012779552715656, |
| "loss": 0.5642, |
| "step": 130 |
| }, |
| { |
| "epoch": 1.2258823529411764, |
| "grad_norm": 0.08196559548377991, |
| "learning_rate": 0.0001194888178913738, |
| "loss": 0.5971, |
| "step": 131 |
| }, |
| { |
| "epoch": 1.2352941176470589, |
| "grad_norm": 0.08850298076868057, |
| "learning_rate": 0.00011884984025559104, |
| "loss": 0.6149, |
| "step": 132 |
| }, |
| { |
| "epoch": 1.244705882352941, |
| "grad_norm": 0.08643686771392822, |
| "learning_rate": 0.00011821086261980832, |
| "loss": 0.6097, |
| "step": 133 |
| }, |
| { |
| "epoch": 1.2541176470588236, |
| "grad_norm": 0.08877662569284439, |
| "learning_rate": 0.00011757188498402556, |
| "loss": 0.6362, |
| "step": 134 |
| }, |
| { |
| "epoch": 1.263529411764706, |
| "grad_norm": 0.08452433347702026, |
| "learning_rate": 0.00011693290734824283, |
| "loss": 0.6331, |
| "step": 135 |
| }, |
| { |
| "epoch": 1.2729411764705882, |
| "grad_norm": 0.08155957609415054, |
| "learning_rate": 0.00011629392971246007, |
| "loss": 0.6214, |
| "step": 136 |
| }, |
| { |
| "epoch": 1.2823529411764705, |
| "grad_norm": 0.08229539543390274, |
| "learning_rate": 0.00011565495207667732, |
| "loss": 0.5981, |
| "step": 137 |
| }, |
| { |
| "epoch": 1.291764705882353, |
| "grad_norm": 0.09338800609111786, |
| "learning_rate": 0.00011501597444089457, |
| "loss": 0.6326, |
| "step": 138 |
| }, |
| { |
| "epoch": 1.3011764705882354, |
| "grad_norm": 0.0837632343173027, |
| "learning_rate": 0.00011437699680511182, |
| "loss": 0.5736, |
| "step": 139 |
| }, |
| { |
| "epoch": 1.3105882352941176, |
| "grad_norm": 0.08500789105892181, |
| "learning_rate": 0.00011373801916932908, |
| "loss": 0.6253, |
| "step": 140 |
| }, |
| { |
| "epoch": 1.32, |
| "grad_norm": 0.0843970999121666, |
| "learning_rate": 0.00011309904153354633, |
| "loss": 0.6053, |
| "step": 141 |
| }, |
| { |
| "epoch": 1.3294117647058823, |
| "grad_norm": 0.08848956227302551, |
| "learning_rate": 0.00011246006389776358, |
| "loss": 0.6184, |
| "step": 142 |
| }, |
| { |
| "epoch": 1.3388235294117647, |
| "grad_norm": 0.10718461126089096, |
| "learning_rate": 0.00011182108626198085, |
| "loss": 0.583, |
| "step": 143 |
| }, |
| { |
| "epoch": 1.348235294117647, |
| "grad_norm": 0.08756856620311737, |
| "learning_rate": 0.00011118210862619809, |
| "loss": 0.5861, |
| "step": 144 |
| }, |
| { |
| "epoch": 1.3576470588235294, |
| "grad_norm": 0.08492905646562576, |
| "learning_rate": 0.00011054313099041533, |
| "loss": 0.5707, |
| "step": 145 |
| }, |
| { |
| "epoch": 1.3670588235294119, |
| "grad_norm": 0.08758281171321869, |
| "learning_rate": 0.0001099041533546326, |
| "loss": 0.555, |
| "step": 146 |
| }, |
| { |
| "epoch": 1.3764705882352941, |
| "grad_norm": 0.09751519560813904, |
| "learning_rate": 0.00010926517571884985, |
| "loss": 0.5976, |
| "step": 147 |
| }, |
| { |
| "epoch": 1.3858823529411763, |
| "grad_norm": 0.09590818732976913, |
| "learning_rate": 0.00010862619808306709, |
| "loss": 0.6015, |
| "step": 148 |
| }, |
| { |
| "epoch": 1.3952941176470588, |
| "grad_norm": 0.08812163025140762, |
| "learning_rate": 0.00010798722044728435, |
| "loss": 0.5905, |
| "step": 149 |
| }, |
| { |
| "epoch": 1.4047058823529412, |
| "grad_norm": 0.09162674099206924, |
| "learning_rate": 0.0001073482428115016, |
| "loss": 0.6063, |
| "step": 150 |
| }, |
| { |
| "epoch": 1.4141176470588235, |
| "grad_norm": 0.08727669715881348, |
| "learning_rate": 0.00010670926517571886, |
| "loss": 0.6181, |
| "step": 151 |
| }, |
| { |
| "epoch": 1.423529411764706, |
| "grad_norm": 0.0928676649928093, |
| "learning_rate": 0.00010607028753993611, |
| "loss": 0.6188, |
| "step": 152 |
| }, |
| { |
| "epoch": 1.4329411764705882, |
| "grad_norm": 0.09681576490402222, |
| "learning_rate": 0.00010543130990415335, |
| "loss": 0.5818, |
| "step": 153 |
| }, |
| { |
| "epoch": 1.4423529411764706, |
| "grad_norm": 0.08795556426048279, |
| "learning_rate": 0.00010479233226837062, |
| "loss": 0.6134, |
| "step": 154 |
| }, |
| { |
| "epoch": 1.4517647058823528, |
| "grad_norm": 0.0861482247710228, |
| "learning_rate": 0.00010415335463258787, |
| "loss": 0.6082, |
| "step": 155 |
| }, |
| { |
| "epoch": 1.4611764705882353, |
| "grad_norm": 0.09977812319993973, |
| "learning_rate": 0.00010351437699680511, |
| "loss": 0.5921, |
| "step": 156 |
| }, |
| { |
| "epoch": 1.4705882352941178, |
| "grad_norm": 0.10190846771001816, |
| "learning_rate": 0.00010287539936102237, |
| "loss": 0.6271, |
| "step": 157 |
| }, |
| { |
| "epoch": 1.48, |
| "grad_norm": 0.09232486039400101, |
| "learning_rate": 0.00010223642172523961, |
| "loss": 0.6067, |
| "step": 158 |
| }, |
| { |
| "epoch": 1.4894117647058824, |
| "grad_norm": 0.10211990028619766, |
| "learning_rate": 0.00010159744408945688, |
| "loss": 0.6238, |
| "step": 159 |
| }, |
| { |
| "epoch": 1.4988235294117647, |
| "grad_norm": 0.08955651521682739, |
| "learning_rate": 0.00010095846645367413, |
| "loss": 0.6002, |
| "step": 160 |
| }, |
| { |
| "epoch": 1.5082352941176471, |
| "grad_norm": 0.10033038258552551, |
| "learning_rate": 0.00010031948881789137, |
| "loss": 0.6129, |
| "step": 161 |
| }, |
| { |
| "epoch": 1.5176470588235293, |
| "grad_norm": 0.09701602160930634, |
| "learning_rate": 9.968051118210863e-05, |
| "loss": 0.5867, |
| "step": 162 |
| }, |
| { |
| "epoch": 1.5270588235294118, |
| "grad_norm": 0.0969882607460022, |
| "learning_rate": 9.904153354632588e-05, |
| "loss": 0.599, |
| "step": 163 |
| }, |
| { |
| "epoch": 1.5364705882352943, |
| "grad_norm": 0.09163233637809753, |
| "learning_rate": 9.840255591054314e-05, |
| "loss": 0.6078, |
| "step": 164 |
| }, |
| { |
| "epoch": 1.5458823529411765, |
| "grad_norm": 0.0925818681716919, |
| "learning_rate": 9.77635782747604e-05, |
| "loss": 0.6231, |
| "step": 165 |
| }, |
| { |
| "epoch": 1.5552941176470587, |
| "grad_norm": 0.09490446001291275, |
| "learning_rate": 9.712460063897764e-05, |
| "loss": 0.6099, |
| "step": 166 |
| }, |
| { |
| "epoch": 1.5647058823529412, |
| "grad_norm": 0.10370604693889618, |
| "learning_rate": 9.648562300319489e-05, |
| "loss": 0.6133, |
| "step": 167 |
| }, |
| { |
| "epoch": 1.5741176470588236, |
| "grad_norm": 0.10043718665838242, |
| "learning_rate": 9.584664536741214e-05, |
| "loss": 0.6002, |
| "step": 168 |
| }, |
| { |
| "epoch": 1.5835294117647059, |
| "grad_norm": 0.10102451592683792, |
| "learning_rate": 9.52076677316294e-05, |
| "loss": 0.6326, |
| "step": 169 |
| }, |
| { |
| "epoch": 1.592941176470588, |
| "grad_norm": 0.09170352667570114, |
| "learning_rate": 9.456869009584664e-05, |
| "loss": 0.5725, |
| "step": 170 |
| }, |
| { |
| "epoch": 1.6023529411764705, |
| "grad_norm": 0.09696916490793228, |
| "learning_rate": 9.39297124600639e-05, |
| "loss": 0.6162, |
| "step": 171 |
| }, |
| { |
| "epoch": 1.611764705882353, |
| "grad_norm": 0.1032591313123703, |
| "learning_rate": 9.329073482428116e-05, |
| "loss": 0.6022, |
| "step": 172 |
| }, |
| { |
| "epoch": 1.6211764705882352, |
| "grad_norm": 0.09963531792163849, |
| "learning_rate": 9.265175718849842e-05, |
| "loss": 0.5918, |
| "step": 173 |
| }, |
| { |
| "epoch": 1.6305882352941177, |
| "grad_norm": 0.10566572844982147, |
| "learning_rate": 9.201277955271566e-05, |
| "loss": 0.6248, |
| "step": 174 |
| }, |
| { |
| "epoch": 1.6400000000000001, |
| "grad_norm": 0.09654838591814041, |
| "learning_rate": 9.137380191693292e-05, |
| "loss": 0.5747, |
| "step": 175 |
| }, |
| { |
| "epoch": 1.6494117647058824, |
| "grad_norm": 0.09412531554698944, |
| "learning_rate": 9.073482428115016e-05, |
| "loss": 0.611, |
| "step": 176 |
| }, |
| { |
| "epoch": 1.6588235294117646, |
| "grad_norm": 0.10707147419452667, |
| "learning_rate": 9.009584664536742e-05, |
| "loss": 0.6371, |
| "step": 177 |
| }, |
| { |
| "epoch": 1.668235294117647, |
| "grad_norm": 0.0976637527346611, |
| "learning_rate": 8.945686900958466e-05, |
| "loss": 0.636, |
| "step": 178 |
| }, |
| { |
| "epoch": 1.6776470588235295, |
| "grad_norm": 0.09769190102815628, |
| "learning_rate": 8.881789137380192e-05, |
| "loss": 0.6379, |
| "step": 179 |
| }, |
| { |
| "epoch": 1.6870588235294117, |
| "grad_norm": 0.1029646173119545, |
| "learning_rate": 8.817891373801918e-05, |
| "loss": 0.6533, |
| "step": 180 |
| }, |
| { |
| "epoch": 1.696470588235294, |
| "grad_norm": 0.10120602697134018, |
| "learning_rate": 8.753993610223643e-05, |
| "loss": 0.6044, |
| "step": 181 |
| }, |
| { |
| "epoch": 1.7058823529411766, |
| "grad_norm": 0.0992206409573555, |
| "learning_rate": 8.690095846645368e-05, |
| "loss": 0.6072, |
| "step": 182 |
| }, |
| { |
| "epoch": 1.7152941176470589, |
| "grad_norm": 0.10124579071998596, |
| "learning_rate": 8.626198083067093e-05, |
| "loss": 0.5923, |
| "step": 183 |
| }, |
| { |
| "epoch": 1.724705882352941, |
| "grad_norm": 0.10673081874847412, |
| "learning_rate": 8.562300319488819e-05, |
| "loss": 0.5964, |
| "step": 184 |
| }, |
| { |
| "epoch": 1.7341176470588235, |
| "grad_norm": 0.10745410621166229, |
| "learning_rate": 8.498402555910544e-05, |
| "loss": 0.6052, |
| "step": 185 |
| }, |
| { |
| "epoch": 1.743529411764706, |
| "grad_norm": 0.10404065251350403, |
| "learning_rate": 8.434504792332268e-05, |
| "loss": 0.5975, |
| "step": 186 |
| }, |
| { |
| "epoch": 1.7529411764705882, |
| "grad_norm": 0.09927040338516235, |
| "learning_rate": 8.370607028753994e-05, |
| "loss": 0.5901, |
| "step": 187 |
| }, |
| { |
| "epoch": 1.7623529411764705, |
| "grad_norm": 0.10322125256061554, |
| "learning_rate": 8.30670926517572e-05, |
| "loss": 0.6154, |
| "step": 188 |
| }, |
| { |
| "epoch": 1.771764705882353, |
| "grad_norm": 0.10899791866540909, |
| "learning_rate": 8.242811501597445e-05, |
| "loss": 0.6255, |
| "step": 189 |
| }, |
| { |
| "epoch": 1.7811764705882354, |
| "grad_norm": 0.1041494831442833, |
| "learning_rate": 8.17891373801917e-05, |
| "loss": 0.5925, |
| "step": 190 |
| }, |
| { |
| "epoch": 1.7905882352941176, |
| "grad_norm": 0.09239751845598221, |
| "learning_rate": 8.115015974440895e-05, |
| "loss": 0.5713, |
| "step": 191 |
| }, |
| { |
| "epoch": 1.8, |
| "grad_norm": 0.09495066851377487, |
| "learning_rate": 8.051118210862621e-05, |
| "loss": 0.5896, |
| "step": 192 |
| }, |
| { |
| "epoch": 1.8, |
| "eval_loss": 0.6196094155311584, |
| "eval_runtime": 5.4299, |
| "eval_samples_per_second": 8.103, |
| "eval_steps_per_second": 1.105, |
| "step": 192 |
| }, |
| { |
| "epoch": 1.8094117647058825, |
| "grad_norm": 0.09602215141057968, |
| "learning_rate": 7.987220447284345e-05, |
| "loss": 0.5979, |
| "step": 193 |
| }, |
| { |
| "epoch": 1.8188235294117647, |
| "grad_norm": 0.10422977060079575, |
| "learning_rate": 7.923322683706071e-05, |
| "loss": 0.5704, |
| "step": 194 |
| }, |
| { |
| "epoch": 1.828235294117647, |
| "grad_norm": 0.09678755700588226, |
| "learning_rate": 7.859424920127795e-05, |
| "loss": 0.5774, |
| "step": 195 |
| }, |
| { |
| "epoch": 1.8376470588235294, |
| "grad_norm": 0.09845128655433655, |
| "learning_rate": 7.795527156549521e-05, |
| "loss": 0.5816, |
| "step": 196 |
| }, |
| { |
| "epoch": 1.8470588235294119, |
| "grad_norm": 0.1065673753619194, |
| "learning_rate": 7.731629392971247e-05, |
| "loss": 0.6135, |
| "step": 197 |
| }, |
| { |
| "epoch": 1.856470588235294, |
| "grad_norm": 0.10570935904979706, |
| "learning_rate": 7.667731629392971e-05, |
| "loss": 0.5908, |
| "step": 198 |
| }, |
| { |
| "epoch": 1.8658823529411763, |
| "grad_norm": 0.09649720042943954, |
| "learning_rate": 7.603833865814697e-05, |
| "loss": 0.599, |
| "step": 199 |
| }, |
| { |
| "epoch": 1.8752941176470588, |
| "grad_norm": 0.10398146510124207, |
| "learning_rate": 7.539936102236423e-05, |
| "loss": 0.5853, |
| "step": 200 |
| }, |
| { |
| "epoch": 1.8847058823529412, |
| "grad_norm": 0.09633094817399979, |
| "learning_rate": 7.476038338658148e-05, |
| "loss": 0.5927, |
| "step": 201 |
| }, |
| { |
| "epoch": 1.8941176470588235, |
| "grad_norm": 0.0993979275226593, |
| "learning_rate": 7.412140575079873e-05, |
| "loss": 0.5928, |
| "step": 202 |
| }, |
| { |
| "epoch": 1.903529411764706, |
| "grad_norm": 0.10687059164047241, |
| "learning_rate": 7.348242811501597e-05, |
| "loss": 0.5882, |
| "step": 203 |
| }, |
| { |
| "epoch": 1.9129411764705884, |
| "grad_norm": 0.10128702223300934, |
| "learning_rate": 7.284345047923323e-05, |
| "loss": 0.6009, |
| "step": 204 |
| }, |
| { |
| "epoch": 1.9223529411764706, |
| "grad_norm": 0.10716967284679413, |
| "learning_rate": 7.220447284345049e-05, |
| "loss": 0.6198, |
| "step": 205 |
| }, |
| { |
| "epoch": 1.9317647058823528, |
| "grad_norm": 0.10603494942188263, |
| "learning_rate": 7.156549520766773e-05, |
| "loss": 0.6222, |
| "step": 206 |
| }, |
| { |
| "epoch": 1.9411764705882353, |
| "grad_norm": 0.10125169903039932, |
| "learning_rate": 7.092651757188499e-05, |
| "loss": 0.5938, |
| "step": 207 |
| }, |
| { |
| "epoch": 1.9505882352941177, |
| "grad_norm": 0.10542251914739609, |
| "learning_rate": 7.028753993610224e-05, |
| "loss": 0.6341, |
| "step": 208 |
| }, |
| { |
| "epoch": 1.96, |
| "grad_norm": 0.09659598022699356, |
| "learning_rate": 6.96485623003195e-05, |
| "loss": 0.59, |
| "step": 209 |
| }, |
| { |
| "epoch": 1.9694117647058822, |
| "grad_norm": 0.1122719794511795, |
| "learning_rate": 6.900958466453674e-05, |
| "loss": 0.6109, |
| "step": 210 |
| }, |
| { |
| "epoch": 1.9788235294117649, |
| "grad_norm": 0.11106933653354645, |
| "learning_rate": 6.8370607028754e-05, |
| "loss": 0.58, |
| "step": 211 |
| }, |
| { |
| "epoch": 1.988235294117647, |
| "grad_norm": 0.09962272644042969, |
| "learning_rate": 6.773162939297125e-05, |
| "loss": 0.6105, |
| "step": 212 |
| }, |
| { |
| "epoch": 1.9976470588235293, |
| "grad_norm": 0.09547895193099976, |
| "learning_rate": 6.70926517571885e-05, |
| "loss": 0.5709, |
| "step": 213 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 0.21181434392929077, |
| "learning_rate": 6.645367412140575e-05, |
| "loss": 0.6164, |
| "step": 214 |
| }, |
| { |
| "epoch": 2.0094117647058822, |
| "grad_norm": 0.11414559185504913, |
| "learning_rate": 6.5814696485623e-05, |
| "loss": 0.5642, |
| "step": 215 |
| }, |
| { |
| "epoch": 2.018823529411765, |
| "grad_norm": 0.11991499364376068, |
| "learning_rate": 6.517571884984026e-05, |
| "loss": 0.5741, |
| "step": 216 |
| }, |
| { |
| "epoch": 2.028235294117647, |
| "grad_norm": 0.10295461863279343, |
| "learning_rate": 6.45367412140575e-05, |
| "loss": 0.5604, |
| "step": 217 |
| }, |
| { |
| "epoch": 2.0376470588235294, |
| "grad_norm": 0.10355547070503235, |
| "learning_rate": 6.389776357827476e-05, |
| "loss": 0.5884, |
| "step": 218 |
| }, |
| { |
| "epoch": 2.0470588235294116, |
| "grad_norm": 0.12203363329172134, |
| "learning_rate": 6.325878594249202e-05, |
| "loss": 0.5903, |
| "step": 219 |
| }, |
| { |
| "epoch": 2.0564705882352943, |
| "grad_norm": 0.12483850866556168, |
| "learning_rate": 6.261980830670928e-05, |
| "loss": 0.5666, |
| "step": 220 |
| }, |
| { |
| "epoch": 2.0658823529411765, |
| "grad_norm": 0.12819188833236694, |
| "learning_rate": 6.198083067092652e-05, |
| "loss": 0.5493, |
| "step": 221 |
| }, |
| { |
| "epoch": 2.0752941176470587, |
| "grad_norm": 0.11613185703754425, |
| "learning_rate": 6.134185303514376e-05, |
| "loss": 0.6121, |
| "step": 222 |
| }, |
| { |
| "epoch": 2.084705882352941, |
| "grad_norm": 0.11060836911201477, |
| "learning_rate": 6.070287539936103e-05, |
| "loss": 0.5748, |
| "step": 223 |
| }, |
| { |
| "epoch": 2.0941176470588236, |
| "grad_norm": 0.13958008587360382, |
| "learning_rate": 6.006389776357828e-05, |
| "loss": 0.62, |
| "step": 224 |
| }, |
| { |
| "epoch": 2.103529411764706, |
| "grad_norm": 0.14533892273902893, |
| "learning_rate": 5.942492012779552e-05, |
| "loss": 0.5897, |
| "step": 225 |
| }, |
| { |
| "epoch": 2.112941176470588, |
| "grad_norm": 0.12220677733421326, |
| "learning_rate": 5.878594249201278e-05, |
| "loss": 0.5508, |
| "step": 226 |
| }, |
| { |
| "epoch": 2.1223529411764708, |
| "grad_norm": 0.11975925415754318, |
| "learning_rate": 5.814696485623004e-05, |
| "loss": 0.5483, |
| "step": 227 |
| }, |
| { |
| "epoch": 2.131764705882353, |
| "grad_norm": 0.12308664619922638, |
| "learning_rate": 5.750798722044729e-05, |
| "loss": 0.5921, |
| "step": 228 |
| }, |
| { |
| "epoch": 2.1411764705882352, |
| "grad_norm": 0.12570548057556152, |
| "learning_rate": 5.686900958466454e-05, |
| "loss": 0.6229, |
| "step": 229 |
| }, |
| { |
| "epoch": 2.1505882352941175, |
| "grad_norm": 0.12253190577030182, |
| "learning_rate": 5.623003194888179e-05, |
| "loss": 0.5758, |
| "step": 230 |
| }, |
| { |
| "epoch": 2.16, |
| "grad_norm": 0.1249895840883255, |
| "learning_rate": 5.5591054313099045e-05, |
| "loss": 0.5633, |
| "step": 231 |
| }, |
| { |
| "epoch": 2.1694117647058824, |
| "grad_norm": 0.1355949193239212, |
| "learning_rate": 5.49520766773163e-05, |
| "loss": 0.5486, |
| "step": 232 |
| }, |
| { |
| "epoch": 2.1788235294117646, |
| "grad_norm": 0.12255489826202393, |
| "learning_rate": 5.4313099041533546e-05, |
| "loss": 0.5906, |
| "step": 233 |
| }, |
| { |
| "epoch": 2.1882352941176473, |
| "grad_norm": 0.12850622832775116, |
| "learning_rate": 5.36741214057508e-05, |
| "loss": 0.5738, |
| "step": 234 |
| }, |
| { |
| "epoch": 2.1976470588235295, |
| "grad_norm": 0.18490703403949738, |
| "learning_rate": 5.3035143769968054e-05, |
| "loss": 0.5737, |
| "step": 235 |
| }, |
| { |
| "epoch": 2.2070588235294117, |
| "grad_norm": 0.135318785905838, |
| "learning_rate": 5.239616613418531e-05, |
| "loss": 0.5963, |
| "step": 236 |
| }, |
| { |
| "epoch": 2.216470588235294, |
| "grad_norm": 0.12884393334388733, |
| "learning_rate": 5.1757188498402555e-05, |
| "loss": 0.5399, |
| "step": 237 |
| }, |
| { |
| "epoch": 2.2258823529411766, |
| "grad_norm": 0.13065044581890106, |
| "learning_rate": 5.1118210862619806e-05, |
| "loss": 0.5611, |
| "step": 238 |
| }, |
| { |
| "epoch": 2.235294117647059, |
| "grad_norm": 0.12230806797742844, |
| "learning_rate": 5.047923322683706e-05, |
| "loss": 0.5446, |
| "step": 239 |
| }, |
| { |
| "epoch": 2.244705882352941, |
| "grad_norm": 0.13359951972961426, |
| "learning_rate": 4.984025559105431e-05, |
| "loss": 0.5652, |
| "step": 240 |
| }, |
| { |
| "epoch": 2.2541176470588233, |
| "grad_norm": 0.1409720480442047, |
| "learning_rate": 4.920127795527157e-05, |
| "loss": 0.5437, |
| "step": 241 |
| }, |
| { |
| "epoch": 2.263529411764706, |
| "grad_norm": 0.1343528777360916, |
| "learning_rate": 4.856230031948882e-05, |
| "loss": 0.5811, |
| "step": 242 |
| }, |
| { |
| "epoch": 2.2729411764705882, |
| "grad_norm": 0.1315775066614151, |
| "learning_rate": 4.792332268370607e-05, |
| "loss": 0.5718, |
| "step": 243 |
| }, |
| { |
| "epoch": 2.2823529411764705, |
| "grad_norm": 0.1300155520439148, |
| "learning_rate": 4.728434504792332e-05, |
| "loss": 0.5629, |
| "step": 244 |
| }, |
| { |
| "epoch": 2.291764705882353, |
| "grad_norm": 0.1364685744047165, |
| "learning_rate": 4.664536741214058e-05, |
| "loss": 0.5641, |
| "step": 245 |
| }, |
| { |
| "epoch": 2.3011764705882354, |
| "grad_norm": 0.13575628399848938, |
| "learning_rate": 4.600638977635783e-05, |
| "loss": 0.5861, |
| "step": 246 |
| }, |
| { |
| "epoch": 2.3105882352941176, |
| "grad_norm": 0.13364288210868835, |
| "learning_rate": 4.536741214057508e-05, |
| "loss": 0.5959, |
| "step": 247 |
| }, |
| { |
| "epoch": 2.32, |
| "grad_norm": 0.129843071103096, |
| "learning_rate": 4.472843450479233e-05, |
| "loss": 0.5631, |
| "step": 248 |
| }, |
| { |
| "epoch": 2.3294117647058825, |
| "grad_norm": 0.13373443484306335, |
| "learning_rate": 4.408945686900959e-05, |
| "loss": 0.557, |
| "step": 249 |
| }, |
| { |
| "epoch": 2.3388235294117647, |
| "grad_norm": 0.14091914892196655, |
| "learning_rate": 4.345047923322684e-05, |
| "loss": 0.577, |
| "step": 250 |
| }, |
| { |
| "epoch": 2.348235294117647, |
| "grad_norm": 0.13922256231307983, |
| "learning_rate": 4.2811501597444096e-05, |
| "loss": 0.5404, |
| "step": 251 |
| }, |
| { |
| "epoch": 2.357647058823529, |
| "grad_norm": 0.13278770446777344, |
| "learning_rate": 4.217252396166134e-05, |
| "loss": 0.5131, |
| "step": 252 |
| }, |
| { |
| "epoch": 2.367058823529412, |
| "grad_norm": 0.1421966701745987, |
| "learning_rate": 4.15335463258786e-05, |
| "loss": 0.5803, |
| "step": 253 |
| }, |
| { |
| "epoch": 2.376470588235294, |
| "grad_norm": 0.13415196537971497, |
| "learning_rate": 4.089456869009585e-05, |
| "loss": 0.5689, |
| "step": 254 |
| }, |
| { |
| "epoch": 2.3858823529411763, |
| "grad_norm": 0.13316908478736877, |
| "learning_rate": 4.0255591054313104e-05, |
| "loss": 0.5741, |
| "step": 255 |
| }, |
| { |
| "epoch": 2.395294117647059, |
| "grad_norm": 0.13901880383491516, |
| "learning_rate": 3.9616613418530355e-05, |
| "loss": 0.5365, |
| "step": 256 |
| }, |
| { |
| "epoch": 2.4047058823529412, |
| "grad_norm": 0.13510987162590027, |
| "learning_rate": 3.8977635782747605e-05, |
| "loss": 0.5511, |
| "step": 257 |
| }, |
| { |
| "epoch": 2.4141176470588235, |
| "grad_norm": 0.140598326921463, |
| "learning_rate": 3.8338658146964856e-05, |
| "loss": 0.5566, |
| "step": 258 |
| }, |
| { |
| "epoch": 2.4235294117647057, |
| "grad_norm": 0.13736282289028168, |
| "learning_rate": 3.769968051118211e-05, |
| "loss": 0.576, |
| "step": 259 |
| }, |
| { |
| "epoch": 2.4329411764705884, |
| "grad_norm": 0.13893315196037292, |
| "learning_rate": 3.7060702875399364e-05, |
| "loss": 0.5839, |
| "step": 260 |
| }, |
| { |
| "epoch": 2.4423529411764706, |
| "grad_norm": 0.13888853788375854, |
| "learning_rate": 3.6421725239616614e-05, |
| "loss": 0.5634, |
| "step": 261 |
| }, |
| { |
| "epoch": 2.451764705882353, |
| "grad_norm": 0.1429632157087326, |
| "learning_rate": 3.5782747603833865e-05, |
| "loss": 0.5627, |
| "step": 262 |
| }, |
| { |
| "epoch": 2.461176470588235, |
| "grad_norm": 0.13526563346385956, |
| "learning_rate": 3.514376996805112e-05, |
| "loss": 0.5538, |
| "step": 263 |
| }, |
| { |
| "epoch": 2.4705882352941178, |
| "grad_norm": 0.1388704627752304, |
| "learning_rate": 3.450479233226837e-05, |
| "loss": 0.5671, |
| "step": 264 |
| }, |
| { |
| "epoch": 2.48, |
| "grad_norm": 0.13550125062465668, |
| "learning_rate": 3.386581469648562e-05, |
| "loss": 0.5431, |
| "step": 265 |
| }, |
| { |
| "epoch": 2.489411764705882, |
| "grad_norm": 0.13878309726715088, |
| "learning_rate": 3.322683706070287e-05, |
| "loss": 0.5771, |
| "step": 266 |
| }, |
| { |
| "epoch": 2.498823529411765, |
| "grad_norm": 0.15442192554473877, |
| "learning_rate": 3.258785942492013e-05, |
| "loss": 0.5689, |
| "step": 267 |
| }, |
| { |
| "epoch": 2.508235294117647, |
| "grad_norm": 0.13992522656917572, |
| "learning_rate": 3.194888178913738e-05, |
| "loss": 0.5514, |
| "step": 268 |
| }, |
| { |
| "epoch": 2.5176470588235293, |
| "grad_norm": 0.13724872469902039, |
| "learning_rate": 3.130990415335464e-05, |
| "loss": 0.5537, |
| "step": 269 |
| }, |
| { |
| "epoch": 2.527058823529412, |
| "grad_norm": 0.13630978763103485, |
| "learning_rate": 3.067092651757188e-05, |
| "loss": 0.5616, |
| "step": 270 |
| }, |
| { |
| "epoch": 2.5364705882352943, |
| "grad_norm": 0.13853025436401367, |
| "learning_rate": 3.003194888178914e-05, |
| "loss": 0.5656, |
| "step": 271 |
| }, |
| { |
| "epoch": 2.5458823529411765, |
| "grad_norm": 0.1420363038778305, |
| "learning_rate": 2.939297124600639e-05, |
| "loss": 0.5826, |
| "step": 272 |
| }, |
| { |
| "epoch": 2.5552941176470587, |
| "grad_norm": 0.1408669352531433, |
| "learning_rate": 2.8753993610223644e-05, |
| "loss": 0.5583, |
| "step": 273 |
| }, |
| { |
| "epoch": 2.564705882352941, |
| "grad_norm": 0.1526196449995041, |
| "learning_rate": 2.8115015974440894e-05, |
| "loss": 0.5917, |
| "step": 274 |
| }, |
| { |
| "epoch": 2.5741176470588236, |
| "grad_norm": 0.1629597693681717, |
| "learning_rate": 2.747603833865815e-05, |
| "loss": 0.5624, |
| "step": 275 |
| }, |
| { |
| "epoch": 2.583529411764706, |
| "grad_norm": 0.14529789984226227, |
| "learning_rate": 2.68370607028754e-05, |
| "loss": 0.5328, |
| "step": 276 |
| }, |
| { |
| "epoch": 2.592941176470588, |
| "grad_norm": 0.1526152491569519, |
| "learning_rate": 2.6198083067092656e-05, |
| "loss": 0.588, |
| "step": 277 |
| }, |
| { |
| "epoch": 2.6023529411764708, |
| "grad_norm": 0.14647038280963898, |
| "learning_rate": 2.5559105431309903e-05, |
| "loss": 0.5696, |
| "step": 278 |
| }, |
| { |
| "epoch": 2.611764705882353, |
| "grad_norm": 0.13880112767219543, |
| "learning_rate": 2.4920127795527157e-05, |
| "loss": 0.5448, |
| "step": 279 |
| }, |
| { |
| "epoch": 2.621176470588235, |
| "grad_norm": 0.1403883844614029, |
| "learning_rate": 2.428115015974441e-05, |
| "loss": 0.5485, |
| "step": 280 |
| }, |
| { |
| "epoch": 2.630588235294118, |
| "grad_norm": 0.1447409838438034, |
| "learning_rate": 2.364217252396166e-05, |
| "loss": 0.5519, |
| "step": 281 |
| }, |
| { |
| "epoch": 2.64, |
| "grad_norm": 0.14351636171340942, |
| "learning_rate": 2.3003194888178915e-05, |
| "loss": 0.5546, |
| "step": 282 |
| }, |
| { |
| "epoch": 2.6494117647058824, |
| "grad_norm": 0.13909754157066345, |
| "learning_rate": 2.2364217252396165e-05, |
| "loss": 0.5576, |
| "step": 283 |
| }, |
| { |
| "epoch": 2.6588235294117646, |
| "grad_norm": 0.14639312028884888, |
| "learning_rate": 2.172523961661342e-05, |
| "loss": 0.5786, |
| "step": 284 |
| }, |
| { |
| "epoch": 2.668235294117647, |
| "grad_norm": 0.1540895700454712, |
| "learning_rate": 2.108626198083067e-05, |
| "loss": 0.6019, |
| "step": 285 |
| }, |
| { |
| "epoch": 2.6776470588235295, |
| "grad_norm": 0.1561094969511032, |
| "learning_rate": 2.0447284345047924e-05, |
| "loss": 0.6138, |
| "step": 286 |
| }, |
| { |
| "epoch": 2.6870588235294117, |
| "grad_norm": 0.14837074279785156, |
| "learning_rate": 1.9808306709265177e-05, |
| "loss": 0.5633, |
| "step": 287 |
| }, |
| { |
| "epoch": 2.696470588235294, |
| "grad_norm": 0.1421661227941513, |
| "learning_rate": 1.9169329073482428e-05, |
| "loss": 0.5308, |
| "step": 288 |
| }, |
| { |
| "epoch": 2.696470588235294, |
| "eval_loss": 0.6246063709259033, |
| "eval_runtime": 5.4441, |
| "eval_samples_per_second": 8.082, |
| "eval_steps_per_second": 1.102, |
| "step": 288 |
| }, |
| { |
| "epoch": 2.7058823529411766, |
| "grad_norm": 0.14756028354167938, |
| "learning_rate": 1.8530351437699682e-05, |
| "loss": 0.5692, |
| "step": 289 |
| }, |
| { |
| "epoch": 2.715294117647059, |
| "grad_norm": 0.1475173979997635, |
| "learning_rate": 1.7891373801916932e-05, |
| "loss": 0.5383, |
| "step": 290 |
| }, |
| { |
| "epoch": 2.724705882352941, |
| "grad_norm": 0.14893798530101776, |
| "learning_rate": 1.7252396166134186e-05, |
| "loss": 0.5603, |
| "step": 291 |
| }, |
| { |
| "epoch": 2.7341176470588238, |
| "grad_norm": 0.14635878801345825, |
| "learning_rate": 1.6613418530351437e-05, |
| "loss": 0.573, |
| "step": 292 |
| }, |
| { |
| "epoch": 2.743529411764706, |
| "grad_norm": 0.14744271337985992, |
| "learning_rate": 1.597444089456869e-05, |
| "loss": 0.5646, |
| "step": 293 |
| }, |
| { |
| "epoch": 2.7529411764705882, |
| "grad_norm": 0.14066603779792786, |
| "learning_rate": 1.533546325878594e-05, |
| "loss": 0.5335, |
| "step": 294 |
| }, |
| { |
| "epoch": 2.7623529411764705, |
| "grad_norm": 0.14420145750045776, |
| "learning_rate": 1.4696485623003195e-05, |
| "loss": 0.5885, |
| "step": 295 |
| }, |
| { |
| "epoch": 2.7717647058823527, |
| "grad_norm": 0.1417757123708725, |
| "learning_rate": 1.4057507987220447e-05, |
| "loss": 0.525, |
| "step": 296 |
| }, |
| { |
| "epoch": 2.7811764705882354, |
| "grad_norm": 0.13913412392139435, |
| "learning_rate": 1.34185303514377e-05, |
| "loss": 0.5355, |
| "step": 297 |
| }, |
| { |
| "epoch": 2.7905882352941176, |
| "grad_norm": 0.14291714131832123, |
| "learning_rate": 1.2779552715654951e-05, |
| "loss": 0.5419, |
| "step": 298 |
| }, |
| { |
| "epoch": 2.8, |
| "grad_norm": 0.15101145207881927, |
| "learning_rate": 1.2140575079872205e-05, |
| "loss": 0.5945, |
| "step": 299 |
| }, |
| { |
| "epoch": 2.8094117647058825, |
| "grad_norm": 0.15145310759544373, |
| "learning_rate": 1.1501597444089457e-05, |
| "loss": 0.5801, |
| "step": 300 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 318, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 3, |
| "save_steps": 300, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 4.1701101665255424e+17, |
| "train_batch_size": 2, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|