{ "best_metric": null, "best_model_checkpoint": null, "epoch": 4.924760601915184, "eval_steps": 500, "global_step": 110, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.04377564979480164, "grad_norm": 10.977038416240942, "learning_rate": 7.272727272727273e-06, "loss": 1.6783, "step": 1 }, { "epoch": 0.08755129958960328, "grad_norm": 10.685982173301461, "learning_rate": 1.4545454545454546e-05, "loss": 1.6756, "step": 2 }, { "epoch": 0.13132694938440492, "grad_norm": 7.560341958816415, "learning_rate": 2.1818181818181818e-05, "loss": 1.5708, "step": 3 }, { "epoch": 0.17510259917920656, "grad_norm": 6.257863682610026, "learning_rate": 2.9090909090909093e-05, "loss": 1.5415, "step": 4 }, { "epoch": 0.2188782489740082, "grad_norm": 9.338746208897016, "learning_rate": 3.6363636363636364e-05, "loss": 1.5506, "step": 5 }, { "epoch": 0.26265389876880985, "grad_norm": 6.7019369753380404, "learning_rate": 4.3636363636363636e-05, "loss": 1.4494, "step": 6 }, { "epoch": 0.3064295485636115, "grad_norm": 2.7387345576832467, "learning_rate": 5.0909090909090914e-05, "loss": 1.3764, "step": 7 }, { "epoch": 0.35020519835841313, "grad_norm": 2.9905508612885416, "learning_rate": 5.8181818181818185e-05, "loss": 1.3124, "step": 8 }, { "epoch": 0.39398084815321477, "grad_norm": 5.142107061812394, "learning_rate": 6.545454545454546e-05, "loss": 1.3263, "step": 9 }, { "epoch": 0.4377564979480164, "grad_norm": 6.498361005724975, "learning_rate": 7.272727272727273e-05, "loss": 1.302, "step": 10 }, { "epoch": 0.48153214774281805, "grad_norm": 4.736561117836272, "learning_rate": 8e-05, "loss": 1.2851, "step": 11 }, { "epoch": 0.5253077975376197, "grad_norm": 4.913666681984767, "learning_rate": 7.997986169532741e-05, "loss": 1.2886, "step": 12 }, { "epoch": 0.5690834473324213, "grad_norm": 3.3337518195815763, "learning_rate": 7.991946705887539e-05, "loss": 1.2546, "step": 13 }, { "epoch": 0.612859097127223, "grad_norm": 2.7069045612640275, "learning_rate": 7.981887690292339e-05, "loss": 1.2314, "step": 14 }, { "epoch": 0.6566347469220246, "grad_norm": 2.8584488966099855, "learning_rate": 7.967819251323182e-05, "loss": 1.2315, "step": 15 }, { "epoch": 0.7004103967168263, "grad_norm": 3.1377249425512765, "learning_rate": 7.949755554705577e-05, "loss": 1.2093, "step": 16 }, { "epoch": 0.7441860465116279, "grad_norm": 1.3936091349070174, "learning_rate": 7.927714789050826e-05, "loss": 1.196, "step": 17 }, { "epoch": 0.7879616963064295, "grad_norm": 3.028923392641176, "learning_rate": 7.90171914754163e-05, "loss": 1.2348, "step": 18 }, { "epoch": 0.8317373461012312, "grad_norm": 2.196134083608347, "learning_rate": 7.871794805585427e-05, "loss": 1.2053, "step": 19 }, { "epoch": 0.8755129958960328, "grad_norm": 1.6611206571415684, "learning_rate": 7.837971894457991e-05, "loss": 1.1834, "step": 20 }, { "epoch": 0.9192886456908345, "grad_norm": 2.681882547825232, "learning_rate": 7.800284470963783e-05, "loss": 1.1831, "step": 21 }, { "epoch": 0.9630642954856361, "grad_norm": 1.5845981079722795, "learning_rate": 7.758770483143634e-05, "loss": 1.1698, "step": 22 }, { "epoch": 1.0341997264021887, "grad_norm": 4.534657915532011, "learning_rate": 7.71347173206429e-05, "loss": 2.0516, "step": 23 }, { "epoch": 1.0779753761969904, "grad_norm": 2.8451498787070433, "learning_rate": 7.664433829728279e-05, "loss": 1.152, "step": 24 }, { "epoch": 1.121751025991792, "grad_norm": 1.9899709698106878, "learning_rate": 7.611706153146486e-05, "loss": 1.1468, "step": 25 }, { "epoch": 1.1655266757865936, "grad_norm": 2.8389265579783345, "learning_rate": 7.555341794619695e-05, "loss": 1.1408, "step": 26 }, { "epoch": 1.2093023255813953, "grad_norm": 2.038952629316326, "learning_rate": 7.49539750827914e-05, "loss": 1.1219, "step": 27 }, { "epoch": 1.253077975376197, "grad_norm": 2.0796847065206743, "learning_rate": 7.431933652939909e-05, "loss": 1.1136, "step": 28 }, { "epoch": 1.2968536251709986, "grad_norm": 1.447940531771876, "learning_rate": 7.365014131324725e-05, "loss": 1.1191, "step": 29 }, { "epoch": 1.3406292749658002, "grad_norm": 1.337447462838519, "learning_rate": 7.294706325719331e-05, "loss": 1.1145, "step": 30 }, { "epoch": 1.3844049247606018, "grad_norm": 2.831107849480652, "learning_rate": 7.221081030124235e-05, "loss": 1.1269, "step": 31 }, { "epoch": 1.4281805745554035, "grad_norm": 1.3913695315986807, "learning_rate": 7.144212378971151e-05, "loss": 1.1065, "step": 32 }, { "epoch": 1.4719562243502051, "grad_norm": 2.9279595696419647, "learning_rate": 7.064177772475912e-05, "loss": 1.1347, "step": 33 }, { "epoch": 1.5157318741450068, "grad_norm": 2.1607556666180012, "learning_rate": 6.98105779870302e-05, "loss": 1.1176, "step": 34 }, { "epoch": 1.5595075239398084, "grad_norm": 2.035595521860542, "learning_rate": 6.89493615242028e-05, "loss": 1.0969, "step": 35 }, { "epoch": 1.60328317373461, "grad_norm": 1.9632755753820994, "learning_rate": 6.805899550825285e-05, "loss": 1.1039, "step": 36 }, { "epoch": 1.6470588235294117, "grad_norm": 1.756863438311583, "learning_rate": 6.714037646228529e-05, "loss": 1.1059, "step": 37 }, { "epoch": 1.6908344733242133, "grad_norm": 1.6660401725071718, "learning_rate": 6.619442935781141e-05, "loss": 1.098, "step": 38 }, { "epoch": 1.734610123119015, "grad_norm": 1.4739841905330868, "learning_rate": 6.52221066833809e-05, "loss": 1.0852, "step": 39 }, { "epoch": 1.7783857729138166, "grad_norm": 1.3135170392231592, "learning_rate": 6.422438748550667e-05, "loss": 1.0901, "step": 40 }, { "epoch": 1.8221614227086183, "grad_norm": 1.3850190455081264, "learning_rate": 6.320227638284793e-05, "loss": 1.0781, "step": 41 }, { "epoch": 1.86593707250342, "grad_norm": 0.9405053138714191, "learning_rate": 6.215680255464442e-05, "loss": 1.0825, "step": 42 }, { "epoch": 1.9097127222982215, "grad_norm": 1.1113532920678866, "learning_rate": 6.10890187044201e-05, "loss": 1.0766, "step": 43 }, { "epoch": 1.9534883720930232, "grad_norm": 1.187498515237833, "learning_rate": 6.000000000000001e-05, "loss": 1.0763, "step": 44 }, { "epoch": 2.024623803009576, "grad_norm": 2.160651041437791, "learning_rate": 5.889084299090732e-05, "loss": 1.9051, "step": 45 }, { "epoch": 2.0683994528043774, "grad_norm": 0.9784307660310307, "learning_rate": 5.776266450423097e-05, "loss": 1.0442, "step": 46 }, { "epoch": 2.112175102599179, "grad_norm": 1.031157632132424, "learning_rate": 5.661660052007547e-05, "loss": 1.0507, "step": 47 }, { "epoch": 2.1559507523939807, "grad_norm": 1.4533738667722713, "learning_rate": 5.5453805027725145e-05, "loss": 1.0499, "step": 48 }, { "epoch": 2.1997264021887823, "grad_norm": 0.875133136635367, "learning_rate": 5.427544886367488e-05, "loss": 1.0381, "step": 49 }, { "epoch": 2.243502051983584, "grad_norm": 0.9817484749723364, "learning_rate": 5.3082718532696874e-05, "loss": 1.0437, "step": 50 }, { "epoch": 2.2872777017783856, "grad_norm": 1.223185197293581, "learning_rate": 5.1876815013131e-05, "loss": 1.0329, "step": 51 }, { "epoch": 2.3310533515731873, "grad_norm": 0.8071734287546062, "learning_rate": 5.06589525476014e-05, "loss": 1.0287, "step": 52 }, { "epoch": 2.374829001367989, "grad_norm": 1.290044073399014, "learning_rate": 4.94303574203771e-05, "loss": 1.04, "step": 53 }, { "epoch": 2.4186046511627906, "grad_norm": 0.6470063533070358, "learning_rate": 4.819226672260763e-05, "loss": 1.0331, "step": 54 }, { "epoch": 2.462380300957592, "grad_norm": 0.9069681051512629, "learning_rate": 4.694592710667723e-05, "loss": 1.0243, "step": 55 }, { "epoch": 2.506155950752394, "grad_norm": 0.8654968713633596, "learning_rate": 4.5692593530931416e-05, "loss": 1.0052, "step": 56 }, { "epoch": 2.5499316005471955, "grad_norm": 1.0033676935022624, "learning_rate": 4.4433527996040443e-05, "loss": 1.0299, "step": 57 }, { "epoch": 2.593707250341997, "grad_norm": 0.6256956675554375, "learning_rate": 4.316999827427154e-05, "loss": 1.0344, "step": 58 }, { "epoch": 2.6374829001367988, "grad_norm": 0.7160307755017002, "learning_rate": 4.19032766329497e-05, "loss": 1.0197, "step": 59 }, { "epoch": 2.6812585499316004, "grad_norm": 0.7016377338794068, "learning_rate": 4.063463855339232e-05, "loss": 1.0142, "step": 60 }, { "epoch": 2.725034199726402, "grad_norm": 0.5412134943726296, "learning_rate": 3.9365361446607684e-05, "loss": 1.0194, "step": 61 }, { "epoch": 2.7688098495212037, "grad_norm": 0.5581535816981518, "learning_rate": 3.809672336705031e-05, "loss": 1.021, "step": 62 }, { "epoch": 2.8125854993160053, "grad_norm": 0.4895297941884866, "learning_rate": 3.683000172572846e-05, "loss": 1.0145, "step": 63 }, { "epoch": 2.856361149110807, "grad_norm": 0.5430942752350344, "learning_rate": 3.556647200395956e-05, "loss": 1.0159, "step": 64 }, { "epoch": 2.9001367989056086, "grad_norm": 0.43778465011409384, "learning_rate": 3.4307406469068604e-05, "loss": 1.0037, "step": 65 }, { "epoch": 2.9439124487004102, "grad_norm": 0.5574066963629686, "learning_rate": 3.305407289332279e-05, "loss": 1.0124, "step": 66 }, { "epoch": 3.015047879616963, "grad_norm": 0.8597898186828361, "learning_rate": 3.180773327739238e-05, "loss": 1.8113, "step": 67 }, { "epoch": 3.0588235294117645, "grad_norm": 0.7562084727521547, "learning_rate": 3.0569642579622905e-05, "loss": 0.9808, "step": 68 }, { "epoch": 3.102599179206566, "grad_norm": 0.5467749248417106, "learning_rate": 2.9341047452398607e-05, "loss": 0.9786, "step": 69 }, { "epoch": 3.146374829001368, "grad_norm": 0.7317552492863292, "learning_rate": 2.8123184986869022e-05, "loss": 0.9807, "step": 70 }, { "epoch": 3.19015047879617, "grad_norm": 0.4988928521059118, "learning_rate": 2.691728146730314e-05, "loss": 0.9801, "step": 71 }, { "epoch": 3.233926128590971, "grad_norm": 0.6214726877620874, "learning_rate": 2.5724551136325132e-05, "loss": 0.9752, "step": 72 }, { "epoch": 3.277701778385773, "grad_norm": 0.5723924398692547, "learning_rate": 2.4546194972274852e-05, "loss": 0.9833, "step": 73 }, { "epoch": 3.3214774281805743, "grad_norm": 0.3802903109677986, "learning_rate": 2.338339947992455e-05, "loss": 0.9704, "step": 74 }, { "epoch": 3.3652530779753764, "grad_norm": 0.5679922591981417, "learning_rate": 2.2237335495769035e-05, "loss": 0.9832, "step": 75 }, { "epoch": 3.4090287277701776, "grad_norm": 0.32921569959729985, "learning_rate": 2.11091570090927e-05, "loss": 0.9832, "step": 76 }, { "epoch": 3.4528043775649797, "grad_norm": 0.41963134099051513, "learning_rate": 2.0000000000000012e-05, "loss": 0.9925, "step": 77 }, { "epoch": 3.496580027359781, "grad_norm": 0.4216326585626274, "learning_rate": 1.8910981295579903e-05, "loss": 0.9693, "step": 78 }, { "epoch": 3.540355677154583, "grad_norm": 0.281097341488956, "learning_rate": 1.7843197445355593e-05, "loss": 0.9687, "step": 79 }, { "epoch": 3.584131326949384, "grad_norm": 0.34537711442895225, "learning_rate": 1.679772361715208e-05, "loss": 0.9792, "step": 80 }, { "epoch": 3.6279069767441863, "grad_norm": 0.38729234396659473, "learning_rate": 1.5775612514493343e-05, "loss": 0.9716, "step": 81 }, { "epoch": 3.6716826265389875, "grad_norm": 0.23216602871427763, "learning_rate": 1.4777893316619114e-05, "loss": 0.9791, "step": 82 }, { "epoch": 3.7154582763337896, "grad_norm": 0.34978409677834293, "learning_rate": 1.3805570642188602e-05, "loss": 0.9686, "step": 83 }, { "epoch": 3.7592339261285908, "grad_norm": 0.2917010744705113, "learning_rate": 1.2859623537714719e-05, "loss": 0.9674, "step": 84 }, { "epoch": 3.803009575923393, "grad_norm": 0.2659722448857111, "learning_rate": 1.1941004491747145e-05, "loss": 0.9639, "step": 85 }, { "epoch": 3.846785225718194, "grad_norm": 0.2936639116335773, "learning_rate": 1.1050638475797193e-05, "loss": 0.9609, "step": 86 }, { "epoch": 3.890560875512996, "grad_norm": 0.22955492375415615, "learning_rate": 1.0189422012969814e-05, "loss": 0.9722, "step": 87 }, { "epoch": 3.9343365253077973, "grad_norm": 0.20270732194600466, "learning_rate": 9.358222275240884e-06, "loss": 0.968, "step": 88 }, { "epoch": 4.00547195622435, "grad_norm": 0.4286144089204306, "learning_rate": 8.55787621028851e-06, "loss": 1.746, "step": 89 }, { "epoch": 4.049247606019152, "grad_norm": 0.156416222673832, "learning_rate": 7.789189698757656e-06, "loss": 0.9564, "step": 90 }, { "epoch": 4.093023255813954, "grad_norm": 0.23495972177315438, "learning_rate": 7.052936742806693e-06, "loss": 0.9593, "step": 91 }, { "epoch": 4.136798905608755, "grad_norm": 0.2226318621452638, "learning_rate": 6.349858686752748e-06, "loss": 0.9627, "step": 92 }, { "epoch": 4.180574555403557, "grad_norm": 0.16215755051665368, "learning_rate": 5.680663470600918e-06, "loss": 0.9486, "step": 93 }, { "epoch": 4.224350205198358, "grad_norm": 0.17528331588918997, "learning_rate": 5.046024917208603e-06, "loss": 0.9564, "step": 94 }, { "epoch": 4.26812585499316, "grad_norm": 0.19247905962597153, "learning_rate": 4.446582053803066e-06, "loss": 0.9532, "step": 95 }, { "epoch": 4.311901504787961, "grad_norm": 0.1497248018118723, "learning_rate": 3.882938468535158e-06, "loss": 0.9441, "step": 96 }, { "epoch": 4.3556771545827635, "grad_norm": 0.1368167463262844, "learning_rate": 3.3556617027172168e-06, "loss": 0.9502, "step": 97 }, { "epoch": 4.399452804377565, "grad_norm": 0.14584282376533797, "learning_rate": 2.8652826793570975e-06, "loss": 0.9499, "step": 98 }, { "epoch": 4.443228454172367, "grad_norm": 0.1507749360921821, "learning_rate": 2.4122951685636674e-06, "loss": 0.9623, "step": 99 }, { "epoch": 4.487004103967168, "grad_norm": 0.12741172341135956, "learning_rate": 1.997155290362187e-06, "loss": 0.9546, "step": 100 }, { "epoch": 4.53077975376197, "grad_norm": 0.11988388634885762, "learning_rate": 1.6202810554201099e-06, "loss": 0.9634, "step": 101 }, { "epoch": 4.574555403556771, "grad_norm": 0.10876564961414978, "learning_rate": 1.2820519441457502e-06, "loss": 0.9489, "step": 102 }, { "epoch": 4.618331053351573, "grad_norm": 0.10280656448879005, "learning_rate": 9.828085245837183e-07, "loss": 0.9481, "step": 103 }, { "epoch": 4.6621067031463745, "grad_norm": 0.10558127777006338, "learning_rate": 7.228521094917318e-07, "loss": 0.9517, "step": 104 }, { "epoch": 4.705882352941177, "grad_norm": 0.10826937864073408, "learning_rate": 5.024444529442285e-07, "loss": 0.9544, "step": 105 }, { "epoch": 4.749658002735978, "grad_norm": 0.10594923812863334, "learning_rate": 3.218074867681864e-07, "loss": 0.9384, "step": 106 }, { "epoch": 4.79343365253078, "grad_norm": 0.10786116991531677, "learning_rate": 1.8112309707661647e-07, "loss": 0.943, "step": 107 }, { "epoch": 4.837209302325581, "grad_norm": 0.11373942157615996, "learning_rate": 8.053294112462696e-08, "loss": 0.9525, "step": 108 }, { "epoch": 4.880984952120383, "grad_norm": 0.10941188396825419, "learning_rate": 2.01383046725967e-08, "loss": 0.9539, "step": 109 }, { "epoch": 4.924760601915184, "grad_norm": 0.10542013778303103, "learning_rate": 0.0, "loss": 0.9416, "step": 110 }, { "epoch": 4.924760601915184, "step": 110, "total_flos": 2.9249198798007173e+18, "train_loss": 1.1118553096597845, "train_runtime": 25822.1288, "train_samples_per_second": 2.263, "train_steps_per_second": 0.004 } ], "logging_steps": 1, "max_steps": 110, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.9249198798007173e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }