{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 171, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.03549639489739324, "grad_norm": 127.17037963867188, "learning_rate": 5.555555555555555e-07, "loss": 3.1202, "step": 2 }, { "epoch": 0.07099278979478647, "grad_norm": 127.35641479492188, "learning_rate": 1.6666666666666667e-06, "loss": 3.1169, "step": 4 }, { "epoch": 0.1064891846921797, "grad_norm": 94.25624084472656, "learning_rate": 2.7777777777777783e-06, "loss": 2.8233, "step": 6 }, { "epoch": 0.14198557958957295, "grad_norm": 17.775108337402344, "learning_rate": 3.88888888888889e-06, "loss": 1.8356, "step": 8 }, { "epoch": 0.17748197448696618, "grad_norm": 10.735865592956543, "learning_rate": 5e-06, "loss": 1.5481, "step": 10 }, { "epoch": 0.2129783693843594, "grad_norm": 7.582458972930908, "learning_rate": 6.111111111111112e-06, "loss": 1.2527, "step": 12 }, { "epoch": 0.24847476428175264, "grad_norm": 5.5655999183654785, "learning_rate": 7.222222222222223e-06, "loss": 1.0648, "step": 14 }, { "epoch": 0.2839711591791459, "grad_norm": 4.178959369659424, "learning_rate": 8.333333333333334e-06, "loss": 0.8722, "step": 16 }, { "epoch": 0.3194675540765391, "grad_norm": 2.8051388263702393, "learning_rate": 9.444444444444445e-06, "loss": 0.6238, "step": 18 }, { "epoch": 0.35496394897393235, "grad_norm": 2.487281322479248, "learning_rate": 9.998945997517957e-06, "loss": 0.493, "step": 20 }, { "epoch": 0.39046034387132555, "grad_norm": 2.143423557281494, "learning_rate": 9.990516643685222e-06, "loss": 0.4248, "step": 22 }, { "epoch": 0.4259567387687188, "grad_norm": 1.8732106685638428, "learning_rate": 9.973672149817232e-06, "loss": 0.2676, "step": 24 }, { "epoch": 0.461453133666112, "grad_norm": 1.6402384042739868, "learning_rate": 9.948440919541277e-06, "loss": 0.2245, "step": 26 }, { "epoch": 0.49694952856350527, "grad_norm": 1.3924551010131836, "learning_rate": 9.91486549841951e-06, "loss": 0.1811, "step": 28 }, { "epoch": 0.5324459234608985, "grad_norm": 2.457998037338257, "learning_rate": 9.873002502207502e-06, "loss": 0.1166, "step": 30 }, { "epoch": 0.5679423183582918, "grad_norm": 1.0437461137771606, "learning_rate": 9.822922521387277e-06, "loss": 0.1204, "step": 32 }, { "epoch": 0.603438713255685, "grad_norm": 3.6355228424072266, "learning_rate": 9.764710002135784e-06, "loss": 0.1161, "step": 34 }, { "epoch": 0.6389351081530782, "grad_norm": 1.3620433807373047, "learning_rate": 9.698463103929542e-06, "loss": 0.1179, "step": 36 }, { "epoch": 0.6744315030504714, "grad_norm": 0.5827310681343079, "learning_rate": 9.62429353402556e-06, "loss": 0.0806, "step": 38 }, { "epoch": 0.7099278979478647, "grad_norm": 0.9769260883331299, "learning_rate": 9.542326359097619e-06, "loss": 0.0941, "step": 40 }, { "epoch": 0.7454242928452579, "grad_norm": 0.6220430731773376, "learning_rate": 9.452699794345583e-06, "loss": 0.0761, "step": 42 }, { "epoch": 0.7809206877426511, "grad_norm": 0.5837266445159912, "learning_rate": 9.355564970433288e-06, "loss": 0.068, "step": 44 }, { "epoch": 0.8164170826400444, "grad_norm": 0.445272296667099, "learning_rate": 9.251085678648072e-06, "loss": 0.0642, "step": 46 }, { "epoch": 0.8519134775374376, "grad_norm": 0.4069404900074005, "learning_rate": 9.13943809471159e-06, "loss": 0.0526, "step": 48 }, { "epoch": 0.8874098724348308, "grad_norm": 0.47115278244018555, "learning_rate": 9.020810481707709e-06, "loss": 0.049, "step": 50 }, { "epoch": 0.922906267332224, "grad_norm": 0.9125522375106812, "learning_rate": 8.895402872628352e-06, "loss": 0.0547, "step": 52 }, { "epoch": 0.9584026622296173, "grad_norm": 0.3842329978942871, "learning_rate": 8.763426733072624e-06, "loss": 0.0449, "step": 54 }, { "epoch": 0.9938990571270105, "grad_norm": 0.7765663266181946, "learning_rate": 8.625104604667965e-06, "loss": 0.0618, "step": 56 }, { "epoch": 1.0177481974486966, "grad_norm": 0.39780569076538086, "learning_rate": 8.480669729814635e-06, "loss": 0.0477, "step": 58 }, { "epoch": 1.0532445923460898, "grad_norm": 0.3841244578361511, "learning_rate": 8.330365658386252e-06, "loss": 0.0412, "step": 60 }, { "epoch": 1.088740987243483, "grad_norm": 0.39046695828437805, "learning_rate": 8.174445837049614e-06, "loss": 0.0366, "step": 62 }, { "epoch": 1.1242373821408762, "grad_norm": 0.35411760210990906, "learning_rate": 8.013173181896283e-06, "loss": 0.0381, "step": 64 }, { "epoch": 1.1597337770382696, "grad_norm": 0.37250956892967224, "learning_rate": 7.846819635106569e-06, "loss": 0.0379, "step": 66 }, { "epoch": 1.1952301719356628, "grad_norm": 0.4131050407886505, "learning_rate": 7.675665706393502e-06, "loss": 0.0381, "step": 68 }, { "epoch": 1.230726566833056, "grad_norm": 0.4184521734714508, "learning_rate": 7.500000000000001e-06, "loss": 0.0348, "step": 70 }, { "epoch": 1.2662229617304492, "grad_norm": 0.3667065501213074, "learning_rate": 7.320118728046818e-06, "loss": 0.0357, "step": 72 }, { "epoch": 1.3017193566278424, "grad_norm": 0.307443231344223, "learning_rate": 7.136325211051905e-06, "loss": 0.0344, "step": 74 }, { "epoch": 1.3372157515252356, "grad_norm": 0.3104756474494934, "learning_rate": 6.948929366463397e-06, "loss": 0.037, "step": 76 }, { "epoch": 1.372712146422629, "grad_norm": 0.4366794228553772, "learning_rate": 6.758247186068684e-06, "loss": 0.0387, "step": 78 }, { "epoch": 1.4082085413200223, "grad_norm": 0.32028865814208984, "learning_rate": 6.5646002031607726e-06, "loss": 0.0313, "step": 80 }, { "epoch": 1.4437049362174155, "grad_norm": 0.2846560776233673, "learning_rate": 6.368314950360416e-06, "loss": 0.0348, "step": 82 }, { "epoch": 1.4792013311148087, "grad_norm": 0.47968789935112, "learning_rate": 6.169722409008244e-06, "loss": 0.0302, "step": 84 }, { "epoch": 1.5146977260122019, "grad_norm": 0.6367243528366089, "learning_rate": 5.9691574510553505e-06, "loss": 0.0335, "step": 86 }, { "epoch": 1.550194120909595, "grad_norm": 0.25928691029548645, "learning_rate": 5.766958274393428e-06, "loss": 0.0334, "step": 88 }, { "epoch": 1.5856905158069883, "grad_norm": 4.997822284698486, "learning_rate": 5.5634658325766066e-06, "loss": 0.0272, "step": 90 }, { "epoch": 1.6211869107043815, "grad_norm": 0.2188279628753662, "learning_rate": 5.359023259896638e-06, "loss": 0.0269, "step": 92 }, { "epoch": 1.6566833056017747, "grad_norm": 0.35858553647994995, "learning_rate": 5.153975292780852e-06, "loss": 0.0352, "step": 94 }, { "epoch": 1.692179700499168, "grad_norm": 0.308819979429245, "learning_rate": 4.948667688488552e-06, "loss": 0.0327, "step": 96 }, { "epoch": 1.7276760953965613, "grad_norm": 0.5604074001312256, "learning_rate": 4.7434466420860515e-06, "loss": 0.033, "step": 98 }, { "epoch": 1.7631724902939545, "grad_norm": 0.31649985909461975, "learning_rate": 4.53865820268349e-06, "loss": 0.0292, "step": 100 }, { "epoch": 1.7986688851913477, "grad_norm": 0.2754175662994385, "learning_rate": 4.334647689917734e-06, "loss": 0.0379, "step": 102 }, { "epoch": 1.8341652800887411, "grad_norm": 0.3243215084075928, "learning_rate": 4.131759111665349e-06, "loss": 0.0315, "step": 104 }, { "epoch": 1.8696616749861343, "grad_norm": 0.2798413336277008, "learning_rate": 3.930334583967514e-06, "loss": 0.0345, "step": 106 }, { "epoch": 1.9051580698835275, "grad_norm": 0.2298552244901657, "learning_rate": 3.730713754144961e-06, "loss": 0.0277, "step": 108 }, { "epoch": 1.9406544647809207, "grad_norm": 0.44037896394729614, "learning_rate": 3.5332332280757706e-06, "loss": 0.0276, "step": 110 }, { "epoch": 1.976150859678314, "grad_norm": 0.41545015573501587, "learning_rate": 3.3382260026017027e-06, "loss": 0.0383, "step": 112 }, { "epoch": 2.0, "grad_norm": 0.3469395339488983, "learning_rate": 3.1460209040201967e-06, "loss": 0.0258, "step": 114 }, { "epoch": 2.035496394897393, "grad_norm": 0.30031341314315796, "learning_rate": 2.956942033608843e-06, "loss": 0.0221, "step": 116 }, { "epoch": 2.0709927897947864, "grad_norm": 0.23123888671398163, "learning_rate": 2.771308221117309e-06, "loss": 0.027, "step": 118 }, { "epoch": 2.1064891846921796, "grad_norm": 0.2226598858833313, "learning_rate": 2.5894324871482557e-06, "loss": 0.0188, "step": 120 }, { "epoch": 2.141985579589573, "grad_norm": 0.24335065484046936, "learning_rate": 2.411621515333788e-06, "loss": 0.0226, "step": 122 }, { "epoch": 2.177481974486966, "grad_norm": 0.4972982108592987, "learning_rate": 2.238175135197471e-06, "loss": 0.0221, "step": 124 }, { "epoch": 2.212978369384359, "grad_norm": 0.199220210313797, "learning_rate": 2.069385816573928e-06, "loss": 0.0182, "step": 126 }, { "epoch": 2.2484747642817524, "grad_norm": 0.2942477762699127, "learning_rate": 1.9055381764385272e-06, "loss": 0.0257, "step": 128 }, { "epoch": 2.283971159179146, "grad_norm": 0.2242165505886078, "learning_rate": 1.746908498978791e-06, "loss": 0.0217, "step": 130 }, { "epoch": 2.3194675540765393, "grad_norm": 0.19704177975654602, "learning_rate": 1.5937642697167288e-06, "loss": 0.0201, "step": 132 }, { "epoch": 2.3549639489739325, "grad_norm": 0.3458758592605591, "learning_rate": 1.4463637244677648e-06, "loss": 0.0214, "step": 134 }, { "epoch": 2.3904603438713257, "grad_norm": 0.20692254602909088, "learning_rate": 1.3049554138967052e-06, "loss": 0.0173, "step": 136 }, { "epoch": 2.425956738768719, "grad_norm": 0.23293040692806244, "learning_rate": 1.1697777844051105e-06, "loss": 0.0217, "step": 138 }, { "epoch": 2.461453133666112, "grad_norm": 0.24933604896068573, "learning_rate": 1.0410587760567104e-06, "loss": 0.0174, "step": 140 }, { "epoch": 2.4969495285635053, "grad_norm": 0.24881727993488312, "learning_rate": 9.190154382188921e-07, "loss": 0.0201, "step": 142 }, { "epoch": 2.5324459234608985, "grad_norm": 0.22158710658550262, "learning_rate": 8.03853563568367e-07, "loss": 0.0194, "step": 144 }, { "epoch": 2.5679423183582917, "grad_norm": 0.2617812156677246, "learning_rate": 6.957673410781617e-07, "loss": 0.0238, "step": 146 }, { "epoch": 2.603438713255685, "grad_norm": 0.23341882228851318, "learning_rate": 5.949390285710777e-07, "loss": 0.0207, "step": 148 }, { "epoch": 2.638935108153078, "grad_norm": 0.20265567302703857, "learning_rate": 5.015386453917742e-07, "loss": 0.0195, "step": 150 }, { "epoch": 2.6744315030504713, "grad_norm": 0.24993731081485748, "learning_rate": 4.15723685715686e-07, "loss": 0.0219, "step": 152 }, { "epoch": 2.709927897947865, "grad_norm": 0.20348049700260162, "learning_rate": 3.3763885297822153e-07, "loss": 0.0172, "step": 154 }, { "epoch": 2.745424292845258, "grad_norm": 0.43210646510124207, "learning_rate": 2.6741581587202747e-07, "loss": 0.0228, "step": 156 }, { "epoch": 2.7809206877426513, "grad_norm": 0.2683945298194885, "learning_rate": 2.0517298632379445e-07, "loss": 0.0194, "step": 158 }, { "epoch": 2.8164170826400445, "grad_norm": 0.2853659689426422, "learning_rate": 1.510153198249531e-07, "loss": 0.0214, "step": 160 }, { "epoch": 2.8519134775374377, "grad_norm": 0.28844180703163147, "learning_rate": 1.0503413845297739e-07, "loss": 0.0217, "step": 162 }, { "epoch": 2.887409872434831, "grad_norm": 0.38325849175453186, "learning_rate": 6.730697688170251e-08, "loss": 0.0246, "step": 164 }, { "epoch": 2.922906267332224, "grad_norm": 0.4123646020889282, "learning_rate": 3.7897451640321326e-08, "loss": 0.0239, "step": 166 }, { "epoch": 2.9584026622296173, "grad_norm": 0.2761909067630768, "learning_rate": 1.6855153841527915e-08, "loss": 0.0198, "step": 168 }, { "epoch": 2.9938990571270105, "grad_norm": 0.28191664814949036, "learning_rate": 4.2155655596809455e-09, "loss": 0.0218, "step": 170 } ], "logging_steps": 2, "max_steps": 171, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 25, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 208500261126144.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }