{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.12977265992858178, "eval_steps": 500, "global_step": 60000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.002162877665476363, "grad_norm": 2.268080472946167, "learning_rate": 4.9945869519559514e-05, "loss": 7.7847, "step": 1000 }, { "epoch": 0.004325755330952726, "grad_norm": 2.551193952560425, "learning_rate": 4.983760855867853e-05, "loss": 5.9343, "step": 2000 }, { "epoch": 0.006488632996429089, "grad_norm": 2.584500551223755, "learning_rate": 4.972934759779754e-05, "loss": 5.3418, "step": 3000 }, { "epoch": 0.008651510661905452, "grad_norm": 2.7399814128875732, "learning_rate": 4.962108663691656e-05, "loss": 4.9453, "step": 4000 }, { "epoch": 0.010814388327381816, "grad_norm": 2.7650134563446045, "learning_rate": 4.9512825676035575e-05, "loss": 4.6277, "step": 5000 }, { "epoch": 0.012977265992858178, "grad_norm": 2.8655128479003906, "learning_rate": 4.9404564715154586e-05, "loss": 4.3648, "step": 6000 }, { "epoch": 0.015140143658334541, "grad_norm": 2.8312480449676514, "learning_rate": 4.92963037542736e-05, "loss": 4.1441, "step": 7000 }, { "epoch": 0.017303021323810903, "grad_norm": 2.842963218688965, "learning_rate": 4.918804279339262e-05, "loss": 3.9787, "step": 8000 }, { "epoch": 0.019465898989287266, "grad_norm": 2.6685407161712646, "learning_rate": 4.907978183251164e-05, "loss": 3.8343, "step": 9000 }, { "epoch": 0.02162877665476363, "grad_norm": 2.7094473838806152, "learning_rate": 4.897152087163065e-05, "loss": 3.7211, "step": 10000 }, { "epoch": 0.023791654320239993, "grad_norm": 2.7870073318481445, "learning_rate": 4.886336817171055e-05, "loss": 3.6358, "step": 11000 }, { "epoch": 0.025954531985716355, "grad_norm": 2.6066818237304688, "learning_rate": 4.8755107210829566e-05, "loss": 3.5526, "step": 12000 }, { "epoch": 0.028117409651192717, "grad_norm": 2.8119983673095703, "learning_rate": 4.864695451090946e-05, "loss": 3.4758, "step": 13000 }, { "epoch": 0.030280287316669083, "grad_norm": 2.7353262901306152, "learning_rate": 4.853869355002847e-05, "loss": 3.4134, "step": 14000 }, { "epoch": 0.032443164982145445, "grad_norm": 2.6471846103668213, "learning_rate": 4.843054085010837e-05, "loss": 3.3547, "step": 15000 }, { "epoch": 0.03460604264762181, "grad_norm": 2.67105770111084, "learning_rate": 4.832227988922739e-05, "loss": 3.3033, "step": 16000 }, { "epoch": 0.03676892031309817, "grad_norm": 2.667982339859009, "learning_rate": 4.8214127189307287e-05, "loss": 3.2608, "step": 17000 }, { "epoch": 0.03893179797857453, "grad_norm": 2.588027238845825, "learning_rate": 4.81058662284263e-05, "loss": 3.2199, "step": 18000 }, { "epoch": 0.0410946756440509, "grad_norm": 2.650073289871216, "learning_rate": 4.799771352850619e-05, "loss": 3.1817, "step": 19000 }, { "epoch": 0.04325755330952726, "grad_norm": 2.4354302883148193, "learning_rate": 4.788945256762521e-05, "loss": 3.1464, "step": 20000 }, { "epoch": 0.045420430975003624, "grad_norm": 2.6227738857269287, "learning_rate": 4.778129986770511e-05, "loss": 3.1134, "step": 21000 }, { "epoch": 0.047583308640479986, "grad_norm": 2.5212135314941406, "learning_rate": 4.767303890682412e-05, "loss": 3.08, "step": 22000 }, { "epoch": 0.04974618630595635, "grad_norm": 2.4768059253692627, "learning_rate": 4.756488620690402e-05, "loss": 3.0511, "step": 23000 }, { "epoch": 0.05190906397143271, "grad_norm": 2.5772910118103027, "learning_rate": 4.7456625246023034e-05, "loss": 3.0225, "step": 24000 }, { "epoch": 0.05407194163690907, "grad_norm": 2.4655566215515137, "learning_rate": 4.7348472546102937e-05, "loss": 3.0053, "step": 25000 }, { "epoch": 0.056234819302385435, "grad_norm": 2.432565212249756, "learning_rate": 4.724021158522195e-05, "loss": 2.9726, "step": 26000 }, { "epoch": 0.058397696967861804, "grad_norm": 2.437964677810669, "learning_rate": 4.7131950624340964e-05, "loss": 2.9542, "step": 27000 }, { "epoch": 0.060560574633338166, "grad_norm": 2.498307466506958, "learning_rate": 4.702379792442086e-05, "loss": 2.9373, "step": 28000 }, { "epoch": 0.06272345229881453, "grad_norm": 2.3338165283203125, "learning_rate": 4.6915536963539876e-05, "loss": 2.9143, "step": 29000 }, { "epoch": 0.06488632996429089, "grad_norm": 2.2597150802612305, "learning_rate": 4.680738426361977e-05, "loss": 2.8947, "step": 30000 }, { "epoch": 0.06704920762976725, "grad_norm": 2.3500025272369385, "learning_rate": 4.669912330273879e-05, "loss": 2.8729, "step": 31000 }, { "epoch": 0.06921208529524361, "grad_norm": 2.592963457107544, "learning_rate": 4.6590970602818684e-05, "loss": 2.8638, "step": 32000 }, { "epoch": 0.07137496296071998, "grad_norm": 2.2916111946105957, "learning_rate": 4.64827096419377e-05, "loss": 2.8423, "step": 33000 }, { "epoch": 0.07353784062619634, "grad_norm": 2.298884153366089, "learning_rate": 4.63745569420176e-05, "loss": 2.8258, "step": 34000 }, { "epoch": 0.0757007182916727, "grad_norm": 2.2187180519104004, "learning_rate": 4.6266295981136614e-05, "loss": 2.8085, "step": 35000 }, { "epoch": 0.07786359595714906, "grad_norm": 2.3649799823760986, "learning_rate": 4.615814328121651e-05, "loss": 2.7946, "step": 36000 }, { "epoch": 0.08002647362262542, "grad_norm": 2.189671754837036, "learning_rate": 4.6049882320335526e-05, "loss": 2.7828, "step": 37000 }, { "epoch": 0.0821893512881018, "grad_norm": 2.341826915740967, "learning_rate": 4.594172962041542e-05, "loss": 2.7706, "step": 38000 }, { "epoch": 0.08435222895357816, "grad_norm": 2.3907382488250732, "learning_rate": 4.583346865953443e-05, "loss": 2.7548, "step": 39000 }, { "epoch": 0.08651510661905452, "grad_norm": 2.2284059524536133, "learning_rate": 4.5725315959614334e-05, "loss": 2.7453, "step": 40000 }, { "epoch": 0.08867798428453089, "grad_norm": 2.254753589630127, "learning_rate": 4.561705499873335e-05, "loss": 2.7335, "step": 41000 }, { "epoch": 0.09084086195000725, "grad_norm": 2.2830395698547363, "learning_rate": 4.550890229881325e-05, "loss": 2.7188, "step": 42000 }, { "epoch": 0.09300373961548361, "grad_norm": 2.2295284271240234, "learning_rate": 4.540064133793226e-05, "loss": 2.7066, "step": 43000 }, { "epoch": 0.09516661728095997, "grad_norm": 2.177199602127075, "learning_rate": 4.529248863801216e-05, "loss": 2.6916, "step": 44000 }, { "epoch": 0.09732949494643633, "grad_norm": 2.0382652282714844, "learning_rate": 4.5184227677131176e-05, "loss": 2.6852, "step": 45000 }, { "epoch": 0.0994923726119127, "grad_norm": 2.240755319595337, "learning_rate": 4.507607497721107e-05, "loss": 2.6793, "step": 46000 }, { "epoch": 0.10165525027738906, "grad_norm": 2.1940410137176514, "learning_rate": 4.496781401633008e-05, "loss": 2.6659, "step": 47000 }, { "epoch": 0.10381812794286542, "grad_norm": 2.203796148300171, "learning_rate": 4.4859661316409984e-05, "loss": 2.6526, "step": 48000 }, { "epoch": 0.10598100560834178, "grad_norm": 2.1186819076538086, "learning_rate": 4.4751400355529e-05, "loss": 2.6463, "step": 49000 }, { "epoch": 0.10814388327381814, "grad_norm": 2.13010573387146, "learning_rate": 4.46432476556089e-05, "loss": 2.6403, "step": 50000 }, { "epoch": 0.11030676093929451, "grad_norm": 2.1970157623291016, "learning_rate": 4.4535094955688785e-05, "loss": 2.6276, "step": 51000 }, { "epoch": 0.11246963860477087, "grad_norm": 2.1742944717407227, "learning_rate": 4.44268339948078e-05, "loss": 2.6178, "step": 52000 }, { "epoch": 0.11463251627024724, "grad_norm": 2.195935010910034, "learning_rate": 4.4318681294887705e-05, "loss": 2.6095, "step": 53000 }, { "epoch": 0.11679539393572361, "grad_norm": 2.0977721214294434, "learning_rate": 4.421042033400672e-05, "loss": 2.6037, "step": 54000 }, { "epoch": 0.11895827160119997, "grad_norm": 2.1214547157287598, "learning_rate": 4.410215937312573e-05, "loss": 2.5962, "step": 55000 }, { "epoch": 0.12112114926667633, "grad_norm": 2.216897964477539, "learning_rate": 4.399400667320563e-05, "loss": 2.5881, "step": 56000 }, { "epoch": 0.1232840269321527, "grad_norm": 2.1751599311828613, "learning_rate": 4.3885745712324644e-05, "loss": 2.5817, "step": 57000 }, { "epoch": 0.12544690459762906, "grad_norm": 2.1866142749786377, "learning_rate": 4.377748475144366e-05, "loss": 2.5691, "step": 58000 }, { "epoch": 0.1276097822631054, "grad_norm": 2.1192381381988525, "learning_rate": 4.366933205152356e-05, "loss": 2.5649, "step": 59000 }, { "epoch": 0.12977265992858178, "grad_norm": 1.9748364686965942, "learning_rate": 4.3561071090642574e-05, "loss": 2.5535, "step": 60000 } ], "logging_steps": 1000, "max_steps": 462347, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 10000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.30908615081984e+18, "train_batch_size": 64, "trial_name": null, "trial_params": null }