{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 0.7111111111111111, "eval_steps": 100, "global_step": 100, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0071111111111111115, "grad_norm": 147.2163848876953, "learning_rate": 0.0, "loss": 15.3229, "step": 1 }, { "epoch": 0.014222222222222223, "grad_norm": 156.96929931640625, "learning_rate": 4.000000000000001e-06, "loss": 15.8117, "step": 2 }, { "epoch": 0.021333333333333333, "grad_norm": 157.3893280029297, "learning_rate": 8.000000000000001e-06, "loss": 15.283, "step": 3 }, { "epoch": 0.028444444444444446, "grad_norm": 158.99237060546875, "learning_rate": 1.2e-05, "loss": 15.268, "step": 4 }, { "epoch": 0.035555555555555556, "grad_norm": 160.12179565429688, "learning_rate": 1.6000000000000003e-05, "loss": 15.4867, "step": 5 }, { "epoch": 0.042666666666666665, "grad_norm": 156.13417053222656, "learning_rate": 2e-05, "loss": 15.1959, "step": 6 }, { "epoch": 0.049777777777777775, "grad_norm": 169.3893585205078, "learning_rate": 1.995215311004785e-05, "loss": 15.1078, "step": 7 }, { "epoch": 0.05688888888888889, "grad_norm": 180.29098510742188, "learning_rate": 1.9904306220095696e-05, "loss": 14.6331, "step": 8 }, { "epoch": 0.064, "grad_norm": 184.67332458496094, "learning_rate": 1.9856459330143542e-05, "loss": 14.1344, "step": 9 }, { "epoch": 0.07111111111111111, "grad_norm": 184.16485595703125, "learning_rate": 1.980861244019139e-05, "loss": 13.8721, "step": 10 }, { "epoch": 0.07822222222222222, "grad_norm": 206.7319793701172, "learning_rate": 1.9760765550239236e-05, "loss": 13.7052, "step": 11 }, { "epoch": 0.08533333333333333, "grad_norm": 205.46336364746094, "learning_rate": 1.9712918660287083e-05, "loss": 13.162, "step": 12 }, { "epoch": 0.09244444444444444, "grad_norm": 213.31185913085938, "learning_rate": 1.966507177033493e-05, "loss": 12.6663, "step": 13 }, { "epoch": 0.09955555555555555, "grad_norm": 211.15602111816406, "learning_rate": 1.9617224880382777e-05, "loss": 12.4509, "step": 14 }, { "epoch": 0.10666666666666667, "grad_norm": 213.89552307128906, "learning_rate": 1.9569377990430624e-05, "loss": 12.0618, "step": 15 }, { "epoch": 0.11377777777777778, "grad_norm": 224.93186950683594, "learning_rate": 1.952153110047847e-05, "loss": 11.5434, "step": 16 }, { "epoch": 0.12088888888888889, "grad_norm": 209.9617462158203, "learning_rate": 1.9473684210526318e-05, "loss": 11.2266, "step": 17 }, { "epoch": 0.128, "grad_norm": 205.28099060058594, "learning_rate": 1.9425837320574165e-05, "loss": 11.1783, "step": 18 }, { "epoch": 0.1351111111111111, "grad_norm": 208.29965209960938, "learning_rate": 1.9377990430622012e-05, "loss": 10.1627, "step": 19 }, { "epoch": 0.14222222222222222, "grad_norm": 214.2541961669922, "learning_rate": 1.933014354066986e-05, "loss": 9.8809, "step": 20 }, { "epoch": 0.14933333333333335, "grad_norm": 204.05067443847656, "learning_rate": 1.9282296650717706e-05, "loss": 9.5277, "step": 21 }, { "epoch": 0.15644444444444444, "grad_norm": 193.07321166992188, "learning_rate": 1.9234449760765553e-05, "loss": 8.9317, "step": 22 }, { "epoch": 0.16355555555555557, "grad_norm": 186.99464416503906, "learning_rate": 1.9186602870813396e-05, "loss": 8.929, "step": 23 }, { "epoch": 0.17066666666666666, "grad_norm": 186.2172088623047, "learning_rate": 1.9138755980861247e-05, "loss": 9.0329, "step": 24 }, { "epoch": 0.17777777777777778, "grad_norm": 173.13241577148438, "learning_rate": 1.9090909090909094e-05, "loss": 8.4785, "step": 25 }, { "epoch": 0.18488888888888888, "grad_norm": 170.62806701660156, "learning_rate": 1.904306220095694e-05, "loss": 8.1484, "step": 26 }, { "epoch": 0.192, "grad_norm": 171.30633544921875, "learning_rate": 1.8995215311004787e-05, "loss": 7.3798, "step": 27 }, { "epoch": 0.1991111111111111, "grad_norm": 150.02285766601562, "learning_rate": 1.894736842105263e-05, "loss": 6.6586, "step": 28 }, { "epoch": 0.20622222222222222, "grad_norm": 143.904541015625, "learning_rate": 1.889952153110048e-05, "loss": 6.9233, "step": 29 }, { "epoch": 0.21333333333333335, "grad_norm": 148.44595336914062, "learning_rate": 1.8851674641148328e-05, "loss": 7.3481, "step": 30 }, { "epoch": 0.22044444444444444, "grad_norm": 157.87283325195312, "learning_rate": 1.8803827751196172e-05, "loss": 6.8199, "step": 31 }, { "epoch": 0.22755555555555557, "grad_norm": 136.0242462158203, "learning_rate": 1.8755980861244022e-05, "loss": 6.6569, "step": 32 }, { "epoch": 0.23466666666666666, "grad_norm": 145.17214965820312, "learning_rate": 1.870813397129187e-05, "loss": 5.6681, "step": 33 }, { "epoch": 0.24177777777777779, "grad_norm": 138.59141540527344, "learning_rate": 1.8660287081339713e-05, "loss": 6.2618, "step": 34 }, { "epoch": 0.24888888888888888, "grad_norm": 126.07086181640625, "learning_rate": 1.8612440191387563e-05, "loss": 5.3403, "step": 35 }, { "epoch": 0.256, "grad_norm": 134.73928833007812, "learning_rate": 1.8564593301435407e-05, "loss": 5.3615, "step": 36 }, { "epoch": 0.26311111111111113, "grad_norm": 131.0662384033203, "learning_rate": 1.8516746411483253e-05, "loss": 4.6661, "step": 37 }, { "epoch": 0.2702222222222222, "grad_norm": 123.21418762207031, "learning_rate": 1.8468899521531104e-05, "loss": 5.0741, "step": 38 }, { "epoch": 0.2773333333333333, "grad_norm": 121.35318756103516, "learning_rate": 1.8421052631578947e-05, "loss": 4.7446, "step": 39 }, { "epoch": 0.28444444444444444, "grad_norm": 102.97701263427734, "learning_rate": 1.8373205741626798e-05, "loss": 4.1012, "step": 40 }, { "epoch": 0.29155555555555557, "grad_norm": 100.17095947265625, "learning_rate": 1.832535885167464e-05, "loss": 3.9313, "step": 41 }, { "epoch": 0.2986666666666667, "grad_norm": 91.87388610839844, "learning_rate": 1.8277511961722488e-05, "loss": 3.8289, "step": 42 }, { "epoch": 0.30577777777777776, "grad_norm": 92.78990936279297, "learning_rate": 1.822966507177034e-05, "loss": 3.9649, "step": 43 }, { "epoch": 0.3128888888888889, "grad_norm": 93.31304168701172, "learning_rate": 1.8181818181818182e-05, "loss": 3.8026, "step": 44 }, { "epoch": 0.32, "grad_norm": 96.7074203491211, "learning_rate": 1.813397129186603e-05, "loss": 3.8662, "step": 45 }, { "epoch": 0.32711111111111113, "grad_norm": 88.96202850341797, "learning_rate": 1.8086124401913876e-05, "loss": 3.5804, "step": 46 }, { "epoch": 0.3342222222222222, "grad_norm": 83.96373748779297, "learning_rate": 1.8038277511961723e-05, "loss": 3.4166, "step": 47 }, { "epoch": 0.3413333333333333, "grad_norm": 86.4635238647461, "learning_rate": 1.799043062200957e-05, "loss": 3.2399, "step": 48 }, { "epoch": 0.34844444444444445, "grad_norm": 85.60812377929688, "learning_rate": 1.7942583732057417e-05, "loss": 2.9237, "step": 49 }, { "epoch": 0.35555555555555557, "grad_norm": 74.39842224121094, "learning_rate": 1.7894736842105264e-05, "loss": 3.1092, "step": 50 }, { "epoch": 0.3626666666666667, "grad_norm": 78.66775512695312, "learning_rate": 1.7846889952153114e-05, "loss": 2.9677, "step": 51 }, { "epoch": 0.36977777777777776, "grad_norm": 73.33041381835938, "learning_rate": 1.7799043062200958e-05, "loss": 2.9961, "step": 52 }, { "epoch": 0.3768888888888889, "grad_norm": 69.15443420410156, "learning_rate": 1.7751196172248805e-05, "loss": 2.5786, "step": 53 }, { "epoch": 0.384, "grad_norm": 69.43597412109375, "learning_rate": 1.770334928229665e-05, "loss": 2.6235, "step": 54 }, { "epoch": 0.39111111111111113, "grad_norm": 67.97980499267578, "learning_rate": 1.76555023923445e-05, "loss": 2.9034, "step": 55 }, { "epoch": 0.3982222222222222, "grad_norm": 65.86293029785156, "learning_rate": 1.7607655502392345e-05, "loss": 2.5154, "step": 56 }, { "epoch": 0.4053333333333333, "grad_norm": 70.66265106201172, "learning_rate": 1.7559808612440192e-05, "loss": 2.4679, "step": 57 }, { "epoch": 0.41244444444444445, "grad_norm": 67.84262084960938, "learning_rate": 1.751196172248804e-05, "loss": 2.2115, "step": 58 }, { "epoch": 0.41955555555555557, "grad_norm": 71.07489013671875, "learning_rate": 1.7464114832535886e-05, "loss": 2.3696, "step": 59 }, { "epoch": 0.4266666666666667, "grad_norm": 66.5838394165039, "learning_rate": 1.7416267942583733e-05, "loss": 2.2369, "step": 60 }, { "epoch": 0.43377777777777776, "grad_norm": 59.810028076171875, "learning_rate": 1.736842105263158e-05, "loss": 1.9434, "step": 61 }, { "epoch": 0.4408888888888889, "grad_norm": 60.31599426269531, "learning_rate": 1.7320574162679427e-05, "loss": 2.1169, "step": 62 }, { "epoch": 0.448, "grad_norm": 63.39937210083008, "learning_rate": 1.7272727272727274e-05, "loss": 2.1662, "step": 63 }, { "epoch": 0.45511111111111113, "grad_norm": 54.780609130859375, "learning_rate": 1.722488038277512e-05, "loss": 1.8301, "step": 64 }, { "epoch": 0.4622222222222222, "grad_norm": 51.63141632080078, "learning_rate": 1.7177033492822968e-05, "loss": 1.8786, "step": 65 }, { "epoch": 0.4693333333333333, "grad_norm": 61.2618293762207, "learning_rate": 1.7129186602870815e-05, "loss": 1.6842, "step": 66 }, { "epoch": 0.47644444444444445, "grad_norm": 54.38351058959961, "learning_rate": 1.7081339712918662e-05, "loss": 1.665, "step": 67 }, { "epoch": 0.48355555555555557, "grad_norm": 55.260860443115234, "learning_rate": 1.703349282296651e-05, "loss": 1.6765, "step": 68 }, { "epoch": 0.49066666666666664, "grad_norm": 62.19572067260742, "learning_rate": 1.6985645933014356e-05, "loss": 1.6315, "step": 69 }, { "epoch": 0.49777777777777776, "grad_norm": 55.4731559753418, "learning_rate": 1.6937799043062203e-05, "loss": 1.6022, "step": 70 }, { "epoch": 0.5048888888888889, "grad_norm": 53.94541931152344, "learning_rate": 1.688995215311005e-05, "loss": 1.6217, "step": 71 }, { "epoch": 0.512, "grad_norm": 48.46171188354492, "learning_rate": 1.6842105263157896e-05, "loss": 1.457, "step": 72 }, { "epoch": 0.5191111111111111, "grad_norm": 55.64387512207031, "learning_rate": 1.6794258373205743e-05, "loss": 1.3321, "step": 73 }, { "epoch": 0.5262222222222223, "grad_norm": 53.93634033203125, "learning_rate": 1.674641148325359e-05, "loss": 1.2584, "step": 74 }, { "epoch": 0.5333333333333333, "grad_norm": 52.08723449707031, "learning_rate": 1.6698564593301437e-05, "loss": 1.3642, "step": 75 }, { "epoch": 0.5404444444444444, "grad_norm": 62.591835021972656, "learning_rate": 1.6650717703349284e-05, "loss": 1.2531, "step": 76 }, { "epoch": 0.5475555555555556, "grad_norm": 53.1253547668457, "learning_rate": 1.6602870813397128e-05, "loss": 1.1524, "step": 77 }, { "epoch": 0.5546666666666666, "grad_norm": 44.78287124633789, "learning_rate": 1.6555023923444978e-05, "loss": 1.0058, "step": 78 }, { "epoch": 0.5617777777777778, "grad_norm": 48.20844268798828, "learning_rate": 1.6507177033492825e-05, "loss": 0.9914, "step": 79 }, { "epoch": 0.5688888888888889, "grad_norm": 47.6641731262207, "learning_rate": 1.6459330143540672e-05, "loss": 1.0974, "step": 80 }, { "epoch": 0.576, "grad_norm": 52.28041458129883, "learning_rate": 1.641148325358852e-05, "loss": 1.1361, "step": 81 }, { "epoch": 0.5831111111111111, "grad_norm": 51.71315002441406, "learning_rate": 1.6363636363636366e-05, "loss": 1.1431, "step": 82 }, { "epoch": 0.5902222222222222, "grad_norm": 44.117557525634766, "learning_rate": 1.6315789473684213e-05, "loss": 0.9574, "step": 83 }, { "epoch": 0.5973333333333334, "grad_norm": 46.53805160522461, "learning_rate": 1.626794258373206e-05, "loss": 0.9035, "step": 84 }, { "epoch": 0.6044444444444445, "grad_norm": 42.801536560058594, "learning_rate": 1.6220095693779903e-05, "loss": 0.8866, "step": 85 }, { "epoch": 0.6115555555555555, "grad_norm": 42.011192321777344, "learning_rate": 1.6172248803827754e-05, "loss": 0.6358, "step": 86 }, { "epoch": 0.6186666666666667, "grad_norm": 39.56203842163086, "learning_rate": 1.61244019138756e-05, "loss": 0.6602, "step": 87 }, { "epoch": 0.6257777777777778, "grad_norm": 45.57675552368164, "learning_rate": 1.6076555023923444e-05, "loss": 0.6939, "step": 88 }, { "epoch": 0.6328888888888888, "grad_norm": 44.18901443481445, "learning_rate": 1.6028708133971294e-05, "loss": 0.6864, "step": 89 }, { "epoch": 0.64, "grad_norm": 40.97324752807617, "learning_rate": 1.5980861244019138e-05, "loss": 0.6563, "step": 90 }, { "epoch": 0.6471111111111111, "grad_norm": 39.4112663269043, "learning_rate": 1.593301435406699e-05, "loss": 0.5773, "step": 91 }, { "epoch": 0.6542222222222223, "grad_norm": 31.436506271362305, "learning_rate": 1.5885167464114835e-05, "loss": 0.5103, "step": 92 }, { "epoch": 0.6613333333333333, "grad_norm": 44.75907897949219, "learning_rate": 1.583732057416268e-05, "loss": 0.8296, "step": 93 }, { "epoch": 0.6684444444444444, "grad_norm": 41.57793426513672, "learning_rate": 1.578947368421053e-05, "loss": 0.5342, "step": 94 }, { "epoch": 0.6755555555555556, "grad_norm": 50.84144592285156, "learning_rate": 1.5741626794258376e-05, "loss": 0.5866, "step": 95 }, { "epoch": 0.6826666666666666, "grad_norm": 35.748844146728516, "learning_rate": 1.569377990430622e-05, "loss": 0.43, "step": 96 }, { "epoch": 0.6897777777777778, "grad_norm": 34.87845993041992, "learning_rate": 1.564593301435407e-05, "loss": 0.5604, "step": 97 }, { "epoch": 0.6968888888888889, "grad_norm": 30.870040893554688, "learning_rate": 1.5598086124401914e-05, "loss": 0.4239, "step": 98 }, { "epoch": 0.704, "grad_norm": 32.96578598022461, "learning_rate": 1.555023923444976e-05, "loss": 0.4253, "step": 99 }, { "epoch": 0.7111111111111111, "grad_norm": 29.156658172607422, "learning_rate": 1.550239234449761e-05, "loss": 0.3539, "step": 100 }, { "epoch": 0.7111111111111111, "eval_loss": 0.3586403727531433, "eval_runtime": 10.0992, "eval_samples_per_second": 24.755, "eval_steps_per_second": 12.377, "step": 100 } ], "logging_steps": 1, "max_steps": 423, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 4160649017613312.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }