| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 10.0, | |
| "global_step": 21790, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.885268471776045e-06, | |
| "loss": 1.4166, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.770536943552089e-06, | |
| "loss": 0.6545, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.655805415328132e-06, | |
| "loss": 0.5315, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.541073887104177e-06, | |
| "loss": 0.4715, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "exact_match": 67.5, | |
| "f1": 75.87103174603179, | |
| "step": 2179 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 4.4263423588802205e-06, | |
| "loss": 0.3414, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 4.311610830656265e-06, | |
| "loss": 0.2945, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 4.196879302432309e-06, | |
| "loss": 0.3019, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 4.0821477742083525e-06, | |
| "loss": 0.3033, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "exact_match": 68.33333333333333, | |
| "f1": 76.34986772486776, | |
| "step": 4358 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 3.967416245984397e-06, | |
| "loss": 0.2622, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 3.852684717760441e-06, | |
| "loss": 0.1801, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 3.7379531895364852e-06, | |
| "loss": 0.1697, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 3.623221661312529e-06, | |
| "loss": 0.2118, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 3.508490133088573e-06, | |
| "loss": 0.1559, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "exact_match": 67.91666666666667, | |
| "f1": 76.26147001147002, | |
| "step": 6537 | |
| }, | |
| { | |
| "epoch": 3.21, | |
| "learning_rate": 3.393758604864617e-06, | |
| "loss": 0.1068, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 3.44, | |
| "learning_rate": 3.2790270766406612e-06, | |
| "loss": 0.0951, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 3.67, | |
| "learning_rate": 3.1642955484167054e-06, | |
| "loss": 0.0986, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 3.9, | |
| "learning_rate": 3.049564020192749e-06, | |
| "loss": 0.1122, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "exact_match": 67.5, | |
| "f1": 74.74419793169795, | |
| "step": 8716 | |
| }, | |
| { | |
| "epoch": 4.13, | |
| "learning_rate": 2.934832491968793e-06, | |
| "loss": 0.0811, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 4.36, | |
| "learning_rate": 2.8201009637448373e-06, | |
| "loss": 0.0546, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 4.59, | |
| "learning_rate": 2.7053694355208814e-06, | |
| "loss": 0.056, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 4.82, | |
| "learning_rate": 2.5906379072969255e-06, | |
| "loss": 0.0731, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "exact_match": 70.41666666666667, | |
| "f1": 77.59175084175087, | |
| "step": 10895 | |
| }, | |
| { | |
| "epoch": 5.05, | |
| "learning_rate": 2.4759063790729696e-06, | |
| "loss": 0.0687, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 5.28, | |
| "learning_rate": 2.3611748508490133e-06, | |
| "loss": 0.0471, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 5.51, | |
| "learning_rate": 2.2464433226250574e-06, | |
| "loss": 0.0426, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 5.74, | |
| "learning_rate": 2.1317117944011015e-06, | |
| "loss": 0.0437, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 5.97, | |
| "learning_rate": 2.0169802661771456e-06, | |
| "loss": 0.0299, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "exact_match": 66.66666666666667, | |
| "f1": 74.826330804272, | |
| "step": 13074 | |
| }, | |
| { | |
| "epoch": 6.2, | |
| "learning_rate": 1.9022487379531897e-06, | |
| "loss": 0.0232, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 6.42, | |
| "learning_rate": 1.7875172097292336e-06, | |
| "loss": 0.0221, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 6.65, | |
| "learning_rate": 1.672785681505278e-06, | |
| "loss": 0.0242, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 6.88, | |
| "learning_rate": 1.5580541532813219e-06, | |
| "loss": 0.0327, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "exact_match": 67.08333333333333, | |
| "f1": 75.72919876963996, | |
| "step": 15253 | |
| }, | |
| { | |
| "epoch": 7.11, | |
| "learning_rate": 1.443322625057366e-06, | |
| "loss": 0.0207, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 7.34, | |
| "learning_rate": 1.3285910968334099e-06, | |
| "loss": 0.0243, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 7.57, | |
| "learning_rate": 1.213859568609454e-06, | |
| "loss": 0.0127, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 7.8, | |
| "learning_rate": 1.099128040385498e-06, | |
| "loss": 0.0189, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "exact_match": 66.25, | |
| "f1": 73.67448496492618, | |
| "step": 17432 | |
| }, | |
| { | |
| "epoch": 8.03, | |
| "learning_rate": 9.84396512161542e-07, | |
| "loss": 0.0234, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 8.26, | |
| "learning_rate": 8.696649839375861e-07, | |
| "loss": 0.0041, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 8.49, | |
| "learning_rate": 7.549334557136302e-07, | |
| "loss": 0.0115, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 8.72, | |
| "learning_rate": 6.402019274896742e-07, | |
| "loss": 0.0159, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 8.95, | |
| "learning_rate": 5.254703992657182e-07, | |
| "loss": 0.0116, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 9.0, | |
| "exact_match": 69.16666666666667, | |
| "f1": 77.33799506593625, | |
| "step": 19611 | |
| }, | |
| { | |
| "epoch": 9.18, | |
| "learning_rate": 4.107388710417623e-07, | |
| "loss": 0.0096, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 9.41, | |
| "learning_rate": 2.9600734281780635e-07, | |
| "loss": 0.0057, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 9.64, | |
| "learning_rate": 1.8127581459385043e-07, | |
| "loss": 0.005, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 9.87, | |
| "learning_rate": 6.654428636989445e-08, | |
| "loss": 0.0042, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "exact_match": 66.66666666666667, | |
| "f1": 73.91331557875677, | |
| "step": 21790 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "step": 21790, | |
| "total_flos": 1.3475301465871872e+17, | |
| "train_runtime": 17438.5747, | |
| "train_samples_per_second": 1.25 | |
| } | |
| ], | |
| "max_steps": 21790, | |
| "num_train_epochs": 10, | |
| "total_flos": 1.3475301465871872e+17, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |