| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 0.3429796355841372, |
| "eval_steps": 500, |
| "global_step": 100, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0, |
| "learning_rate": 0.00019931271477663232, |
| "loss": 2.5587, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.01, |
| "learning_rate": 0.0001986254295532646, |
| "loss": 2.3914, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.01, |
| "learning_rate": 0.00019793814432989693, |
| "loss": 2.4218, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.01, |
| "learning_rate": 0.00019725085910652924, |
| "loss": 2.3414, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 0.0001965635738831615, |
| "loss": 2.2469, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 0.00019587628865979381, |
| "loss": 2.3241, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 0.00019518900343642613, |
| "loss": 2.3266, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.03, |
| "learning_rate": 0.00019450171821305842, |
| "loss": 2.1856, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.03, |
| "learning_rate": 0.00019381443298969073, |
| "loss": 2.3247, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.03, |
| "learning_rate": 0.00019312714776632305, |
| "loss": 2.3245, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 0.00019243986254295533, |
| "loss": 2.2591, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 0.00019175257731958765, |
| "loss": 2.1767, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 0.00019106529209621996, |
| "loss": 2.3478, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.05, |
| "learning_rate": 0.00019037800687285222, |
| "loss": 2.3339, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.05, |
| "learning_rate": 0.00018969072164948454, |
| "loss": 2.234, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.05, |
| "learning_rate": 0.00018900343642611685, |
| "loss": 2.2651, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 0.00018831615120274914, |
| "loss": 2.1831, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 0.00018762886597938145, |
| "loss": 2.216, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 0.00018694158075601377, |
| "loss": 2.1359, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 0.00018625429553264605, |
| "loss": 2.1215, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 0.00018556701030927837, |
| "loss": 2.2179, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 0.00018487972508591068, |
| "loss": 2.2598, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 0.00018419243986254294, |
| "loss": 2.1813, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 0.00018350515463917526, |
| "loss": 2.2006, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 0.00018281786941580757, |
| "loss": 2.1564, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 0.00018213058419243986, |
| "loss": 2.2537, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 0.00018144329896907217, |
| "loss": 2.1975, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 0.0001807560137457045, |
| "loss": 2.2566, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 0.00018006872852233677, |
| "loss": 2.1464, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 0.0001793814432989691, |
| "loss": 2.1421, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 0.0001786941580756014, |
| "loss": 2.1276, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 0.00017800687285223366, |
| "loss": 2.0649, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 0.00017731958762886598, |
| "loss": 2.1835, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 0.0001766323024054983, |
| "loss": 2.1711, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 0.00017594501718213058, |
| "loss": 2.2591, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 0.0001752577319587629, |
| "loss": 2.1471, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 0.0001745704467353952, |
| "loss": 2.0861, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 0.0001738831615120275, |
| "loss": 2.0702, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 0.0001731958762886598, |
| "loss": 2.1096, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 0.00017250859106529212, |
| "loss": 2.1062, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 0.00017182130584192438, |
| "loss": 2.2545, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 0.0001711340206185567, |
| "loss": 2.1572, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 0.000170446735395189, |
| "loss": 2.0749, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 0.0001697594501718213, |
| "loss": 2.1922, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 0.00016907216494845361, |
| "loss": 2.1915, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 0.00016838487972508593, |
| "loss": 2.1594, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 0.00016769759450171822, |
| "loss": 2.176, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 0.00016701030927835053, |
| "loss": 2.1223, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 0.00016632302405498285, |
| "loss": 2.1263, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 0.00016563573883161513, |
| "loss": 2.0481, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 0.00016494845360824742, |
| "loss": 2.1043, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 0.00016426116838487973, |
| "loss": 2.1678, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 0.00016357388316151202, |
| "loss": 2.1602, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 0.00016288659793814434, |
| "loss": 2.1448, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 0.00016219931271477665, |
| "loss": 2.1536, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 0.00016151202749140894, |
| "loss": 2.0339, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 0.00016082474226804125, |
| "loss": 2.023, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 0.00016013745704467357, |
| "loss": 2.1407, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 0.00015945017182130585, |
| "loss": 2.1134, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.21, |
| "learning_rate": 0.00015876288659793814, |
| "loss": 2.1652, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.21, |
| "learning_rate": 0.00015807560137457046, |
| "loss": 2.0051, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.21, |
| "learning_rate": 0.00015738831615120274, |
| "loss": 2.0604, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 0.00015670103092783506, |
| "loss": 2.1708, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 0.00015601374570446737, |
| "loss": 2.1106, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 0.00015532646048109966, |
| "loss": 2.1445, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 0.00015463917525773197, |
| "loss": 2.0879, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 0.0001539518900343643, |
| "loss": 2.1498, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 0.00015326460481099657, |
| "loss": 2.0719, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 0.00015257731958762886, |
| "loss": 2.2167, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 0.00015189003436426118, |
| "loss": 2.0811, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 0.00015120274914089346, |
| "loss": 2.1058, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 0.00015051546391752578, |
| "loss": 2.0392, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 0.0001498281786941581, |
| "loss": 2.0957, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 0.00014914089347079038, |
| "loss": 1.9994, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 0.0001484536082474227, |
| "loss": 2.0464, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 0.000147766323024055, |
| "loss": 2.0417, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 0.0001470790378006873, |
| "loss": 2.105, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.27, |
| "learning_rate": 0.00014639175257731958, |
| "loss": 2.1147, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.27, |
| "learning_rate": 0.0001457044673539519, |
| "loss": 1.9964, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.27, |
| "learning_rate": 0.00014501718213058418, |
| "loss": 1.9723, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 0.0001443298969072165, |
| "loss": 2.0621, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 0.00014364261168384881, |
| "loss": 2.2703, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 0.0001429553264604811, |
| "loss": 2.0815, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 0.00014226804123711342, |
| "loss": 2.0774, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 0.00014158075601374573, |
| "loss": 2.066, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 0.00014089347079037802, |
| "loss": 2.03, |
| "step": 86 |
| }, |
| { |
| "epoch": 0.3, |
| "learning_rate": 0.0001402061855670103, |
| "loss": 2.1433, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.3, |
| "learning_rate": 0.00013951890034364262, |
| "loss": 2.0811, |
| "step": 88 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 0.0001388316151202749, |
| "loss": 1.9791, |
| "step": 89 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 0.00013814432989690722, |
| "loss": 2.0876, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 0.00013745704467353953, |
| "loss": 2.0314, |
| "step": 91 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 0.00013676975945017182, |
| "loss": 1.9485, |
| "step": 92 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 0.00013608247422680414, |
| "loss": 2.078, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 0.00013539518900343645, |
| "loss": 2.1251, |
| "step": 94 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 0.00013470790378006874, |
| "loss": 1.9736, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 0.00013402061855670103, |
| "loss": 2.0189, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 0.00013333333333333334, |
| "loss": 2.0061, |
| "step": 97 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 0.00013264604810996563, |
| "loss": 1.9595, |
| "step": 98 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 0.00013195876288659794, |
| "loss": 1.9702, |
| "step": 99 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 0.00013127147766323026, |
| "loss": 2.0322, |
| "step": 100 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 291, |
| "num_train_epochs": 1, |
| "save_steps": 100, |
| "total_flos": 6.083398113601536e+16, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|