| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.0, | |
| "eval_steps": 500, | |
| "global_step": 434, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.002304147465437788, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.998846730480915e-06, | |
| "loss": 3.6297, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.004608294930875576, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.997691064419303e-06, | |
| "loss": 3.7821, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.0069124423963133645, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.996532994337224e-06, | |
| "loss": 3.0545, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.009216589861751152, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.99537251272559e-06, | |
| "loss": 3.3776, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.01152073732718894, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.994209612044008e-06, | |
| "loss": 2.6629, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.013824884792626729, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.993044284720612e-06, | |
| "loss": 2.9764, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.016129032258064516, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.99187652315191e-06, | |
| "loss": 3.0443, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.018433179723502304, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.990706319702603e-06, | |
| "loss": 2.9777, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.020737327188940093, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.989533666705432e-06, | |
| "loss": 2.6634, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.02304147465437788, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.988358556461003e-06, | |
| "loss": 2.5688, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.02534562211981567, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.98718098123762e-06, | |
| "loss": 2.7999, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.027649769585253458, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.986000933271115e-06, | |
| "loss": 2.8395, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.029953917050691243, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.984818404764686e-06, | |
| "loss": 2.9181, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.03225806451612903, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.983633387888708e-06, | |
| "loss": 2.3633, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.03456221198156682, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.982445874780575e-06, | |
| "loss": 2.6805, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.03686635944700461, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.981255857544517e-06, | |
| "loss": 3.2891, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.03917050691244239, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.980063328251437e-06, | |
| "loss": 2.3438, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.041474654377880185, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.97886827893872e-06, | |
| "loss": 3.0867, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.04377880184331797, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.97767070161006e-06, | |
| "loss": 3.0397, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.04608294930875576, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.976470588235294e-06, | |
| "loss": 2.675, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.04838709677419355, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.975267930750206e-06, | |
| "loss": 2.588, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.05069124423963134, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.974062721056355e-06, | |
| "loss": 2.5917, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.052995391705069124, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.97285495102089e-06, | |
| "loss": 2.6421, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.055299539170506916, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.971644612476371e-06, | |
| "loss": 2.8325, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.0576036866359447, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.970431697220581e-06, | |
| "loss": 2.7928, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.059907834101382486, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.96921619701634e-06, | |
| "loss": 2.7043, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.06221198156682028, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.967998103591326e-06, | |
| "loss": 2.5606, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.06451612903225806, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.966777408637875e-06, | |
| "loss": 2.5944, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.06682027649769585, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.965554103812804e-06, | |
| "loss": 2.0101, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.06912442396313365, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.964328180737219e-06, | |
| "loss": 2.669, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.07142857142857142, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.963099630996312e-06, | |
| "loss": 2.5637, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.07373271889400922, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.961868446139181e-06, | |
| "loss": 2.8749, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.07603686635944701, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.960634617678637e-06, | |
| "loss": 2.8018, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.07834101382488479, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.959398137090996e-06, | |
| "loss": 2.5725, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.08064516129032258, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.958158995815901e-06, | |
| "loss": 2.8989, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.08294930875576037, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.956917185256105e-06, | |
| "loss": 2.4909, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.08525345622119816, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.955672696777287e-06, | |
| "loss": 2.5061, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.08755760368663594, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.954425521707844e-06, | |
| "loss": 2.3851, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.08986175115207373, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.953175651338697e-06, | |
| "loss": 3.0129, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.09216589861751152, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.951923076923079e-06, | |
| "loss": 2.8212, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.0944700460829493, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.950667789676334e-06, | |
| "loss": 2.952, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.0967741935483871, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.949409780775717e-06, | |
| "loss": 2.2857, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.09907834101382489, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.948149041360185e-06, | |
| "loss": 2.2268, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.10138248847926268, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.94688556253018e-06, | |
| "loss": 2.7615, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.10368663594470046, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.945619335347432e-06, | |
| "loss": 3.1013, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.10599078341013825, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.944350350834746e-06, | |
| "loss": 3.0616, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.10829493087557604, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.943078599975779e-06, | |
| "loss": 2.2084, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.11059907834101383, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.941804073714841e-06, | |
| "loss": 2.1988, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.11290322580645161, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.94052676295667e-06, | |
| "loss": 2.5741, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.1152073732718894, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.939246658566223e-06, | |
| "loss": 2.4302, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.1175115207373272, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.937963751368447e-06, | |
| "loss": 2.5109, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.11981566820276497, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.936678032148077e-06, | |
| "loss": 2.2233, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.12211981566820276, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.935389491649398e-06, | |
| "loss": 3.1745, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.12442396313364056, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.934098120576033e-06, | |
| "loss": 2.6837, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.12672811059907835, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.932803909590715e-06, | |
| "loss": 2.5496, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.12903225806451613, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.931506849315069e-06, | |
| "loss": 2.6769, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.1313364055299539, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.930206930329376e-06, | |
| "loss": 2.4358, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.1336405529953917, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.928904143172348e-06, | |
| "loss": 2.6133, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.1359447004608295, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.927598478340902e-06, | |
| "loss": 2.6126, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.1382488479262673, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.926289926289927e-06, | |
| "loss": 2.3671, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.14055299539170507, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.92497847743205e-06, | |
| "loss": 2.3528, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.14285714285714285, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.923664122137405e-06, | |
| "loss": 2.894, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.14516129032258066, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.922346850733393e-06, | |
| "loss": 2.944, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.14746543778801843, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.921026653504443e-06, | |
| "loss": 2.495, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.1497695852534562, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.919703520691786e-06, | |
| "loss": 2.2919, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.15207373271889402, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.918377442493198e-06, | |
| "loss": 2.6164, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.1543778801843318, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.917048409062771e-06, | |
| "loss": 2.5487, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.15668202764976957, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.91571641051066e-06, | |
| "loss": 2.6602, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.15898617511520738, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.914381436902843e-06, | |
| "loss": 2.5056, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.16129032258064516, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.913043478260871e-06, | |
| "loss": 2.8503, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.16359447004608296, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.911702524561623e-06, | |
| "loss": 2.6178, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.16589861751152074, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.910358565737052e-06, | |
| "loss": 2.7173, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.16820276497695852, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.909011591673938e-06, | |
| "loss": 2.4166, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.17050691244239632, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.907661592213627e-06, | |
| "loss": 2.5658, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.1728110599078341, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.906308557151781e-06, | |
| "loss": 2.3656, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.17511520737327188, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.90495247623812e-06, | |
| "loss": 2.4611, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.1774193548387097, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.903593339176162e-06, | |
| "loss": 2.2937, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.17972350230414746, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.902231135622965e-06, | |
| "loss": 2.172, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.18202764976958524, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.900865855188858e-06, | |
| "loss": 2.4489, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.18433179723502305, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.899497487437187e-06, | |
| "loss": 2.7153, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.18663594470046083, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.89812602188404e-06, | |
| "loss": 2.7078, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.1889400921658986, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.896751447997987e-06, | |
| "loss": 2.1448, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.1912442396313364, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.8953737551998e-06, | |
| "loss": 2.4494, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.1935483870967742, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.893992932862191e-06, | |
| "loss": 2.5461, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.195852534562212, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.89260897030954e-06, | |
| "loss": 2.219, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.19815668202764977, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.891221856817608e-06, | |
| "loss": 2.6866, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.20046082949308755, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.889831581613272e-06, | |
| "loss": 2.2086, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.20276497695852536, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.88843813387424e-06, | |
| "loss": 2.658, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.20506912442396313, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.887041502728773e-06, | |
| "loss": 2.2961, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.2073732718894009, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.885641677255401e-06, | |
| "loss": 2.5561, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.20967741935483872, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.884238646482637e-06, | |
| "loss": 2.5409, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.2119815668202765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.882832399388692e-06, | |
| "loss": 2.2584, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.21428571428571427, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.881422924901188e-06, | |
| "loss": 2.5798, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.21658986175115208, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.88001021189686e-06, | |
| "loss": 2.6547, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.21889400921658986, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.878594249201278e-06, | |
| "loss": 1.9719, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.22119815668202766, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.877175025588538e-06, | |
| "loss": 2.6163, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.22350230414746544, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.875752529780967e-06, | |
| "loss": 2.5005, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.22580645161290322, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.874326750448835e-06, | |
| "loss": 2.4969, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.22811059907834103, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.872897676210041e-06, | |
| "loss": 2.4532, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.2304147465437788, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.871465295629821e-06, | |
| "loss": 2.7851, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.23271889400921658, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.870029597220437e-06, | |
| "loss": 2.4056, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.2350230414746544, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.868590569440867e-06, | |
| "loss": 2.4184, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.23732718894009217, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.867148200696505e-06, | |
| "loss": 2.4005, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.23963133640552994, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.865702479338843e-06, | |
| "loss": 2.3063, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.24193548387096775, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.86425339366516e-06, | |
| "loss": 2.309, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.24423963133640553, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.8628009319182e-06, | |
| "loss": 2.2872, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.2465437788018433, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.861345082285862e-06, | |
| "loss": 2.5453, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.2488479262672811, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.859885832900883e-06, | |
| "loss": 2.4237, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.2511520737327189, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.8584231718405e-06, | |
| "loss": 2.7325, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.2534562211981567, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.856957087126139e-06, | |
| "loss": 2.8046, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.2557603686635945, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.855487566723084e-06, | |
| "loss": 2.3187, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.25806451612903225, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.854014598540148e-06, | |
| "loss": 2.1758, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.26036866359447003, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.852538170429336e-06, | |
| "loss": 2.3946, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.2626728110599078, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.851058270185526e-06, | |
| "loss": 2.5586, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.26497695852534564, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.849574885546111e-06, | |
| "loss": 2.4507, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.2672811059907834, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.848088004190675e-06, | |
| "loss": 2.2848, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.2695852534562212, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.846597613740658e-06, | |
| "loss": 2.7136, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.271889400921659, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.845103701758992e-06, | |
| "loss": 2.2407, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.27419354838709675, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.843606255749772e-06, | |
| "loss": 1.8917, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.2764976958525346, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.842105263157896e-06, | |
| "loss": 2.49, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.27880184331797236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.840600711368727e-06, | |
| "loss": 2.2081, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.28110599078341014, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.839092587707729e-06, | |
| "loss": 2.5349, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.2834101382488479, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.837580879440117e-06, | |
| "loss": 2.2057, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.2857142857142857, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.836065573770494e-06, | |
| "loss": 2.6525, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.2880184331797235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.834546657842488e-06, | |
| "loss": 2.466, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.2903225806451613, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.833024118738406e-06, | |
| "loss": 2.3602, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.2926267281105991, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.831497943478838e-06, | |
| "loss": 1.9779, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.29493087557603687, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.829968119022318e-06, | |
| "loss": 2.0746, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.29723502304147464, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.82843463226493e-06, | |
| "loss": 2.5047, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.2995391705069124, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.826897470039947e-06, | |
| "loss": 2.3455, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.30184331797235026, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.825356619117452e-06, | |
| "loss": 2.7969, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.30414746543778803, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.823812066203951e-06, | |
| "loss": 2.1092, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.3064516129032258, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.822263797942004e-06, | |
| "loss": 2.454, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.3087557603686636, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.82071180090982e-06, | |
| "loss": 2.2478, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.31105990783410137, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.819156061620899e-06, | |
| "loss": 2.1831, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.31336405529953915, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.817596566523606e-06, | |
| "loss": 2.199, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.315668202764977, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.816033302000808e-06, | |
| "loss": 2.3337, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.31797235023041476, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.814466254369455e-06, | |
| "loss": 2.3119, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.32027649769585254, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.8128954098802e-06, | |
| "loss": 2.319, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.3225806451612903, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.811320754716981e-06, | |
| "loss": 2.37, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.3248847926267281, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.809742274996628e-06, | |
| "loss": 2.3041, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.3271889400921659, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.808159956768442e-06, | |
| "loss": 2.7213, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.3294930875576037, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.806573786013798e-06, | |
| "loss": 2.6252, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.3317972350230415, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.80498374864572e-06, | |
| "loss": 1.8312, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.33410138248847926, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.803389830508476e-06, | |
| "loss": 2.4053, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.33640552995391704, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.80179201737714e-06, | |
| "loss": 2.3841, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.3387096774193548, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.800190294957183e-06, | |
| "loss": 2.4041, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.34101382488479265, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.79858464888405e-06, | |
| "loss": 2.0689, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.3433179723502304, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.796975064722715e-06, | |
| "loss": 2.0356, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.3456221198156682, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.795361527967258e-06, | |
| "loss": 2.0666, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.347926267281106, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.793744024040433e-06, | |
| "loss": 2.5768, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.35023041474654376, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.792122538293217e-06, | |
| "loss": 2.1563, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.35253456221198154, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.790497056004383e-06, | |
| "loss": 2.3727, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.3548387096774194, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.78886756238004e-06, | |
| "loss": 1.9836, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.35714285714285715, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.787234042553192e-06, | |
| "loss": 2.0014, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.35944700460829493, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.785596481583288e-06, | |
| "loss": 2.128, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.3617511520737327, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.78395486445576e-06, | |
| "loss": 1.9376, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.3640552995391705, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.782309176081566e-06, | |
| "loss": 2.6088, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.3663594470046083, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.780659401296732e-06, | |
| "loss": 2.476, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.3686635944700461, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.779005524861878e-06, | |
| "loss": 2.2163, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.3709677419354839, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.777347531461764e-06, | |
| "loss": 2.6048, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.37327188940092165, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.775685405704793e-06, | |
| "loss": 2.2061, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.37557603686635943, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.774019132122557e-06, | |
| "loss": 2.4101, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.3778801843317972, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.77234869516935e-06, | |
| "loss": 2.5347, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.38018433179723504, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.770674079221682e-06, | |
| "loss": 2.1867, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.3824884792626728, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.768995268577792e-06, | |
| "loss": 2.3605, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.3847926267281106, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.767312247457155e-06, | |
| "loss": 2.1918, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.3870967741935484, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.765625e-06, | |
| "loss": 2.4631, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.38940092165898615, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.763933510266798e-06, | |
| "loss": 2.077, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.391705069124424, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.762237762237764e-06, | |
| "loss": 2.274, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.39400921658986177, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.760537739812353e-06, | |
| "loss": 2.5203, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.39631336405529954, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.75883342680875e-06, | |
| "loss": 2.2468, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.3986175115207373, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.757124806963358e-06, | |
| "loss": 2.2579, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.4009216589861751, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.755411863930279e-06, | |
| "loss": 2.3655, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.4032258064516129, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.75369458128079e-06, | |
| "loss": 2.353, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.4055299539170507, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.75197294250282e-06, | |
| "loss": 2.2856, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.4078341013824885, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.750246931000424e-06, | |
| "loss": 2.2884, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.41013824884792627, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.748516530093248e-06, | |
| "loss": 2.1927, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.41244239631336405, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.746781723015986e-06, | |
| "loss": 2.4978, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.4147465437788018, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.745042492917848e-06, | |
| "loss": 2.1364, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.41705069124423966, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.743298822862007e-06, | |
| "loss": 2.4476, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.41935483870967744, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.74155069582505e-06, | |
| "loss": 1.8542, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.4216589861751152, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.739798094696433e-06, | |
| "loss": 2.2825, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.423963133640553, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.738041002277906e-06, | |
| "loss": 2.504, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.42626728110599077, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.736279401282966e-06, | |
| "loss": 2.1668, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.42857142857142855, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.734513274336284e-06, | |
| "loss": 2.2124, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.4308755760368664, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.732742603973132e-06, | |
| "loss": 2.0533, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.43317972350230416, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.73096737263881e-06, | |
| "loss": 2.2371, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.43548387096774194, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.729187562688066e-06, | |
| "loss": 2.577, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.4377880184331797, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.727403156384506e-06, | |
| "loss": 2.0062, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.4400921658986175, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.725614135900015e-06, | |
| "loss": 2.5623, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.4423963133640553, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.723820483314155e-06, | |
| "loss": 2.3879, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.4447004608294931, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.722022180613568e-06, | |
| "loss": 2.3884, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.4470046082949309, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.720219209691377e-06, | |
| "loss": 2.1677, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.44930875576036866, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.71841155234657e-06, | |
| "loss": 2.1849, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.45161290322580644, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.7165991902834e-06, | |
| "loss": 2.4744, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.4539170506912442, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.714782105110758e-06, | |
| "loss": 2.1845, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.45622119815668205, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.71296027834155e-06, | |
| "loss": 2.1075, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.45852534562211983, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.711133691392076e-06, | |
| "loss": 2.2594, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.4608294930875576, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.709302325581395e-06, | |
| "loss": 2.0445, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.4631336405529954, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.707466162130695e-06, | |
| "loss": 1.8306, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.46543778801843316, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.705625182162635e-06, | |
| "loss": 2.0827, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.46774193548387094, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.703779366700715e-06, | |
| "loss": 2.4998, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.4700460829493088, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.701928696668615e-06, | |
| "loss": 1.9692, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.47235023041474655, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.70007315288954e-06, | |
| "loss": 2.2138, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.47465437788018433, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.698212716085556e-06, | |
| "loss": 2.3699, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.4769585253456221, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.696347366876926e-06, | |
| "loss": 2.2658, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.4792626728110599, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.694477085781434e-06, | |
| "loss": 2.3592, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.4815668202764977, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.692601853213708e-06, | |
| "loss": 2.1424, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.4838709677419355, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.690721649484536e-06, | |
| "loss": 2.1127, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.4861751152073733, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.688836454800178e-06, | |
| "loss": 2.39, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.48847926267281105, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.686946249261665e-06, | |
| "loss": 2.2204, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.49078341013824883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.685051012864114e-06, | |
| "loss": 2.2297, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.4930875576036866, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.683150725496004e-06, | |
| "loss": 2.2213, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.49539170506912444, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.681245366938474e-06, | |
| "loss": 2.3086, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.4976958525345622, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.679334916864607e-06, | |
| "loss": 2.1777, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.67741935483871e-06, | |
| "loss": 2.2379, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.5023041474654378, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.675498660315571e-06, | |
| "loss": 2.3132, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.5046082949308756, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.673572812639739e-06, | |
| "loss": 2.0896, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.5069124423963134, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.671641791044779e-06, | |
| "loss": 2.1555, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.5092165898617511, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.669705574652519e-06, | |
| "loss": 2.2675, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.511520737327189, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.667764142472315e-06, | |
| "loss": 2.0952, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.5138248847926268, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.665817473400271e-06, | |
| "loss": 2.1719, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.5161290322580645, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.663865546218488e-06, | |
| "loss": 2.269, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.5184331797235023, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.66190833959429e-06, | |
| "loss": 2.3643, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.5207373271889401, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.659945832079446e-06, | |
| "loss": 2.2851, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.5230414746543779, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.657978002109388e-06, | |
| "loss": 2.2314, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.5253456221198156, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.656004828002415e-06, | |
| "loss": 2.306, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.5276497695852534, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.654026287958907e-06, | |
| "loss": 2.72, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.5299539170506913, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.652042360060516e-06, | |
| "loss": 2.2472, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.532258064516129, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.650053022269354e-06, | |
| "loss": 2.14, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.5345622119815668, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.648058252427186e-06, | |
| "loss": 2.1352, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.5368663594470046, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.646058028254597e-06, | |
| "loss": 2.1396, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.5391705069124424, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.644052327350167e-06, | |
| "loss": 2.0091, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.5414746543778802, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.642041127189643e-06, | |
| "loss": 2.3559, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.543778801843318, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.640024405125077e-06, | |
| "loss": 2.2764, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.5460829493087558, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.638002138383995e-06, | |
| "loss": 2.4431, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.5483870967741935, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.635974304068522e-06, | |
| "loss": 2.2692, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.5506912442396313, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.633940879154541e-06, | |
| "loss": 2.2558, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.5529953917050692, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.631901840490798e-06, | |
| "loss": 1.8841, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.5552995391705069, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.629857164798035e-06, | |
| "loss": 2.243, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.5576036866359447, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.627806828668103e-06, | |
| "loss": 2.7055, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.5599078341013825, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.625750808563067e-06, | |
| "loss": 2.0709, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.5622119815668203, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.623689080814313e-06, | |
| "loss": 2.2729, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.5645161290322581, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.621621621621622e-06, | |
| "loss": 2.1815, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.5668202764976958, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.619548407052275e-06, | |
| "loss": 2.3074, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.5691244239631337, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.617469413040112e-06, | |
| "loss": 2.2204, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.5714285714285714, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.615384615384616e-06, | |
| "loss": 2.2254, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.5737327188940092, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.613293989749963e-06, | |
| "loss": 2.2028, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.576036866359447, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.611197511664076e-06, | |
| "loss": 1.9041, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.5783410138248848, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.609095156517678e-06, | |
| "loss": 2.1567, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.5806451612903226, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.60698689956332e-06, | |
| "loss": 2.2442, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.5829493087557603, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.604872715914416e-06, | |
| "loss": 2.244, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.5852534562211982, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.602752580544262e-06, | |
| "loss": 2.3902, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.5875576036866359, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.600626468285045e-06, | |
| "loss": 2.3403, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.5898617511520737, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.598494353826851e-06, | |
| "loss": 2.3575, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.5921658986175116, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.596356211716666e-06, | |
| "loss": 2.3505, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.5944700460829493, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.594212016357346e-06, | |
| "loss": 2.1773, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.5967741935483871, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.592061742006616e-06, | |
| "loss": 2.296, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.5990783410138248, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.589905362776025e-06, | |
| "loss": 1.9344, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.6013824884792627, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.587742852629917e-06, | |
| "loss": 2.5543, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.6036866359447005, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.585574185384373e-06, | |
| "loss": 2.7177, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.6059907834101382, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.583399334706163e-06, | |
| "loss": 2.3868, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.6082949308755761, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.581218274111676e-06, | |
| "loss": 2.3185, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.6105990783410138, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.579030976965846e-06, | |
| "loss": 1.9385, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.6129032258064516, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.57683741648107e-06, | |
| "loss": 2.3687, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.6152073732718893, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.574637565716107e-06, | |
| "loss": 2.2045, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.6175115207373272, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.572431397574987e-06, | |
| "loss": 2.4644, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.619815668202765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.57021888480588e-06, | |
| "loss": 2.3897, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.6221198156682027, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.568e-06, | |
| "loss": 2.4176, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.6244239631336406, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.56577471559045e-06, | |
| "loss": 2.3342, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.6267281105990783, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.563543003851093e-06, | |
| "loss": 2.3815, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.6290322580645161, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.561304836895388e-06, | |
| "loss": 2.435, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.631336405529954, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.55906018667525e-06, | |
| "loss": 2.4491, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.6336405529953917, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.556809024979856e-06, | |
| "loss": 2.3206, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.6359447004608295, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.554551323434475e-06, | |
| "loss": 2.1149, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.6382488479262672, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.552287053499273e-06, | |
| "loss": 2.2707, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.6405529953917051, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.550016186468114e-06, | |
| "loss": 2.3246, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.6428571428571429, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.547738693467338e-06, | |
| "loss": 2.2557, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.6451612903225806, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.545454545454547e-06, | |
| "loss": 2.3648, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.6474654377880185, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.543163713217365e-06, | |
| "loss": 2.3793, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.6497695852534562, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.540866167372192e-06, | |
| "loss": 2.2765, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.652073732718894, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.538561878362955e-06, | |
| "loss": 2.3297, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.6543778801843319, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.53625081645983e-06, | |
| "loss": 2.2636, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.6566820276497696, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.533932951757974e-06, | |
| "loss": 2.271, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.6589861751152074, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.53160825417622e-06, | |
| "loss": 2.497, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.6612903225806451, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.529276693455799e-06, | |
| "loss": 1.994, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.663594470046083, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.526938239159003e-06, | |
| "loss": 1.9041, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.6658986175115207, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.524592860667874e-06, | |
| "loss": 2.3384, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.6682027649769585, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.522240527182867e-06, | |
| "loss": 2.0346, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.6705069124423964, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.519881207721498e-06, | |
| "loss": 2.1251, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.6728110599078341, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.517514871116987e-06, | |
| "loss": 2.3156, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.6751152073732719, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.51514148601688e-06, | |
| "loss": 2.3411, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.6774193548387096, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.512761020881673e-06, | |
| "loss": 2.5454, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.6797235023041475, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.510373443983403e-06, | |
| "loss": 1.864, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.6820276497695853, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.507978723404256e-06, | |
| "loss": 2.2479, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.684331797235023, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.505576827035126e-06, | |
| "loss": 2.2826, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.6866359447004609, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.503167722574192e-06, | |
| "loss": 1.8908, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.6889400921658986, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.500751377525463e-06, | |
| "loss": 2.0887, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.6912442396313364, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.498327759197325e-06, | |
| "loss": 1.9868, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.6935483870967742, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.495896834701056e-06, | |
| "loss": 2.03, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.695852534562212, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.493458570949348e-06, | |
| "loss": 2.413, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.6981566820276498, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.491012934654798e-06, | |
| "loss": 2.2828, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.7004608294930875, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.488559892328399e-06, | |
| "loss": 2.4673, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.7027649769585254, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.486099410278012e-06, | |
| "loss": 2.1999, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.7050691244239631, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.483631454606818e-06, | |
| "loss": 2.1067, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.7073732718894009, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.481155991211765e-06, | |
| "loss": 2.0962, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.7096774193548387, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.47867298578199e-06, | |
| "loss": 2.4821, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.7119815668202765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.476182403797255e-06, | |
| "loss": 2.4688, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.7142857142857143, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.473684210526317e-06, | |
| "loss": 2.1085, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.716589861751152, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.471178371025337e-06, | |
| "loss": 2.435, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.7188940092165899, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.46866485013624e-06, | |
| "loss": 2.3131, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.7211981566820277, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.466143612485077e-06, | |
| "loss": 2.5537, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 0.7235023041474654, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.463614622480356e-06, | |
| "loss": 2.1758, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.7258064516129032, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.461077844311378e-06, | |
| "loss": 2.2752, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.728110599078341, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.45853324194654e-06, | |
| "loss": 2.0488, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.7304147465437788, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.45598077913163e-06, | |
| "loss": 2.3105, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 0.7327188940092166, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.453420419388106e-06, | |
| "loss": 1.936, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.7350230414746544, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.450852126011364e-06, | |
| "loss": 2.1634, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 0.7373271889400922, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.448275862068967e-06, | |
| "loss": 1.9953, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.7396313364055299, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.445691590398895e-06, | |
| "loss": 2.3994, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 0.7419354838709677, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.44309927360775e-06, | |
| "loss": 1.837, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 0.7442396313364056, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.440498874068943e-06, | |
| "loss": 2.0258, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 0.7465437788018433, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.43789035392089e-06, | |
| "loss": 2.5129, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.7488479262672811, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.435273675065161e-06, | |
| "loss": 2.0436, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.7511520737327189, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.432648799164637e-06, | |
| "loss": 2.4449, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 0.7534562211981567, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.430015687641625e-06, | |
| "loss": 2.0828, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 0.7557603686635944, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.42737430167598e-06, | |
| "loss": 2.3661, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.7580645161290323, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.424724602203184e-06, | |
| "loss": 1.9816, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 0.7603686635944701, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.422066549912434e-06, | |
| "loss": 2.3207, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.7626728110599078, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.419400105244694e-06, | |
| "loss": 2.1479, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 0.7649769585253456, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.416725228390726e-06, | |
| "loss": 2.211, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.7672811059907834, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.41404187928911e-06, | |
| "loss": 1.8804, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 0.7695852534562212, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.411350017624251e-06, | |
| "loss": 2.1775, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 0.771889400921659, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.40864960282436e-06, | |
| "loss": 2.2466, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.7741935483870968, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.405940594059407e-06, | |
| "loss": 2.0984, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.7764976958525346, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.403222950239065e-06, | |
| "loss": 2.5951, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 0.7788018433179723, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.400496630010643e-06, | |
| "loss": 2.1321, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.7811059907834101, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.397761591756973e-06, | |
| "loss": 1.8965, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 0.783410138248848, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.395017793594307e-06, | |
| "loss": 2.4059, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.7857142857142857, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.392265193370167e-06, | |
| "loss": 2.4522, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 0.7880184331797235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.389503748661194e-06, | |
| "loss": 2.4031, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 0.7903225806451613, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.386733416770964e-06, | |
| "loss": 2.3547, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 0.7926267281105991, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.383954154727794e-06, | |
| "loss": 2.0233, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 0.7949308755760369, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.381165919282512e-06, | |
| "loss": 2.3673, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.7972350230414746, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.378368666906218e-06, | |
| "loss": 2.4832, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 0.7995391705069125, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.375562353788017e-06, | |
| "loss": 2.0684, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 0.8018433179723502, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.372746935832732e-06, | |
| "loss": 2.1924, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.804147465437788, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.369922368658603e-06, | |
| "loss": 2.3609, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 0.8064516129032258, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.367088607594939e-06, | |
| "loss": 2.2614, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.8087557603686636, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.364245607679769e-06, | |
| "loss": 2.5841, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 0.8110599078341014, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.361393323657475e-06, | |
| "loss": 2.038, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.8133640552995391, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.358531709976377e-06, | |
| "loss": 2.2917, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 0.815668202764977, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.355660720786313e-06, | |
| "loss": 2.0844, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 0.8179723502304147, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.352780309936191e-06, | |
| "loss": 1.7098, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.8202764976958525, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.349890430971511e-06, | |
| "loss": 2.4264, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 0.8225806451612904, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.346991037131882e-06, | |
| "loss": 2.4563, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 0.8248847926267281, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.34408208134848e-06, | |
| "loss": 1.9052, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 0.8271889400921659, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.341163516241514e-06, | |
| "loss": 2.1467, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 0.8294930875576036, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.338235294117646e-06, | |
| "loss": 2.2446, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.8317972350230415, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.33529736696741e-06, | |
| "loss": 2.1227, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 0.8341013824884793, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.332349686462561e-06, | |
| "loss": 2.2938, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 0.836405529953917, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.329392203953447e-06, | |
| "loss": 2.2272, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 0.8387096774193549, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.326424870466322e-06, | |
| "loss": 1.9074, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.8410138248847926, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.32344763670065e-06, | |
| "loss": 1.7826, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.8433179723502304, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.320460453026365e-06, | |
| "loss": 2.4799, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 0.8456221198156681, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.317463269481124e-06, | |
| "loss": 1.8732, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 0.847926267281106, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.314456035767513e-06, | |
| "loss": 2.4681, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 0.8502304147465438, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.311438701250233e-06, | |
| "loss": 1.7413, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 0.8525345622119815, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.308411214953271e-06, | |
| "loss": 2.2242, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.8548387096774194, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.305373525557013e-06, | |
| "loss": 2.6895, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 0.8571428571428571, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.30232558139535e-06, | |
| "loss": 1.8005, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 0.8594470046082949, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.299267330452753e-06, | |
| "loss": 1.9613, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 0.8617511520737328, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.29619872036131e-06, | |
| "loss": 2.441, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 0.8640552995391705, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.29311969839774e-06, | |
| "loss": 2.0829, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.8663594470046083, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.290030211480364e-06, | |
| "loss": 2.0612, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 0.868663594470046, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.286930206166069e-06, | |
| "loss": 2.1892, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 0.8709677419354839, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.283819628647214e-06, | |
| "loss": 2.562, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 0.8732718894009217, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.280698424748529e-06, | |
| "loss": 2.2855, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 0.8755760368663594, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.277566539923955e-06, | |
| "loss": 2.4445, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.8778801843317973, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.274423919253477e-06, | |
| "loss": 2.1253, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 0.880184331797235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.271270507439909e-06, | |
| "loss": 2.4146, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 0.8824884792626728, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.268106248805657e-06, | |
| "loss": 2.4259, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 0.8847926267281107, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.264931087289435e-06, | |
| "loss": 2.227, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 0.8870967741935484, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.261744966442955e-06, | |
| "loss": 2.437, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.8894009216589862, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.258547829427584e-06, | |
| "loss": 2.3971, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 0.8917050691244239, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.255339619010967e-06, | |
| "loss": 2.164, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 0.8940092165898618, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.252120277563609e-06, | |
| "loss": 2.22, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 0.8963133640552995, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.248889747055418e-06, | |
| "loss": 2.1684, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 0.8986175115207373, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.245647969052225e-06, | |
| "loss": 2.0864, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.9009216589861752, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.242394884712266e-06, | |
| "loss": 2.3236, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 0.9032258064516129, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.23913043478261e-06, | |
| "loss": 2.205, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 0.9055299539170507, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.235854559595567e-06, | |
| "loss": 1.9739, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 0.9078341013824884, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.232567199065057e-06, | |
| "loss": 2.1698, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 0.9101382488479263, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.229268292682927e-06, | |
| "loss": 2.1516, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.9124423963133641, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.225957779515246e-06, | |
| "loss": 1.8965, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 0.9147465437788018, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.222635598198551e-06, | |
| "loss": 2.0203, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 0.9170506912442397, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.219301686936056e-06, | |
| "loss": 2.5318, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 0.9193548387096774, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.21595598349381e-06, | |
| "loss": 2.1496, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 0.9216589861751152, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.21259842519685e-06, | |
| "loss": 1.8457, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.923963133640553, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.209228948925261e-06, | |
| "loss": 2.0916, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 0.9262672811059908, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.205847491110234e-06, | |
| "loss": 2.4179, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 0.9285714285714286, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.202453987730062e-06, | |
| "loss": 1.8332, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 0.9308755760368663, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.199048374306107e-06, | |
| "loss": 2.5119, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 0.9331797235023042, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.19563058589871e-06, | |
| "loss": 1.9103, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.9354838709677419, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.192200557103064e-06, | |
| "loss": 2.5436, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 0.9377880184331797, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.188758222045046e-06, | |
| "loss": 2.1996, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 0.9400921658986175, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.185303514376996e-06, | |
| "loss": 2.1928, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 0.9423963133640553, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.181836367273456e-06, | |
| "loss": 2.39, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 0.9447004608294931, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.178356713426856e-06, | |
| "loss": 1.9289, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.9470046082949308, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.174864485043165e-06, | |
| "loss": 1.6711, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 0.9493087557603687, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.17135961383749e-06, | |
| "loss": 1.9521, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 0.9516129032258065, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.16784203102962e-06, | |
| "loss": 2.1426, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 0.9539170506912442, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.164311667339525e-06, | |
| "loss": 2.3366, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 0.956221198156682, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.160768452982811e-06, | |
| "loss": 2.5255, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.9585253456221198, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.157212317666126e-06, | |
| "loss": 2.3075, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.9608294930875576, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.153643190582505e-06, | |
| "loss": 2.2583, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 0.9631336405529954, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.15006100040667e-06, | |
| "loss": 2.0314, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 0.9654377880184332, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.146465675290284e-06, | |
| "loss": 2.3014, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 0.967741935483871, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.142857142857142e-06, | |
| "loss": 2.5585, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.9700460829493087, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.139235330198323e-06, | |
| "loss": 2.2314, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 0.9723502304147466, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.135600163867268e-06, | |
| "loss": 2.3196, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 0.9746543778801844, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.131951569874822e-06, | |
| "loss": 2.2138, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 0.9769585253456221, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.128289473684213e-06, | |
| "loss": 2.3115, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 0.9792626728110599, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.124613800205974e-06, | |
| "loss": 2.3028, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.9815668202764977, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.120924473792819e-06, | |
| "loss": 2.3505, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 0.9838709677419355, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.117221418234443e-06, | |
| "loss": 2.2614, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 0.9861751152073732, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.113504556752279e-06, | |
| "loss": 2.0659, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 0.988479262672811, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.10977381199419e-06, | |
| "loss": 2.0262, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 0.9907834101382489, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.106029106029106e-06, | |
| "loss": 2.4228, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.9930875576036866, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.102270360341596e-06, | |
| "loss": 1.9768, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 0.9953917050691244, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.098497495826378e-06, | |
| "loss": 1.817, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 0.9976958525345622, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.094710432782772e-06, | |
| "loss": 2.1856, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.090909090909091e-06, | |
| "loss": 2.3401, | |
| "step": 434 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 868, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 2, | |
| "save_steps": 217, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.0952722129291837e+18, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |