| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.0, | |
| "eval_steps": 500, | |
| "global_step": 868, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.002304147465437788, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.998846730480915e-06, | |
| "loss": 3.6297, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.004608294930875576, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.997691064419303e-06, | |
| "loss": 3.7821, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.0069124423963133645, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.996532994337224e-06, | |
| "loss": 3.0545, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.009216589861751152, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.99537251272559e-06, | |
| "loss": 3.3776, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.01152073732718894, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.994209612044008e-06, | |
| "loss": 2.6629, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.013824884792626729, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.993044284720612e-06, | |
| "loss": 2.9764, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.016129032258064516, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.99187652315191e-06, | |
| "loss": 3.0443, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.018433179723502304, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.990706319702603e-06, | |
| "loss": 2.9777, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.020737327188940093, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.989533666705432e-06, | |
| "loss": 2.6634, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.02304147465437788, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.988358556461003e-06, | |
| "loss": 2.5688, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.02534562211981567, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.98718098123762e-06, | |
| "loss": 2.7999, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.027649769585253458, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.986000933271115e-06, | |
| "loss": 2.8395, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.029953917050691243, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.984818404764686e-06, | |
| "loss": 2.9181, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.03225806451612903, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.983633387888708e-06, | |
| "loss": 2.3633, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.03456221198156682, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.982445874780575e-06, | |
| "loss": 2.6805, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.03686635944700461, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.981255857544517e-06, | |
| "loss": 3.2891, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.03917050691244239, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.980063328251437e-06, | |
| "loss": 2.3438, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.041474654377880185, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.97886827893872e-06, | |
| "loss": 3.0867, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.04377880184331797, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.97767070161006e-06, | |
| "loss": 3.0397, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.04608294930875576, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.976470588235294e-06, | |
| "loss": 2.675, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.04838709677419355, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.975267930750206e-06, | |
| "loss": 2.588, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.05069124423963134, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.974062721056355e-06, | |
| "loss": 2.5917, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.052995391705069124, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.97285495102089e-06, | |
| "loss": 2.6421, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.055299539170506916, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.971644612476371e-06, | |
| "loss": 2.8325, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.0576036866359447, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.970431697220581e-06, | |
| "loss": 2.7928, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.059907834101382486, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.96921619701634e-06, | |
| "loss": 2.7043, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.06221198156682028, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.967998103591326e-06, | |
| "loss": 2.5606, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.06451612903225806, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.966777408637875e-06, | |
| "loss": 2.5944, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.06682027649769585, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.965554103812804e-06, | |
| "loss": 2.0101, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.06912442396313365, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.964328180737219e-06, | |
| "loss": 2.669, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.07142857142857142, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.963099630996312e-06, | |
| "loss": 2.5637, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.07373271889400922, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.961868446139181e-06, | |
| "loss": 2.8749, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.07603686635944701, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.960634617678637e-06, | |
| "loss": 2.8018, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.07834101382488479, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.959398137090996e-06, | |
| "loss": 2.5725, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.08064516129032258, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.958158995815901e-06, | |
| "loss": 2.8989, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.08294930875576037, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.956917185256105e-06, | |
| "loss": 2.4909, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.08525345622119816, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.955672696777287e-06, | |
| "loss": 2.5061, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.08755760368663594, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.954425521707844e-06, | |
| "loss": 2.3851, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.08986175115207373, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.953175651338697e-06, | |
| "loss": 3.0129, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.09216589861751152, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.951923076923079e-06, | |
| "loss": 2.8212, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.0944700460829493, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.950667789676334e-06, | |
| "loss": 2.952, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.0967741935483871, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.949409780775717e-06, | |
| "loss": 2.2857, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.09907834101382489, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.948149041360185e-06, | |
| "loss": 2.2268, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.10138248847926268, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.94688556253018e-06, | |
| "loss": 2.7615, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.10368663594470046, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.945619335347432e-06, | |
| "loss": 3.1013, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.10599078341013825, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.944350350834746e-06, | |
| "loss": 3.0616, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.10829493087557604, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.943078599975779e-06, | |
| "loss": 2.2084, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.11059907834101383, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.941804073714841e-06, | |
| "loss": 2.1988, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.11290322580645161, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.94052676295667e-06, | |
| "loss": 2.5741, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.1152073732718894, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.939246658566223e-06, | |
| "loss": 2.4302, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.1175115207373272, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.937963751368447e-06, | |
| "loss": 2.5109, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.11981566820276497, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.936678032148077e-06, | |
| "loss": 2.2233, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.12211981566820276, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.935389491649398e-06, | |
| "loss": 3.1745, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.12442396313364056, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.934098120576033e-06, | |
| "loss": 2.6837, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.12672811059907835, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.932803909590715e-06, | |
| "loss": 2.5496, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.12903225806451613, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.931506849315069e-06, | |
| "loss": 2.6769, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.1313364055299539, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.930206930329376e-06, | |
| "loss": 2.4358, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.1336405529953917, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.928904143172348e-06, | |
| "loss": 2.6133, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.1359447004608295, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.927598478340902e-06, | |
| "loss": 2.6126, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.1382488479262673, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.926289926289927e-06, | |
| "loss": 2.3671, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.14055299539170507, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.92497847743205e-06, | |
| "loss": 2.3528, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.14285714285714285, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.923664122137405e-06, | |
| "loss": 2.894, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.14516129032258066, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.922346850733393e-06, | |
| "loss": 2.944, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.14746543778801843, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.921026653504443e-06, | |
| "loss": 2.495, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.1497695852534562, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.919703520691786e-06, | |
| "loss": 2.2919, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.15207373271889402, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.918377442493198e-06, | |
| "loss": 2.6164, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.1543778801843318, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.917048409062771e-06, | |
| "loss": 2.5487, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.15668202764976957, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.91571641051066e-06, | |
| "loss": 2.6602, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.15898617511520738, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.914381436902843e-06, | |
| "loss": 2.5056, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.16129032258064516, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.913043478260871e-06, | |
| "loss": 2.8503, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.16359447004608296, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.911702524561623e-06, | |
| "loss": 2.6178, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.16589861751152074, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.910358565737052e-06, | |
| "loss": 2.7173, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.16820276497695852, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.909011591673938e-06, | |
| "loss": 2.4166, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.17050691244239632, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.907661592213627e-06, | |
| "loss": 2.5658, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.1728110599078341, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.906308557151781e-06, | |
| "loss": 2.3656, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.17511520737327188, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.90495247623812e-06, | |
| "loss": 2.4611, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.1774193548387097, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.903593339176162e-06, | |
| "loss": 2.2937, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.17972350230414746, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.902231135622965e-06, | |
| "loss": 2.172, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.18202764976958524, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.900865855188858e-06, | |
| "loss": 2.4489, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.18433179723502305, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.899497487437187e-06, | |
| "loss": 2.7153, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.18663594470046083, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.89812602188404e-06, | |
| "loss": 2.7078, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.1889400921658986, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.896751447997987e-06, | |
| "loss": 2.1448, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.1912442396313364, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.8953737551998e-06, | |
| "loss": 2.4494, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.1935483870967742, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.893992932862191e-06, | |
| "loss": 2.5461, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.195852534562212, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.89260897030954e-06, | |
| "loss": 2.219, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.19815668202764977, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.891221856817608e-06, | |
| "loss": 2.6866, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.20046082949308755, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.889831581613272e-06, | |
| "loss": 2.2086, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.20276497695852536, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.88843813387424e-06, | |
| "loss": 2.658, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.20506912442396313, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.887041502728773e-06, | |
| "loss": 2.2961, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.2073732718894009, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.885641677255401e-06, | |
| "loss": 2.5561, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.20967741935483872, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.884238646482637e-06, | |
| "loss": 2.5409, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.2119815668202765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.882832399388692e-06, | |
| "loss": 2.2584, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.21428571428571427, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.881422924901188e-06, | |
| "loss": 2.5798, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.21658986175115208, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.88001021189686e-06, | |
| "loss": 2.6547, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.21889400921658986, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.878594249201278e-06, | |
| "loss": 1.9719, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.22119815668202766, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.877175025588538e-06, | |
| "loss": 2.6163, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.22350230414746544, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.875752529780967e-06, | |
| "loss": 2.5005, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.22580645161290322, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.874326750448835e-06, | |
| "loss": 2.4969, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.22811059907834103, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.872897676210041e-06, | |
| "loss": 2.4532, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.2304147465437788, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.871465295629821e-06, | |
| "loss": 2.7851, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.23271889400921658, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.870029597220437e-06, | |
| "loss": 2.4056, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.2350230414746544, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.868590569440867e-06, | |
| "loss": 2.4184, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.23732718894009217, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.867148200696505e-06, | |
| "loss": 2.4005, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.23963133640552994, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.865702479338843e-06, | |
| "loss": 2.3063, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.24193548387096775, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.86425339366516e-06, | |
| "loss": 2.309, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.24423963133640553, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.8628009319182e-06, | |
| "loss": 2.2872, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.2465437788018433, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.861345082285862e-06, | |
| "loss": 2.5453, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.2488479262672811, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.859885832900883e-06, | |
| "loss": 2.4237, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.2511520737327189, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.8584231718405e-06, | |
| "loss": 2.7325, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.2534562211981567, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.856957087126139e-06, | |
| "loss": 2.8046, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.2557603686635945, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.855487566723084e-06, | |
| "loss": 2.3187, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.25806451612903225, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.854014598540148e-06, | |
| "loss": 2.1758, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.26036866359447003, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.852538170429336e-06, | |
| "loss": 2.3946, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.2626728110599078, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.851058270185526e-06, | |
| "loss": 2.5586, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.26497695852534564, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.849574885546111e-06, | |
| "loss": 2.4507, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.2672811059907834, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.848088004190675e-06, | |
| "loss": 2.2848, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.2695852534562212, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.846597613740658e-06, | |
| "loss": 2.7136, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.271889400921659, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.845103701758992e-06, | |
| "loss": 2.2407, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.27419354838709675, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.843606255749772e-06, | |
| "loss": 1.8917, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.2764976958525346, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.842105263157896e-06, | |
| "loss": 2.49, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.27880184331797236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.840600711368727e-06, | |
| "loss": 2.2081, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.28110599078341014, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.839092587707729e-06, | |
| "loss": 2.5349, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.2834101382488479, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.837580879440117e-06, | |
| "loss": 2.2057, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.2857142857142857, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.836065573770494e-06, | |
| "loss": 2.6525, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.2880184331797235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.834546657842488e-06, | |
| "loss": 2.466, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.2903225806451613, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.833024118738406e-06, | |
| "loss": 2.3602, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.2926267281105991, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.831497943478838e-06, | |
| "loss": 1.9779, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.29493087557603687, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.829968119022318e-06, | |
| "loss": 2.0746, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.29723502304147464, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.82843463226493e-06, | |
| "loss": 2.5047, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.2995391705069124, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.826897470039947e-06, | |
| "loss": 2.3455, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.30184331797235026, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.825356619117452e-06, | |
| "loss": 2.7969, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.30414746543778803, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.823812066203951e-06, | |
| "loss": 2.1092, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.3064516129032258, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.822263797942004e-06, | |
| "loss": 2.454, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.3087557603686636, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.82071180090982e-06, | |
| "loss": 2.2478, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.31105990783410137, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.819156061620899e-06, | |
| "loss": 2.1831, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.31336405529953915, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.817596566523606e-06, | |
| "loss": 2.199, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.315668202764977, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.816033302000808e-06, | |
| "loss": 2.3337, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.31797235023041476, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.814466254369455e-06, | |
| "loss": 2.3119, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.32027649769585254, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.8128954098802e-06, | |
| "loss": 2.319, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.3225806451612903, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.811320754716981e-06, | |
| "loss": 2.37, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.3248847926267281, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.809742274996628e-06, | |
| "loss": 2.3041, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.3271889400921659, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.808159956768442e-06, | |
| "loss": 2.7213, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.3294930875576037, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.806573786013798e-06, | |
| "loss": 2.6252, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.3317972350230415, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.80498374864572e-06, | |
| "loss": 1.8312, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.33410138248847926, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.803389830508476e-06, | |
| "loss": 2.4053, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.33640552995391704, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.80179201737714e-06, | |
| "loss": 2.3841, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.3387096774193548, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.800190294957183e-06, | |
| "loss": 2.4041, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.34101382488479265, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.79858464888405e-06, | |
| "loss": 2.0689, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.3433179723502304, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.796975064722715e-06, | |
| "loss": 2.0356, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.3456221198156682, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.795361527967258e-06, | |
| "loss": 2.0666, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.347926267281106, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.793744024040433e-06, | |
| "loss": 2.5768, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.35023041474654376, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.792122538293217e-06, | |
| "loss": 2.1563, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.35253456221198154, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.790497056004383e-06, | |
| "loss": 2.3727, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.3548387096774194, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.78886756238004e-06, | |
| "loss": 1.9836, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.35714285714285715, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.787234042553192e-06, | |
| "loss": 2.0014, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.35944700460829493, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.785596481583288e-06, | |
| "loss": 2.128, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.3617511520737327, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.78395486445576e-06, | |
| "loss": 1.9376, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.3640552995391705, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.782309176081566e-06, | |
| "loss": 2.6088, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.3663594470046083, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.780659401296732e-06, | |
| "loss": 2.476, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.3686635944700461, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.779005524861878e-06, | |
| "loss": 2.2163, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.3709677419354839, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.777347531461764e-06, | |
| "loss": 2.6048, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.37327188940092165, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.775685405704793e-06, | |
| "loss": 2.2061, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.37557603686635943, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.774019132122557e-06, | |
| "loss": 2.4101, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.3778801843317972, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.77234869516935e-06, | |
| "loss": 2.5347, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.38018433179723504, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.770674079221682e-06, | |
| "loss": 2.1867, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.3824884792626728, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.768995268577792e-06, | |
| "loss": 2.3605, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.3847926267281106, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.767312247457155e-06, | |
| "loss": 2.1918, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.3870967741935484, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.765625e-06, | |
| "loss": 2.4631, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.38940092165898615, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.763933510266798e-06, | |
| "loss": 2.077, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.391705069124424, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.762237762237764e-06, | |
| "loss": 2.274, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.39400921658986177, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.760537739812353e-06, | |
| "loss": 2.5203, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.39631336405529954, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.75883342680875e-06, | |
| "loss": 2.2468, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.3986175115207373, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.757124806963358e-06, | |
| "loss": 2.2579, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.4009216589861751, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.755411863930279e-06, | |
| "loss": 2.3655, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.4032258064516129, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.75369458128079e-06, | |
| "loss": 2.353, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.4055299539170507, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.75197294250282e-06, | |
| "loss": 2.2856, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.4078341013824885, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.750246931000424e-06, | |
| "loss": 2.2884, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.41013824884792627, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.748516530093248e-06, | |
| "loss": 2.1927, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.41244239631336405, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.746781723015986e-06, | |
| "loss": 2.4978, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.4147465437788018, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.745042492917848e-06, | |
| "loss": 2.1364, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.41705069124423966, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.743298822862007e-06, | |
| "loss": 2.4476, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.41935483870967744, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.74155069582505e-06, | |
| "loss": 1.8542, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.4216589861751152, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.739798094696433e-06, | |
| "loss": 2.2825, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.423963133640553, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.738041002277906e-06, | |
| "loss": 2.504, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.42626728110599077, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.736279401282966e-06, | |
| "loss": 2.1668, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.42857142857142855, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.734513274336284e-06, | |
| "loss": 2.2124, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.4308755760368664, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.732742603973132e-06, | |
| "loss": 2.0533, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.43317972350230416, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.73096737263881e-06, | |
| "loss": 2.2371, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.43548387096774194, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.729187562688066e-06, | |
| "loss": 2.577, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.4377880184331797, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.727403156384506e-06, | |
| "loss": 2.0062, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.4400921658986175, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.725614135900015e-06, | |
| "loss": 2.5623, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.4423963133640553, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.723820483314155e-06, | |
| "loss": 2.3879, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.4447004608294931, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.722022180613568e-06, | |
| "loss": 2.3884, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.4470046082949309, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.720219209691377e-06, | |
| "loss": 2.1677, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.44930875576036866, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.71841155234657e-06, | |
| "loss": 2.1849, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.45161290322580644, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.7165991902834e-06, | |
| "loss": 2.4744, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.4539170506912442, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.714782105110758e-06, | |
| "loss": 2.1845, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.45622119815668205, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.71296027834155e-06, | |
| "loss": 2.1075, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.45852534562211983, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.711133691392076e-06, | |
| "loss": 2.2594, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.4608294930875576, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.709302325581395e-06, | |
| "loss": 2.0445, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.4631336405529954, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.707466162130695e-06, | |
| "loss": 1.8306, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.46543778801843316, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.705625182162635e-06, | |
| "loss": 2.0827, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.46774193548387094, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.703779366700715e-06, | |
| "loss": 2.4998, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.4700460829493088, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.701928696668615e-06, | |
| "loss": 1.9692, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.47235023041474655, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.70007315288954e-06, | |
| "loss": 2.2138, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.47465437788018433, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.698212716085556e-06, | |
| "loss": 2.3699, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.4769585253456221, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.696347366876926e-06, | |
| "loss": 2.2658, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.4792626728110599, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.694477085781434e-06, | |
| "loss": 2.3592, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.4815668202764977, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.692601853213708e-06, | |
| "loss": 2.1424, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.4838709677419355, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.690721649484536e-06, | |
| "loss": 2.1127, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.4861751152073733, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.688836454800178e-06, | |
| "loss": 2.39, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.48847926267281105, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.686946249261665e-06, | |
| "loss": 2.2204, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.49078341013824883, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.685051012864114e-06, | |
| "loss": 2.2297, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.4930875576036866, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.683150725496004e-06, | |
| "loss": 2.2213, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.49539170506912444, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.681245366938474e-06, | |
| "loss": 2.3086, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.4976958525345622, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.679334916864607e-06, | |
| "loss": 2.1777, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.67741935483871e-06, | |
| "loss": 2.2379, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.5023041474654378, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.675498660315571e-06, | |
| "loss": 2.3132, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.5046082949308756, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.673572812639739e-06, | |
| "loss": 2.0896, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.5069124423963134, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.671641791044779e-06, | |
| "loss": 2.1555, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.5092165898617511, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.669705574652519e-06, | |
| "loss": 2.2675, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.511520737327189, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.667764142472315e-06, | |
| "loss": 2.0952, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.5138248847926268, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.665817473400271e-06, | |
| "loss": 2.1719, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.5161290322580645, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.663865546218488e-06, | |
| "loss": 2.269, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.5184331797235023, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.66190833959429e-06, | |
| "loss": 2.3643, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.5207373271889401, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.659945832079446e-06, | |
| "loss": 2.2851, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.5230414746543779, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.657978002109388e-06, | |
| "loss": 2.2314, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.5253456221198156, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.656004828002415e-06, | |
| "loss": 2.306, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.5276497695852534, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.654026287958907e-06, | |
| "loss": 2.72, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.5299539170506913, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.652042360060516e-06, | |
| "loss": 2.2472, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.532258064516129, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.650053022269354e-06, | |
| "loss": 2.14, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.5345622119815668, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.648058252427186e-06, | |
| "loss": 2.1352, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.5368663594470046, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.646058028254597e-06, | |
| "loss": 2.1396, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.5391705069124424, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.644052327350167e-06, | |
| "loss": 2.0091, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.5414746543778802, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.642041127189643e-06, | |
| "loss": 2.3559, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.543778801843318, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.640024405125077e-06, | |
| "loss": 2.2764, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.5460829493087558, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.638002138383995e-06, | |
| "loss": 2.4431, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.5483870967741935, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.635974304068522e-06, | |
| "loss": 2.2692, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.5506912442396313, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.633940879154541e-06, | |
| "loss": 2.2558, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.5529953917050692, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.631901840490798e-06, | |
| "loss": 1.8841, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.5552995391705069, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.629857164798035e-06, | |
| "loss": 2.243, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.5576036866359447, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.627806828668103e-06, | |
| "loss": 2.7055, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.5599078341013825, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.625750808563067e-06, | |
| "loss": 2.0709, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.5622119815668203, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.623689080814313e-06, | |
| "loss": 2.2729, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.5645161290322581, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.621621621621622e-06, | |
| "loss": 2.1815, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.5668202764976958, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.619548407052275e-06, | |
| "loss": 2.3074, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.5691244239631337, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.617469413040112e-06, | |
| "loss": 2.2204, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.5714285714285714, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.615384615384616e-06, | |
| "loss": 2.2254, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.5737327188940092, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.613293989749963e-06, | |
| "loss": 2.2028, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.576036866359447, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.611197511664076e-06, | |
| "loss": 1.9041, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.5783410138248848, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.609095156517678e-06, | |
| "loss": 2.1567, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.5806451612903226, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.60698689956332e-06, | |
| "loss": 2.2442, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.5829493087557603, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.604872715914416e-06, | |
| "loss": 2.244, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.5852534562211982, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.602752580544262e-06, | |
| "loss": 2.3902, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.5875576036866359, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.600626468285045e-06, | |
| "loss": 2.3403, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.5898617511520737, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.598494353826851e-06, | |
| "loss": 2.3575, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.5921658986175116, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.596356211716666e-06, | |
| "loss": 2.3505, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.5944700460829493, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.594212016357346e-06, | |
| "loss": 2.1773, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.5967741935483871, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.592061742006616e-06, | |
| "loss": 2.296, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.5990783410138248, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.589905362776025e-06, | |
| "loss": 1.9344, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.6013824884792627, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.587742852629917e-06, | |
| "loss": 2.5543, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.6036866359447005, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.585574185384373e-06, | |
| "loss": 2.7177, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.6059907834101382, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.583399334706163e-06, | |
| "loss": 2.3868, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.6082949308755761, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.581218274111676e-06, | |
| "loss": 2.3185, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.6105990783410138, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.579030976965846e-06, | |
| "loss": 1.9385, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.6129032258064516, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.57683741648107e-06, | |
| "loss": 2.3687, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.6152073732718893, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.574637565716107e-06, | |
| "loss": 2.2045, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.6175115207373272, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.572431397574987e-06, | |
| "loss": 2.4644, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.619815668202765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.57021888480588e-06, | |
| "loss": 2.3897, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.6221198156682027, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.568e-06, | |
| "loss": 2.4176, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.6244239631336406, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.56577471559045e-06, | |
| "loss": 2.3342, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.6267281105990783, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.563543003851093e-06, | |
| "loss": 2.3815, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.6290322580645161, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.561304836895388e-06, | |
| "loss": 2.435, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.631336405529954, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.55906018667525e-06, | |
| "loss": 2.4491, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.6336405529953917, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.556809024979856e-06, | |
| "loss": 2.3206, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.6359447004608295, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.554551323434475e-06, | |
| "loss": 2.1149, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.6382488479262672, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.552287053499273e-06, | |
| "loss": 2.2707, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.6405529953917051, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.550016186468114e-06, | |
| "loss": 2.3246, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.6428571428571429, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.547738693467338e-06, | |
| "loss": 2.2557, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.6451612903225806, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.545454545454547e-06, | |
| "loss": 2.3648, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.6474654377880185, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.543163713217365e-06, | |
| "loss": 2.3793, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.6497695852534562, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.540866167372192e-06, | |
| "loss": 2.2765, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.652073732718894, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.538561878362955e-06, | |
| "loss": 2.3297, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.6543778801843319, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.53625081645983e-06, | |
| "loss": 2.2636, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.6566820276497696, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.533932951757974e-06, | |
| "loss": 2.271, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.6589861751152074, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.53160825417622e-06, | |
| "loss": 2.497, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.6612903225806451, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.529276693455799e-06, | |
| "loss": 1.994, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.663594470046083, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.526938239159003e-06, | |
| "loss": 1.9041, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.6658986175115207, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.524592860667874e-06, | |
| "loss": 2.3384, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.6682027649769585, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.522240527182867e-06, | |
| "loss": 2.0346, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.6705069124423964, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.519881207721498e-06, | |
| "loss": 2.1251, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.6728110599078341, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.517514871116987e-06, | |
| "loss": 2.3156, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.6751152073732719, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.51514148601688e-06, | |
| "loss": 2.3411, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.6774193548387096, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.512761020881673e-06, | |
| "loss": 2.5454, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.6797235023041475, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.510373443983403e-06, | |
| "loss": 1.864, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.6820276497695853, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.507978723404256e-06, | |
| "loss": 2.2479, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.684331797235023, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.505576827035126e-06, | |
| "loss": 2.2826, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.6866359447004609, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.503167722574192e-06, | |
| "loss": 1.8908, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.6889400921658986, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.500751377525463e-06, | |
| "loss": 2.0887, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.6912442396313364, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.498327759197325e-06, | |
| "loss": 1.9868, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.6935483870967742, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.495896834701056e-06, | |
| "loss": 2.03, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.695852534562212, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.493458570949348e-06, | |
| "loss": 2.413, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.6981566820276498, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.491012934654798e-06, | |
| "loss": 2.2828, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.7004608294930875, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.488559892328399e-06, | |
| "loss": 2.4673, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.7027649769585254, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.486099410278012e-06, | |
| "loss": 2.1999, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.7050691244239631, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.483631454606818e-06, | |
| "loss": 2.1067, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.7073732718894009, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.481155991211765e-06, | |
| "loss": 2.0962, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.7096774193548387, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.47867298578199e-06, | |
| "loss": 2.4821, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.7119815668202765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.476182403797255e-06, | |
| "loss": 2.4688, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.7142857142857143, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.473684210526317e-06, | |
| "loss": 2.1085, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.716589861751152, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.471178371025337e-06, | |
| "loss": 2.435, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.7188940092165899, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.46866485013624e-06, | |
| "loss": 2.3131, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.7211981566820277, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.466143612485077e-06, | |
| "loss": 2.5537, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 0.7235023041474654, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.463614622480356e-06, | |
| "loss": 2.1758, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.7258064516129032, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.461077844311378e-06, | |
| "loss": 2.2752, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.728110599078341, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.45853324194654e-06, | |
| "loss": 2.0488, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.7304147465437788, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.45598077913163e-06, | |
| "loss": 2.3105, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 0.7327188940092166, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.453420419388106e-06, | |
| "loss": 1.936, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.7350230414746544, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.450852126011364e-06, | |
| "loss": 2.1634, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 0.7373271889400922, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.448275862068967e-06, | |
| "loss": 1.9953, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.7396313364055299, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.445691590398895e-06, | |
| "loss": 2.3994, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 0.7419354838709677, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.44309927360775e-06, | |
| "loss": 1.837, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 0.7442396313364056, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.440498874068943e-06, | |
| "loss": 2.0258, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 0.7465437788018433, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.43789035392089e-06, | |
| "loss": 2.5129, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.7488479262672811, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.435273675065161e-06, | |
| "loss": 2.0436, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.7511520737327189, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.432648799164637e-06, | |
| "loss": 2.4449, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 0.7534562211981567, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.430015687641625e-06, | |
| "loss": 2.0828, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 0.7557603686635944, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.42737430167598e-06, | |
| "loss": 2.3661, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.7580645161290323, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.424724602203184e-06, | |
| "loss": 1.9816, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 0.7603686635944701, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.422066549912434e-06, | |
| "loss": 2.3207, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.7626728110599078, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.419400105244694e-06, | |
| "loss": 2.1479, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 0.7649769585253456, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.416725228390726e-06, | |
| "loss": 2.211, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.7672811059907834, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.41404187928911e-06, | |
| "loss": 1.8804, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 0.7695852534562212, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.411350017624251e-06, | |
| "loss": 2.1775, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 0.771889400921659, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.40864960282436e-06, | |
| "loss": 2.2466, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.7741935483870968, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.405940594059407e-06, | |
| "loss": 2.0984, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.7764976958525346, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.403222950239065e-06, | |
| "loss": 2.5951, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 0.7788018433179723, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.400496630010643e-06, | |
| "loss": 2.1321, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.7811059907834101, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.397761591756973e-06, | |
| "loss": 1.8965, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 0.783410138248848, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.395017793594307e-06, | |
| "loss": 2.4059, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.7857142857142857, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.392265193370167e-06, | |
| "loss": 2.4522, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 0.7880184331797235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.389503748661194e-06, | |
| "loss": 2.4031, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 0.7903225806451613, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.386733416770964e-06, | |
| "loss": 2.3547, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 0.7926267281105991, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.383954154727794e-06, | |
| "loss": 2.0233, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 0.7949308755760369, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.381165919282512e-06, | |
| "loss": 2.3673, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.7972350230414746, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.378368666906218e-06, | |
| "loss": 2.4832, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 0.7995391705069125, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.375562353788017e-06, | |
| "loss": 2.0684, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 0.8018433179723502, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.372746935832732e-06, | |
| "loss": 2.1924, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.804147465437788, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.369922368658603e-06, | |
| "loss": 2.3609, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 0.8064516129032258, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.367088607594939e-06, | |
| "loss": 2.2614, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.8087557603686636, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.364245607679769e-06, | |
| "loss": 2.5841, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 0.8110599078341014, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.361393323657475e-06, | |
| "loss": 2.038, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.8133640552995391, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.358531709976377e-06, | |
| "loss": 2.2917, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 0.815668202764977, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.355660720786313e-06, | |
| "loss": 2.0844, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 0.8179723502304147, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.352780309936191e-06, | |
| "loss": 1.7098, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.8202764976958525, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.349890430971511e-06, | |
| "loss": 2.4264, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 0.8225806451612904, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.346991037131882e-06, | |
| "loss": 2.4563, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 0.8248847926267281, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.34408208134848e-06, | |
| "loss": 1.9052, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 0.8271889400921659, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.341163516241514e-06, | |
| "loss": 2.1467, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 0.8294930875576036, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.338235294117646e-06, | |
| "loss": 2.2446, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.8317972350230415, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.33529736696741e-06, | |
| "loss": 2.1227, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 0.8341013824884793, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.332349686462561e-06, | |
| "loss": 2.2938, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 0.836405529953917, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.329392203953447e-06, | |
| "loss": 2.2272, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 0.8387096774193549, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.326424870466322e-06, | |
| "loss": 1.9074, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.8410138248847926, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.32344763670065e-06, | |
| "loss": 1.7826, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.8433179723502304, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.320460453026365e-06, | |
| "loss": 2.4799, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 0.8456221198156681, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.317463269481124e-06, | |
| "loss": 1.8732, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 0.847926267281106, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.314456035767513e-06, | |
| "loss": 2.4681, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 0.8502304147465438, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.311438701250233e-06, | |
| "loss": 1.7413, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 0.8525345622119815, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.308411214953271e-06, | |
| "loss": 2.2242, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.8548387096774194, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.305373525557013e-06, | |
| "loss": 2.6895, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 0.8571428571428571, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.30232558139535e-06, | |
| "loss": 1.8005, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 0.8594470046082949, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.299267330452753e-06, | |
| "loss": 1.9613, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 0.8617511520737328, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.29619872036131e-06, | |
| "loss": 2.441, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 0.8640552995391705, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.29311969839774e-06, | |
| "loss": 2.0829, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.8663594470046083, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.290030211480364e-06, | |
| "loss": 2.0612, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 0.868663594470046, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.286930206166069e-06, | |
| "loss": 2.1892, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 0.8709677419354839, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.283819628647214e-06, | |
| "loss": 2.562, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 0.8732718894009217, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.280698424748529e-06, | |
| "loss": 2.2855, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 0.8755760368663594, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.277566539923955e-06, | |
| "loss": 2.4445, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.8778801843317973, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.274423919253477e-06, | |
| "loss": 2.1253, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 0.880184331797235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.271270507439909e-06, | |
| "loss": 2.4146, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 0.8824884792626728, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.268106248805657e-06, | |
| "loss": 2.4259, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 0.8847926267281107, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.264931087289435e-06, | |
| "loss": 2.227, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 0.8870967741935484, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.261744966442955e-06, | |
| "loss": 2.437, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.8894009216589862, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.258547829427584e-06, | |
| "loss": 2.3971, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 0.8917050691244239, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.255339619010967e-06, | |
| "loss": 2.164, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 0.8940092165898618, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.252120277563609e-06, | |
| "loss": 2.22, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 0.8963133640552995, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.248889747055418e-06, | |
| "loss": 2.1684, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 0.8986175115207373, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.245647969052225e-06, | |
| "loss": 2.0864, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.9009216589861752, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.242394884712266e-06, | |
| "loss": 2.3236, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 0.9032258064516129, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.23913043478261e-06, | |
| "loss": 2.205, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 0.9055299539170507, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.235854559595567e-06, | |
| "loss": 1.9739, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 0.9078341013824884, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.232567199065057e-06, | |
| "loss": 2.1698, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 0.9101382488479263, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.229268292682927e-06, | |
| "loss": 2.1516, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.9124423963133641, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.225957779515246e-06, | |
| "loss": 1.8965, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 0.9147465437788018, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.222635598198551e-06, | |
| "loss": 2.0203, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 0.9170506912442397, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.219301686936056e-06, | |
| "loss": 2.5318, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 0.9193548387096774, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.21595598349381e-06, | |
| "loss": 2.1496, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 0.9216589861751152, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.21259842519685e-06, | |
| "loss": 1.8457, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.923963133640553, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.209228948925261e-06, | |
| "loss": 2.0916, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 0.9262672811059908, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.205847491110234e-06, | |
| "loss": 2.4179, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 0.9285714285714286, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.202453987730062e-06, | |
| "loss": 1.8332, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 0.9308755760368663, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.199048374306107e-06, | |
| "loss": 2.5119, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 0.9331797235023042, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.19563058589871e-06, | |
| "loss": 1.9103, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.9354838709677419, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.192200557103064e-06, | |
| "loss": 2.5436, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 0.9377880184331797, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.188758222045046e-06, | |
| "loss": 2.1996, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 0.9400921658986175, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.185303514376996e-06, | |
| "loss": 2.1928, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 0.9423963133640553, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.181836367273456e-06, | |
| "loss": 2.39, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 0.9447004608294931, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.178356713426856e-06, | |
| "loss": 1.9289, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.9470046082949308, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.174864485043165e-06, | |
| "loss": 1.6711, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 0.9493087557603687, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.17135961383749e-06, | |
| "loss": 1.9521, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 0.9516129032258065, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.16784203102962e-06, | |
| "loss": 2.1426, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 0.9539170506912442, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.164311667339525e-06, | |
| "loss": 2.3366, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 0.956221198156682, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.160768452982811e-06, | |
| "loss": 2.5255, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.9585253456221198, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.157212317666126e-06, | |
| "loss": 2.3075, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.9608294930875576, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.153643190582505e-06, | |
| "loss": 2.2583, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 0.9631336405529954, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.15006100040667e-06, | |
| "loss": 2.0314, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 0.9654377880184332, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.146465675290284e-06, | |
| "loss": 2.3014, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 0.967741935483871, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.142857142857142e-06, | |
| "loss": 2.5585, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.9700460829493087, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.139235330198323e-06, | |
| "loss": 2.2314, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 0.9723502304147466, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.135600163867268e-06, | |
| "loss": 2.3196, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 0.9746543778801844, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.131951569874822e-06, | |
| "loss": 2.2138, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 0.9769585253456221, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.128289473684213e-06, | |
| "loss": 2.3115, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 0.9792626728110599, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.124613800205974e-06, | |
| "loss": 2.3028, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.9815668202764977, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.120924473792819e-06, | |
| "loss": 2.3505, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 0.9838709677419355, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.117221418234443e-06, | |
| "loss": 2.2614, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 0.9861751152073732, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.113504556752279e-06, | |
| "loss": 2.0659, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 0.988479262672811, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.10977381199419e-06, | |
| "loss": 2.0262, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 0.9907834101382489, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.106029106029106e-06, | |
| "loss": 2.4228, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.9930875576036866, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.102270360341596e-06, | |
| "loss": 1.9768, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 0.9953917050691244, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.098497495826378e-06, | |
| "loss": 1.817, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 0.9976958525345622, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.094710432782772e-06, | |
| "loss": 2.1856, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.090909090909091e-06, | |
| "loss": 2.3401, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 1.0023041474654377, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.087093389296958e-06, | |
| "loss": 2.5513, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 1.0046082949308757, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.083263246425568e-06, | |
| "loss": 2.1191, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 1.0069124423963134, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.079418580155889e-06, | |
| "loss": 2.1874, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 1.0092165898617511, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.075559307724779e-06, | |
| "loss": 2.2957, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 1.0115207373271888, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.071685345739058e-06, | |
| "loss": 2.178, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 1.0138248847926268, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.067796610169493e-06, | |
| "loss": 2.1927, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 1.0161290322580645, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.063893016344726e-06, | |
| "loss": 2.1682, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 1.0184331797235022, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.05997447894513e-06, | |
| "loss": 2.2347, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 1.0207373271889402, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.056040911996592e-06, | |
| "loss": 2.3404, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 1.023041474654378, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.052092228864219e-06, | |
| "loss": 2.2901, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 1.0253456221198156, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.04812834224599e-06, | |
| "loss": 1.9081, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 1.0276497695852536, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.044149164166311e-06, | |
| "loss": 2.0529, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 1.0299539170506913, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.040154605969508e-06, | |
| "loss": 2.3772, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 1.032258064516129, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.036144578313254e-06, | |
| "loss": 2.3574, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 1.0345622119815667, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.032118991161889e-06, | |
| "loss": 2.0081, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 1.0368663594470047, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.028077753779698e-06, | |
| "loss": 2.0889, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 1.0391705069124424, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.024020774724086e-06, | |
| "loss": 2.5457, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 1.0414746543778801, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.019947961838683e-06, | |
| "loss": 1.9219, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 1.043778801843318, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.01585922224636e-06, | |
| "loss": 2.3704, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 1.0460829493087558, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.011754462342185e-06, | |
| "loss": 2.1011, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 1.0483870967741935, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.007633587786259e-06, | |
| "loss": 2.1355, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 1.0506912442396312, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.003496503496504e-06, | |
| "loss": 2.5515, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 1.0529953917050692, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.99934311364134e-06, | |
| "loss": 2.3604, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 1.055299539170507, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.995173321632296e-06, | |
| "loss": 2.2622, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 1.0576036866359446, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.990987030116509e-06, | |
| "loss": 2.4929, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 1.0599078341013826, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.986784140969164e-06, | |
| "loss": 2.1313, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 1.0622119815668203, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.982564555285809e-06, | |
| "loss": 2.2448, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 1.064516129032258, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.978328173374614e-06, | |
| "loss": 2.1678, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 1.066820276497696, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.974074894748505e-06, | |
| "loss": 2.3513, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 1.0691244239631337, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.969804618117229e-06, | |
| "loss": 2.2733, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 1.0714285714285714, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.96551724137931e-06, | |
| "loss": 2.2442, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 1.0737327188940091, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.961212661613909e-06, | |
| "loss": 1.8666, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 1.076036866359447, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.956890775072592e-06, | |
| "loss": 2.0399, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 1.0783410138248848, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.952551477170994e-06, | |
| "loss": 2.0874, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 1.0806451612903225, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.948194662480377e-06, | |
| "loss": 2.0166, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 1.0829493087557605, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.943820224719103e-06, | |
| "loss": 2.1944, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 1.0852534562211982, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.939428056743977e-06, | |
| "loss": 2.1455, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 1.087557603686636, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.935018050541517e-06, | |
| "loss": 2.539, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 1.0898617511520738, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.930590097219082e-06, | |
| "loss": 2.2455, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 1.0921658986175116, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.926144086995923e-06, | |
| "loss": 2.1535, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 1.0944700460829493, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.921679909194099e-06, | |
| "loss": 2.2458, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 1.096774193548387, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.9171974522293e-06, | |
| "loss": 2.1534, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 1.099078341013825, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.91269660360155e-06, | |
| "loss": 2.2254, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 1.1013824884792627, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.908177249885793e-06, | |
| "loss": 2.0497, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 1.1036866359447004, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.903639276722361e-06, | |
| "loss": 2.1932, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 1.1059907834101383, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.899082568807339e-06, | |
| "loss": 2.2727, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 1.108294930875576, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.894507009882784e-06, | |
| "loss": 2.0313, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 1.1105990783410138, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.889912482726854e-06, | |
| "loss": 2.0227, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 1.1129032258064515, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.88529886914378e-06, | |
| "loss": 2.1115, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 1.1152073732718895, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.880666049953746e-06, | |
| "loss": 2.3122, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 1.1175115207373272, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.87601390498262e-06, | |
| "loss": 2.1055, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 1.119815668202765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.871342313051557e-06, | |
| "loss": 2.2331, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 1.1221198156682028, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.86665115196649e-06, | |
| "loss": 2.1796, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 1.1244239631336406, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.861940298507463e-06, | |
| "loss": 2.304, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 1.1267281105990783, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.857209628417855e-06, | |
| "loss": 2.354, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 1.129032258064516, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.852459016393443e-06, | |
| "loss": 2.3015, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 1.131336405529954, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.847688336071346e-06, | |
| "loss": 2.2448, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 1.1336405529953917, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.842897460018816e-06, | |
| "loss": 2.3596, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 1.1359447004608294, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.838086259721895e-06, | |
| "loss": 1.9896, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 1.1382488479262673, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.833254605573925e-06, | |
| "loss": 2.1158, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 1.140552995391705, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.828402366863905e-06, | |
| "loss": 2.4531, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 1.1428571428571428, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.823529411764707e-06, | |
| "loss": 2.0021, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 1.1451612903225807, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.81863560732113e-06, | |
| "loss": 2.1852, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 1.1474654377880185, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.813720819437828e-06, | |
| "loss": 1.98, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 1.1497695852534562, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.808784912867032e-06, | |
| "loss": 1.9642, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 1.1520737327188941, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.803827751196173e-06, | |
| "loss": 2.2951, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 1.1543778801843319, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.798849196835292e-06, | |
| "loss": 2.087, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 1.1566820276497696, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.793849111004326e-06, | |
| "loss": 2.2937, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 1.1589861751152073, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.788827353720203e-06, | |
| "loss": 2.4084, | |
| "step": 503 | |
| }, | |
| { | |
| "epoch": 1.1612903225806452, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.783783783783785e-06, | |
| "loss": 2.0154, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 1.163594470046083, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.778718258766626e-06, | |
| "loss": 2.0608, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 1.1658986175115207, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.773630634997577e-06, | |
| "loss": 2.0001, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 1.1682027649769586, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.768520767549187e-06, | |
| "loss": 1.861, | |
| "step": 507 | |
| }, | |
| { | |
| "epoch": 1.1705069124423964, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.763388510223954e-06, | |
| "loss": 2.2712, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 1.172811059907834, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.758233715540376e-06, | |
| "loss": 2.2442, | |
| "step": 509 | |
| }, | |
| { | |
| "epoch": 1.1751152073732718, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.753056234718826e-06, | |
| "loss": 1.9768, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 1.1774193548387097, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.747855917667238e-06, | |
| "loss": 2.2363, | |
| "step": 511 | |
| }, | |
| { | |
| "epoch": 1.1797235023041475, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.742632612966601e-06, | |
| "loss": 1.926, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 1.1820276497695852, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.737386167856265e-06, | |
| "loss": 1.9019, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 1.1843317972350231, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.732116428219042e-06, | |
| "loss": 1.8069, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 1.1866359447004609, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.726823238566132e-06, | |
| "loss": 2.3622, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 1.1889400921658986, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.721506442021803e-06, | |
| "loss": 2.2355, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 1.1912442396313363, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.716165880307922e-06, | |
| "loss": 1.9082, | |
| "step": 517 | |
| }, | |
| { | |
| "epoch": 1.1935483870967742, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.710801393728223e-06, | |
| "loss": 2.0509, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 1.195852534562212, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.705412821152409e-06, | |
| "loss": 2.2892, | |
| "step": 519 | |
| }, | |
| { | |
| "epoch": 1.1981566820276497, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.700000000000001e-06, | |
| "loss": 2.4489, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 1.2004608294930876, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.694562766224006e-06, | |
| "loss": 2.1379, | |
| "step": 521 | |
| }, | |
| { | |
| "epoch": 1.2027649769585254, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.689100954294325e-06, | |
| "loss": 2.3232, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 1.205069124423963, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.683614397180973e-06, | |
| "loss": 2.2577, | |
| "step": 523 | |
| }, | |
| { | |
| "epoch": 1.2073732718894008, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.678102926337033e-06, | |
| "loss": 1.9628, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 1.2096774193548387, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.672566371681416e-06, | |
| "loss": 2.2783, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 1.2119815668202765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.667004561581348e-06, | |
| "loss": 2.0486, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 1.2142857142857142, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.661417322834646e-06, | |
| "loss": 2.3865, | |
| "step": 527 | |
| }, | |
| { | |
| "epoch": 1.2165898617511521, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.655804480651731e-06, | |
| "loss": 2.0669, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 1.2188940092165899, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.650165858637407e-06, | |
| "loss": 2.321, | |
| "step": 529 | |
| }, | |
| { | |
| "epoch": 1.2211981566820276, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.64450127877238e-06, | |
| "loss": 2.2825, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 1.2235023041474655, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.638810561394516e-06, | |
| "loss": 2.1408, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 1.2258064516129032, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.633093525179856e-06, | |
| "loss": 2.0595, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 1.228110599078341, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.62734998712336e-06, | |
| "loss": 2.4359, | |
| "step": 533 | |
| }, | |
| { | |
| "epoch": 1.230414746543779, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.62157976251936e-06, | |
| "loss": 2.0414, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 1.2327188940092166, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.615782664941786e-06, | |
| "loss": 2.1725, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 1.2350230414746544, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.609958506224068e-06, | |
| "loss": 2.0361, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 1.237327188940092, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.604107096438784e-06, | |
| "loss": 2.1823, | |
| "step": 537 | |
| }, | |
| { | |
| "epoch": 1.23963133640553, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.59822824387702e-06, | |
| "loss": 2.1377, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 1.2419354838709677, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.592321755027422e-06, | |
| "loss": 2.2211, | |
| "step": 539 | |
| }, | |
| { | |
| "epoch": 1.2442396313364055, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.586387434554974e-06, | |
| "loss": 2.2804, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 1.2465437788018434, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.580425085279455e-06, | |
| "loss": 2.0535, | |
| "step": 541 | |
| }, | |
| { | |
| "epoch": 1.2488479262672811, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.574434508153603e-06, | |
| "loss": 2.2902, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 1.2511520737327189, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.56841550224097e-06, | |
| "loss": 2.2136, | |
| "step": 543 | |
| }, | |
| { | |
| "epoch": 1.2534562211981566, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.562367864693448e-06, | |
| "loss": 2.1156, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 1.2557603686635945, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.556291390728475e-06, | |
| "loss": 1.8844, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 1.2580645161290323, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.55018587360595e-06, | |
| "loss": 2.2077, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 1.26036866359447, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.544051104604738e-06, | |
| "loss": 2.0569, | |
| "step": 547 | |
| }, | |
| { | |
| "epoch": 1.262672811059908, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.537886872998934e-06, | |
| "loss": 2.3563, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 1.2649769585253456, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.5316929660337e-06, | |
| "loss": 1.7421, | |
| "step": 549 | |
| }, | |
| { | |
| "epoch": 1.2672811059907834, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.525469168900806e-06, | |
| "loss": 2.539, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 1.269585253456221, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.519215264713787e-06, | |
| "loss": 2.1152, | |
| "step": 551 | |
| }, | |
| { | |
| "epoch": 1.271889400921659, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.512931034482758e-06, | |
| "loss": 2.1897, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 1.2741935483870968, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.506616257088848e-06, | |
| "loss": 2.3483, | |
| "step": 553 | |
| }, | |
| { | |
| "epoch": 1.2764976958525347, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.500270709258256e-06, | |
| "loss": 2.1675, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 1.2788018433179724, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.493894165535958e-06, | |
| "loss": 2.3407, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 1.2811059907834101, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.487486398258977e-06, | |
| "loss": 2.3402, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 1.2834101382488479, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.481047177529316e-06, | |
| "loss": 2.1818, | |
| "step": 557 | |
| }, | |
| { | |
| "epoch": 1.2857142857142856, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.47457627118644e-06, | |
| "loss": 2.0397, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 1.2880184331797235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.468073444779393e-06, | |
| "loss": 2.3515, | |
| "step": 559 | |
| }, | |
| { | |
| "epoch": 1.2903225806451613, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.46153846153846e-06, | |
| "loss": 2.1762, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.2926267281105992, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.454971082346463e-06, | |
| "loss": 2.0294, | |
| "step": 561 | |
| }, | |
| { | |
| "epoch": 1.294930875576037, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.448371065709553e-06, | |
| "loss": 2.1765, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 1.2972350230414746, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.441738167727651e-06, | |
| "loss": 2.2303, | |
| "step": 563 | |
| }, | |
| { | |
| "epoch": 1.2995391705069124, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.435072142064373e-06, | |
| "loss": 2.1646, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 1.3018433179723503, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.428372739916551e-06, | |
| "loss": 2.3313, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 1.304147465437788, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.42163970998327e-06, | |
| "loss": 2.1818, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 1.3064516129032258, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.414872798434444e-06, | |
| "loss": 1.9629, | |
| "step": 567 | |
| }, | |
| { | |
| "epoch": 1.3087557603686637, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.408071748878924e-06, | |
| "loss": 2.284, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 1.3110599078341014, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.401236302332116e-06, | |
| "loss": 2.4839, | |
| "step": 569 | |
| }, | |
| { | |
| "epoch": 1.3133640552995391, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.3943661971831e-06, | |
| "loss": 2.4827, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 1.3156682027649769, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.387461169161254e-06, | |
| "loss": 1.8538, | |
| "step": 571 | |
| }, | |
| { | |
| "epoch": 1.3179723502304148, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.38052095130238e-06, | |
| "loss": 2.1597, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 1.3202764976958525, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.373545273914277e-06, | |
| "loss": 2.2918, | |
| "step": 573 | |
| }, | |
| { | |
| "epoch": 1.3225806451612903, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.366533864541834e-06, | |
| "loss": 2.257, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 1.3248847926267282, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.359486447931526e-06, | |
| "loss": 2.2058, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 1.327188940092166, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.352402745995425e-06, | |
| "loss": 2.3933, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 1.3294930875576036, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.34528247777459e-06, | |
| "loss": 1.9175, | |
| "step": 577 | |
| }, | |
| { | |
| "epoch": 1.3317972350230414, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.338125359401957e-06, | |
| "loss": 2.0542, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 1.3341013824884793, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.330931104064572e-06, | |
| "loss": 2.3477, | |
| "step": 579 | |
| }, | |
| { | |
| "epoch": 1.336405529953917, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.32369942196532e-06, | |
| "loss": 1.8106, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 1.3387096774193548, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.316430020283975e-06, | |
| "loss": 2.2048, | |
| "step": 581 | |
| }, | |
| { | |
| "epoch": 1.3410138248847927, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.309122603137711e-06, | |
| "loss": 2.2727, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 1.3433179723502304, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.301776871540926e-06, | |
| "loss": 2.3041, | |
| "step": 583 | |
| }, | |
| { | |
| "epoch": 1.3456221198156681, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.294392523364487e-06, | |
| "loss": 2.2184, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 1.3479262672811059, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.28696925329429e-06, | |
| "loss": 2.1741, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 1.3502304147465438, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.279506752789194e-06, | |
| "loss": 2.3027, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 1.3525345622119815, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.272004710038269e-06, | |
| "loss": 1.8337, | |
| "step": 587 | |
| }, | |
| { | |
| "epoch": 1.3548387096774195, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.264462809917356e-06, | |
| "loss": 2.2419, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 1.3571428571428572, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.256880733944954e-06, | |
| "loss": 2.0455, | |
| "step": 589 | |
| }, | |
| { | |
| "epoch": 1.359447004608295, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.249258160237388e-06, | |
| "loss": 2.2288, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 1.3617511520737327, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.241594763463255e-06, | |
| "loss": 2.3845, | |
| "step": 591 | |
| }, | |
| { | |
| "epoch": 1.3640552995391704, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.233890214797137e-06, | |
| "loss": 1.8509, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 1.3663594470046083, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.226144181872571e-06, | |
| "loss": 2.3071, | |
| "step": 593 | |
| }, | |
| { | |
| "epoch": 1.368663594470046, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.218356328734255e-06, | |
| "loss": 2.3412, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 1.370967741935484, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.210526315789475e-06, | |
| "loss": 2.3244, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 1.3732718894009217, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.202653799758746e-06, | |
| "loss": 2.2093, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 1.3755760368663594, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.194738433625643e-06, | |
| "loss": 2.245, | |
| "step": 597 | |
| }, | |
| { | |
| "epoch": 1.3778801843317972, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.18677986658581e-06, | |
| "loss": 1.8521, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 1.380184331797235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.178777743995136e-06, | |
| "loss": 2.327, | |
| "step": 599 | |
| }, | |
| { | |
| "epoch": 1.3824884792626728, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.170731707317073e-06, | |
| "loss": 1.9856, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.3847926267281105, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.162641394069091e-06, | |
| "loss": 1.8187, | |
| "step": 601 | |
| }, | |
| { | |
| "epoch": 1.3870967741935485, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.154506437768241e-06, | |
| "loss": 1.9824, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 1.3894009216589862, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.146326467875806e-06, | |
| "loss": 1.7706, | |
| "step": 603 | |
| }, | |
| { | |
| "epoch": 1.391705069124424, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.13810110974106e-06, | |
| "loss": 2.423, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 1.3940092165898617, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.129829984544051e-06, | |
| "loss": 2.2774, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 1.3963133640552996, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.121512709237446e-06, | |
| "loss": 2.2349, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 1.3986175115207373, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.113148896487411e-06, | |
| "loss": 2.1761, | |
| "step": 607 | |
| }, | |
| { | |
| "epoch": 1.400921658986175, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.104738154613467e-06, | |
| "loss": 1.6387, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 1.403225806451613, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.096280087527354e-06, | |
| "loss": 2.0971, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 1.4055299539170507, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.087774294670848e-06, | |
| "loss": 2.3251, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 1.4078341013824884, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.079220370952531e-06, | |
| "loss": 1.9343, | |
| "step": 611 | |
| }, | |
| { | |
| "epoch": 1.4101382488479262, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.070617906683481e-06, | |
| "loss": 2.2662, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 1.412442396313364, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.061966487511855e-06, | |
| "loss": 2.2092, | |
| "step": 613 | |
| }, | |
| { | |
| "epoch": 1.4147465437788018, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.053265694356374e-06, | |
| "loss": 2.0375, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 1.4170506912442398, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.044515103338633e-06, | |
| "loss": 2.3473, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 1.4193548387096775, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.035714285714286e-06, | |
| "loss": 2.5178, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 1.4216589861751152, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.026862807803005e-06, | |
| "loss": 2.1105, | |
| "step": 617 | |
| }, | |
| { | |
| "epoch": 1.423963133640553, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.017960230917255e-06, | |
| "loss": 1.9589, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 1.4262672811059907, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.009006111289803e-06, | |
| "loss": 2.2071, | |
| "step": 619 | |
| }, | |
| { | |
| "epoch": 1.4285714285714286, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.000000000000001e-06, | |
| "loss": 2.3059, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 1.4308755760368663, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.990941442898738e-06, | |
| "loss": 2.3726, | |
| "step": 621 | |
| }, | |
| { | |
| "epoch": 1.4331797235023043, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.981829980532122e-06, | |
| "loss": 2.0598, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 1.435483870967742, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.97266514806378e-06, | |
| "loss": 2.1011, | |
| "step": 623 | |
| }, | |
| { | |
| "epoch": 1.4377880184331797, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.963446475195824e-06, | |
| "loss": 2.0975, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 1.4400921658986174, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.95417348608838e-06, | |
| "loss": 2.1737, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 1.4423963133640554, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.944845699277742e-06, | |
| "loss": 2.1402, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 1.444700460829493, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.935462627593021e-06, | |
| "loss": 2.194, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 1.4470046082949308, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.926023778071336e-06, | |
| "loss": 2.0133, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 1.4493087557603688, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.916528651871482e-06, | |
| "loss": 2.3248, | |
| "step": 629 | |
| }, | |
| { | |
| "epoch": 1.4516129032258065, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.906976744186046e-06, | |
| "loss": 2.114, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 1.4539170506912442, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.89736754415195e-06, | |
| "loss": 2.2423, | |
| "step": 631 | |
| }, | |
| { | |
| "epoch": 1.456221198156682, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.887700534759358e-06, | |
| "loss": 2.4931, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 1.4585253456221199, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.877975192758967e-06, | |
| "loss": 2.1069, | |
| "step": 633 | |
| }, | |
| { | |
| "epoch": 1.4608294930875576, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.868190988567587e-06, | |
| "loss": 2.4308, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 1.4631336405529953, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.858347386172008e-06, | |
| "loss": 2.1255, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 1.4654377880184333, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.848443843031123e-06, | |
| "loss": 2.1629, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 1.467741935483871, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.838479809976249e-06, | |
| "loss": 2.1334, | |
| "step": 637 | |
| }, | |
| { | |
| "epoch": 1.4700460829493087, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.8284547311096e-06, | |
| "loss": 2.2989, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 1.4723502304147464, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.818368043700922e-06, | |
| "loss": 2.0108, | |
| "step": 639 | |
| }, | |
| { | |
| "epoch": 1.4746543778801844, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.808219178082192e-06, | |
| "loss": 1.9985, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 1.476958525345622, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.798007557540365e-06, | |
| "loss": 2.3147, | |
| "step": 641 | |
| }, | |
| { | |
| "epoch": 1.4792626728110598, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.787732598208133e-06, | |
| "loss": 2.327, | |
| "step": 642 | |
| }, | |
| { | |
| "epoch": 1.4815668202764978, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.777393708952645e-06, | |
| "loss": 2.0685, | |
| "step": 643 | |
| }, | |
| { | |
| "epoch": 1.4838709677419355, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.766990291262136e-06, | |
| "loss": 1.9718, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 1.4861751152073732, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.756521739130437e-06, | |
| "loss": 2.3679, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 1.488479262672811, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.745987438939288e-06, | |
| "loss": 1.9236, | |
| "step": 646 | |
| }, | |
| { | |
| "epoch": 1.4907834101382489, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.735386769338468e-06, | |
| "loss": 2.2692, | |
| "step": 647 | |
| }, | |
| { | |
| "epoch": 1.4930875576036866, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.724719101123595e-06, | |
| "loss": 2.076, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 1.4953917050691246, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.713983797111659e-06, | |
| "loss": 2.1628, | |
| "step": 649 | |
| }, | |
| { | |
| "epoch": 1.4976958525345623, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.703180212014134e-06, | |
| "loss": 2.4163, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.692307692307692e-06, | |
| "loss": 2.0675, | |
| "step": 651 | |
| }, | |
| { | |
| "epoch": 1.5023041474654377, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.681365576102418e-06, | |
| "loss": 2.2859, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 1.5046082949308754, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.670353193007492e-06, | |
| "loss": 2.2192, | |
| "step": 653 | |
| }, | |
| { | |
| "epoch": 1.5069124423963134, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.659269863994274e-06, | |
| "loss": 2.104, | |
| "step": 654 | |
| }, | |
| { | |
| "epoch": 1.5092165898617511, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.648114901256734e-06, | |
| "loss": 2.3686, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 1.511520737327189, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.636887608069165e-06, | |
| "loss": 2.2697, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 1.5138248847926268, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.625587278641128e-06, | |
| "loss": 2.4146, | |
| "step": 657 | |
| }, | |
| { | |
| "epoch": 1.5161290322580645, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.614213197969545e-06, | |
| "loss": 2.4442, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 1.5184331797235022, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.6027646416878876e-06, | |
| "loss": 2.3551, | |
| "step": 659 | |
| }, | |
| { | |
| "epoch": 1.52073732718894, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.591240875912409e-06, | |
| "loss": 2.5825, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 1.523041474654378, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.579641157085317e-06, | |
| "loss": 2.2146, | |
| "step": 661 | |
| }, | |
| { | |
| "epoch": 1.5253456221198156, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.567964731814842e-06, | |
| "loss": 2.3835, | |
| "step": 662 | |
| }, | |
| { | |
| "epoch": 1.5276497695852536, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.5562108367121265e-06, | |
| "loss": 1.9557, | |
| "step": 663 | |
| }, | |
| { | |
| "epoch": 1.5299539170506913, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.544378698224851e-06, | |
| "loss": 2.2251, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 1.532258064516129, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.532467532467532e-06, | |
| "loss": 2.2089, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 1.5345622119815667, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.5204765450484e-06, | |
| "loss": 2.3537, | |
| "step": 666 | |
| }, | |
| { | |
| "epoch": 1.5368663594470044, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.508404930892791e-06, | |
| "loss": 2.2675, | |
| "step": 667 | |
| }, | |
| { | |
| "epoch": 1.5391705069124424, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.496251874062969e-06, | |
| "loss": 2.2828, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 1.5414746543778803, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.484016547574278e-06, | |
| "loss": 2.5155, | |
| "step": 669 | |
| }, | |
| { | |
| "epoch": 1.543778801843318, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.471698113207549e-06, | |
| "loss": 2.3364, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 1.5460829493087558, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.459295721317683e-06, | |
| "loss": 1.7703, | |
| "step": 671 | |
| }, | |
| { | |
| "epoch": 1.5483870967741935, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.446808510638298e-06, | |
| "loss": 2.331, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 1.5506912442396312, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.434235608082349e-06, | |
| "loss": 2.2976, | |
| "step": 673 | |
| }, | |
| { | |
| "epoch": 1.5529953917050692, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.421576128538639e-06, | |
| "loss": 2.2165, | |
| "step": 674 | |
| }, | |
| { | |
| "epoch": 1.555299539170507, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.408829174664108e-06, | |
| "loss": 2.3928, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 1.5576036866359448, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.395993836671803e-06, | |
| "loss": 2.3778, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 1.5599078341013826, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.383069192114418e-06, | |
| "loss": 2.5435, | |
| "step": 677 | |
| }, | |
| { | |
| "epoch": 1.5622119815668203, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.370054305663304e-06, | |
| "loss": 2.3359, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 1.564516129032258, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.3569482288828335e-06, | |
| "loss": 2.1116, | |
| "step": 679 | |
| }, | |
| { | |
| "epoch": 1.5668202764976957, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.34375e-06, | |
| "loss": 2.2972, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.5691244239631337, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.3304586436691505e-06, | |
| "loss": 2.0481, | |
| "step": 681 | |
| }, | |
| { | |
| "epoch": 1.5714285714285714, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.317073170731709e-06, | |
| "loss": 2.0658, | |
| "step": 682 | |
| }, | |
| { | |
| "epoch": 1.5737327188940093, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.303592577970787e-06, | |
| "loss": 1.8831, | |
| "step": 683 | |
| }, | |
| { | |
| "epoch": 1.576036866359447, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.29001584786054e-06, | |
| "loss": 2.4854, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 1.5783410138248848, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.2763419483101405e-06, | |
| "loss": 2.5698, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 1.5806451612903225, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.262569832402235e-06, | |
| "loss": 1.8958, | |
| "step": 686 | |
| }, | |
| { | |
| "epoch": 1.5829493087557602, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.2486984381257516e-06, | |
| "loss": 2.0024, | |
| "step": 687 | |
| }, | |
| { | |
| "epoch": 1.5852534562211982, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.234726688102894e-06, | |
| "loss": 2.306, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 1.587557603686636, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.220653489310205e-06, | |
| "loss": 2.3266, | |
| "step": 689 | |
| }, | |
| { | |
| "epoch": 1.5898617511520738, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.206477732793522e-06, | |
| "loss": 2.123, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 1.5921658986175116, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.192198293376676e-06, | |
| "loss": 2.0752, | |
| "step": 691 | |
| }, | |
| { | |
| "epoch": 1.5944700460829493, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.177814029363785e-06, | |
| "loss": 2.1435, | |
| "step": 692 | |
| }, | |
| { | |
| "epoch": 1.596774193548387, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.1633237822349565e-06, | |
| "loss": 2.1723, | |
| "step": 693 | |
| }, | |
| { | |
| "epoch": 1.5990783410138247, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.148726376335251e-06, | |
| "loss": 2.1988, | |
| "step": 694 | |
| }, | |
| { | |
| "epoch": 1.6013824884792627, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.134020618556702e-06, | |
| "loss": 1.8426, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 1.6036866359447006, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.119205298013245e-06, | |
| "loss": 2.2582, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 1.6059907834101383, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.104279185708352e-06, | |
| "loss": 2.3248, | |
| "step": 697 | |
| }, | |
| { | |
| "epoch": 1.608294930875576, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.089241034195163e-06, | |
| "loss": 2.0316, | |
| "step": 698 | |
| }, | |
| { | |
| "epoch": 1.6105990783410138, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.074089577228968e-06, | |
| "loss": 2.4332, | |
| "step": 699 | |
| }, | |
| { | |
| "epoch": 1.6129032258064515, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.058823529411766e-06, | |
| "loss": 2.1161, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.6152073732718892, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.0434415858287656e-06, | |
| "loss": 2.2994, | |
| "step": 701 | |
| }, | |
| { | |
| "epoch": 1.6175115207373272, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.027942421676546e-06, | |
| "loss": 2.5017, | |
| "step": 702 | |
| }, | |
| { | |
| "epoch": 1.6198156682027651, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.012324691882704e-06, | |
| "loss": 2.23, | |
| "step": 703 | |
| }, | |
| { | |
| "epoch": 1.6221198156682028, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.9965870307167235e-06, | |
| "loss": 1.8913, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 1.6244239631336406, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.980728051391863e-06, | |
| "loss": 1.853, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 1.6267281105990783, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.9647463456577815e-06, | |
| "loss": 2.2723, | |
| "step": 706 | |
| }, | |
| { | |
| "epoch": 1.629032258064516, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.948640483383685e-06, | |
| "loss": 2.071, | |
| "step": 707 | |
| }, | |
| { | |
| "epoch": 1.631336405529954, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.932409012131715e-06, | |
| "loss": 2.2638, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 1.6336405529953917, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.916050456720314e-06, | |
| "loss": 2.1189, | |
| "step": 709 | |
| }, | |
| { | |
| "epoch": 1.6359447004608296, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.899563318777293e-06, | |
| "loss": 2.348, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 1.6382488479262673, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.882946076282333e-06, | |
| "loss": 2.4884, | |
| "step": 711 | |
| }, | |
| { | |
| "epoch": 1.640552995391705, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.866197183098592e-06, | |
| "loss": 2.2421, | |
| "step": 712 | |
| }, | |
| { | |
| "epoch": 1.6428571428571428, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.849315068493151e-06, | |
| "loss": 2.0634, | |
| "step": 713 | |
| }, | |
| { | |
| "epoch": 1.6451612903225805, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.832298136645963e-06, | |
| "loss": 2.3555, | |
| "step": 714 | |
| }, | |
| { | |
| "epoch": 1.6474654377880185, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.815144766146993e-06, | |
| "loss": 2.0054, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 1.6497695852534562, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.797853309481217e-06, | |
| "loss": 2.0496, | |
| "step": 716 | |
| }, | |
| { | |
| "epoch": 1.6520737327188941, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.780422092501123e-06, | |
| "loss": 2.3211, | |
| "step": 717 | |
| }, | |
| { | |
| "epoch": 1.6543778801843319, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.762849413886385e-06, | |
| "loss": 1.9154, | |
| "step": 718 | |
| }, | |
| { | |
| "epoch": 1.6566820276497696, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.745133544590313e-06, | |
| "loss": 2.1494, | |
| "step": 719 | |
| }, | |
| { | |
| "epoch": 1.6589861751152073, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.7272727272727275e-06, | |
| "loss": 2.1532, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 1.661290322580645, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.70926517571885e-06, | |
| "loss": 1.9703, | |
| "step": 721 | |
| }, | |
| { | |
| "epoch": 1.663594470046083, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.691109074243812e-06, | |
| "loss": 2.4101, | |
| "step": 722 | |
| }, | |
| { | |
| "epoch": 1.6658986175115207, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.672802577082374e-06, | |
| "loss": 2.3254, | |
| "step": 723 | |
| }, | |
| { | |
| "epoch": 1.6682027649769586, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.654343807763401e-06, | |
| "loss": 2.2264, | |
| "step": 724 | |
| }, | |
| { | |
| "epoch": 1.6705069124423964, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.6357308584686775e-06, | |
| "loss": 2.2232, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 1.672811059907834, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.616961789375583e-06, | |
| "loss": 2.228, | |
| "step": 726 | |
| }, | |
| { | |
| "epoch": 1.6751152073732718, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.598034627983154e-06, | |
| "loss": 2.5189, | |
| "step": 727 | |
| }, | |
| { | |
| "epoch": 1.6774193548387095, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.578947368421052e-06, | |
| "loss": 2.0719, | |
| "step": 728 | |
| }, | |
| { | |
| "epoch": 1.6797235023041475, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.559697970740916e-06, | |
| "loss": 1.9752, | |
| "step": 729 | |
| }, | |
| { | |
| "epoch": 1.6820276497695854, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.5402843601895735e-06, | |
| "loss": 2.1016, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 1.6843317972350231, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.52070442646359e-06, | |
| "loss": 2.1034, | |
| "step": 731 | |
| }, | |
| { | |
| "epoch": 1.6866359447004609, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.500956022944551e-06, | |
| "loss": 2.2459, | |
| "step": 732 | |
| }, | |
| { | |
| "epoch": 1.6889400921658986, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.481036965914547e-06, | |
| "loss": 2.0818, | |
| "step": 733 | |
| }, | |
| { | |
| "epoch": 1.6912442396313363, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.460945033751206e-06, | |
| "loss": 2.4337, | |
| "step": 734 | |
| }, | |
| { | |
| "epoch": 1.6935483870967742, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.440677966101695e-06, | |
| "loss": 2.3903, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 1.695852534562212, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.42023346303502e-06, | |
| "loss": 2.1265, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 1.69815668202765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.3996091841719595e-06, | |
| "loss": 2.1477, | |
| "step": 737 | |
| }, | |
| { | |
| "epoch": 1.7004608294930876, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.3788027477919535e-06, | |
| "loss": 2.4987, | |
| "step": 738 | |
| }, | |
| { | |
| "epoch": 1.7027649769585254, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.357811729916214e-06, | |
| "loss": 2.423, | |
| "step": 739 | |
| }, | |
| { | |
| "epoch": 1.705069124423963, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.336633663366336e-06, | |
| "loss": 2.3027, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 1.7073732718894008, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.315266036797612e-06, | |
| "loss": 2.4029, | |
| "step": 741 | |
| }, | |
| { | |
| "epoch": 1.7096774193548387, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.2937062937062944e-06, | |
| "loss": 2.0332, | |
| "step": 742 | |
| }, | |
| { | |
| "epoch": 1.7119815668202765, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.271951831409935e-06, | |
| "loss": 2.2298, | |
| "step": 743 | |
| }, | |
| { | |
| "epoch": 1.7142857142857144, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.25e-06, | |
| "loss": 2.1978, | |
| "step": 744 | |
| }, | |
| { | |
| "epoch": 1.7165898617511521, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.227848101265823e-06, | |
| "loss": 2.148, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 1.7188940092165899, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.205493387589013e-06, | |
| "loss": 1.7415, | |
| "step": 746 | |
| }, | |
| { | |
| "epoch": 1.7211981566820276, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.182933060807359e-06, | |
| "loss": 2.1957, | |
| "step": 747 | |
| }, | |
| { | |
| "epoch": 1.7235023041474653, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.1601642710472284e-06, | |
| "loss": 2.2568, | |
| "step": 748 | |
| }, | |
| { | |
| "epoch": 1.7258064516129032, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.137184115523466e-06, | |
| "loss": 2.2825, | |
| "step": 749 | |
| }, | |
| { | |
| "epoch": 1.728110599078341, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.113989637305699e-06, | |
| "loss": 2.5317, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 1.730414746543779, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.090577824049974e-06, | |
| "loss": 2.4529, | |
| "step": 751 | |
| }, | |
| { | |
| "epoch": 1.7327188940092166, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.066945606694561e-06, | |
| "loss": 2.4748, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 1.7350230414746544, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.04308985811876e-06, | |
| "loss": 2.3451, | |
| "step": 753 | |
| }, | |
| { | |
| "epoch": 1.737327188940092, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.0190073917634635e-06, | |
| "loss": 2.0164, | |
| "step": 754 | |
| }, | |
| { | |
| "epoch": 1.7396313364055298, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.994694960212202e-06, | |
| "loss": 1.9073, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 1.7419354838709677, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.970149253731343e-06, | |
| "loss": 2.2839, | |
| "step": 756 | |
| }, | |
| { | |
| "epoch": 1.7442396313364057, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.945366898768078e-06, | |
| "loss": 2.1681, | |
| "step": 757 | |
| }, | |
| { | |
| "epoch": 1.7465437788018434, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.920344456404736e-06, | |
| "loss": 2.3917, | |
| "step": 758 | |
| }, | |
| { | |
| "epoch": 1.7488479262672811, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.895078420767983e-06, | |
| "loss": 1.9239, | |
| "step": 759 | |
| }, | |
| { | |
| "epoch": 1.7511520737327189, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.869565217391304e-06, | |
| "loss": 2.3452, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 1.7534562211981566, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.843801201529219e-06, | |
| "loss": 2.2105, | |
| "step": 761 | |
| }, | |
| { | |
| "epoch": 1.7557603686635943, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.817782656421514e-06, | |
| "loss": 2.2953, | |
| "step": 762 | |
| }, | |
| { | |
| "epoch": 1.7580645161290323, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.791505791505793e-06, | |
| "loss": 2.3598, | |
| "step": 763 | |
| }, | |
| { | |
| "epoch": 1.7603686635944702, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.7649667405764965e-06, | |
| "loss": 2.2871, | |
| "step": 764 | |
| }, | |
| { | |
| "epoch": 1.762672811059908, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.7381615598885795e-06, | |
| "loss": 2.3103, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 1.7649769585253456, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.711086226203808e-06, | |
| "loss": 2.3354, | |
| "step": 766 | |
| }, | |
| { | |
| "epoch": 1.7672811059907834, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.683736634777714e-06, | |
| "loss": 2.0412, | |
| "step": 767 | |
| }, | |
| { | |
| "epoch": 1.769585253456221, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.656108597285067e-06, | |
| "loss": 2.3042, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 1.771889400921659, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.628197839681638e-06, | |
| "loss": 2.301, | |
| "step": 769 | |
| }, | |
| { | |
| "epoch": 1.7741935483870968, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.600000000000001e-06, | |
| "loss": 2.3394, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 1.7764976958525347, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.571510626076968e-06, | |
| "loss": 1.991, | |
| "step": 771 | |
| }, | |
| { | |
| "epoch": 1.7788018433179724, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.5427251732101625e-06, | |
| "loss": 2.0793, | |
| "step": 772 | |
| }, | |
| { | |
| "epoch": 1.7811059907834101, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.513639001741149e-06, | |
| "loss": 1.9698, | |
| "step": 773 | |
| }, | |
| { | |
| "epoch": 1.7834101382488479, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.4842473745624265e-06, | |
| "loss": 2.305, | |
| "step": 774 | |
| }, | |
| { | |
| "epoch": 1.7857142857142856, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.4545454545454545e-06, | |
| "loss": 2.5145, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 1.7880184331797235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.424528301886792e-06, | |
| "loss": 2.2537, | |
| "step": 776 | |
| }, | |
| { | |
| "epoch": 1.7903225806451613, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.394190871369295e-06, | |
| "loss": 2.1169, | |
| "step": 777 | |
| }, | |
| { | |
| "epoch": 1.7926267281105992, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.363528009535162e-06, | |
| "loss": 2.1254, | |
| "step": 778 | |
| }, | |
| { | |
| "epoch": 1.794930875576037, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.332534451767526e-06, | |
| "loss": 1.9744, | |
| "step": 779 | |
| }, | |
| { | |
| "epoch": 1.7972350230414746, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.301204819277109e-06, | |
| "loss": 1.861, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 1.7995391705069124, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.269533615990309e-06, | |
| "loss": 2.1956, | |
| "step": 781 | |
| }, | |
| { | |
| "epoch": 1.80184331797235, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.237515225334957e-06, | |
| "loss": 2.0532, | |
| "step": 782 | |
| }, | |
| { | |
| "epoch": 1.804147465437788, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.20514390691978e-06, | |
| "loss": 2.402, | |
| "step": 783 | |
| }, | |
| { | |
| "epoch": 1.8064516129032258, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.172413793103448e-06, | |
| "loss": 2.6081, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 1.8087557603686637, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.139318885448916e-06, | |
| "loss": 2.4214, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 1.8110599078341014, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.1058530510585306e-06, | |
| "loss": 1.9568, | |
| "step": 786 | |
| }, | |
| { | |
| "epoch": 1.8133640552995391, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.072010018785223e-06, | |
| "loss": 2.3105, | |
| "step": 787 | |
| }, | |
| { | |
| "epoch": 1.8156682027649769, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.037783375314862e-06, | |
| "loss": 2.113, | |
| "step": 788 | |
| }, | |
| { | |
| "epoch": 1.8179723502304146, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.00316656111463e-06, | |
| "loss": 2.2062, | |
| "step": 789 | |
| }, | |
| { | |
| "epoch": 1.8202764976958525, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.968152866242038e-06, | |
| "loss": 2.2599, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 1.8225806451612905, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.932735426008968e-06, | |
| "loss": 2.1384, | |
| "step": 791 | |
| }, | |
| { | |
| "epoch": 1.8248847926267282, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.896907216494845e-06, | |
| "loss": 1.8991, | |
| "step": 792 | |
| }, | |
| { | |
| "epoch": 1.827188940092166, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.8606610499027876e-06, | |
| "loss": 2.3475, | |
| "step": 793 | |
| }, | |
| { | |
| "epoch": 1.8294930875576036, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.823989569752283e-06, | |
| "loss": 1.9257, | |
| "step": 794 | |
| }, | |
| { | |
| "epoch": 1.8317972350230414, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.78688524590164e-06, | |
| "loss": 1.7366, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 1.8341013824884793, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.74934036939314e-06, | |
| "loss": 2.0084, | |
| "step": 796 | |
| }, | |
| { | |
| "epoch": 1.836405529953917, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.711347047113471e-06, | |
| "loss": 2.2387, | |
| "step": 797 | |
| }, | |
| { | |
| "epoch": 1.838709677419355, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.672897196261682e-06, | |
| "loss": 2.065, | |
| "step": 798 | |
| }, | |
| { | |
| "epoch": 1.8410138248847927, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.633982538616521e-06, | |
| "loss": 2.0072, | |
| "step": 799 | |
| }, | |
| { | |
| "epoch": 1.8433179723502304, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.594594594594596e-06, | |
| "loss": 2.233, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 1.8456221198156681, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.554724677090415e-06, | |
| "loss": 2.2987, | |
| "step": 801 | |
| }, | |
| { | |
| "epoch": 1.8479262672811059, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.5143638850889195e-06, | |
| "loss": 2.0798, | |
| "step": 802 | |
| }, | |
| { | |
| "epoch": 1.8502304147465438, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.473503097040606e-06, | |
| "loss": 2.1042, | |
| "step": 803 | |
| }, | |
| { | |
| "epoch": 1.8525345622119815, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.43213296398892e-06, | |
| "loss": 1.9728, | |
| "step": 804 | |
| }, | |
| { | |
| "epoch": 1.8548387096774195, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.390243902439025e-06, | |
| "loss": 2.1751, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 1.8571428571428572, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.347826086956522e-06, | |
| "loss": 2.2449, | |
| "step": 806 | |
| }, | |
| { | |
| "epoch": 1.859447004608295, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.304869442484122e-06, | |
| "loss": 2.041, | |
| "step": 807 | |
| }, | |
| { | |
| "epoch": 1.8617511520737327, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.2613636363636365e-06, | |
| "loss": 1.8616, | |
| "step": 808 | |
| }, | |
| { | |
| "epoch": 1.8640552995391704, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.217298070050036e-06, | |
| "loss": 2.4581, | |
| "step": 809 | |
| }, | |
| { | |
| "epoch": 1.8663594470046083, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.172661870503597e-06, | |
| "loss": 2.0801, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 1.868663594470046, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.127443881245474e-06, | |
| "loss": 1.9342, | |
| "step": 811 | |
| }, | |
| { | |
| "epoch": 1.870967741935484, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.081632653061224e-06, | |
| "loss": 1.948, | |
| "step": 812 | |
| }, | |
| { | |
| "epoch": 1.8732718894009217, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.035216434336023e-06, | |
| "loss": 2.3085, | |
| "step": 813 | |
| }, | |
| { | |
| "epoch": 1.8755760368663594, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.988183161004432e-06, | |
| "loss": 1.9853, | |
| "step": 814 | |
| }, | |
| { | |
| "epoch": 1.8778801843317972, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.9405204460966546e-06, | |
| "loss": 1.929, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 1.8801843317972349, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.892215568862276e-06, | |
| "loss": 2.3427, | |
| "step": 816 | |
| }, | |
| { | |
| "epoch": 1.8824884792626728, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.843255463451395e-06, | |
| "loss": 2.5295, | |
| "step": 817 | |
| }, | |
| { | |
| "epoch": 1.8847926267281108, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.793626707132018e-06, | |
| "loss": 2.1848, | |
| "step": 818 | |
| }, | |
| { | |
| "epoch": 1.8870967741935485, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.7433155080213903e-06, | |
| "loss": 1.9892, | |
| "step": 819 | |
| }, | |
| { | |
| "epoch": 1.8894009216589862, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.692307692307692e-06, | |
| "loss": 1.9724, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 1.891705069124424, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.640588690937258e-06, | |
| "loss": 2.2537, | |
| "step": 821 | |
| }, | |
| { | |
| "epoch": 1.8940092165898617, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.5881435257410295e-06, | |
| "loss": 2.278, | |
| "step": 822 | |
| }, | |
| { | |
| "epoch": 1.8963133640552994, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.5349567949725054e-06, | |
| "loss": 2.2778, | |
| "step": 823 | |
| }, | |
| { | |
| "epoch": 1.8986175115207373, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.4810126582278482e-06, | |
| "loss": 2.572, | |
| "step": 824 | |
| }, | |
| { | |
| "epoch": 1.9009216589861753, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.426294820717132e-06, | |
| "loss": 2.3863, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 1.903225806451613, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.3707865168539327e-06, | |
| "loss": 2.0672, | |
| "step": 826 | |
| }, | |
| { | |
| "epoch": 1.9055299539170507, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.3144704931285366e-06, | |
| "loss": 2.2825, | |
| "step": 827 | |
| }, | |
| { | |
| "epoch": 1.9078341013824884, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.2573289902280133e-06, | |
| "loss": 1.962, | |
| "step": 828 | |
| }, | |
| { | |
| "epoch": 1.9101382488479262, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.1993437243642333e-06, | |
| "loss": 2.27, | |
| "step": 829 | |
| }, | |
| { | |
| "epoch": 1.912442396313364, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.140495867768595e-06, | |
| "loss": 1.9069, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 1.9147465437788018, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.080766028309742e-06, | |
| "loss": 1.9966, | |
| "step": 831 | |
| }, | |
| { | |
| "epoch": 1.9170506912442398, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.02013422818792e-06, | |
| "loss": 2.3821, | |
| "step": 832 | |
| }, | |
| { | |
| "epoch": 1.9193548387096775, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.958579881656805e-06, | |
| "loss": 2.5374, | |
| "step": 833 | |
| }, | |
| { | |
| "epoch": 1.9216589861751152, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.896081771720613e-06, | |
| "loss": 2.5009, | |
| "step": 834 | |
| }, | |
| { | |
| "epoch": 1.923963133640553, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.8326180257510733e-06, | |
| "loss": 2.4116, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 1.9262672811059907, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.768166089965398e-06, | |
| "loss": 2.0421, | |
| "step": 836 | |
| }, | |
| { | |
| "epoch": 1.9285714285714286, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.7027027027027026e-06, | |
| "loss": 2.5362, | |
| "step": 837 | |
| }, | |
| { | |
| "epoch": 1.9308755760368663, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.6362038664323376e-06, | |
| "loss": 2.173, | |
| "step": 838 | |
| }, | |
| { | |
| "epoch": 1.9331797235023043, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.5686448184233835e-06, | |
| "loss": 2.3125, | |
| "step": 839 | |
| }, | |
| { | |
| "epoch": 1.935483870967742, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.5e-06, | |
| "loss": 2.2148, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 1.9377880184331797, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.4302430243024304e-06, | |
| "loss": 2.2172, | |
| "step": 841 | |
| }, | |
| { | |
| "epoch": 1.9400921658986174, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.3593466424682394e-06, | |
| "loss": 2.1904, | |
| "step": 842 | |
| }, | |
| { | |
| "epoch": 1.9423963133640552, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.2872827081427264e-06, | |
| "loss": 2.1084, | |
| "step": 843 | |
| }, | |
| { | |
| "epoch": 1.944700460829493, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.2140221402214023e-06, | |
| "loss": 2.2455, | |
| "step": 844 | |
| }, | |
| { | |
| "epoch": 1.9470046082949308, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.1395348837209305e-06, | |
| "loss": 2.2483, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 1.9493087557603688, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.0637898686679174e-06, | |
| "loss": 2.305, | |
| "step": 846 | |
| }, | |
| { | |
| "epoch": 1.9516129032258065, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.9867549668874175e-06, | |
| "loss": 2.3151, | |
| "step": 847 | |
| }, | |
| { | |
| "epoch": 1.9539170506912442, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.908396946564886e-06, | |
| "loss": 2.1444, | |
| "step": 848 | |
| }, | |
| { | |
| "epoch": 1.956221198156682, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.828681424446583e-06, | |
| "loss": 2.3103, | |
| "step": 849 | |
| }, | |
| { | |
| "epoch": 1.9585253456221197, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.7475728155339805e-06, | |
| "loss": 2.2019, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 1.9608294930875576, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.6650342801175317e-06, | |
| "loss": 2.297, | |
| "step": 851 | |
| }, | |
| { | |
| "epoch": 1.9631336405529956, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.5810276679841899e-06, | |
| "loss": 2.261, | |
| "step": 852 | |
| }, | |
| { | |
| "epoch": 1.9654377880184333, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.4955134596211368e-06, | |
| "loss": 2.0714, | |
| "step": 853 | |
| }, | |
| { | |
| "epoch": 1.967741935483871, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.4084507042253523e-06, | |
| "loss": 2.0849, | |
| "step": 854 | |
| }, | |
| { | |
| "epoch": 1.9700460829493087, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.3197969543147206e-06, | |
| "loss": 1.8001, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 1.9723502304147464, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.2295081967213118e-06, | |
| "loss": 1.8648, | |
| "step": 856 | |
| }, | |
| { | |
| "epoch": 1.9746543778801844, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.1375387797311274e-06, | |
| "loss": 2.1359, | |
| "step": 857 | |
| }, | |
| { | |
| "epoch": 1.976958525345622, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.0438413361169103e-06, | |
| "loss": 2.278, | |
| "step": 858 | |
| }, | |
| { | |
| "epoch": 1.97926267281106, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.483667017913594e-07, | |
| "loss": 2.2074, | |
| "step": 859 | |
| }, | |
| { | |
| "epoch": 1.9815668202764978, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.510638297872341e-07, | |
| "loss": 2.1091, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 1.9838709677419355, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.518796992481203e-07, | |
| "loss": 2.358, | |
| "step": 861 | |
| }, | |
| { | |
| "epoch": 1.9861751152073732, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.507592190889371e-07, | |
| "loss": 1.9049, | |
| "step": 862 | |
| }, | |
| { | |
| "epoch": 1.988479262672811, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.47645125958379e-07, | |
| "loss": 2.2344, | |
| "step": 863 | |
| }, | |
| { | |
| "epoch": 1.9907834101382489, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.4247787610619474e-07, | |
| "loss": 2.163, | |
| "step": 864 | |
| }, | |
| { | |
| "epoch": 1.9930875576036866, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.3519553072625703e-07, | |
| "loss": 2.1836, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 1.9953917050691246, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.257336343115124e-07, | |
| "loss": 2.2758, | |
| "step": 866 | |
| }, | |
| { | |
| "epoch": 1.9976958525345623, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.1402508551881413e-07, | |
| "loss": 2.1508, | |
| "step": 867 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "grad_norm": 0.0, | |
| "learning_rate": 0, | |
| "loss": 2.1981, | |
| "step": 868 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 868, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 2, | |
| "save_steps": 217, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 2.1905444258583675e+18, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |