| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 3.0, |
| "global_step": 795, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0, |
| "learning_rate": 6.25e-07, |
| "loss": 0.856, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.01, |
| "learning_rate": 1.25e-06, |
| "loss": 0.792, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.01, |
| "learning_rate": 1.8750000000000003e-06, |
| "loss": 0.835, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 2.5e-06, |
| "loss": 0.8315, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 3.125e-06, |
| "loss": 0.8115, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 3.7500000000000005e-06, |
| "loss": 0.8032, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.03, |
| "learning_rate": 4.3750000000000005e-06, |
| "loss": 0.7256, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.03, |
| "learning_rate": 5e-06, |
| "loss": 0.7007, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.03, |
| "learning_rate": 5.625e-06, |
| "loss": 0.647, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 6.25e-06, |
| "loss": 0.6431, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 6.875e-06, |
| "loss": 0.6455, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.05, |
| "learning_rate": 7.500000000000001e-06, |
| "loss": 0.5786, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.05, |
| "learning_rate": 8.125000000000001e-06, |
| "loss": 0.6245, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.05, |
| "learning_rate": 8.750000000000001e-06, |
| "loss": 0.5962, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 9.375000000000001e-06, |
| "loss": 0.5854, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 1e-05, |
| "loss": 0.5693, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 1.0625e-05, |
| "loss": 0.5615, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 1.125e-05, |
| "loss": 0.5354, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 1.1875e-05, |
| "loss": 0.563, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 1.25e-05, |
| "loss": 0.5596, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 1.3125e-05, |
| "loss": 0.543, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 1.375e-05, |
| "loss": 0.5403, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 1.4375e-05, |
| "loss": 0.5437, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 1.5000000000000002e-05, |
| "loss": 0.5286, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 1.5625e-05, |
| "loss": 0.5417, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 1.6250000000000002e-05, |
| "loss": 0.5151, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 1.6875e-05, |
| "loss": 0.5154, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 1.7500000000000002e-05, |
| "loss": 0.5271, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 1.8125e-05, |
| "loss": 0.5283, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 1.8750000000000002e-05, |
| "loss": 0.5127, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 1.9375e-05, |
| "loss": 0.5125, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 2e-05, |
| "loss": 0.5166, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 1.9999915234318064e-05, |
| "loss": 0.5288, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 1.999966093870929e-05, |
| "loss": 0.5264, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 1.99992371174848e-05, |
| "loss": 0.5229, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 1.9998643777829674e-05, |
| "loss": 0.5239, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 1.9997880929802895e-05, |
| "loss": 0.4995, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 1.9996948586337127e-05, |
| "loss": 0.4934, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 1.9995846763238514e-05, |
| "loss": 0.5022, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 1.999457547918641e-05, |
| "loss": 0.4658, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 1.9993134755733075e-05, |
| "loss": 0.4968, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 1.9991524617303282e-05, |
| "loss": 0.5078, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 1.998974509119393e-05, |
| "loss": 0.5183, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 1.9987796207573573e-05, |
| "loss": 0.4844, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 1.9985677999481898e-05, |
| "loss": 0.5, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 1.9983390502829168e-05, |
| "loss": 0.488, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 1.9980933756395635e-05, |
| "loss": 0.4744, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 1.9978307801830855e-05, |
| "loss": 0.5098, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 1.9975512683652985e-05, |
| "loss": 0.4851, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 1.997254844924806e-05, |
| "loss": 0.4861, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 1.996941514886914e-05, |
| "loss": 0.5112, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 1.9966112835635493e-05, |
| "loss": 0.4678, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 1.9962641565531694e-05, |
| "loss": 0.4897, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 1.995900139740665e-05, |
| "loss": 0.4563, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.21, |
| "learning_rate": 1.9955192392972628e-05, |
| "loss": 0.488, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.21, |
| "learning_rate": 1.9951214616804203e-05, |
| "loss": 0.4812, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 1.994706813633716e-05, |
| "loss": 0.4617, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 1.994275302186734e-05, |
| "loss": 0.4836, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 1.9938269346549473e-05, |
| "loss": 0.4846, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 1.9933617186395917e-05, |
| "loss": 0.4749, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 1.9928796620275377e-05, |
| "loss": 0.4788, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 1.9923807729911567e-05, |
| "loss": 0.4802, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 1.9918650599881828e-05, |
| "loss": 0.468, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 1.9913325317615684e-05, |
| "loss": 0.4854, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 1.9907831973393377e-05, |
| "loss": 0.4731, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 1.9902170660344323e-05, |
| "loss": 0.5017, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 1.9896341474445526e-05, |
| "loss": 0.4915, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 1.9890344514519974e-05, |
| "loss": 0.4868, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 1.9884179882234946e-05, |
| "loss": 0.457, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 1.9877847682100294e-05, |
| "loss": 0.48, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.27, |
| "learning_rate": 1.9871348021466673e-05, |
| "loss": 0.4707, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.27, |
| "learning_rate": 1.986468101052371e-05, |
| "loss": 0.4912, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 1.9857846762298157e-05, |
| "loss": 0.4768, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 1.985084539265195e-05, |
| "loss": 0.4492, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 1.984367702028027e-05, |
| "loss": 0.4761, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 1.983634176670951e-05, |
| "loss": 0.437, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 1.9828839756295223e-05, |
| "loss": 0.4619, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 1.982117111622001e-05, |
| "loss": 0.4719, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.3, |
| "learning_rate": 1.9813335976491387e-05, |
| "loss": 0.4753, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.3, |
| "learning_rate": 1.980533446993953e-05, |
| "loss": 0.4697, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 1.9797166732215078e-05, |
| "loss": 0.4648, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 1.97888329017868e-05, |
| "loss": 0.4534, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 1.9780333119939264e-05, |
| "loss": 0.4553, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 1.9771667530770427e-05, |
| "loss": 0.4714, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 1.9762836281189207e-05, |
| "loss": 0.4385, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 1.9753839520912984e-05, |
| "loss": 0.4697, |
| "step": 86 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 1.9744677402465053e-05, |
| "loss": 0.4775, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 1.973535008117207e-05, |
| "loss": 0.4287, |
| "step": 88 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 1.9725857715161375e-05, |
| "loss": 0.4473, |
| "step": 89 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 1.9716200465358352e-05, |
| "loss": 0.4531, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 1.9706378495483664e-05, |
| "loss": 0.4524, |
| "step": 91 |
| }, |
| { |
| "epoch": 0.35, |
| "learning_rate": 1.9696391972050516e-05, |
| "loss": 0.4756, |
| "step": 92 |
| }, |
| { |
| "epoch": 0.35, |
| "learning_rate": 1.9686241064361792e-05, |
| "loss": 0.436, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.35, |
| "learning_rate": 1.9675925944507226e-05, |
| "loss": 0.4265, |
| "step": 94 |
| }, |
| { |
| "epoch": 0.36, |
| "learning_rate": 1.9665446787360444e-05, |
| "loss": 0.4766, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.36, |
| "learning_rate": 1.965480377057603e-05, |
| "loss": 0.4683, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 1.964399707458649e-05, |
| "loss": 0.4602, |
| "step": 97 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 1.9633026882599228e-05, |
| "loss": 0.4497, |
| "step": 98 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 1.9621893380593398e-05, |
| "loss": 0.4358, |
| "step": 99 |
| }, |
| { |
| "epoch": 0.38, |
| "learning_rate": 1.961059675731678e-05, |
| "loss": 0.4641, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.38, |
| "learning_rate": 1.9599137204282566e-05, |
| "loss": 0.4429, |
| "step": 101 |
| }, |
| { |
| "epoch": 0.38, |
| "learning_rate": 1.9587514915766124e-05, |
| "loss": 0.4343, |
| "step": 102 |
| }, |
| { |
| "epoch": 0.39, |
| "learning_rate": 1.9575730088801696e-05, |
| "loss": 0.4553, |
| "step": 103 |
| }, |
| { |
| "epoch": 0.39, |
| "learning_rate": 1.9563782923179063e-05, |
| "loss": 0.4294, |
| "step": 104 |
| }, |
| { |
| "epoch": 0.4, |
| "learning_rate": 1.955167362144015e-05, |
| "loss": 0.4282, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.4, |
| "learning_rate": 1.9539402388875598e-05, |
| "loss": 0.4333, |
| "step": 106 |
| }, |
| { |
| "epoch": 0.4, |
| "learning_rate": 1.9526969433521298e-05, |
| "loss": 0.4631, |
| "step": 107 |
| }, |
| { |
| "epoch": 0.41, |
| "learning_rate": 1.9514374966154826e-05, |
| "loss": 0.4358, |
| "step": 108 |
| }, |
| { |
| "epoch": 0.41, |
| "learning_rate": 1.950161920029191e-05, |
| "loss": 0.4507, |
| "step": 109 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 1.948870235218279e-05, |
| "loss": 0.4512, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 1.9475624640808542e-05, |
| "loss": 0.4495, |
| "step": 111 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 1.946238628787741e-05, |
| "loss": 0.4546, |
| "step": 112 |
| }, |
| { |
| "epoch": 0.43, |
| "learning_rate": 1.9448987517820982e-05, |
| "loss": 0.4451, |
| "step": 113 |
| }, |
| { |
| "epoch": 0.43, |
| "learning_rate": 1.943542855779044e-05, |
| "loss": 0.4504, |
| "step": 114 |
| }, |
| { |
| "epoch": 0.43, |
| "learning_rate": 1.9421709637652683e-05, |
| "loss": 0.4504, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.44, |
| "learning_rate": 1.940783098998643e-05, |
| "loss": 0.4219, |
| "step": 116 |
| }, |
| { |
| "epoch": 0.44, |
| "learning_rate": 1.9393792850078294e-05, |
| "loss": 0.4253, |
| "step": 117 |
| }, |
| { |
| "epoch": 0.45, |
| "learning_rate": 1.9379595455918773e-05, |
| "loss": 0.4309, |
| "step": 118 |
| }, |
| { |
| "epoch": 0.45, |
| "learning_rate": 1.9365239048198227e-05, |
| "loss": 0.4587, |
| "step": 119 |
| }, |
| { |
| "epoch": 0.45, |
| "learning_rate": 1.935072387030279e-05, |
| "loss": 0.4404, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.46, |
| "learning_rate": 1.933605016831026e-05, |
| "loss": 0.4568, |
| "step": 121 |
| }, |
| { |
| "epoch": 0.46, |
| "learning_rate": 1.9321218190985906e-05, |
| "loss": 0.4399, |
| "step": 122 |
| }, |
| { |
| "epoch": 0.46, |
| "learning_rate": 1.9306228189778255e-05, |
| "loss": 0.429, |
| "step": 123 |
| }, |
| { |
| "epoch": 0.47, |
| "learning_rate": 1.9291080418814852e-05, |
| "loss": 0.4341, |
| "step": 124 |
| }, |
| { |
| "epoch": 0.47, |
| "learning_rate": 1.927577513489792e-05, |
| "loss": 0.4377, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.48, |
| "learning_rate": 1.926031259750002e-05, |
| "loss": 0.4258, |
| "step": 126 |
| }, |
| { |
| "epoch": 0.48, |
| "learning_rate": 1.9244693068759668e-05, |
| "loss": 0.4233, |
| "step": 127 |
| }, |
| { |
| "epoch": 0.48, |
| "learning_rate": 1.9228916813476855e-05, |
| "loss": 0.4243, |
| "step": 128 |
| }, |
| { |
| "epoch": 0.49, |
| "learning_rate": 1.9212984099108594e-05, |
| "loss": 0.467, |
| "step": 129 |
| }, |
| { |
| "epoch": 0.49, |
| "learning_rate": 1.9196895195764363e-05, |
| "loss": 0.4299, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.49, |
| "learning_rate": 1.9180650376201536e-05, |
| "loss": 0.4329, |
| "step": 131 |
| }, |
| { |
| "epoch": 0.5, |
| "learning_rate": 1.9164249915820753e-05, |
| "loss": 0.4695, |
| "step": 132 |
| }, |
| { |
| "epoch": 0.5, |
| "learning_rate": 1.9147694092661254e-05, |
| "loss": 0.4431, |
| "step": 133 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 1.9130983187396174e-05, |
| "loss": 0.3972, |
| "step": 134 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 1.911411748332776e-05, |
| "loss": 0.4617, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 1.9097097266382598e-05, |
| "loss": 0.4087, |
| "step": 136 |
| }, |
| { |
| "epoch": 0.52, |
| "learning_rate": 1.907992282510675e-05, |
| "loss": 0.4275, |
| "step": 137 |
| }, |
| { |
| "epoch": 0.52, |
| "learning_rate": 1.9062594450660857e-05, |
| "loss": 0.4285, |
| "step": 138 |
| }, |
| { |
| "epoch": 0.52, |
| "learning_rate": 1.9045112436815217e-05, |
| "loss": 0.4099, |
| "step": 139 |
| }, |
| { |
| "epoch": 0.53, |
| "learning_rate": 1.902747707994479e-05, |
| "loss": 0.4299, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.53, |
| "learning_rate": 1.900968867902419e-05, |
| "loss": 0.4216, |
| "step": 141 |
| }, |
| { |
| "epoch": 0.54, |
| "learning_rate": 1.8991747535622607e-05, |
| "loss": 0.426, |
| "step": 142 |
| }, |
| { |
| "epoch": 0.54, |
| "learning_rate": 1.897365395389869e-05, |
| "loss": 0.449, |
| "step": 143 |
| }, |
| { |
| "epoch": 0.54, |
| "learning_rate": 1.8955408240595396e-05, |
| "loss": 0.4387, |
| "step": 144 |
| }, |
| { |
| "epoch": 0.55, |
| "learning_rate": 1.8937010705034788e-05, |
| "loss": 0.4146, |
| "step": 145 |
| }, |
| { |
| "epoch": 0.55, |
| "learning_rate": 1.8918461659112805e-05, |
| "loss": 0.4326, |
| "step": 146 |
| }, |
| { |
| "epoch": 0.55, |
| "learning_rate": 1.8899761417293944e-05, |
| "loss": 0.4514, |
| "step": 147 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 1.8880910296605956e-05, |
| "loss": 0.4436, |
| "step": 148 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 1.8861908616634465e-05, |
| "loss": 0.4062, |
| "step": 149 |
| }, |
| { |
| "epoch": 0.57, |
| "learning_rate": 1.884275669951754e-05, |
| "loss": 0.4158, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.57, |
| "learning_rate": 1.8823454869940243e-05, |
| "loss": 0.3972, |
| "step": 151 |
| }, |
| { |
| "epoch": 0.57, |
| "learning_rate": 1.880400345512913e-05, |
| "loss": 0.408, |
| "step": 152 |
| }, |
| { |
| "epoch": 0.58, |
| "learning_rate": 1.8784402784846683e-05, |
| "loss": 0.4277, |
| "step": 153 |
| }, |
| { |
| "epoch": 0.58, |
| "learning_rate": 1.8764653191385737e-05, |
| "loss": 0.4163, |
| "step": 154 |
| }, |
| { |
| "epoch": 0.58, |
| "learning_rate": 1.874475500956385e-05, |
| "loss": 0.4158, |
| "step": 155 |
| }, |
| { |
| "epoch": 0.59, |
| "learning_rate": 1.8724708576717607e-05, |
| "loss": 0.4512, |
| "step": 156 |
| }, |
| { |
| "epoch": 0.59, |
| "learning_rate": 1.870451423269692e-05, |
| "loss": 0.4094, |
| "step": 157 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 1.8684172319859258e-05, |
| "loss": 0.408, |
| "step": 158 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 1.8663683183063846e-05, |
| "loss": 0.4189, |
| "step": 159 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 1.864304716966581e-05, |
| "loss": 0.4333, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.61, |
| "learning_rate": 1.86222646295103e-05, |
| "loss": 0.4329, |
| "step": 161 |
| }, |
| { |
| "epoch": 0.61, |
| "learning_rate": 1.8601335914926558e-05, |
| "loss": 0.437, |
| "step": 162 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 1.8580261380721932e-05, |
| "loss": 0.407, |
| "step": 163 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 1.8559041384175874e-05, |
| "loss": 0.4077, |
| "step": 164 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 1.8537676285033886e-05, |
| "loss": 0.4346, |
| "step": 165 |
| }, |
| { |
| "epoch": 0.63, |
| "learning_rate": 1.8516166445501405e-05, |
| "loss": 0.4246, |
| "step": 166 |
| }, |
| { |
| "epoch": 0.63, |
| "learning_rate": 1.849451223023767e-05, |
| "loss": 0.4456, |
| "step": 167 |
| }, |
| { |
| "epoch": 0.63, |
| "learning_rate": 1.8472714006349554e-05, |
| "loss": 0.4209, |
| "step": 168 |
| }, |
| { |
| "epoch": 0.64, |
| "learning_rate": 1.845077214338531e-05, |
| "loss": 0.4275, |
| "step": 169 |
| }, |
| { |
| "epoch": 0.64, |
| "learning_rate": 1.8428687013328338e-05, |
| "loss": 0.4268, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 1.840645899059086e-05, |
| "loss": 0.4041, |
| "step": 171 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 1.838408845200758e-05, |
| "loss": 0.3972, |
| "step": 172 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 1.836157577682928e-05, |
| "loss": 0.4041, |
| "step": 173 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 1.8338921346716426e-05, |
| "loss": 0.3882, |
| "step": 174 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 1.8316125545732653e-05, |
| "loss": 0.3889, |
| "step": 175 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 1.8293188760338285e-05, |
| "loss": 0.3784, |
| "step": 176 |
| }, |
| { |
| "epoch": 0.67, |
| "learning_rate": 1.8270111379383773e-05, |
| "loss": 0.386, |
| "step": 177 |
| }, |
| { |
| "epoch": 0.67, |
| "learning_rate": 1.8246893794103113e-05, |
| "loss": 0.3862, |
| "step": 178 |
| }, |
| { |
| "epoch": 0.68, |
| "learning_rate": 1.8223536398107177e-05, |
| "loss": 0.3838, |
| "step": 179 |
| }, |
| { |
| "epoch": 0.68, |
| "learning_rate": 1.82000395873771e-05, |
| "loss": 0.3848, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.68, |
| "learning_rate": 1.817640376025751e-05, |
| "loss": 0.4333, |
| "step": 181 |
| }, |
| { |
| "epoch": 0.69, |
| "learning_rate": 1.8152629317449814e-05, |
| "loss": 0.3643, |
| "step": 182 |
| }, |
| { |
| "epoch": 0.69, |
| "learning_rate": 1.8128716662005385e-05, |
| "loss": 0.3896, |
| "step": 183 |
| }, |
| { |
| "epoch": 0.69, |
| "learning_rate": 1.810466619931872e-05, |
| "loss": 0.4285, |
| "step": 184 |
| }, |
| { |
| "epoch": 0.7, |
| "learning_rate": 1.8080478337120604e-05, |
| "loss": 0.4031, |
| "step": 185 |
| }, |
| { |
| "epoch": 0.7, |
| "learning_rate": 1.8056153485471167e-05, |
| "loss": 0.4028, |
| "step": 186 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 1.8031692056752926e-05, |
| "loss": 0.4155, |
| "step": 187 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 1.8007094465663823e-05, |
| "loss": 0.3867, |
| "step": 188 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 1.7982361129210172e-05, |
| "loss": 0.4058, |
| "step": 189 |
| }, |
| { |
| "epoch": 0.72, |
| "learning_rate": 1.7957492466699606e-05, |
| "loss": 0.394, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.72, |
| "learning_rate": 1.793248889973395e-05, |
| "loss": 0.4062, |
| "step": 191 |
| }, |
| { |
| "epoch": 0.72, |
| "learning_rate": 1.7907350852202078e-05, |
| "loss": 0.3909, |
| "step": 192 |
| }, |
| { |
| "epoch": 0.73, |
| "learning_rate": 1.788207875027274e-05, |
| "loss": 0.394, |
| "step": 193 |
| }, |
| { |
| "epoch": 0.73, |
| "learning_rate": 1.785667302238734e-05, |
| "loss": 0.4043, |
| "step": 194 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 1.7831134099252633e-05, |
| "loss": 0.4062, |
| "step": 195 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 1.780546241383347e-05, |
| "loss": 0.3811, |
| "step": 196 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 1.7779658401345437e-05, |
| "loss": 0.3818, |
| "step": 197 |
| }, |
| { |
| "epoch": 0.75, |
| "learning_rate": 1.775372249924748e-05, |
| "loss": 0.3811, |
| "step": 198 |
| }, |
| { |
| "epoch": 0.75, |
| "learning_rate": 1.772765514723448e-05, |
| "loss": 0.3726, |
| "step": 199 |
| }, |
| { |
| "epoch": 0.75, |
| "learning_rate": 1.7701456787229805e-05, |
| "loss": 0.3767, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.76, |
| "learning_rate": 1.767512786337784e-05, |
| "loss": 0.3813, |
| "step": 201 |
| }, |
| { |
| "epoch": 0.76, |
| "learning_rate": 1.764866882203641e-05, |
| "loss": 0.3943, |
| "step": 202 |
| }, |
| { |
| "epoch": 0.77, |
| "learning_rate": 1.7622080111769257e-05, |
| "loss": 0.376, |
| "step": 203 |
| }, |
| { |
| "epoch": 0.77, |
| "learning_rate": 1.759536218333841e-05, |
| "loss": 0.3779, |
| "step": 204 |
| }, |
| { |
| "epoch": 0.77, |
| "learning_rate": 1.7568515489696558e-05, |
| "loss": 0.3838, |
| "step": 205 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 1.7541540485979357e-05, |
| "loss": 0.3579, |
| "step": 206 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 1.751443762949772e-05, |
| "loss": 0.4128, |
| "step": 207 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 1.7487207379730078e-05, |
| "loss": 0.3806, |
| "step": 208 |
| }, |
| { |
| "epoch": 0.79, |
| "learning_rate": 1.7459850198314562e-05, |
| "loss": 0.3726, |
| "step": 209 |
| }, |
| { |
| "epoch": 0.79, |
| "learning_rate": 1.7432366549041203e-05, |
| "loss": 0.3682, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.8, |
| "learning_rate": 1.7404756897844054e-05, |
| "loss": 0.3828, |
| "step": 211 |
| }, |
| { |
| "epoch": 0.8, |
| "learning_rate": 1.73770217127933e-05, |
| "loss": 0.3622, |
| "step": 212 |
| }, |
| { |
| "epoch": 0.8, |
| "learning_rate": 1.7349161464087312e-05, |
| "loss": 0.4253, |
| "step": 213 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 1.732117662404469e-05, |
| "loss": 0.3638, |
| "step": 214 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 1.729306766709624e-05, |
| "loss": 0.4026, |
| "step": 215 |
| }, |
| { |
| "epoch": 0.82, |
| "learning_rate": 1.7264835069776945e-05, |
| "loss": 0.4021, |
| "step": 216 |
| }, |
| { |
| "epoch": 0.82, |
| "learning_rate": 1.7236479310717878e-05, |
| "loss": 0.3625, |
| "step": 217 |
| }, |
| { |
| "epoch": 0.82, |
| "learning_rate": 1.7208000870638094e-05, |
| "loss": 0.3806, |
| "step": 218 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 1.7179400232336462e-05, |
| "loss": 0.3914, |
| "step": 219 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 1.7150677880683515e-05, |
| "loss": 0.3923, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 1.712183430261319e-05, |
| "loss": 0.3843, |
| "step": 221 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 1.70928699871146e-05, |
| "loss": 0.3945, |
| "step": 222 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 1.706378542522374e-05, |
| "loss": 0.3948, |
| "step": 223 |
| }, |
| { |
| "epoch": 0.85, |
| "learning_rate": 1.7034581110015156e-05, |
| "loss": 0.3804, |
| "step": 224 |
| }, |
| { |
| "epoch": 0.85, |
| "learning_rate": 1.7005257536593577e-05, |
| "loss": 0.3738, |
| "step": 225 |
| }, |
| { |
| "epoch": 0.85, |
| "learning_rate": 1.6975815202085556e-05, |
| "loss": 0.3582, |
| "step": 226 |
| }, |
| { |
| "epoch": 0.86, |
| "learning_rate": 1.6946254605630995e-05, |
| "loss": 0.3545, |
| "step": 227 |
| }, |
| { |
| "epoch": 0.86, |
| "learning_rate": 1.691657624837472e-05, |
| "loss": 0.3586, |
| "step": 228 |
| }, |
| { |
| "epoch": 0.86, |
| "learning_rate": 1.6886780633457975e-05, |
| "loss": 0.3726, |
| "step": 229 |
| }, |
| { |
| "epoch": 0.87, |
| "learning_rate": 1.6856868266009874e-05, |
| "loss": 0.3818, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.87, |
| "learning_rate": 1.6826839653138872e-05, |
| "loss": 0.3713, |
| "step": 231 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 1.679669530392413e-05, |
| "loss": 0.3711, |
| "step": 232 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 1.6766435729406913e-05, |
| "loss": 0.3855, |
| "step": 233 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 1.6736061442581922e-05, |
| "loss": 0.3943, |
| "step": 234 |
| }, |
| { |
| "epoch": 0.89, |
| "learning_rate": 1.6705572958388576e-05, |
| "loss": 0.3867, |
| "step": 235 |
| }, |
| { |
| "epoch": 0.89, |
| "learning_rate": 1.667497079370231e-05, |
| "loss": 0.3679, |
| "step": 236 |
| }, |
| { |
| "epoch": 0.89, |
| "learning_rate": 1.6644255467325793e-05, |
| "loss": 0.3945, |
| "step": 237 |
| }, |
| { |
| "epoch": 0.9, |
| "learning_rate": 1.6613427499980143e-05, |
| "loss": 0.3906, |
| "step": 238 |
| }, |
| { |
| "epoch": 0.9, |
| "learning_rate": 1.6582487414296097e-05, |
| "loss": 0.3706, |
| "step": 239 |
| }, |
| { |
| "epoch": 0.91, |
| "learning_rate": 1.655143573480515e-05, |
| "loss": 0.3794, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.91, |
| "learning_rate": 1.6520272987930652e-05, |
| "loss": 0.376, |
| "step": 241 |
| }, |
| { |
| "epoch": 0.91, |
| "learning_rate": 1.6488999701978905e-05, |
| "loss": 0.3486, |
| "step": 242 |
| }, |
| { |
| "epoch": 0.92, |
| "learning_rate": 1.645761640713019e-05, |
| "loss": 0.3806, |
| "step": 243 |
| }, |
| { |
| "epoch": 0.92, |
| "learning_rate": 1.6426123635429787e-05, |
| "loss": 0.3521, |
| "step": 244 |
| }, |
| { |
| "epoch": 0.92, |
| "learning_rate": 1.639452192077895e-05, |
| "loss": 0.3962, |
| "step": 245 |
| }, |
| { |
| "epoch": 0.93, |
| "learning_rate": 1.6362811798925852e-05, |
| "loss": 0.3372, |
| "step": 246 |
| }, |
| { |
| "epoch": 0.93, |
| "learning_rate": 1.633099380745652e-05, |
| "loss": 0.3545, |
| "step": 247 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 1.62990684857857e-05, |
| "loss": 0.3774, |
| "step": 248 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 1.6267036375147728e-05, |
| "loss": 0.3672, |
| "step": 249 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 1.6234898018587336e-05, |
| "loss": 0.3418, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.95, |
| "learning_rate": 1.6202653960950474e-05, |
| "loss": 0.3789, |
| "step": 251 |
| }, |
| { |
| "epoch": 0.95, |
| "learning_rate": 1.617030474887505e-05, |
| "loss": 0.3457, |
| "step": 252 |
| }, |
| { |
| "epoch": 0.95, |
| "learning_rate": 1.613785093078166e-05, |
| "loss": 0.3298, |
| "step": 253 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 1.6105293056864314e-05, |
| "loss": 0.3906, |
| "step": 254 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 1.607263167908109e-05, |
| "loss": 0.3518, |
| "step": 255 |
| }, |
| { |
| "epoch": 0.97, |
| "learning_rate": 1.6039867351144778e-05, |
| "loss": 0.3496, |
| "step": 256 |
| }, |
| { |
| "epoch": 0.97, |
| "learning_rate": 1.6007000628513498e-05, |
| "loss": 0.3816, |
| "step": 257 |
| }, |
| { |
| "epoch": 0.97, |
| "learning_rate": 1.597403206838128e-05, |
| "loss": 0.385, |
| "step": 258 |
| }, |
| { |
| "epoch": 0.98, |
| "learning_rate": 1.5940962229668625e-05, |
| "loss": 0.3774, |
| "step": 259 |
| }, |
| { |
| "epoch": 0.98, |
| "learning_rate": 1.5907791673013016e-05, |
| "loss": 0.3354, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.98, |
| "learning_rate": 1.5874520960759423e-05, |
| "loss": 0.3489, |
| "step": 261 |
| }, |
| { |
| "epoch": 0.99, |
| "learning_rate": 1.584115065695077e-05, |
| "loss": 0.3491, |
| "step": 262 |
| }, |
| { |
| "epoch": 0.99, |
| "learning_rate": 1.5807681327318372e-05, |
| "loss": 0.3892, |
| "step": 263 |
| }, |
| { |
| "epoch": 1.0, |
| "learning_rate": 1.5774113539272332e-05, |
| "loss": 0.3838, |
| "step": 264 |
| }, |
| { |
| "epoch": 1.0, |
| "learning_rate": 1.5740447861891946e-05, |
| "loss": 0.3872, |
| "step": 265 |
| }, |
| { |
| "epoch": 1.0, |
| "learning_rate": 1.5706684865916025e-05, |
| "loss": 0.324, |
| "step": 266 |
| }, |
| { |
| "epoch": 1.01, |
| "learning_rate": 1.5672825123733257e-05, |
| "loss": 0.3145, |
| "step": 267 |
| }, |
| { |
| "epoch": 1.01, |
| "learning_rate": 1.563886920937246e-05, |
| "loss": 0.2955, |
| "step": 268 |
| }, |
| { |
| "epoch": 1.02, |
| "learning_rate": 1.5604817698492886e-05, |
| "loss": 0.3337, |
| "step": 269 |
| }, |
| { |
| "epoch": 1.02, |
| "learning_rate": 1.557067116837444e-05, |
| "loss": 0.3225, |
| "step": 270 |
| }, |
| { |
| "epoch": 1.02, |
| "learning_rate": 1.5536430197907904e-05, |
| "loss": 0.3074, |
| "step": 271 |
| }, |
| { |
| "epoch": 1.03, |
| "learning_rate": 1.5502095367585124e-05, |
| "loss": 0.2902, |
| "step": 272 |
| }, |
| { |
| "epoch": 1.03, |
| "learning_rate": 1.5467667259489157e-05, |
| "loss": 0.3196, |
| "step": 273 |
| }, |
| { |
| "epoch": 1.03, |
| "learning_rate": 1.543314645728442e-05, |
| "loss": 0.3165, |
| "step": 274 |
| }, |
| { |
| "epoch": 1.04, |
| "learning_rate": 1.539853354620678e-05, |
| "loss": 0.3162, |
| "step": 275 |
| }, |
| { |
| "epoch": 1.04, |
| "learning_rate": 1.5363829113053633e-05, |
| "loss": 0.307, |
| "step": 276 |
| }, |
| { |
| "epoch": 1.05, |
| "learning_rate": 1.5329033746173975e-05, |
| "loss": 0.3289, |
| "step": 277 |
| }, |
| { |
| "epoch": 1.05, |
| "learning_rate": 1.5294148035458406e-05, |
| "loss": 0.3269, |
| "step": 278 |
| }, |
| { |
| "epoch": 1.05, |
| "learning_rate": 1.5259172572329132e-05, |
| "loss": 0.3296, |
| "step": 279 |
| }, |
| { |
| "epoch": 1.06, |
| "learning_rate": 1.5224107949729952e-05, |
| "loss": 0.3134, |
| "step": 280 |
| }, |
| { |
| "epoch": 1.06, |
| "learning_rate": 1.5188954762116197e-05, |
| "loss": 0.3328, |
| "step": 281 |
| }, |
| { |
| "epoch": 1.06, |
| "learning_rate": 1.515371360544465e-05, |
| "loss": 0.2966, |
| "step": 282 |
| }, |
| { |
| "epoch": 1.07, |
| "learning_rate": 1.5118385077163446e-05, |
| "loss": 0.3091, |
| "step": 283 |
| }, |
| { |
| "epoch": 1.07, |
| "learning_rate": 1.5082969776201948e-05, |
| "loss": 0.3149, |
| "step": 284 |
| }, |
| { |
| "epoch": 1.08, |
| "learning_rate": 1.5047468302960577e-05, |
| "loss": 0.2905, |
| "step": 285 |
| }, |
| { |
| "epoch": 1.08, |
| "learning_rate": 1.5011881259300654e-05, |
| "loss": 0.3082, |
| "step": 286 |
| }, |
| { |
| "epoch": 1.08, |
| "learning_rate": 1.4976209248534183e-05, |
| "loss": 0.3025, |
| "step": 287 |
| }, |
| { |
| "epoch": 1.09, |
| "learning_rate": 1.4940452875413627e-05, |
| "loss": 0.3101, |
| "step": 288 |
| }, |
| { |
| "epoch": 1.09, |
| "learning_rate": 1.4904612746121657e-05, |
| "loss": 0.304, |
| "step": 289 |
| }, |
| { |
| "epoch": 1.09, |
| "learning_rate": 1.4868689468260876e-05, |
| "loss": 0.3428, |
| "step": 290 |
| }, |
| { |
| "epoch": 1.1, |
| "learning_rate": 1.483268365084351e-05, |
| "loss": 0.2863, |
| "step": 291 |
| }, |
| { |
| "epoch": 1.1, |
| "learning_rate": 1.479659590428109e-05, |
| "loss": 0.3323, |
| "step": 292 |
| }, |
| { |
| "epoch": 1.11, |
| "learning_rate": 1.476042684037411e-05, |
| "loss": 0.3051, |
| "step": 293 |
| }, |
| { |
| "epoch": 1.11, |
| "learning_rate": 1.4724177072301642e-05, |
| "loss": 0.2874, |
| "step": 294 |
| }, |
| { |
| "epoch": 1.11, |
| "learning_rate": 1.4687847214610944e-05, |
| "loss": 0.3176, |
| "step": 295 |
| }, |
| { |
| "epoch": 1.12, |
| "learning_rate": 1.4651437883207056e-05, |
| "loss": 0.3197, |
| "step": 296 |
| }, |
| { |
| "epoch": 1.12, |
| "learning_rate": 1.4614949695342335e-05, |
| "loss": 0.3259, |
| "step": 297 |
| }, |
| { |
| "epoch": 1.12, |
| "learning_rate": 1.4578383269606004e-05, |
| "loss": 0.2832, |
| "step": 298 |
| }, |
| { |
| "epoch": 1.13, |
| "learning_rate": 1.4541739225913669e-05, |
| "loss": 0.3047, |
| "step": 299 |
| }, |
| { |
| "epoch": 1.13, |
| "learning_rate": 1.4505018185496802e-05, |
| "loss": 0.318, |
| "step": 300 |
| }, |
| { |
| "epoch": 1.14, |
| "learning_rate": 1.4468220770892208e-05, |
| "loss": 0.2836, |
| "step": 301 |
| }, |
| { |
| "epoch": 1.14, |
| "learning_rate": 1.443134760593147e-05, |
| "loss": 0.275, |
| "step": 302 |
| }, |
| { |
| "epoch": 1.14, |
| "learning_rate": 1.4394399315730389e-05, |
| "loss": 0.3, |
| "step": 303 |
| }, |
| { |
| "epoch": 1.15, |
| "learning_rate": 1.4357376526678368e-05, |
| "loss": 0.2927, |
| "step": 304 |
| }, |
| { |
| "epoch": 1.15, |
| "learning_rate": 1.4320279866427798e-05, |
| "loss": 0.3098, |
| "step": 305 |
| }, |
| { |
| "epoch": 1.15, |
| "learning_rate": 1.428310996388342e-05, |
| "loss": 0.2933, |
| "step": 306 |
| }, |
| { |
| "epoch": 1.16, |
| "learning_rate": 1.424586744919166e-05, |
| "loss": 0.2913, |
| "step": 307 |
| }, |
| { |
| "epoch": 1.16, |
| "learning_rate": 1.4208552953729949e-05, |
| "loss": 0.2844, |
| "step": 308 |
| }, |
| { |
| "epoch": 1.17, |
| "learning_rate": 1.4171167110096017e-05, |
| "loss": 0.2993, |
| "step": 309 |
| }, |
| { |
| "epoch": 1.17, |
| "learning_rate": 1.4133710552097175e-05, |
| "loss": 0.2803, |
| "step": 310 |
| }, |
| { |
| "epoch": 1.17, |
| "learning_rate": 1.4096183914739554e-05, |
| "loss": 0.3308, |
| "step": 311 |
| }, |
| { |
| "epoch": 1.18, |
| "learning_rate": 1.4058587834217356e-05, |
| "loss": 0.3281, |
| "step": 312 |
| }, |
| { |
| "epoch": 1.18, |
| "learning_rate": 1.4020922947902067e-05, |
| "loss": 0.3112, |
| "step": 313 |
| }, |
| { |
| "epoch": 1.18, |
| "learning_rate": 1.3983189894331636e-05, |
| "loss": 0.3196, |
| "step": 314 |
| }, |
| { |
| "epoch": 1.19, |
| "learning_rate": 1.3945389313199669e-05, |
| "loss": 0.313, |
| "step": 315 |
| }, |
| { |
| "epoch": 1.19, |
| "learning_rate": 1.3907521845344571e-05, |
| "loss": 0.2998, |
| "step": 316 |
| }, |
| { |
| "epoch": 1.2, |
| "learning_rate": 1.3869588132738696e-05, |
| "loss": 0.317, |
| "step": 317 |
| }, |
| { |
| "epoch": 1.2, |
| "learning_rate": 1.3831588818477437e-05, |
| "loss": 0.3042, |
| "step": 318 |
| }, |
| { |
| "epoch": 1.2, |
| "learning_rate": 1.3793524546768358e-05, |
| "loss": 0.3219, |
| "step": 319 |
| }, |
| { |
| "epoch": 1.21, |
| "learning_rate": 1.375539596292025e-05, |
| "loss": 0.3298, |
| "step": 320 |
| }, |
| { |
| "epoch": 1.21, |
| "learning_rate": 1.371720371333219e-05, |
| "loss": 0.2942, |
| "step": 321 |
| }, |
| { |
| "epoch": 1.22, |
| "learning_rate": 1.3678948445482598e-05, |
| "loss": 0.2849, |
| "step": 322 |
| }, |
| { |
| "epoch": 1.22, |
| "learning_rate": 1.3640630807918246e-05, |
| "loss": 0.2966, |
| "step": 323 |
| }, |
| { |
| "epoch": 1.22, |
| "learning_rate": 1.3602251450243273e-05, |
| "loss": 0.2968, |
| "step": 324 |
| }, |
| { |
| "epoch": 1.23, |
| "learning_rate": 1.3563811023108157e-05, |
| "loss": 0.2969, |
| "step": 325 |
| }, |
| { |
| "epoch": 1.23, |
| "learning_rate": 1.3525310178198707e-05, |
| "loss": 0.3016, |
| "step": 326 |
| }, |
| { |
| "epoch": 1.23, |
| "learning_rate": 1.3486749568225002e-05, |
| "loss": 0.2727, |
| "step": 327 |
| }, |
| { |
| "epoch": 1.24, |
| "learning_rate": 1.3448129846910312e-05, |
| "loss": 0.2738, |
| "step": 328 |
| }, |
| { |
| "epoch": 1.24, |
| "learning_rate": 1.3409451668980047e-05, |
| "loss": 0.2787, |
| "step": 329 |
| }, |
| { |
| "epoch": 1.25, |
| "learning_rate": 1.3370715690150631e-05, |
| "loss": 0.2789, |
| "step": 330 |
| }, |
| { |
| "epoch": 1.25, |
| "learning_rate": 1.3331922567118394e-05, |
| "loss": 0.285, |
| "step": 331 |
| }, |
| { |
| "epoch": 1.25, |
| "learning_rate": 1.3293072957548443e-05, |
| "loss": 0.2745, |
| "step": 332 |
| }, |
| { |
| "epoch": 1.26, |
| "learning_rate": 1.325416752006351e-05, |
| "loss": 0.3013, |
| "step": 333 |
| }, |
| { |
| "epoch": 1.26, |
| "learning_rate": 1.321520691423278e-05, |
| "loss": 0.2968, |
| "step": 334 |
| }, |
| { |
| "epoch": 1.26, |
| "learning_rate": 1.317619180056072e-05, |
| "loss": 0.2772, |
| "step": 335 |
| }, |
| { |
| "epoch": 1.27, |
| "learning_rate": 1.313712284047587e-05, |
| "loss": 0.2599, |
| "step": 336 |
| }, |
| { |
| "epoch": 1.27, |
| "learning_rate": 1.3098000696319642e-05, |
| "loss": 0.283, |
| "step": 337 |
| }, |
| { |
| "epoch": 1.28, |
| "learning_rate": 1.305882603133508e-05, |
| "loss": 0.2819, |
| "step": 338 |
| }, |
| { |
| "epoch": 1.28, |
| "learning_rate": 1.301959950965562e-05, |
| "loss": 0.2982, |
| "step": 339 |
| }, |
| { |
| "epoch": 1.28, |
| "learning_rate": 1.2980321796293838e-05, |
| "loss": 0.2552, |
| "step": 340 |
| }, |
| { |
| "epoch": 1.29, |
| "learning_rate": 1.2940993557130166e-05, |
| "loss": 0.2743, |
| "step": 341 |
| }, |
| { |
| "epoch": 1.29, |
| "learning_rate": 1.2901615458901602e-05, |
| "loss": 0.2858, |
| "step": 342 |
| }, |
| { |
| "epoch": 1.29, |
| "learning_rate": 1.2862188169190419e-05, |
| "loss": 0.2767, |
| "step": 343 |
| }, |
| { |
| "epoch": 1.3, |
| "learning_rate": 1.282271235641284e-05, |
| "loss": 0.3038, |
| "step": 344 |
| }, |
| { |
| "epoch": 1.3, |
| "learning_rate": 1.2783188689807697e-05, |
| "loss": 0.2587, |
| "step": 345 |
| }, |
| { |
| "epoch": 1.31, |
| "learning_rate": 1.27436178394251e-05, |
| "loss": 0.2593, |
| "step": 346 |
| }, |
| { |
| "epoch": 1.31, |
| "learning_rate": 1.2704000476115079e-05, |
| "loss": 0.2947, |
| "step": 347 |
| }, |
| { |
| "epoch": 1.31, |
| "learning_rate": 1.2664337271516194e-05, |
| "loss": 0.3076, |
| "step": 348 |
| }, |
| { |
| "epoch": 1.32, |
| "learning_rate": 1.262462889804416e-05, |
| "loss": 0.2772, |
| "step": 349 |
| }, |
| { |
| "epoch": 1.32, |
| "learning_rate": 1.2584876028880455e-05, |
| "loss": 0.2466, |
| "step": 350 |
| }, |
| { |
| "epoch": 1.32, |
| "learning_rate": 1.2545079337960883e-05, |
| "loss": 0.2823, |
| "step": 351 |
| }, |
| { |
| "epoch": 1.33, |
| "learning_rate": 1.2505239499964179e-05, |
| "loss": 0.3049, |
| "step": 352 |
| }, |
| { |
| "epoch": 1.33, |
| "learning_rate": 1.246535719030055e-05, |
| "loss": 0.2764, |
| "step": 353 |
| }, |
| { |
| "epoch": 1.34, |
| "learning_rate": 1.2425433085100224e-05, |
| "loss": 0.2744, |
| "step": 354 |
| }, |
| { |
| "epoch": 1.34, |
| "learning_rate": 1.238546786120201e-05, |
| "loss": 0.2751, |
| "step": 355 |
| }, |
| { |
| "epoch": 1.34, |
| "learning_rate": 1.2345462196141797e-05, |
| "loss": 0.2932, |
| "step": 356 |
| }, |
| { |
| "epoch": 1.35, |
| "learning_rate": 1.2305416768141082e-05, |
| "loss": 0.2756, |
| "step": 357 |
| }, |
| { |
| "epoch": 1.35, |
| "learning_rate": 1.2265332256095463e-05, |
| "loss": 0.2816, |
| "step": 358 |
| }, |
| { |
| "epoch": 1.35, |
| "learning_rate": 1.2225209339563144e-05, |
| "loss": 0.3037, |
| "step": 359 |
| }, |
| { |
| "epoch": 1.36, |
| "learning_rate": 1.2185048698753403e-05, |
| "loss": 0.2765, |
| "step": 360 |
| }, |
| { |
| "epoch": 1.36, |
| "learning_rate": 1.2144851014515055e-05, |
| "loss": 0.274, |
| "step": 361 |
| }, |
| { |
| "epoch": 1.37, |
| "learning_rate": 1.2104616968324928e-05, |
| "loss": 0.2786, |
| "step": 362 |
| }, |
| { |
| "epoch": 1.37, |
| "learning_rate": 1.2064347242276293e-05, |
| "loss": 0.2832, |
| "step": 363 |
| }, |
| { |
| "epoch": 1.37, |
| "learning_rate": 1.202404251906731e-05, |
| "loss": 0.2985, |
| "step": 364 |
| }, |
| { |
| "epoch": 1.38, |
| "learning_rate": 1.1983703481989443e-05, |
| "loss": 0.2792, |
| "step": 365 |
| }, |
| { |
| "epoch": 1.38, |
| "learning_rate": 1.1943330814915897e-05, |
| "loss": 0.3042, |
| "step": 366 |
| }, |
| { |
| "epoch": 1.38, |
| "learning_rate": 1.1902925202289997e-05, |
| "loss": 0.2667, |
| "step": 367 |
| }, |
| { |
| "epoch": 1.39, |
| "learning_rate": 1.1862487329113606e-05, |
| "loss": 0.2931, |
| "step": 368 |
| }, |
| { |
| "epoch": 1.39, |
| "learning_rate": 1.1822017880935507e-05, |
| "loss": 0.2621, |
| "step": 369 |
| }, |
| { |
| "epoch": 1.4, |
| "learning_rate": 1.178151754383977e-05, |
| "loss": 0.2799, |
| "step": 370 |
| }, |
| { |
| "epoch": 1.4, |
| "learning_rate": 1.1740987004434137e-05, |
| "loss": 0.302, |
| "step": 371 |
| }, |
| { |
| "epoch": 1.4, |
| "learning_rate": 1.1700426949838364e-05, |
| "loss": 0.2913, |
| "step": 372 |
| }, |
| { |
| "epoch": 1.41, |
| "learning_rate": 1.16598380676726e-05, |
| "loss": 0.276, |
| "step": 373 |
| }, |
| { |
| "epoch": 1.41, |
| "learning_rate": 1.1619221046045688e-05, |
| "loss": 0.2812, |
| "step": 374 |
| }, |
| { |
| "epoch": 1.42, |
| "learning_rate": 1.1578576573543541e-05, |
| "loss": 0.291, |
| "step": 375 |
| }, |
| { |
| "epoch": 1.42, |
| "learning_rate": 1.1537905339217448e-05, |
| "loss": 0.3162, |
| "step": 376 |
| }, |
| { |
| "epoch": 1.42, |
| "learning_rate": 1.1497208032572385e-05, |
| "loss": 0.2717, |
| "step": 377 |
| }, |
| { |
| "epoch": 1.43, |
| "learning_rate": 1.1456485343555344e-05, |
| "loss": 0.2635, |
| "step": 378 |
| }, |
| { |
| "epoch": 1.43, |
| "learning_rate": 1.141573796254363e-05, |
| "loss": 0.2666, |
| "step": 379 |
| }, |
| { |
| "epoch": 1.43, |
| "learning_rate": 1.1374966580333147e-05, |
| "loss": 0.2942, |
| "step": 380 |
| }, |
| { |
| "epoch": 1.44, |
| "learning_rate": 1.1334171888126698e-05, |
| "loss": 0.2633, |
| "step": 381 |
| }, |
| { |
| "epoch": 1.44, |
| "learning_rate": 1.1293354577522264e-05, |
| "loss": 0.2731, |
| "step": 382 |
| }, |
| { |
| "epoch": 1.45, |
| "learning_rate": 1.1252515340501282e-05, |
| "loss": 0.238, |
| "step": 383 |
| }, |
| { |
| "epoch": 1.45, |
| "learning_rate": 1.1211654869416901e-05, |
| "loss": 0.2789, |
| "step": 384 |
| }, |
| { |
| "epoch": 1.45, |
| "learning_rate": 1.1170773856982268e-05, |
| "loss": 0.2583, |
| "step": 385 |
| }, |
| { |
| "epoch": 1.46, |
| "learning_rate": 1.1129872996258757e-05, |
| "loss": 0.2806, |
| "step": 386 |
| }, |
| { |
| "epoch": 1.46, |
| "learning_rate": 1.1088952980644242e-05, |
| "loss": 0.2756, |
| "step": 387 |
| }, |
| { |
| "epoch": 1.46, |
| "learning_rate": 1.1048014503861321e-05, |
| "loss": 0.2637, |
| "step": 388 |
| }, |
| { |
| "epoch": 1.47, |
| "learning_rate": 1.1007058259945584e-05, |
| "loss": 0.25, |
| "step": 389 |
| }, |
| { |
| "epoch": 1.47, |
| "learning_rate": 1.0966084943233818e-05, |
| "loss": 0.2723, |
| "step": 390 |
| }, |
| { |
| "epoch": 1.48, |
| "learning_rate": 1.092509524835224e-05, |
| "loss": 0.2329, |
| "step": 391 |
| }, |
| { |
| "epoch": 1.48, |
| "learning_rate": 1.0884089870204751e-05, |
| "loss": 0.2628, |
| "step": 392 |
| }, |
| { |
| "epoch": 1.48, |
| "learning_rate": 1.0843069503961112e-05, |
| "loss": 0.2418, |
| "step": 393 |
| }, |
| { |
| "epoch": 1.49, |
| "learning_rate": 1.0802034845045189e-05, |
| "loss": 0.2628, |
| "step": 394 |
| }, |
| { |
| "epoch": 1.49, |
| "learning_rate": 1.0760986589123145e-05, |
| "loss": 0.2655, |
| "step": 395 |
| }, |
| { |
| "epoch": 1.49, |
| "learning_rate": 1.0719925432091671e-05, |
| "loss": 0.2441, |
| "step": 396 |
| }, |
| { |
| "epoch": 1.5, |
| "learning_rate": 1.067885207006616e-05, |
| "loss": 0.2496, |
| "step": 397 |
| }, |
| { |
| "epoch": 1.5, |
| "learning_rate": 1.0637767199368911e-05, |
| "loss": 0.2797, |
| "step": 398 |
| }, |
| { |
| "epoch": 1.51, |
| "learning_rate": 1.0596671516517356e-05, |
| "loss": 0.2412, |
| "step": 399 |
| }, |
| { |
| "epoch": 1.51, |
| "learning_rate": 1.0555565718212198e-05, |
| "loss": 0.2534, |
| "step": 400 |
| }, |
| { |
| "epoch": 1.51, |
| "learning_rate": 1.0514450501325646e-05, |
| "loss": 0.2886, |
| "step": 401 |
| }, |
| { |
| "epoch": 1.52, |
| "learning_rate": 1.0473326562889583e-05, |
| "loss": 0.2683, |
| "step": 402 |
| }, |
| { |
| "epoch": 1.52, |
| "learning_rate": 1.043219460008374e-05, |
| "loss": 0.2771, |
| "step": 403 |
| }, |
| { |
| "epoch": 1.52, |
| "learning_rate": 1.0391055310223899e-05, |
| "loss": 0.2427, |
| "step": 404 |
| }, |
| { |
| "epoch": 1.53, |
| "learning_rate": 1.0349909390750046e-05, |
| "loss": 0.2482, |
| "step": 405 |
| }, |
| { |
| "epoch": 1.53, |
| "learning_rate": 1.0308757539214573e-05, |
| "loss": 0.2629, |
| "step": 406 |
| }, |
| { |
| "epoch": 1.54, |
| "learning_rate": 1.0267600453270422e-05, |
| "loss": 0.2706, |
| "step": 407 |
| }, |
| { |
| "epoch": 1.54, |
| "learning_rate": 1.022643883065929e-05, |
| "loss": 0.2333, |
| "step": 408 |
| }, |
| { |
| "epoch": 1.54, |
| "learning_rate": 1.0185273369199781e-05, |
| "loss": 0.2584, |
| "step": 409 |
| }, |
| { |
| "epoch": 1.55, |
| "learning_rate": 1.0144104766775574e-05, |
| "loss": 0.2981, |
| "step": 410 |
| }, |
| { |
| "epoch": 1.55, |
| "learning_rate": 1.01029337213236e-05, |
| "loss": 0.267, |
| "step": 411 |
| }, |
| { |
| "epoch": 1.55, |
| "learning_rate": 1.006176093082221e-05, |
| "loss": 0.2473, |
| "step": 412 |
| }, |
| { |
| "epoch": 1.56, |
| "learning_rate": 1.0020587093279339e-05, |
| "loss": 0.2633, |
| "step": 413 |
| }, |
| { |
| "epoch": 1.56, |
| "learning_rate": 9.979412906720663e-06, |
| "loss": 0.2687, |
| "step": 414 |
| }, |
| { |
| "epoch": 1.57, |
| "learning_rate": 9.938239069177792e-06, |
| "loss": 0.2538, |
| "step": 415 |
| }, |
| { |
| "epoch": 1.57, |
| "learning_rate": 9.897066278676405e-06, |
| "loss": 0.2701, |
| "step": 416 |
| }, |
| { |
| "epoch": 1.57, |
| "learning_rate": 9.855895233224431e-06, |
| "loss": 0.2322, |
| "step": 417 |
| }, |
| { |
| "epoch": 1.58, |
| "learning_rate": 9.81472663080022e-06, |
| "loss": 0.2654, |
| "step": 418 |
| }, |
| { |
| "epoch": 1.58, |
| "learning_rate": 9.773561169340711e-06, |
| "loss": 0.2854, |
| "step": 419 |
| }, |
| { |
| "epoch": 1.58, |
| "learning_rate": 9.732399546729578e-06, |
| "loss": 0.2571, |
| "step": 420 |
| }, |
| { |
| "epoch": 1.59, |
| "learning_rate": 9.691242460785433e-06, |
| "loss": 0.2462, |
| "step": 421 |
| }, |
| { |
| "epoch": 1.59, |
| "learning_rate": 9.650090609249957e-06, |
| "loss": 0.2491, |
| "step": 422 |
| }, |
| { |
| "epoch": 1.6, |
| "learning_rate": 9.608944689776104e-06, |
| "loss": 0.2666, |
| "step": 423 |
| }, |
| { |
| "epoch": 1.6, |
| "learning_rate": 9.56780539991626e-06, |
| "loss": 0.2628, |
| "step": 424 |
| }, |
| { |
| "epoch": 1.6, |
| "learning_rate": 9.52667343711042e-06, |
| "loss": 0.2853, |
| "step": 425 |
| }, |
| { |
| "epoch": 1.61, |
| "learning_rate": 9.485549498674357e-06, |
| "loss": 0.2388, |
| "step": 426 |
| }, |
| { |
| "epoch": 1.61, |
| "learning_rate": 9.444434281787806e-06, |
| "loss": 0.2714, |
| "step": 427 |
| }, |
| { |
| "epoch": 1.62, |
| "learning_rate": 9.403328483482647e-06, |
| "loss": 0.2351, |
| "step": 428 |
| }, |
| { |
| "epoch": 1.62, |
| "learning_rate": 9.362232800631087e-06, |
| "loss": 0.2605, |
| "step": 429 |
| }, |
| { |
| "epoch": 1.62, |
| "learning_rate": 9.321147929933847e-06, |
| "loss": 0.2599, |
| "step": 430 |
| }, |
| { |
| "epoch": 1.63, |
| "learning_rate": 9.28007456790833e-06, |
| "loss": 0.2548, |
| "step": 431 |
| }, |
| { |
| "epoch": 1.63, |
| "learning_rate": 9.239013410876856e-06, |
| "loss": 0.2678, |
| "step": 432 |
| }, |
| { |
| "epoch": 1.63, |
| "learning_rate": 9.197965154954815e-06, |
| "loss": 0.2654, |
| "step": 433 |
| }, |
| { |
| "epoch": 1.64, |
| "learning_rate": 9.15693049603889e-06, |
| "loss": 0.2583, |
| "step": 434 |
| }, |
| { |
| "epoch": 1.64, |
| "learning_rate": 9.115910129795252e-06, |
| "loss": 0.264, |
| "step": 435 |
| }, |
| { |
| "epoch": 1.65, |
| "learning_rate": 9.074904751647762e-06, |
| "loss": 0.2699, |
| "step": 436 |
| }, |
| { |
| "epoch": 1.65, |
| "learning_rate": 9.033915056766187e-06, |
| "loss": 0.2476, |
| "step": 437 |
| }, |
| { |
| "epoch": 1.65, |
| "learning_rate": 8.992941740054418e-06, |
| "loss": 0.2108, |
| "step": 438 |
| }, |
| { |
| "epoch": 1.66, |
| "learning_rate": 8.951985496138679e-06, |
| "loss": 0.2362, |
| "step": 439 |
| }, |
| { |
| "epoch": 1.66, |
| "learning_rate": 8.911047019355763e-06, |
| "loss": 0.2617, |
| "step": 440 |
| }, |
| { |
| "epoch": 1.66, |
| "learning_rate": 8.870127003741245e-06, |
| "loss": 0.2454, |
| "step": 441 |
| }, |
| { |
| "epoch": 1.67, |
| "learning_rate": 8.829226143017735e-06, |
| "loss": 0.2444, |
| "step": 442 |
| }, |
| { |
| "epoch": 1.67, |
| "learning_rate": 8.788345130583099e-06, |
| "loss": 0.2322, |
| "step": 443 |
| }, |
| { |
| "epoch": 1.68, |
| "learning_rate": 8.747484659498725e-06, |
| "loss": 0.2427, |
| "step": 444 |
| }, |
| { |
| "epoch": 1.68, |
| "learning_rate": 8.706645422477739e-06, |
| "loss": 0.2528, |
| "step": 445 |
| }, |
| { |
| "epoch": 1.68, |
| "learning_rate": 8.665828111873307e-06, |
| "loss": 0.2742, |
| "step": 446 |
| }, |
| { |
| "epoch": 1.69, |
| "learning_rate": 8.625033419666858e-06, |
| "loss": 0.2197, |
| "step": 447 |
| }, |
| { |
| "epoch": 1.69, |
| "learning_rate": 8.584262037456374e-06, |
| "loss": 0.244, |
| "step": 448 |
| }, |
| { |
| "epoch": 1.69, |
| "learning_rate": 8.54351465644466e-06, |
| "loss": 0.2678, |
| "step": 449 |
| }, |
| { |
| "epoch": 1.7, |
| "learning_rate": 8.50279196742762e-06, |
| "loss": 0.2372, |
| "step": 450 |
| }, |
| { |
| "epoch": 1.7, |
| "learning_rate": 8.462094660782555e-06, |
| "loss": 0.2557, |
| "step": 451 |
| }, |
| { |
| "epoch": 1.71, |
| "learning_rate": 8.42142342645646e-06, |
| "loss": 0.2335, |
| "step": 452 |
| }, |
| { |
| "epoch": 1.71, |
| "learning_rate": 8.380778953954314e-06, |
| "loss": 0.2389, |
| "step": 453 |
| }, |
| { |
| "epoch": 1.71, |
| "learning_rate": 8.340161932327405e-06, |
| "loss": 0.2683, |
| "step": 454 |
| }, |
| { |
| "epoch": 1.72, |
| "learning_rate": 8.299573050161637e-06, |
| "loss": 0.2378, |
| "step": 455 |
| }, |
| { |
| "epoch": 1.72, |
| "learning_rate": 8.259012995565868e-06, |
| "loss": 0.2603, |
| "step": 456 |
| }, |
| { |
| "epoch": 1.72, |
| "learning_rate": 8.21848245616023e-06, |
| "loss": 0.2318, |
| "step": 457 |
| }, |
| { |
| "epoch": 1.73, |
| "learning_rate": 8.177982119064497e-06, |
| "loss": 0.2061, |
| "step": 458 |
| }, |
| { |
| "epoch": 1.73, |
| "learning_rate": 8.137512670886397e-06, |
| "loss": 0.2288, |
| "step": 459 |
| }, |
| { |
| "epoch": 1.74, |
| "learning_rate": 8.097074797710007e-06, |
| "loss": 0.2244, |
| "step": 460 |
| }, |
| { |
| "epoch": 1.74, |
| "learning_rate": 8.056669185084108e-06, |
| "loss": 0.2323, |
| "step": 461 |
| }, |
| { |
| "epoch": 1.74, |
| "learning_rate": 8.016296518010558e-06, |
| "loss": 0.2546, |
| "step": 462 |
| }, |
| { |
| "epoch": 1.75, |
| "learning_rate": 7.975957480932695e-06, |
| "loss": 0.2544, |
| "step": 463 |
| }, |
| { |
| "epoch": 1.75, |
| "learning_rate": 7.935652757723709e-06, |
| "loss": 0.2522, |
| "step": 464 |
| }, |
| { |
| "epoch": 1.75, |
| "learning_rate": 7.895383031675074e-06, |
| "loss": 0.2426, |
| "step": 465 |
| }, |
| { |
| "epoch": 1.76, |
| "learning_rate": 7.855148985484946e-06, |
| "loss": 0.225, |
| "step": 466 |
| }, |
| { |
| "epoch": 1.76, |
| "learning_rate": 7.814951301246598e-06, |
| "loss": 0.2477, |
| "step": 467 |
| }, |
| { |
| "epoch": 1.77, |
| "learning_rate": 7.774790660436857e-06, |
| "loss": 0.2391, |
| "step": 468 |
| }, |
| { |
| "epoch": 1.77, |
| "learning_rate": 7.73466774390454e-06, |
| "loss": 0.2243, |
| "step": 469 |
| }, |
| { |
| "epoch": 1.77, |
| "learning_rate": 7.694583231858921e-06, |
| "loss": 0.2365, |
| "step": 470 |
| }, |
| { |
| "epoch": 1.78, |
| "learning_rate": 7.654537803858205e-06, |
| "loss": 0.2336, |
| "step": 471 |
| }, |
| { |
| "epoch": 1.78, |
| "learning_rate": 7.614532138797994e-06, |
| "loss": 0.2428, |
| "step": 472 |
| }, |
| { |
| "epoch": 1.78, |
| "learning_rate": 7.574566914899779e-06, |
| "loss": 0.2498, |
| "step": 473 |
| }, |
| { |
| "epoch": 1.79, |
| "learning_rate": 7.534642809699455e-06, |
| "loss": 0.2444, |
| "step": 474 |
| }, |
| { |
| "epoch": 1.79, |
| "learning_rate": 7.494760500035824e-06, |
| "loss": 0.2194, |
| "step": 475 |
| }, |
| { |
| "epoch": 1.8, |
| "learning_rate": 7.454920662039118e-06, |
| "loss": 0.2544, |
| "step": 476 |
| }, |
| { |
| "epoch": 1.8, |
| "learning_rate": 7.415123971119549e-06, |
| "loss": 0.2506, |
| "step": 477 |
| }, |
| { |
| "epoch": 1.8, |
| "learning_rate": 7.375371101955842e-06, |
| "loss": 0.2413, |
| "step": 478 |
| }, |
| { |
| "epoch": 1.81, |
| "learning_rate": 7.335662728483808e-06, |
| "loss": 0.2262, |
| "step": 479 |
| }, |
| { |
| "epoch": 1.81, |
| "learning_rate": 7.295999523884921e-06, |
| "loss": 0.2096, |
| "step": 480 |
| }, |
| { |
| "epoch": 1.82, |
| "learning_rate": 7.256382160574902e-06, |
| "loss": 0.2446, |
| "step": 481 |
| }, |
| { |
| "epoch": 1.82, |
| "learning_rate": 7.2168113101923085e-06, |
| "loss": 0.2427, |
| "step": 482 |
| }, |
| { |
| "epoch": 1.82, |
| "learning_rate": 7.177287643587164e-06, |
| "loss": 0.2389, |
| "step": 483 |
| }, |
| { |
| "epoch": 1.83, |
| "learning_rate": 7.1378118308095835e-06, |
| "loss": 0.2334, |
| "step": 484 |
| }, |
| { |
| "epoch": 1.83, |
| "learning_rate": 7.0983845410984e-06, |
| "loss": 0.2528, |
| "step": 485 |
| }, |
| { |
| "epoch": 1.83, |
| "learning_rate": 7.05900644286984e-06, |
| "loss": 0.2317, |
| "step": 486 |
| }, |
| { |
| "epoch": 1.84, |
| "learning_rate": 7.019678203706164e-06, |
| "loss": 0.2352, |
| "step": 487 |
| }, |
| { |
| "epoch": 1.84, |
| "learning_rate": 6.980400490344383e-06, |
| "loss": 0.2289, |
| "step": 488 |
| }, |
| { |
| "epoch": 1.85, |
| "learning_rate": 6.941173968664923e-06, |
| "loss": 0.2273, |
| "step": 489 |
| }, |
| { |
| "epoch": 1.85, |
| "learning_rate": 6.901999303680359e-06, |
| "loss": 0.1824, |
| "step": 490 |
| }, |
| { |
| "epoch": 1.85, |
| "learning_rate": 6.862877159524133e-06, |
| "loss": 0.2325, |
| "step": 491 |
| }, |
| { |
| "epoch": 1.86, |
| "learning_rate": 6.8238081994392836e-06, |
| "loss": 0.2399, |
| "step": 492 |
| }, |
| { |
| "epoch": 1.86, |
| "learning_rate": 6.7847930857672205e-06, |
| "loss": 0.2269, |
| "step": 493 |
| }, |
| { |
| "epoch": 1.86, |
| "learning_rate": 6.745832479936492e-06, |
| "loss": 0.2161, |
| "step": 494 |
| }, |
| { |
| "epoch": 1.87, |
| "learning_rate": 6.706927042451561e-06, |
| "loss": 0.2291, |
| "step": 495 |
| }, |
| { |
| "epoch": 1.87, |
| "learning_rate": 6.66807743288161e-06, |
| "loss": 0.2207, |
| "step": 496 |
| }, |
| { |
| "epoch": 1.88, |
| "learning_rate": 6.629284309849373e-06, |
| "loss": 0.2136, |
| "step": 497 |
| }, |
| { |
| "epoch": 1.88, |
| "learning_rate": 6.590548331019957e-06, |
| "loss": 0.2098, |
| "step": 498 |
| }, |
| { |
| "epoch": 1.88, |
| "learning_rate": 6.55187015308969e-06, |
| "loss": 0.2432, |
| "step": 499 |
| }, |
| { |
| "epoch": 1.89, |
| "learning_rate": 6.513250431775003e-06, |
| "loss": 0.2526, |
| "step": 500 |
| }, |
| { |
| "epoch": 1.89, |
| "eval_loss": 0.2714967727661133, |
| "eval_runtime": 477.7572, |
| "eval_samples_per_second": 1.448, |
| "eval_steps_per_second": 0.362, |
| "step": 500 |
| }, |
| { |
| "epoch": 1.89, |
| "learning_rate": 6.474689821801295e-06, |
| "loss": 0.2352, |
| "step": 501 |
| }, |
| { |
| "epoch": 1.89, |
| "learning_rate": 6.436188976891846e-06, |
| "loss": 0.2379, |
| "step": 502 |
| }, |
| { |
| "epoch": 1.9, |
| "learning_rate": 6.39774854975673e-06, |
| "loss": 0.217, |
| "step": 503 |
| }, |
| { |
| "epoch": 1.9, |
| "learning_rate": 6.359369192081756e-06, |
| "loss": 0.2463, |
| "step": 504 |
| }, |
| { |
| "epoch": 1.91, |
| "learning_rate": 6.321051554517406e-06, |
| "loss": 0.241, |
| "step": 505 |
| }, |
| { |
| "epoch": 1.91, |
| "learning_rate": 6.282796286667814e-06, |
| "loss": 0.2594, |
| "step": 506 |
| }, |
| { |
| "epoch": 1.91, |
| "learning_rate": 6.244604037079754e-06, |
| "loss": 0.2195, |
| "step": 507 |
| }, |
| { |
| "epoch": 1.92, |
| "learning_rate": 6.206475453231644e-06, |
| "loss": 0.2463, |
| "step": 508 |
| }, |
| { |
| "epoch": 1.92, |
| "learning_rate": 6.168411181522569e-06, |
| "loss": 0.239, |
| "step": 509 |
| }, |
| { |
| "epoch": 1.92, |
| "learning_rate": 6.13041186726131e-06, |
| "loss": 0.2115, |
| "step": 510 |
| }, |
| { |
| "epoch": 1.93, |
| "learning_rate": 6.092478154655431e-06, |
| "loss": 0.1979, |
| "step": 511 |
| }, |
| { |
| "epoch": 1.93, |
| "learning_rate": 6.054610686800333e-06, |
| "loss": 0.215, |
| "step": 512 |
| }, |
| { |
| "epoch": 1.94, |
| "learning_rate": 6.016810105668365e-06, |
| "loss": 0.2428, |
| "step": 513 |
| }, |
| { |
| "epoch": 1.94, |
| "learning_rate": 5.979077052097936e-06, |
| "loss": 0.2042, |
| "step": 514 |
| }, |
| { |
| "epoch": 1.94, |
| "learning_rate": 5.941412165782645e-06, |
| "loss": 0.2126, |
| "step": 515 |
| }, |
| { |
| "epoch": 1.95, |
| "learning_rate": 5.903816085260447e-06, |
| "loss": 0.223, |
| "step": 516 |
| }, |
| { |
| "epoch": 1.95, |
| "learning_rate": 5.866289447902829e-06, |
| "loss": 0.2163, |
| "step": 517 |
| }, |
| { |
| "epoch": 1.95, |
| "learning_rate": 5.828832889903983e-06, |
| "loss": 0.2302, |
| "step": 518 |
| }, |
| { |
| "epoch": 1.96, |
| "learning_rate": 5.791447046270055e-06, |
| "loss": 0.2329, |
| "step": 519 |
| }, |
| { |
| "epoch": 1.96, |
| "learning_rate": 5.754132550808345e-06, |
| "loss": 0.1956, |
| "step": 520 |
| }, |
| { |
| "epoch": 1.97, |
| "learning_rate": 5.716890036116582e-06, |
| "loss": 0.2402, |
| "step": 521 |
| }, |
| { |
| "epoch": 1.97, |
| "learning_rate": 5.6797201335722064e-06, |
| "loss": 0.2158, |
| "step": 522 |
| }, |
| { |
| "epoch": 1.97, |
| "learning_rate": 5.642623473321638e-06, |
| "loss": 0.2141, |
| "step": 523 |
| }, |
| { |
| "epoch": 1.98, |
| "learning_rate": 5.6056006842696145e-06, |
| "loss": 0.2346, |
| "step": 524 |
| }, |
| { |
| "epoch": 1.98, |
| "learning_rate": 5.568652394068532e-06, |
| "loss": 0.1821, |
| "step": 525 |
| }, |
| { |
| "epoch": 1.98, |
| "learning_rate": 5.531779229107797e-06, |
| "loss": 0.21, |
| "step": 526 |
| }, |
| { |
| "epoch": 1.99, |
| "learning_rate": 5.494981814503199e-06, |
| "loss": 0.1902, |
| "step": 527 |
| }, |
| { |
| "epoch": 1.99, |
| "learning_rate": 5.458260774086332e-06, |
| "loss": 0.1962, |
| "step": 528 |
| }, |
| { |
| "epoch": 2.0, |
| "learning_rate": 5.421616730394e-06, |
| "loss": 0.2277, |
| "step": 529 |
| }, |
| { |
| "epoch": 2.0, |
| "learning_rate": 5.38505030465767e-06, |
| "loss": 0.2076, |
| "step": 530 |
| }, |
| { |
| "epoch": 2.0, |
| "learning_rate": 5.348562116792946e-06, |
| "loss": 0.2212, |
| "step": 531 |
| }, |
| { |
| "epoch": 2.01, |
| "learning_rate": 5.312152785389056e-06, |
| "loss": 0.1982, |
| "step": 532 |
| }, |
| { |
| "epoch": 2.01, |
| "learning_rate": 5.275822927698362e-06, |
| "loss": 0.2003, |
| "step": 533 |
| }, |
| { |
| "epoch": 2.02, |
| "learning_rate": 5.2395731596258925e-06, |
| "loss": 0.184, |
| "step": 534 |
| }, |
| { |
| "epoch": 2.02, |
| "learning_rate": 5.20340409571891e-06, |
| "loss": 0.1873, |
| "step": 535 |
| }, |
| { |
| "epoch": 2.02, |
| "learning_rate": 5.167316349156495e-06, |
| "loss": 0.1909, |
| "step": 536 |
| }, |
| { |
| "epoch": 2.03, |
| "learning_rate": 5.131310531739129e-06, |
| "loss": 0.1779, |
| "step": 537 |
| }, |
| { |
| "epoch": 2.03, |
| "learning_rate": 5.095387253878346e-06, |
| "loss": 0.1965, |
| "step": 538 |
| }, |
| { |
| "epoch": 2.03, |
| "learning_rate": 5.0595471245863745e-06, |
| "loss": 0.184, |
| "step": 539 |
| }, |
| { |
| "epoch": 2.04, |
| "learning_rate": 5.023790751465818e-06, |
| "loss": 0.1933, |
| "step": 540 |
| }, |
| { |
| "epoch": 2.04, |
| "learning_rate": 4.9881187406993455e-06, |
| "loss": 0.158, |
| "step": 541 |
| }, |
| { |
| "epoch": 2.05, |
| "learning_rate": 4.952531697039424e-06, |
| "loss": 0.1793, |
| "step": 542 |
| }, |
| { |
| "epoch": 2.05, |
| "learning_rate": 4.917030223798057e-06, |
| "loss": 0.1794, |
| "step": 543 |
| }, |
| { |
| "epoch": 2.05, |
| "learning_rate": 4.881614922836555e-06, |
| "loss": 0.1873, |
| "step": 544 |
| }, |
| { |
| "epoch": 2.06, |
| "learning_rate": 4.846286394555352e-06, |
| "loss": 0.1944, |
| "step": 545 |
| }, |
| { |
| "epoch": 2.06, |
| "learning_rate": 4.811045237883803e-06, |
| "loss": 0.1942, |
| "step": 546 |
| }, |
| { |
| "epoch": 2.06, |
| "learning_rate": 4.775892050270051e-06, |
| "loss": 0.1942, |
| "step": 547 |
| }, |
| { |
| "epoch": 2.07, |
| "learning_rate": 4.740827427670871e-06, |
| "loss": 0.183, |
| "step": 548 |
| }, |
| { |
| "epoch": 2.07, |
| "learning_rate": 4.7058519645416004e-06, |
| "loss": 0.1706, |
| "step": 549 |
| }, |
| { |
| "epoch": 2.08, |
| "learning_rate": 4.670966253826027e-06, |
| "loss": 0.1782, |
| "step": 550 |
| }, |
| { |
| "epoch": 2.08, |
| "learning_rate": 4.636170886946371e-06, |
| "loss": 0.215, |
| "step": 551 |
| }, |
| { |
| "epoch": 2.08, |
| "learning_rate": 4.601466453793224e-06, |
| "loss": 0.1782, |
| "step": 552 |
| }, |
| { |
| "epoch": 2.09, |
| "learning_rate": 4.5668535427155816e-06, |
| "loss": 0.1744, |
| "step": 553 |
| }, |
| { |
| "epoch": 2.09, |
| "learning_rate": 4.532332740510842e-06, |
| "loss": 0.1895, |
| "step": 554 |
| }, |
| { |
| "epoch": 2.09, |
| "learning_rate": 4.497904632414879e-06, |
| "loss": 0.1639, |
| "step": 555 |
| }, |
| { |
| "epoch": 2.1, |
| "learning_rate": 4.4635698020921016e-06, |
| "loss": 0.1809, |
| "step": 556 |
| }, |
| { |
| "epoch": 2.1, |
| "learning_rate": 4.429328831625565e-06, |
| "loss": 0.183, |
| "step": 557 |
| }, |
| { |
| "epoch": 2.11, |
| "learning_rate": 4.3951823015071186e-06, |
| "loss": 0.1825, |
| "step": 558 |
| }, |
| { |
| "epoch": 2.11, |
| "learning_rate": 4.361130790627541e-06, |
| "loss": 0.2015, |
| "step": 559 |
| }, |
| { |
| "epoch": 2.11, |
| "learning_rate": 4.327174876266743e-06, |
| "loss": 0.1918, |
| "step": 560 |
| }, |
| { |
| "epoch": 2.12, |
| "learning_rate": 4.293315134083975e-06, |
| "loss": 0.1882, |
| "step": 561 |
| }, |
| { |
| "epoch": 2.12, |
| "learning_rate": 4.259552138108061e-06, |
| "loss": 0.215, |
| "step": 562 |
| }, |
| { |
| "epoch": 2.12, |
| "learning_rate": 4.225886460727671e-06, |
| "loss": 0.1973, |
| "step": 563 |
| }, |
| { |
| "epoch": 2.13, |
| "learning_rate": 4.192318672681631e-06, |
| "loss": 0.1719, |
| "step": 564 |
| }, |
| { |
| "epoch": 2.13, |
| "learning_rate": 4.158849343049233e-06, |
| "loss": 0.1976, |
| "step": 565 |
| }, |
| { |
| "epoch": 2.14, |
| "learning_rate": 4.12547903924058e-06, |
| "loss": 0.2042, |
| "step": 566 |
| }, |
| { |
| "epoch": 2.14, |
| "learning_rate": 4.092208326986986e-06, |
| "loss": 0.1572, |
| "step": 567 |
| }, |
| { |
| "epoch": 2.14, |
| "learning_rate": 4.059037770331379e-06, |
| "loss": 0.1841, |
| "step": 568 |
| }, |
| { |
| "epoch": 2.15, |
| "learning_rate": 4.025967931618722e-06, |
| "loss": 0.1952, |
| "step": 569 |
| }, |
| { |
| "epoch": 2.15, |
| "learning_rate": 3.992999371486508e-06, |
| "loss": 0.1924, |
| "step": 570 |
| }, |
| { |
| "epoch": 2.15, |
| "learning_rate": 3.960132648855226e-06, |
| "loss": 0.1885, |
| "step": 571 |
| }, |
| { |
| "epoch": 2.16, |
| "learning_rate": 3.9273683209189115e-06, |
| "loss": 0.1849, |
| "step": 572 |
| }, |
| { |
| "epoch": 2.16, |
| "learning_rate": 3.894706943135686e-06, |
| "loss": 0.1721, |
| "step": 573 |
| }, |
| { |
| "epoch": 2.17, |
| "learning_rate": 3.862149069218343e-06, |
| "loss": 0.181, |
| "step": 574 |
| }, |
| { |
| "epoch": 2.17, |
| "learning_rate": 3.829695251124953e-06, |
| "loss": 0.2257, |
| "step": 575 |
| }, |
| { |
| "epoch": 2.17, |
| "learning_rate": 3.797346039049529e-06, |
| "loss": 0.1974, |
| "step": 576 |
| }, |
| { |
| "epoch": 2.18, |
| "learning_rate": 3.7651019814126656e-06, |
| "loss": 0.1716, |
| "step": 577 |
| }, |
| { |
| "epoch": 2.18, |
| "learning_rate": 3.732963624852275e-06, |
| "loss": 0.1873, |
| "step": 578 |
| }, |
| { |
| "epoch": 2.18, |
| "learning_rate": 3.7009315142143033e-06, |
| "loss": 0.1727, |
| "step": 579 |
| }, |
| { |
| "epoch": 2.19, |
| "learning_rate": 3.6690061925434817e-06, |
| "loss": 0.2046, |
| "step": 580 |
| }, |
| { |
| "epoch": 2.19, |
| "learning_rate": 3.637188201074149e-06, |
| "loss": 0.1947, |
| "step": 581 |
| }, |
| { |
| "epoch": 2.2, |
| "learning_rate": 3.6054780792210542e-06, |
| "loss": 0.1555, |
| "step": 582 |
| }, |
| { |
| "epoch": 2.2, |
| "learning_rate": 3.5738763645702145e-06, |
| "loss": 0.1697, |
| "step": 583 |
| }, |
| { |
| "epoch": 2.2, |
| "learning_rate": 3.5423835928698126e-06, |
| "loss": 0.2059, |
| "step": 584 |
| }, |
| { |
| "epoch": 2.21, |
| "learning_rate": 3.511000298021098e-06, |
| "loss": 0.1851, |
| "step": 585 |
| }, |
| { |
| "epoch": 2.21, |
| "learning_rate": 3.479727012069349e-06, |
| "loss": 0.1702, |
| "step": 586 |
| }, |
| { |
| "epoch": 2.22, |
| "learning_rate": 3.4485642651948516e-06, |
| "loss": 0.1691, |
| "step": 587 |
| }, |
| { |
| "epoch": 2.22, |
| "learning_rate": 3.4175125857039027e-06, |
| "loss": 0.1613, |
| "step": 588 |
| }, |
| { |
| "epoch": 2.22, |
| "learning_rate": 3.38657250001986e-06, |
| "loss": 0.1934, |
| "step": 589 |
| }, |
| { |
| "epoch": 2.23, |
| "learning_rate": 3.355744532674211e-06, |
| "loss": 0.1686, |
| "step": 590 |
| }, |
| { |
| "epoch": 2.23, |
| "learning_rate": 3.325029206297694e-06, |
| "loss": 0.1669, |
| "step": 591 |
| }, |
| { |
| "epoch": 2.23, |
| "learning_rate": 3.2944270416114256e-06, |
| "loss": 0.1832, |
| "step": 592 |
| }, |
| { |
| "epoch": 2.24, |
| "learning_rate": 3.2639385574180825e-06, |
| "loss": 0.1459, |
| "step": 593 |
| }, |
| { |
| "epoch": 2.24, |
| "learning_rate": 3.2335642705930883e-06, |
| "loss": 0.1718, |
| "step": 594 |
| }, |
| { |
| "epoch": 2.25, |
| "learning_rate": 3.2033046960758763e-06, |
| "loss": 0.1847, |
| "step": 595 |
| }, |
| { |
| "epoch": 2.25, |
| "learning_rate": 3.173160346861134e-06, |
| "loss": 0.1511, |
| "step": 596 |
| }, |
| { |
| "epoch": 2.25, |
| "learning_rate": 3.1431317339901267e-06, |
| "loss": 0.18, |
| "step": 597 |
| }, |
| { |
| "epoch": 2.26, |
| "learning_rate": 3.1132193665420306e-06, |
| "loss": 0.1697, |
| "step": 598 |
| }, |
| { |
| "epoch": 2.26, |
| "learning_rate": 3.0834237516252817e-06, |
| "loss": 0.1908, |
| "step": 599 |
| }, |
| { |
| "epoch": 2.26, |
| "learning_rate": 3.0537453943690076e-06, |
| "loss": 0.166, |
| "step": 600 |
| }, |
| { |
| "epoch": 2.27, |
| "learning_rate": 3.024184797914449e-06, |
| "loss": 0.174, |
| "step": 601 |
| }, |
| { |
| "epoch": 2.27, |
| "learning_rate": 2.994742463406427e-06, |
| "loss": 0.1539, |
| "step": 602 |
| }, |
| { |
| "epoch": 2.28, |
| "learning_rate": 2.96541888998485e-06, |
| "loss": 0.1866, |
| "step": 603 |
| }, |
| { |
| "epoch": 2.28, |
| "learning_rate": 2.9362145747762626e-06, |
| "loss": 0.1911, |
| "step": 604 |
| }, |
| { |
| "epoch": 2.28, |
| "learning_rate": 2.9071300128854007e-06, |
| "loss": 0.1734, |
| "step": 605 |
| }, |
| { |
| "epoch": 2.29, |
| "learning_rate": 2.878165697386812e-06, |
| "loss": 0.1763, |
| "step": 606 |
| }, |
| { |
| "epoch": 2.29, |
| "learning_rate": 2.8493221193164886e-06, |
| "loss": 0.1606, |
| "step": 607 |
| }, |
| { |
| "epoch": 2.29, |
| "learning_rate": 2.82059976766354e-06, |
| "loss": 0.1938, |
| "step": 608 |
| }, |
| { |
| "epoch": 2.3, |
| "learning_rate": 2.79199912936191e-06, |
| "loss": 0.1426, |
| "step": 609 |
| }, |
| { |
| "epoch": 2.3, |
| "learning_rate": 2.763520689282122e-06, |
| "loss": 0.1766, |
| "step": 610 |
| }, |
| { |
| "epoch": 2.31, |
| "learning_rate": 2.7351649302230553e-06, |
| "loss": 0.1729, |
| "step": 611 |
| }, |
| { |
| "epoch": 2.31, |
| "learning_rate": 2.7069323329037632e-06, |
| "loss": 0.1669, |
| "step": 612 |
| }, |
| { |
| "epoch": 2.31, |
| "learning_rate": 2.678823375955314e-06, |
| "loss": 0.1874, |
| "step": 613 |
| }, |
| { |
| "epoch": 2.32, |
| "learning_rate": 2.650838535912692e-06, |
| "loss": 0.1873, |
| "step": 614 |
| }, |
| { |
| "epoch": 2.32, |
| "learning_rate": 2.6229782872067042e-06, |
| "loss": 0.1581, |
| "step": 615 |
| }, |
| { |
| "epoch": 2.32, |
| "learning_rate": 2.595243102155951e-06, |
| "loss": 0.1828, |
| "step": 616 |
| }, |
| { |
| "epoch": 2.33, |
| "learning_rate": 2.567633450958801e-06, |
| "loss": 0.1967, |
| "step": 617 |
| }, |
| { |
| "epoch": 2.33, |
| "learning_rate": 2.540149801685441e-06, |
| "loss": 0.1593, |
| "step": 618 |
| }, |
| { |
| "epoch": 2.34, |
| "learning_rate": 2.512792620269924e-06, |
| "loss": 0.1614, |
| "step": 619 |
| }, |
| { |
| "epoch": 2.34, |
| "learning_rate": 2.485562370502279e-06, |
| "loss": 0.1665, |
| "step": 620 |
| }, |
| { |
| "epoch": 2.34, |
| "learning_rate": 2.4584595140206457e-06, |
| "loss": 0.1723, |
| "step": 621 |
| }, |
| { |
| "epoch": 2.35, |
| "learning_rate": 2.4314845103034456e-06, |
| "loss": 0.1884, |
| "step": 622 |
| }, |
| { |
| "epoch": 2.35, |
| "learning_rate": 2.40463781666159e-06, |
| "loss": 0.202, |
| "step": 623 |
| }, |
| { |
| "epoch": 2.35, |
| "learning_rate": 2.3779198882307443e-06, |
| "loss": 0.1616, |
| "step": 624 |
| }, |
| { |
| "epoch": 2.36, |
| "learning_rate": 2.3513311779635904e-06, |
| "loss": 0.2026, |
| "step": 625 |
| }, |
| { |
| "epoch": 2.36, |
| "learning_rate": 2.324872136622164e-06, |
| "loss": 0.1823, |
| "step": 626 |
| }, |
| { |
| "epoch": 2.37, |
| "learning_rate": 2.2985432127701945e-06, |
| "loss": 0.1833, |
| "step": 627 |
| }, |
| { |
| "epoch": 2.37, |
| "learning_rate": 2.2723448527655267e-06, |
| "loss": 0.1729, |
| "step": 628 |
| }, |
| { |
| "epoch": 2.37, |
| "learning_rate": 2.246277500752524e-06, |
| "loss": 0.1671, |
| "step": 629 |
| }, |
| { |
| "epoch": 2.38, |
| "learning_rate": 2.220341598654565e-06, |
| "loss": 0.1648, |
| "step": 630 |
| }, |
| { |
| "epoch": 2.38, |
| "learning_rate": 2.194537586166532e-06, |
| "loss": 0.1545, |
| "step": 631 |
| }, |
| { |
| "epoch": 2.38, |
| "learning_rate": 2.16886590074737e-06, |
| "loss": 0.1763, |
| "step": 632 |
| }, |
| { |
| "epoch": 2.39, |
| "learning_rate": 2.143326977612662e-06, |
| "loss": 0.1793, |
| "step": 633 |
| }, |
| { |
| "epoch": 2.39, |
| "learning_rate": 2.1179212497272582e-06, |
| "loss": 0.1722, |
| "step": 634 |
| }, |
| { |
| "epoch": 2.4, |
| "learning_rate": 2.0926491477979272e-06, |
| "loss": 0.1847, |
| "step": 635 |
| }, |
| { |
| "epoch": 2.4, |
| "learning_rate": 2.0675111002660566e-06, |
| "loss": 0.1674, |
| "step": 636 |
| }, |
| { |
| "epoch": 2.4, |
| "learning_rate": 2.042507533300395e-06, |
| "loss": 0.1815, |
| "step": 637 |
| }, |
| { |
| "epoch": 2.41, |
| "learning_rate": 2.0176388707898274e-06, |
| "loss": 0.1676, |
| "step": 638 |
| }, |
| { |
| "epoch": 2.41, |
| "learning_rate": 1.992905534336178e-06, |
| "loss": 0.1753, |
| "step": 639 |
| }, |
| { |
| "epoch": 2.42, |
| "learning_rate": 1.9683079432470774e-06, |
| "loss": 0.1696, |
| "step": 640 |
| }, |
| { |
| "epoch": 2.42, |
| "learning_rate": 1.9438465145288377e-06, |
| "loss": 0.189, |
| "step": 641 |
| }, |
| { |
| "epoch": 2.42, |
| "learning_rate": 1.9195216628793956e-06, |
| "loss": 0.1935, |
| "step": 642 |
| }, |
| { |
| "epoch": 2.43, |
| "learning_rate": 1.8953338006812805e-06, |
| "loss": 0.1814, |
| "step": 643 |
| }, |
| { |
| "epoch": 2.43, |
| "learning_rate": 1.8712833379946217e-06, |
| "loss": 0.1982, |
| "step": 644 |
| }, |
| { |
| "epoch": 2.43, |
| "learning_rate": 1.847370682550187e-06, |
| "loss": 0.182, |
| "step": 645 |
| }, |
| { |
| "epoch": 2.44, |
| "learning_rate": 1.82359623974249e-06, |
| "loss": 0.1696, |
| "step": 646 |
| }, |
| { |
| "epoch": 2.44, |
| "learning_rate": 1.7999604126229043e-06, |
| "loss": 0.1931, |
| "step": 647 |
| }, |
| { |
| "epoch": 2.45, |
| "learning_rate": 1.7764636018928249e-06, |
| "loss": 0.1597, |
| "step": 648 |
| }, |
| { |
| "epoch": 2.45, |
| "learning_rate": 1.753106205896895e-06, |
| "loss": 0.1479, |
| "step": 649 |
| }, |
| { |
| "epoch": 2.45, |
| "learning_rate": 1.729888620616228e-06, |
| "loss": 0.148, |
| "step": 650 |
| }, |
| { |
| "epoch": 2.46, |
| "learning_rate": 1.7068112396617164e-06, |
| "loss": 0.1808, |
| "step": 651 |
| }, |
| { |
| "epoch": 2.46, |
| "learning_rate": 1.6838744542673492e-06, |
| "loss": 0.1599, |
| "step": 652 |
| }, |
| { |
| "epoch": 2.46, |
| "learning_rate": 1.6610786532835776e-06, |
| "loss": 0.1671, |
| "step": 653 |
| }, |
| { |
| "epoch": 2.47, |
| "learning_rate": 1.6384242231707203e-06, |
| "loss": 0.1345, |
| "step": 654 |
| }, |
| { |
| "epoch": 2.47, |
| "learning_rate": 1.6159115479924259e-06, |
| "loss": 0.1743, |
| "step": 655 |
| }, |
| { |
| "epoch": 2.48, |
| "learning_rate": 1.593541009409143e-06, |
| "loss": 0.1677, |
| "step": 656 |
| }, |
| { |
| "epoch": 2.48, |
| "learning_rate": 1.5713129866716647e-06, |
| "loss": 0.1418, |
| "step": 657 |
| }, |
| { |
| "epoch": 2.48, |
| "learning_rate": 1.5492278566146945e-06, |
| "loss": 0.1627, |
| "step": 658 |
| }, |
| { |
| "epoch": 2.49, |
| "learning_rate": 1.5272859936504513e-06, |
| "loss": 0.1649, |
| "step": 659 |
| }, |
| { |
| "epoch": 2.49, |
| "learning_rate": 1.5054877697623305e-06, |
| "loss": 0.1418, |
| "step": 660 |
| }, |
| { |
| "epoch": 2.49, |
| "learning_rate": 1.4838335544985982e-06, |
| "loss": 0.1819, |
| "step": 661 |
| }, |
| { |
| "epoch": 2.5, |
| "learning_rate": 1.462323714966114e-06, |
| "loss": 0.1536, |
| "step": 662 |
| }, |
| { |
| "epoch": 2.5, |
| "learning_rate": 1.4409586158241272e-06, |
| "loss": 0.1654, |
| "step": 663 |
| }, |
| { |
| "epoch": 2.51, |
| "learning_rate": 1.4197386192780715e-06, |
| "loss": 0.16, |
| "step": 664 |
| }, |
| { |
| "epoch": 2.51, |
| "learning_rate": 1.3986640850734444e-06, |
| "loss": 0.1573, |
| "step": 665 |
| }, |
| { |
| "epoch": 2.51, |
| "learning_rate": 1.3777353704897002e-06, |
| "loss": 0.1831, |
| "step": 666 |
| }, |
| { |
| "epoch": 2.52, |
| "learning_rate": 1.3569528303341927e-06, |
| "loss": 0.1633, |
| "step": 667 |
| }, |
| { |
| "epoch": 2.52, |
| "learning_rate": 1.3363168169361574e-06, |
| "loss": 0.1652, |
| "step": 668 |
| }, |
| { |
| "epoch": 2.52, |
| "learning_rate": 1.3158276801407432e-06, |
| "loss": 0.1839, |
| "step": 669 |
| }, |
| { |
| "epoch": 2.53, |
| "learning_rate": 1.2954857673030807e-06, |
| "loss": 0.1727, |
| "step": 670 |
| }, |
| { |
| "epoch": 2.53, |
| "learning_rate": 1.2752914232823942e-06, |
| "loss": 0.1725, |
| "step": 671 |
| }, |
| { |
| "epoch": 2.54, |
| "learning_rate": 1.255244990436153e-06, |
| "loss": 0.201, |
| "step": 672 |
| }, |
| { |
| "epoch": 2.54, |
| "learning_rate": 1.2353468086142639e-06, |
| "loss": 0.1599, |
| "step": 673 |
| }, |
| { |
| "epoch": 2.54, |
| "learning_rate": 1.2155972151533225e-06, |
| "loss": 0.1528, |
| "step": 674 |
| }, |
| { |
| "epoch": 2.55, |
| "learning_rate": 1.1959965448708731e-06, |
| "loss": 0.2054, |
| "step": 675 |
| }, |
| { |
| "epoch": 2.55, |
| "learning_rate": 1.1765451300597574e-06, |
| "loss": 0.1611, |
| "step": 676 |
| }, |
| { |
| "epoch": 2.55, |
| "learning_rate": 1.1572433004824635e-06, |
| "loss": 0.162, |
| "step": 677 |
| }, |
| { |
| "epoch": 2.56, |
| "learning_rate": 1.1380913833655383e-06, |
| "loss": 0.1448, |
| "step": 678 |
| }, |
| { |
| "epoch": 2.56, |
| "learning_rate": 1.1190897033940461e-06, |
| "loss": 0.1481, |
| "step": 679 |
| }, |
| { |
| "epoch": 2.57, |
| "learning_rate": 1.1002385827060602e-06, |
| "loss": 0.1831, |
| "step": 680 |
| }, |
| { |
| "epoch": 2.57, |
| "learning_rate": 1.0815383408871983e-06, |
| "loss": 0.1454, |
| "step": 681 |
| }, |
| { |
| "epoch": 2.57, |
| "learning_rate": 1.0629892949652133e-06, |
| "loss": 0.1602, |
| "step": 682 |
| }, |
| { |
| "epoch": 2.58, |
| "learning_rate": 1.0445917594046073e-06, |
| "loss": 0.1445, |
| "step": 683 |
| }, |
| { |
| "epoch": 2.58, |
| "learning_rate": 1.026346046101312e-06, |
| "loss": 0.1743, |
| "step": 684 |
| }, |
| { |
| "epoch": 2.58, |
| "learning_rate": 1.0082524643773916e-06, |
| "loss": 0.1913, |
| "step": 685 |
| }, |
| { |
| "epoch": 2.59, |
| "learning_rate": 9.903113209758098e-07, |
| "loss": 0.1639, |
| "step": 686 |
| }, |
| { |
| "epoch": 2.59, |
| "learning_rate": 9.725229200552123e-07, |
| "loss": 0.1642, |
| "step": 687 |
| }, |
| { |
| "epoch": 2.6, |
| "learning_rate": 9.548875631847875e-07, |
| "loss": 0.1589, |
| "step": 688 |
| }, |
| { |
| "epoch": 2.6, |
| "learning_rate": 9.374055493391455e-07, |
| "loss": 0.1874, |
| "step": 689 |
| }, |
| { |
| "epoch": 2.6, |
| "learning_rate": 9.200771748932513e-07, |
| "loss": 0.1727, |
| "step": 690 |
| }, |
| { |
| "epoch": 2.61, |
| "learning_rate": 9.029027336174023e-07, |
| "loss": 0.1633, |
| "step": 691 |
| }, |
| { |
| "epoch": 2.61, |
| "learning_rate": 8.858825166722418e-07, |
| "loss": 0.1641, |
| "step": 692 |
| }, |
| { |
| "epoch": 2.62, |
| "learning_rate": 8.690168126038301e-07, |
| "loss": 0.1526, |
| "step": 693 |
| }, |
| { |
| "epoch": 2.62, |
| "learning_rate": 8.523059073387474e-07, |
| "loss": 0.1612, |
| "step": 694 |
| }, |
| { |
| "epoch": 2.62, |
| "learning_rate": 8.35750084179251e-07, |
| "loss": 0.1459, |
| "step": 695 |
| }, |
| { |
| "epoch": 2.63, |
| "learning_rate": 8.193496237984677e-07, |
| "loss": 0.179, |
| "step": 696 |
| }, |
| { |
| "epoch": 2.63, |
| "learning_rate": 8.031048042356393e-07, |
| "loss": 0.2097, |
| "step": 697 |
| }, |
| { |
| "epoch": 2.63, |
| "learning_rate": 7.870159008914069e-07, |
| "loss": 0.1656, |
| "step": 698 |
| }, |
| { |
| "epoch": 2.64, |
| "learning_rate": 7.710831865231461e-07, |
| "loss": 0.1652, |
| "step": 699 |
| }, |
| { |
| "epoch": 2.64, |
| "learning_rate": 7.55306931240335e-07, |
| "loss": 0.1738, |
| "step": 700 |
| }, |
| { |
| "epoch": 2.65, |
| "learning_rate": 7.396874024999811e-07, |
| "loss": 0.1501, |
| "step": 701 |
| }, |
| { |
| "epoch": 2.65, |
| "learning_rate": 7.242248651020845e-07, |
| "loss": 0.1627, |
| "step": 702 |
| }, |
| { |
| "epoch": 2.65, |
| "learning_rate": 7.089195811851502e-07, |
| "loss": 0.1574, |
| "step": 703 |
| }, |
| { |
| "epoch": 2.66, |
| "learning_rate": 6.937718102217461e-07, |
| "loss": 0.1866, |
| "step": 704 |
| }, |
| { |
| "epoch": 2.66, |
| "learning_rate": 6.787818090140985e-07, |
| "loss": 0.1519, |
| "step": 705 |
| }, |
| { |
| "epoch": 2.66, |
| "learning_rate": 6.639498316897419e-07, |
| "loss": 0.1476, |
| "step": 706 |
| }, |
| { |
| "epoch": 2.67, |
| "learning_rate": 6.492761296972117e-07, |
| "loss": 0.1718, |
| "step": 707 |
| }, |
| { |
| "epoch": 2.67, |
| "learning_rate": 6.347609518017761e-07, |
| "loss": 0.1735, |
| "step": 708 |
| }, |
| { |
| "epoch": 2.68, |
| "learning_rate": 6.2040454408123e-07, |
| "loss": 0.1581, |
| "step": 709 |
| }, |
| { |
| "epoch": 2.68, |
| "learning_rate": 6.062071499217081e-07, |
| "loss": 0.1473, |
| "step": 710 |
| }, |
| { |
| "epoch": 2.68, |
| "learning_rate": 5.921690100135713e-07, |
| "loss": 0.1606, |
| "step": 711 |
| }, |
| { |
| "epoch": 2.69, |
| "learning_rate": 5.782903623473202e-07, |
| "loss": 0.1654, |
| "step": 712 |
| }, |
| { |
| "epoch": 2.69, |
| "learning_rate": 5.645714422095627e-07, |
| "loss": 0.1842, |
| "step": 713 |
| }, |
| { |
| "epoch": 2.69, |
| "learning_rate": 5.510124821790208e-07, |
| "loss": 0.178, |
| "step": 714 |
| }, |
| { |
| "epoch": 2.7, |
| "learning_rate": 5.376137121225933e-07, |
| "loss": 0.1729, |
| "step": 715 |
| }, |
| { |
| "epoch": 2.7, |
| "learning_rate": 5.24375359191458e-07, |
| "loss": 0.1461, |
| "step": 716 |
| }, |
| { |
| "epoch": 2.71, |
| "learning_rate": 5.11297647817216e-07, |
| "loss": 0.1383, |
| "step": 717 |
| }, |
| { |
| "epoch": 2.71, |
| "learning_rate": 4.983807997080925e-07, |
| "loss": 0.1639, |
| "step": 718 |
| }, |
| { |
| "epoch": 2.71, |
| "learning_rate": 4.856250338451763e-07, |
| "loss": 0.1607, |
| "step": 719 |
| }, |
| { |
| "epoch": 2.72, |
| "learning_rate": 4.7303056647870605e-07, |
| "loss": 0.1604, |
| "step": 720 |
| }, |
| { |
| "epoch": 2.72, |
| "learning_rate": 4.605976111244015e-07, |
| "loss": 0.1653, |
| "step": 721 |
| }, |
| { |
| "epoch": 2.72, |
| "learning_rate": 4.483263785598524e-07, |
| "loss": 0.1677, |
| "step": 722 |
| }, |
| { |
| "epoch": 2.73, |
| "learning_rate": 4.3621707682094063e-07, |
| "loss": 0.1464, |
| "step": 723 |
| }, |
| { |
| "epoch": 2.73, |
| "learning_rate": 4.242699111983051e-07, |
| "loss": 0.1339, |
| "step": 724 |
| }, |
| { |
| "epoch": 2.74, |
| "learning_rate": 4.124850842338779e-07, |
| "loss": 0.1628, |
| "step": 725 |
| }, |
| { |
| "epoch": 2.74, |
| "learning_rate": 4.0086279571743715e-07, |
| "loss": 0.1604, |
| "step": 726 |
| }, |
| { |
| "epoch": 2.74, |
| "learning_rate": 3.8940324268322285e-07, |
| "loss": 0.1587, |
| "step": 727 |
| }, |
| { |
| "epoch": 2.75, |
| "learning_rate": 3.781066194066052e-07, |
| "loss": 0.1534, |
| "step": 728 |
| }, |
| { |
| "epoch": 2.75, |
| "learning_rate": 3.66973117400774e-07, |
| "loss": 0.1775, |
| "step": 729 |
| }, |
| { |
| "epoch": 2.75, |
| "learning_rate": 3.5600292541351e-07, |
| "loss": 0.1565, |
| "step": 730 |
| }, |
| { |
| "epoch": 2.76, |
| "learning_rate": 3.451962294239741e-07, |
| "loss": 0.1609, |
| "step": 731 |
| }, |
| { |
| "epoch": 2.76, |
| "learning_rate": 3.345532126395579e-07, |
| "loss": 0.1426, |
| "step": 732 |
| }, |
| { |
| "epoch": 2.77, |
| "learning_rate": 3.2407405549277683e-07, |
| "loss": 0.1815, |
| "step": 733 |
| }, |
| { |
| "epoch": 2.77, |
| "learning_rate": 3.137589356382076e-07, |
| "loss": 0.1801, |
| "step": 734 |
| }, |
| { |
| "epoch": 2.77, |
| "learning_rate": 3.0360802794948687e-07, |
| "loss": 0.1709, |
| "step": 735 |
| }, |
| { |
| "epoch": 2.78, |
| "learning_rate": 2.936215045163371e-07, |
| "loss": 0.1714, |
| "step": 736 |
| }, |
| { |
| "epoch": 2.78, |
| "learning_rate": 2.8379953464165334e-07, |
| "loss": 0.1487, |
| "step": 737 |
| }, |
| { |
| "epoch": 2.78, |
| "learning_rate": 2.741422848386266e-07, |
| "loss": 0.1533, |
| "step": 738 |
| }, |
| { |
| "epoch": 2.79, |
| "learning_rate": 2.646499188279328e-07, |
| "loss": 0.1635, |
| "step": 739 |
| }, |
| { |
| "epoch": 2.79, |
| "learning_rate": 2.5532259753494825e-07, |
| "loss": 0.1524, |
| "step": 740 |
| }, |
| { |
| "epoch": 2.8, |
| "learning_rate": 2.461604790870209e-07, |
| "loss": 0.1764, |
| "step": 741 |
| }, |
| { |
| "epoch": 2.8, |
| "learning_rate": 2.3716371881079558e-07, |
| "loss": 0.1636, |
| "step": 742 |
| }, |
| { |
| "epoch": 2.8, |
| "learning_rate": 2.2833246922957408e-07, |
| "loss": 0.1606, |
| "step": 743 |
| }, |
| { |
| "epoch": 2.81, |
| "learning_rate": 2.196668800607382e-07, |
| "loss": 0.1533, |
| "step": 744 |
| }, |
| { |
| "epoch": 2.81, |
| "learning_rate": 2.111670982132008e-07, |
| "loss": 0.1442, |
| "step": 745 |
| }, |
| { |
| "epoch": 2.82, |
| "learning_rate": 2.028332677849254e-07, |
| "loss": 0.1649, |
| "step": 746 |
| }, |
| { |
| "epoch": 2.82, |
| "learning_rate": 1.9466553006047383e-07, |
| "loss": 0.1686, |
| "step": 747 |
| }, |
| { |
| "epoch": 2.82, |
| "learning_rate": 1.8666402350861701e-07, |
| "loss": 0.1672, |
| "step": 748 |
| }, |
| { |
| "epoch": 2.83, |
| "learning_rate": 1.7882888377998787e-07, |
| "loss": 0.1764, |
| "step": 749 |
| }, |
| { |
| "epoch": 2.83, |
| "learning_rate": 1.711602437047788e-07, |
| "loss": 0.142, |
| "step": 750 |
| }, |
| { |
| "epoch": 2.83, |
| "learning_rate": 1.6365823329049124e-07, |
| "loss": 0.1835, |
| "step": 751 |
| }, |
| { |
| "epoch": 2.84, |
| "learning_rate": 1.5632297971972966e-07, |
| "loss": 0.1624, |
| "step": 752 |
| }, |
| { |
| "epoch": 2.84, |
| "learning_rate": 1.49154607348051e-07, |
| "loss": 0.1537, |
| "step": 753 |
| }, |
| { |
| "epoch": 2.85, |
| "learning_rate": 1.4215323770184642e-07, |
| "loss": 0.1417, |
| "step": 754 |
| }, |
| { |
| "epoch": 2.85, |
| "learning_rate": 1.3531898947629296e-07, |
| "loss": 0.1468, |
| "step": 755 |
| }, |
| { |
| "epoch": 2.85, |
| "learning_rate": 1.2865197853333179e-07, |
| "loss": 0.1514, |
| "step": 756 |
| }, |
| { |
| "epoch": 2.86, |
| "learning_rate": 1.221523178997075e-07, |
| "loss": 0.149, |
| "step": 757 |
| }, |
| { |
| "epoch": 2.86, |
| "learning_rate": 1.158201177650553e-07, |
| "loss": 0.1629, |
| "step": 758 |
| }, |
| { |
| "epoch": 2.86, |
| "learning_rate": 1.0965548548002802e-07, |
| "loss": 0.1525, |
| "step": 759 |
| }, |
| { |
| "epoch": 2.87, |
| "learning_rate": 1.0365852555447642e-07, |
| "loss": 0.16, |
| "step": 760 |
| }, |
| { |
| "epoch": 2.87, |
| "learning_rate": 9.782933965567953e-08, |
| "loss": 0.1482, |
| "step": 761 |
| }, |
| { |
| "epoch": 2.88, |
| "learning_rate": 9.216802660662161e-08, |
| "loss": 0.1664, |
| "step": 762 |
| }, |
| { |
| "epoch": 2.88, |
| "learning_rate": 8.667468238431453e-08, |
| "loss": 0.1436, |
| "step": 763 |
| }, |
| { |
| "epoch": 2.88, |
| "learning_rate": 8.13494001181725e-08, |
| "loss": 0.1717, |
| "step": 764 |
| }, |
| { |
| "epoch": 2.89, |
| "learning_rate": 7.619227008843322e-08, |
| "loss": 0.1703, |
| "step": 765 |
| }, |
| { |
| "epoch": 2.89, |
| "learning_rate": 7.120337972462365e-08, |
| "loss": 0.1708, |
| "step": 766 |
| }, |
| { |
| "epoch": 2.89, |
| "learning_rate": 6.638281360408339e-08, |
| "loss": 0.1758, |
| "step": 767 |
| }, |
| { |
| "epoch": 2.9, |
| "learning_rate": 6.173065345052687e-08, |
| "loss": 0.1775, |
| "step": 768 |
| }, |
| { |
| "epoch": 2.9, |
| "learning_rate": 5.7246978132659e-08, |
| "loss": 0.1589, |
| "step": 769 |
| }, |
| { |
| "epoch": 2.91, |
| "learning_rate": 5.2931863662841755e-08, |
| "loss": 0.1775, |
| "step": 770 |
| }, |
| { |
| "epoch": 2.91, |
| "learning_rate": 4.878538319579629e-08, |
| "loss": 0.1509, |
| "step": 771 |
| }, |
| { |
| "epoch": 2.91, |
| "learning_rate": 4.480760702737286e-08, |
| "loss": 0.1893, |
| "step": 772 |
| }, |
| { |
| "epoch": 2.92, |
| "learning_rate": 4.099860259335287e-08, |
| "loss": 0.1917, |
| "step": 773 |
| }, |
| { |
| "epoch": 2.92, |
| "learning_rate": 3.735843446830867e-08, |
| "loss": 0.1607, |
| "step": 774 |
| }, |
| { |
| "epoch": 2.92, |
| "learning_rate": 3.3887164364506676e-08, |
| "loss": 0.1372, |
| "step": 775 |
| }, |
| { |
| "epoch": 2.93, |
| "learning_rate": 3.0584851130861516e-08, |
| "loss": 0.1535, |
| "step": 776 |
| }, |
| { |
| "epoch": 2.93, |
| "learning_rate": 2.7451550751941282e-08, |
| "loss": 0.1685, |
| "step": 777 |
| }, |
| { |
| "epoch": 2.94, |
| "learning_rate": 2.4487316347013845e-08, |
| "loss": 0.1531, |
| "step": 778 |
| }, |
| { |
| "epoch": 2.94, |
| "learning_rate": 2.169219816914869e-08, |
| "loss": 0.1663, |
| "step": 779 |
| }, |
| { |
| "epoch": 2.94, |
| "learning_rate": 1.9066243604367595e-08, |
| "loss": 0.1647, |
| "step": 780 |
| }, |
| { |
| "epoch": 2.95, |
| "learning_rate": 1.6609497170834154e-08, |
| "loss": 0.1571, |
| "step": 781 |
| }, |
| { |
| "epoch": 2.95, |
| "learning_rate": 1.4322000518106616e-08, |
| "loss": 0.1685, |
| "step": 782 |
| }, |
| { |
| "epoch": 2.95, |
| "learning_rate": 1.220379242642844e-08, |
| "loss": 0.1361, |
| "step": 783 |
| }, |
| { |
| "epoch": 2.96, |
| "learning_rate": 1.0254908806068831e-08, |
| "loss": 0.1469, |
| "step": 784 |
| }, |
| { |
| "epoch": 2.96, |
| "learning_rate": 8.47538269671988e-09, |
| "loss": 0.1719, |
| "step": 785 |
| }, |
| { |
| "epoch": 2.97, |
| "learning_rate": 6.8652442669281394e-09, |
| "loss": 0.1594, |
| "step": 786 |
| }, |
| { |
| "epoch": 2.97, |
| "learning_rate": 5.4245208135905725e-09, |
| "loss": 0.1523, |
| "step": 787 |
| }, |
| { |
| "epoch": 2.97, |
| "learning_rate": 4.153236761488266e-09, |
| "loss": 0.1614, |
| "step": 788 |
| }, |
| { |
| "epoch": 2.98, |
| "learning_rate": 3.0514136628745363e-09, |
| "loss": 0.1537, |
| "step": 789 |
| }, |
| { |
| "epoch": 2.98, |
| "learning_rate": 2.1190701971052218e-09, |
| "loss": 0.1628, |
| "step": 790 |
| }, |
| { |
| "epoch": 2.98, |
| "learning_rate": 1.3562221703267153e-09, |
| "loss": 0.1766, |
| "step": 791 |
| }, |
| { |
| "epoch": 2.99, |
| "learning_rate": 7.628825152050656e-10, |
| "loss": 0.1805, |
| "step": 792 |
| }, |
| { |
| "epoch": 2.99, |
| "learning_rate": 3.390612907094859e-10, |
| "loss": 0.1617, |
| "step": 793 |
| }, |
| { |
| "epoch": 3.0, |
| "learning_rate": 8.476568193804824e-11, |
| "loss": 0.1549, |
| "step": 794 |
| }, |
| { |
| "epoch": 3.0, |
| "learning_rate": 0.0, |
| "loss": 0.1619, |
| "step": 795 |
| } |
| ], |
| "max_steps": 795, |
| "num_train_epochs": 3, |
| "total_flos": 1006074152878080.0, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|