| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.9989235737351991, | |
| "eval_steps": 500, | |
| "global_step": 464, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.4285714285714286e-06, | |
| "loss": 1.6806, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 2.8571428571428573e-06, | |
| "loss": 1.7444, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.2857142857142855e-06, | |
| "loss": 1.7017, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 5.7142857142857145e-06, | |
| "loss": 1.6305, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 7.1428571428571436e-06, | |
| "loss": 1.5227, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 8.571428571428571e-06, | |
| "loss": 1.3905, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1e-05, | |
| "loss": 1.269, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.1428571428571429e-05, | |
| "loss": 1.1847, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.2857142857142859e-05, | |
| "loss": 1.2008, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.4285714285714287e-05, | |
| "loss": 1.2259, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.5714285714285715e-05, | |
| "loss": 1.0843, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.7142857142857142e-05, | |
| "loss": 1.1093, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.8571428571428575e-05, | |
| "loss": 1.0567, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 2e-05, | |
| "loss": 1.0045, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9999756307053947e-05, | |
| "loss": 0.9836, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9999025240093045e-05, | |
| "loss": 0.9945, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9997806834748455e-05, | |
| "loss": 1.012, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9996101150403543e-05, | |
| "loss": 1.0002, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.999390827019096e-05, | |
| "loss": 1.0176, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.9991228300988586e-05, | |
| "loss": 1.0494, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9988061373414342e-05, | |
| "loss": 1.0395, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9984407641819812e-05, | |
| "loss": 0.9945, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9980267284282718e-05, | |
| "loss": 1.0104, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9975640502598243e-05, | |
| "loss": 0.9753, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 1.9970527522269204e-05, | |
| "loss": 0.9829, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9964928592495046e-05, | |
| "loss": 0.9406, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9958843986159705e-05, | |
| "loss": 0.9735, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9952273999818312e-05, | |
| "loss": 0.9919, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9945218953682736e-05, | |
| "loss": 0.9872, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 1.9937679191605964e-05, | |
| "loss": 0.93, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.992965508106537e-05, | |
| "loss": 0.9179, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.9921147013144782e-05, | |
| "loss": 0.9876, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.991215540251542e-05, | |
| "loss": 0.9884, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 1.9902680687415704e-05, | |
| "loss": 0.9792, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9892723329629885e-05, | |
| "loss": 0.9702, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.988228381446553e-05, | |
| "loss": 0.9575, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.987136265072988e-05, | |
| "loss": 0.9605, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.985996037070505e-05, | |
| "loss": 0.9465, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 1.9848077530122083e-05, | |
| "loss": 0.9438, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.983571470813386e-05, | |
| "loss": 0.8716, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.982287250728689e-05, | |
| "loss": 0.8425, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9809551553491918e-05, | |
| "loss": 0.8671, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.979575249599344e-05, | |
| "loss": 0.9382, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.9781476007338058e-05, | |
| "loss": 0.9426, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9766722783341682e-05, | |
| "loss": 0.9615, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9751493543055634e-05, | |
| "loss": 0.9307, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9735789028731603e-05, | |
| "loss": 0.9455, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.9719610005785466e-05, | |
| "loss": 0.915, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9702957262759964e-05, | |
| "loss": 0.8733, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9685831611286312e-05, | |
| "loss": 0.9304, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9668233886044597e-05, | |
| "loss": 0.8837, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9650164944723116e-05, | |
| "loss": 0.9278, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.9631625667976584e-05, | |
| "loss": 0.8601, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.961261695938319e-05, | |
| "loss": 0.8639, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9593139745400575e-05, | |
| "loss": 0.8201, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9573194975320672e-05, | |
| "loss": 0.8901, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9552783621223437e-05, | |
| "loss": 0.8912, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.9531906677929472e-05, | |
| "loss": 0.9046, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9510565162951538e-05, | |
| "loss": 0.8824, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9488760116444966e-05, | |
| "loss": 0.8882, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.9466492601156964e-05, | |
| "loss": 0.8873, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.944376370237481e-05, | |
| "loss": 0.9555, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.942057452787297e-05, | |
| "loss": 0.9429, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9396926207859085e-05, | |
| "loss": 0.9233, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.937281989491892e-05, | |
| "loss": 0.9213, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9348256763960146e-05, | |
| "loss": 0.8389, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.9323238012155125e-05, | |
| "loss": 0.8301, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.9297764858882516e-05, | |
| "loss": 0.9072, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.9271838545667876e-05, | |
| "loss": 0.8944, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.9245460336123136e-05, | |
| "loss": 0.8538, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.9218631515885007e-05, | |
| "loss": 0.8686, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.9191353392552346e-05, | |
| "loss": 0.8861, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.9163627295622397e-05, | |
| "loss": 0.8764, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.913545457642601e-05, | |
| "loss": 0.9213, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.910683660806177e-05, | |
| "loss": 0.8744, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.907777478532909e-05, | |
| "loss": 0.9065, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.9048270524660197e-05, | |
| "loss": 0.8504, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.901832526405114e-05, | |
| "loss": 0.8971, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8987940462991673e-05, | |
| "loss": 0.806, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.895711760239413e-05, | |
| "loss": 0.9, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.892585818452126e-05, | |
| "loss": 0.9391, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.889416373291298e-05, | |
| "loss": 0.8273, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.8862035792312148e-05, | |
| "loss": 0.8936, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.8829475928589272e-05, | |
| "loss": 0.7711, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.879648572866617e-05, | |
| "loss": 0.8493, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.8763066800438638e-05, | |
| "loss": 0.873, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.8729220772698096e-05, | |
| "loss": 0.8214, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.869494929505219e-05, | |
| "loss": 0.86, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.866025403784439e-05, | |
| "loss": 0.8474, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.8625136692072577e-05, | |
| "loss": 0.8508, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.8589598969306646e-05, | |
| "loss": 0.8444, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.855364260160507e-05, | |
| "loss": 0.8501, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.851726934143048e-05, | |
| "loss": 0.839, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.848048096156426e-05, | |
| "loss": 0.9062, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.8443279255020153e-05, | |
| "loss": 0.9258, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.8405666034956842e-05, | |
| "loss": 0.9218, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.836764313458962e-05, | |
| "loss": 0.8922, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.8329212407100996e-05, | |
| "loss": 0.8281, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.8290375725550417e-05, | |
| "loss": 0.8925, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.8251134982782952e-05, | |
| "loss": 0.886, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.821149209133704e-05, | |
| "loss": 0.8394, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.8171448983351284e-05, | |
| "loss": 0.8578, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.8131007610470278e-05, | |
| "loss": 0.8895, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.8090169943749477e-05, | |
| "loss": 0.8173, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.804893797355914e-05, | |
| "loss": 0.8727, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.8007313709487334e-05, | |
| "loss": 0.8706, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.7965299180241963e-05, | |
| "loss": 0.8391, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.792289643355191e-05, | |
| "loss": 0.8228, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.788010753606722e-05, | |
| "loss": 0.8564, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.78369345732584e-05, | |
| "loss": 0.8927, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.7793379649314743e-05, | |
| "loss": 0.7994, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.7749444887041797e-05, | |
| "loss": 0.8048, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.7705132427757895e-05, | |
| "loss": 0.8455, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.766044443118978e-05, | |
| "loss": 0.8032, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.761538307536737e-05, | |
| "loss": 0.8482, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.7569950556517566e-05, | |
| "loss": 0.8006, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.7524149088957244e-05, | |
| "loss": 0.8869, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.747798090498532e-05, | |
| "loss": 0.7997, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.7431448254773943e-05, | |
| "loss": 0.8333, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.7384553406258842e-05, | |
| "loss": 0.8226, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.7337298645028764e-05, | |
| "loss": 0.7886, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.7289686274214116e-05, | |
| "loss": 0.8996, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.7241718614374678e-05, | |
| "loss": 0.8477, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.7193398003386514e-05, | |
| "loss": 0.878, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.7144726796328034e-05, | |
| "loss": 0.8428, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.709570736536521e-05, | |
| "loss": 0.8092, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.7046342099635948e-05, | |
| "loss": 0.8697, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.6996633405133656e-05, | |
| "loss": 0.7754, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.6946583704589973e-05, | |
| "loss": 0.7986, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.68961954373567e-05, | |
| "loss": 0.8031, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.684547105928689e-05, | |
| "loss": 0.8354, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.6794413042615168e-05, | |
| "loss": 0.8739, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.6743023875837233e-05, | |
| "loss": 0.8048, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.6691306063588583e-05, | |
| "loss": 0.8386, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.6639262126522417e-05, | |
| "loss": 0.7554, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.6586894601186804e-05, | |
| "loss": 0.8475, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.6534206039901057e-05, | |
| "loss": 0.8155, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.6481199010631312e-05, | |
| "loss": 0.8429, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.6427876096865394e-05, | |
| "loss": 0.7883, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.63742398974869e-05, | |
| "loss": 0.77, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.632029302664851e-05, | |
| "loss": 0.7894, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.6266038113644605e-05, | |
| "loss": 0.8344, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.6211477802783105e-05, | |
| "loss": 0.8601, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.6156614753256583e-05, | |
| "loss": 0.9022, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.610145163901268e-05, | |
| "loss": 0.8067, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.6045991148623752e-05, | |
| "loss": 0.8004, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.599023598515586e-05, | |
| "loss": 0.8638, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.5934188866037017e-05, | |
| "loss": 0.8481, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.5877852522924733e-05, | |
| "loss": 0.8726, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.5821229701572897e-05, | |
| "loss": 0.8407, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.5764323161697933e-05, | |
| "loss": 0.7786, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.570713567684432e-05, | |
| "loss": 0.8233, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.564967003424938e-05, | |
| "loss": 0.8965, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.5591929034707468e-05, | |
| "loss": 0.7953, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.553391549243344e-05, | |
| "loss": 0.8382, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.5475632234925505e-05, | |
| "loss": 0.8256, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.54170821028274e-05, | |
| "loss": 0.8488, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.5358267949789968e-05, | |
| "loss": 0.8986, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.529919264233205e-05, | |
| "loss": 0.8572, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.5239859059700794e-05, | |
| "loss": 0.8046, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.5180270093731305e-05, | |
| "loss": 0.8461, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.5120428648705716e-05, | |
| "loss": 0.8132, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.5060337641211637e-05, | |
| "loss": 0.8497, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.5000000000000002e-05, | |
| "loss": 0.8151, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.493941866584231e-05, | |
| "loss": 0.8352, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.4878596591387329e-05, | |
| "loss": 0.8662, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.4817536741017153e-05, | |
| "loss": 0.8644, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.4756242090702756e-05, | |
| "loss": 0.8058, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.469471562785891e-05, | |
| "loss": 0.8268, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.463296035119862e-05, | |
| "loss": 0.7497, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.4570979270586944e-05, | |
| "loss": 0.7812, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.4508775406894308e-05, | |
| "loss": 0.7519, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.4446351791849276e-05, | |
| "loss": 0.9014, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.4383711467890776e-05, | |
| "loss": 0.7682, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.4320857488019826e-05, | |
| "loss": 0.8379, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.4257792915650728e-05, | |
| "loss": 0.8332, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.4194520824461773e-05, | |
| "loss": 0.7577, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.413104429824542e-05, | |
| "loss": 0.8506, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.4067366430758004e-05, | |
| "loss": 0.8465, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.4003490325568953e-05, | |
| "loss": 0.7704, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.3939419095909513e-05, | |
| "loss": 0.8035, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.3875155864521031e-05, | |
| "loss": 0.8743, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.3810703763502744e-05, | |
| "loss": 0.8431, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.3746065934159123e-05, | |
| "loss": 0.797, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.3681245526846782e-05, | |
| "loss": 0.7441, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.3616245700820922e-05, | |
| "loss": 0.7876, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.3551069624081372e-05, | |
| "loss": 0.7772, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.3485720473218153e-05, | |
| "loss": 0.7836, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.342020143325669e-05, | |
| "loss": 0.7715, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.3354515697502552e-05, | |
| "loss": 0.7837, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.3288666467385834e-05, | |
| "loss": 0.8376, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.3222656952305113e-05, | |
| "loss": 0.7995, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.3156490369471026e-05, | |
| "loss": 0.8232, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.3090169943749475e-05, | |
| "loss": 0.8618, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.3023698907504447e-05, | |
| "loss": 0.8094, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.2957080500440469e-05, | |
| "loss": 0.7832, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.2890317969444716e-05, | |
| "loss": 0.8129, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.2823414568428767e-05, | |
| "loss": 0.8238, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.2756373558169992e-05, | |
| "loss": 0.8246, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.2689198206152657e-05, | |
| "loss": 0.8304, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.2621891786408648e-05, | |
| "loss": 0.7911, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.2554457579357906e-05, | |
| "loss": 0.8427, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.2486898871648552e-05, | |
| "loss": 0.7555, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.2419218955996677e-05, | |
| "loss": 0.7625, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.23514211310259e-05, | |
| "loss": 0.7994, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.2283508701106559e-05, | |
| "loss": 0.8183, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.2215484976194675e-05, | |
| "loss": 0.7922, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.2147353271670634e-05, | |
| "loss": 0.7845, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.2079116908177592e-05, | |
| "loss": 0.8034, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.2010779211459649e-05, | |
| "loss": 0.83, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.194234351219972e-05, | |
| "loss": 0.7644, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.187381314585725e-05, | |
| "loss": 0.8275, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.1805191452505602e-05, | |
| "loss": 0.8256, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.1736481776669307e-05, | |
| "loss": 0.796, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.1667687467161025e-05, | |
| "loss": 0.7525, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.159881187691835e-05, | |
| "loss": 0.788, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.1529858362840383e-05, | |
| "loss": 0.8071, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.1460830285624119e-05, | |
| "loss": 0.7687, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.1391731009600655e-05, | |
| "loss": 0.7793, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.1322563902571227e-05, | |
| "loss": 0.7792, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.1253332335643043e-05, | |
| "loss": 0.753, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.1184039683065014e-05, | |
| "loss": 0.8115, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.1114689322063255e-05, | |
| "loss": 0.7863, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.1045284632676535e-05, | |
| "loss": 0.7875, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.0975828997591496e-05, | |
| "loss": 0.7481, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.0906325801977804e-05, | |
| "loss": 0.7768, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.083677843332316e-05, | |
| "loss": 0.7882, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.0767190281268187e-05, | |
| "loss": 0.7586, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.0697564737441254e-05, | |
| "loss": 0.8614, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.0627905195293135e-05, | |
| "loss": 0.8085, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.055821504993164e-05, | |
| "loss": 0.7757, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.0488497697956134e-05, | |
| "loss": 0.7611, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.0418756537291996e-05, | |
| "loss": 0.7021, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.0348994967025012e-05, | |
| "loss": 0.7667, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 1.0279216387235691e-05, | |
| "loss": 0.7475, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 1.0209424198833571e-05, | |
| "loss": 0.8029, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 1.0139621803391454e-05, | |
| "loss": 0.7237, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 1.0069812602979617e-05, | |
| "loss": 0.7464, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 1e-05, | |
| "loss": 0.7393, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 9.930187397020385e-06, | |
| "loss": 0.7582, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 9.860378196608549e-06, | |
| "loss": 0.7435, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 9.790575801166432e-06, | |
| "loss": 0.7691, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 9.720783612764314e-06, | |
| "loss": 0.7905, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 9.651005032974994e-06, | |
| "loss": 0.8642, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 9.581243462708007e-06, | |
| "loss": 0.779, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 9.511502302043867e-06, | |
| "loss": 0.7765, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 9.441784950068362e-06, | |
| "loss": 0.781, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 9.372094804706867e-06, | |
| "loss": 0.674, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 9.302435262558748e-06, | |
| "loss": 0.7532, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 9.232809718731815e-06, | |
| "loss": 0.8772, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 9.163221566676847e-06, | |
| "loss": 0.7803, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 9.093674198022201e-06, | |
| "loss": 0.7963, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 9.024171002408507e-06, | |
| "loss": 0.7205, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 8.954715367323468e-06, | |
| "loss": 0.8772, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 8.885310677936746e-06, | |
| "loss": 0.8081, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 8.815960316934991e-06, | |
| "loss": 0.733, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 8.746667664356957e-06, | |
| "loss": 0.7366, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 8.677436097428775e-06, | |
| "loss": 0.7818, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 8.60826899039935e-06, | |
| "loss": 0.7873, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 8.539169714375885e-06, | |
| "loss": 0.786, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 8.47014163715962e-06, | |
| "loss": 0.7075, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 8.401188123081653e-06, | |
| "loss": 0.8245, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 8.332312532838978e-06, | |
| "loss": 0.8094, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 8.263518223330698e-06, | |
| "loss": 0.7889, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 8.194808547494401e-06, | |
| "loss": 0.7979, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 8.126186854142752e-06, | |
| "loss": 0.7844, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 8.057656487800283e-06, | |
| "loss": 0.809, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 7.989220788540356e-06, | |
| "loss": 0.8085, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 7.92088309182241e-06, | |
| "loss": 0.7639, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 7.852646728329368e-06, | |
| "loss": 0.736, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 7.784515023805328e-06, | |
| "loss": 0.7265, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 7.716491298893443e-06, | |
| "loss": 0.749, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 7.6485788689741e-06, | |
| "loss": 0.8023, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 7.580781044003324e-06, | |
| "loss": 0.7135, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 7.513101128351454e-06, | |
| "loss": 0.806, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 7.445542420642097e-06, | |
| "loss": 0.8008, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 7.378108213591355e-06, | |
| "loss": 0.7742, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 7.310801793847344e-06, | |
| "loss": 0.7987, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 7.243626441830009e-06, | |
| "loss": 0.7829, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 7.176585431571235e-06, | |
| "loss": 0.7326, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 7.109682030555283e-06, | |
| "loss": 0.7526, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 7.042919499559538e-06, | |
| "loss": 0.6843, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 6.976301092495556e-06, | |
| "loss": 0.762, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 6.909830056250527e-06, | |
| "loss": 0.6768, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 6.843509630528977e-06, | |
| "loss": 0.737, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 6.777343047694891e-06, | |
| "loss": 0.832, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 6.711333532614168e-06, | |
| "loss": 0.8041, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 6.645484302497452e-06, | |
| "loss": 0.7919, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 6.579798566743314e-06, | |
| "loss": 0.7579, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 6.5142795267818505e-06, | |
| "loss": 0.7704, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 6.448930375918632e-06, | |
| "loss": 0.7676, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 6.383754299179079e-06, | |
| "loss": 0.7982, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 6.318754473153221e-06, | |
| "loss": 0.8056, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 6.25393406584088e-06, | |
| "loss": 0.8174, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 6.18929623649726e-06, | |
| "loss": 0.842, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 6.124844135478971e-06, | |
| "loss": 0.7873, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 6.06058090409049e-06, | |
| "loss": 0.79, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 5.996509674431053e-06, | |
| "loss": 0.7586, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 5.932633569242e-06, | |
| "loss": 0.7266, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 5.868955701754584e-06, | |
| "loss": 0.7449, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 5.8054791755382286e-06, | |
| "loss": 0.77, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 5.742207084349274e-06, | |
| "loss": 0.7869, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 5.679142511980176e-06, | |
| "loss": 0.7365, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 5.616288532109225e-06, | |
| "loss": 0.8115, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 5.553648208150728e-06, | |
| "loss": 0.7988, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 5.491224593105695e-06, | |
| "loss": 0.7903, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 5.429020729413062e-06, | |
| "loss": 0.7548, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 5.367039648801386e-06, | |
| "loss": 0.84, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 5.305284372141095e-06, | |
| "loss": 0.7513, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 5.243757909297247e-06, | |
| "loss": 0.7355, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 5.1824632589828465e-06, | |
| "loss": 0.7603, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 5.121403408612672e-06, | |
| "loss": 0.756, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 5.060581334157693e-06, | |
| "loss": 0.797, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 5.000000000000003e-06, | |
| "loss": 0.8073, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.939662358788364e-06, | |
| "loss": 0.7516, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.879571351294287e-06, | |
| "loss": 0.7729, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.8197299062687e-06, | |
| "loss": 0.7715, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.76014094029921e-06, | |
| "loss": 0.77, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.700807357667953e-06, | |
| "loss": 0.789, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.641732050210032e-06, | |
| "loss": 0.7728, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.582917897172603e-06, | |
| "loss": 0.7791, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.524367765074499e-06, | |
| "loss": 0.7511, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.46608450756656e-06, | |
| "loss": 0.7562, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.408070965292534e-06, | |
| "loss": 0.7865, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.350329965750622e-06, | |
| "loss": 0.7174, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.292864323155684e-06, | |
| "loss": 0.7345, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.235676838302069e-06, | |
| "loss": 0.778, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.178770298427107e-06, | |
| "loss": 0.795, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.12214747707527e-06, | |
| "loss": 0.8197, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.065811133962987e-06, | |
| "loss": 0.7793, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.009764014844143e-06, | |
| "loss": 0.7559, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 3.954008851376252e-06, | |
| "loss": 0.7528, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 3.898548360987325e-06, | |
| "loss": 0.7598, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 3.8433852467434175e-06, | |
| "loss": 0.8453, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 3.7885221972168974e-06, | |
| "loss": 0.7612, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 3.7339618863553983e-06, | |
| "loss": 0.7027, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 3.679706973351491e-06, | |
| "loss": 0.8072, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 3.625760102513103e-06, | |
| "loss": 0.7634, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 3.5721239031346067e-06, | |
| "loss": 0.7478, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 3.5188009893686916e-06, | |
| "loss": 0.7218, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 3.4657939600989453e-06, | |
| "loss": 0.7411, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.4131053988131947e-06, | |
| "loss": 0.7689, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.360737873477584e-06, | |
| "loss": 0.7517, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.308693936411421e-06, | |
| "loss": 0.7562, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.2569761241627694e-06, | |
| "loss": 0.7675, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 3.2055869573848374e-06, | |
| "loss": 0.7715, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.1545289407131128e-06, | |
| "loss": 0.7528, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.103804562643302e-06, | |
| "loss": 0.7722, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.0534162954100264e-06, | |
| "loss": 0.7861, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 3.003366594866345e-06, | |
| "loss": 0.7526, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 2.953657900364053e-06, | |
| "loss": 0.7173, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 2.9042926346347932e-06, | |
| "loss": 0.7175, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 2.855273203671969e-06, | |
| "loss": 0.7912, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 2.8066019966134907e-06, | |
| "loss": 0.8047, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 2.7582813856253276e-06, | |
| "loss": 0.771, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 2.7103137257858867e-06, | |
| "loss": 0.7895, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 2.6627013549712355e-06, | |
| "loss": 0.6981, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 2.615446593741161e-06, | |
| "loss": 0.7507, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 2.5685517452260566e-06, | |
| "loss": 0.6642, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 2.522019095014683e-06, | |
| "loss": 0.759, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 2.4758509110427576e-06, | |
| "loss": 0.7535, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 2.4300494434824373e-06, | |
| "loss": 0.7649, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 2.3846169246326345e-06, | |
| "loss": 0.7926, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 2.339555568810221e-06, | |
| "loss": 0.7479, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 2.2948675722421086e-06, | |
| "loss": 0.7626, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 2.2505551129582047e-06, | |
| "loss": 0.7163, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 2.206620350685257e-06, | |
| "loss": 0.8096, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 2.163065426741603e-06, | |
| "loss": 0.7981, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 2.119892463932781e-06, | |
| "loss": 0.7507, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 2.0771035664480944e-06, | |
| "loss": 0.7346, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 2.0347008197580376e-06, | |
| "loss": 0.7388, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.9926862905126663e-06, | |
| "loss": 0.768, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.95106202644086e-06, | |
| "loss": 0.7726, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 1.9098300562505266e-06, | |
| "loss": 0.7567, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 1.8689923895297247e-06, | |
| "loss": 0.7588, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 1.8285510166487154e-06, | |
| "loss": 0.6654, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 1.7885079086629598e-06, | |
| "loss": 0.7637, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 1.7488650172170496e-06, | |
| "loss": 0.7029, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.709624274449584e-06, | |
| "loss": 0.7585, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.6707875928990059e-06, | |
| "loss": 0.7061, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.6323568654103838e-06, | |
| "loss": 0.7368, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.5943339650431578e-06, | |
| "loss": 0.7132, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.5567207449798517e-06, | |
| "loss": 0.7341, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.5195190384357405e-06, | |
| "loss": 0.73, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.4827306585695234e-06, | |
| "loss": 0.7187, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.446357398394934e-06, | |
| "loss": 0.6724, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.4104010306933558e-06, | |
| "loss": 0.6546, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.3748633079274254e-06, | |
| "loss": 0.7822, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.339745962155613e-06, | |
| "loss": 0.7428, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.30505070494781e-06, | |
| "loss": 0.8222, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.2707792273019049e-06, | |
| "loss": 0.6602, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.2369331995613664e-06, | |
| "loss": 0.7266, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.2035142713338366e-06, | |
| "loss": 0.7973, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.1705240714107301e-06, | |
| "loss": 0.69, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.1379642076878528e-06, | |
| "loss": 0.7544, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.1058362670870248e-06, | |
| "loss": 0.8371, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.0741418154787443e-06, | |
| "loss": 0.7257, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 1.042882397605871e-06, | |
| "loss": 0.7834, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 1.012059537008332e-06, | |
| "loss": 0.7095, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 9.816747359488632e-07, | |
| "loss": 0.7552, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 9.517294753398066e-07, | |
| "loss": 0.6726, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 9.222252146709143e-07, | |
| "loss": 0.7404, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 8.931633919382299e-07, | |
| "loss": 0.7394, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 8.645454235739903e-07, | |
| "loss": 0.7101, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 8.363727043776037e-07, | |
| "loss": 0.7446, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 8.086466074476562e-07, | |
| "loss": 0.7659, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 7.81368484114996e-07, | |
| "loss": 0.7543, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 7.545396638768698e-07, | |
| "loss": 0.7727, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 7.281614543321269e-07, | |
| "loss": 0.7143, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 7.022351411174866e-07, | |
| "loss": 0.7251, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 6.767619878448783e-07, | |
| "loss": 0.7893, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 6.517432360398556e-07, | |
| "loss": 0.7487, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 6.271801050810856e-07, | |
| "loss": 0.7105, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 6.030737921409169e-07, | |
| "loss": 0.7616, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 5.794254721270331e-07, | |
| "loss": 0.7849, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 5.562362976251901e-07, | |
| "loss": 0.6783, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 5.335073988430373e-07, | |
| "loss": 0.6506, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 5.112398835550348e-07, | |
| "loss": 0.7298, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.894348370484648e-07, | |
| "loss": 0.7401, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.6809332207053083e-07, | |
| "loss": 0.7586, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.4721637877656377e-07, | |
| "loss": 0.6923, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.268050246793276e-07, | |
| "loss": 0.7255, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.068602545994249e-07, | |
| "loss": 0.7007, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 3.8738304061681107e-07, | |
| "loss": 0.7424, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 3.68374332023419e-07, | |
| "loss": 0.7398, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.498350552768859e-07, | |
| "loss": 0.705, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.3176611395540625e-07, | |
| "loss": 0.6938, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.1416838871368925e-07, | |
| "loss": 0.7828, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 2.970427372400353e-07, | |
| "loss": 0.7369, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 2.8038999421453827e-07, | |
| "loss": 0.7793, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 2.6421097126839714e-07, | |
| "loss": 0.7145, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 2.4850645694436736e-07, | |
| "loss": 0.713, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 2.332772166583208e-07, | |
| "loss": 0.7325, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 2.1852399266194312e-07, | |
| "loss": 0.7161, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 2.0424750400655947e-07, | |
| "loss": 0.7902, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.9044844650808468e-07, | |
| "loss": 0.705, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.7712749271311392e-07, | |
| "loss": 0.7681, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.6428529186614195e-07, | |
| "loss": 0.8181, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.519224698779198e-07, | |
| "loss": 0.6704, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.400396292949513e-07, | |
| "loss": 0.7066, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.2863734927012094e-07, | |
| "loss": 0.8501, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.1771618553447217e-07, | |
| "loss": 0.7451, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.0727667037011668e-07, | |
| "loss": 0.7253, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 9.731931258429638e-08, | |
| "loss": 0.7257, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 8.784459748458318e-08, | |
| "loss": 0.6626, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 7.885298685522235e-08, | |
| "loss": 0.7767, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 7.034491893463059e-08, | |
| "loss": 0.7786, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 6.232080839403631e-08, | |
| "loss": 0.7652, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 5.4781046317267103e-08, | |
| "loss": 0.6881, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 4.772600018168816e-08, | |
| "loss": 0.8155, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 4.115601384029666e-08, | |
| "loss": 0.7628, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 3.50714075049563e-08, | |
| "loss": 0.7743, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 2.947247773079753e-08, | |
| "loss": 0.7125, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 2.4359497401758026e-08, | |
| "loss": 0.7575, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.973271571728441e-08, | |
| "loss": 0.7412, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.5592358180189782e-08, | |
| "loss": 0.7395, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.1938626585660252e-08, | |
| "loss": 0.7093, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 8.771699011416169e-09, | |
| "loss": 0.7632, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 6.091729809042379e-09, | |
| "loss": 0.7242, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 3.898849596456477e-09, | |
| "loss": 0.7168, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 2.193165251545004e-09, | |
| "loss": 0.7389, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 9.74759906957612e-10, | |
| "loss": 0.7733, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 2.436929460525317e-10, | |
| "loss": 0.8064, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 0.0, | |
| "loss": 0.7507, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "step": 464, | |
| "total_flos": 312662581837824.0, | |
| "train_loss": 0.8204254103374893, | |
| "train_runtime": 6409.4411, | |
| "train_samples_per_second": 4.638, | |
| "train_steps_per_second": 0.072 | |
| } | |
| ], | |
| "logging_steps": 1.0, | |
| "max_steps": 464, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 50000, | |
| "total_flos": 312662581837824.0, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |