| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 2.99904, |
| "global_step": 2343, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0, |
| "learning_rate": 2.8169014084507043e-07, |
| "loss": 1.8921, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.0, |
| "learning_rate": 5.633802816901409e-07, |
| "loss": 1.8096, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.0, |
| "learning_rate": 8.450704225352114e-07, |
| "loss": 1.7, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.01, |
| "learning_rate": 1.1267605633802817e-06, |
| "loss": 1.7214, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.01, |
| "learning_rate": 1.4084507042253523e-06, |
| "loss": 1.6331, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.01, |
| "learning_rate": 1.6901408450704227e-06, |
| "loss": 1.6166, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.01, |
| "learning_rate": 1.971830985915493e-06, |
| "loss": 1.6235, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.01, |
| "learning_rate": 2.2535211267605635e-06, |
| "loss": 1.5503, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.01, |
| "learning_rate": 2.535211267605634e-06, |
| "loss": 1.7446, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.01, |
| "learning_rate": 2.8169014084507046e-06, |
| "loss": 1.7047, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.01, |
| "learning_rate": 3.0985915492957746e-06, |
| "loss": 1.5921, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 3.3802816901408454e-06, |
| "loss": 1.6417, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 3.6619718309859158e-06, |
| "loss": 1.5161, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 3.943661971830986e-06, |
| "loss": 1.6622, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 4.225352112676057e-06, |
| "loss": 1.5799, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 4.507042253521127e-06, |
| "loss": 1.6811, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 4.788732394366197e-06, |
| "loss": 1.7073, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 5.070422535211268e-06, |
| "loss": 1.6215, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.02, |
| "learning_rate": 5.352112676056338e-06, |
| "loss": 1.6431, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.03, |
| "learning_rate": 5.633802816901409e-06, |
| "loss": 1.4463, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.03, |
| "learning_rate": 5.915492957746479e-06, |
| "loss": 1.6029, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.03, |
| "learning_rate": 6.197183098591549e-06, |
| "loss": 1.5853, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.03, |
| "learning_rate": 6.478873239436621e-06, |
| "loss": 1.6092, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.03, |
| "learning_rate": 6.760563380281691e-06, |
| "loss": 1.5076, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.03, |
| "learning_rate": 7.042253521126761e-06, |
| "loss": 1.6526, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.03, |
| "learning_rate": 7.3239436619718316e-06, |
| "loss": 1.6948, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.03, |
| "learning_rate": 7.6056338028169015e-06, |
| "loss": 1.5392, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 7.887323943661972e-06, |
| "loss": 1.5999, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 8.169014084507043e-06, |
| "loss": 1.565, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 8.450704225352114e-06, |
| "loss": 1.4908, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 8.732394366197183e-06, |
| "loss": 1.6403, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 9.014084507042254e-06, |
| "loss": 1.6085, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 9.295774647887325e-06, |
| "loss": 1.5261, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 9.577464788732394e-06, |
| "loss": 1.6884, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.04, |
| "learning_rate": 9.859154929577466e-06, |
| "loss": 1.6382, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.05, |
| "learning_rate": 1.0140845070422535e-05, |
| "loss": 1.5463, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.05, |
| "learning_rate": 1.0422535211267606e-05, |
| "loss": 1.6388, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.05, |
| "learning_rate": 1.0704225352112675e-05, |
| "loss": 1.565, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.05, |
| "learning_rate": 1.0985915492957748e-05, |
| "loss": 1.6066, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.05, |
| "learning_rate": 1.1267605633802819e-05, |
| "loss": 1.6559, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.05, |
| "learning_rate": 1.1549295774647888e-05, |
| "loss": 1.6585, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.05, |
| "learning_rate": 1.1830985915492958e-05, |
| "loss": 1.6534, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 1.211267605633803e-05, |
| "loss": 1.5766, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 1.2394366197183098e-05, |
| "loss": 1.6348, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 1.2676056338028171e-05, |
| "loss": 1.4892, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 1.2957746478873242e-05, |
| "loss": 1.6072, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 1.323943661971831e-05, |
| "loss": 1.5432, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 1.3521126760563382e-05, |
| "loss": 1.5705, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 1.380281690140845e-05, |
| "loss": 1.5671, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 1.4084507042253522e-05, |
| "loss": 1.3851, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 1.4366197183098594e-05, |
| "loss": 1.7054, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 1.4647887323943663e-05, |
| "loss": 1.587, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 1.4929577464788734e-05, |
| "loss": 1.5249, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 1.5211267605633803e-05, |
| "loss": 1.5869, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 1.5492957746478872e-05, |
| "loss": 1.5268, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 1.5774647887323945e-05, |
| "loss": 1.6697, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 1.6056338028169017e-05, |
| "loss": 1.4727, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.07, |
| "learning_rate": 1.6338028169014086e-05, |
| "loss": 1.5529, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 1.6619718309859155e-05, |
| "loss": 1.5541, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 1.6901408450704228e-05, |
| "loss": 1.5959, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 1.7183098591549297e-05, |
| "loss": 1.5616, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 1.7464788732394366e-05, |
| "loss": 1.6742, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 1.774647887323944e-05, |
| "loss": 1.64, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 1.8028169014084508e-05, |
| "loss": 1.5498, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 1.830985915492958e-05, |
| "loss": 1.5754, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 1.859154929577465e-05, |
| "loss": 1.5734, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 1.887323943661972e-05, |
| "loss": 1.5026, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 1.9154929577464788e-05, |
| "loss": 1.5985, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 1.943661971830986e-05, |
| "loss": 1.5921, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 1.9718309859154933e-05, |
| "loss": 1.5652, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 2e-05, |
| "loss": 1.6493, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 1.9999990440107112e-05, |
| "loss": 1.4978, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 1.9999961760446717e-05, |
| "loss": 1.4696, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 1.999991396107366e-05, |
| "loss": 1.528, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 1.999984704207932e-05, |
| "loss": 1.6531, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 1.9999761003591655e-05, |
| "loss": 1.4925, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 1.999965584577516e-05, |
| "loss": 1.6071, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 1.9999531568830904e-05, |
| "loss": 1.6808, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 1.9999388172996495e-05, |
| "loss": 1.569, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 1.9999225658546105e-05, |
| "loss": 1.616, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 1.9999044025790462e-05, |
| "loss": 1.534, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.1, |
| "learning_rate": 1.9998843275076836e-05, |
| "loss": 1.5611, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 1.9998623406789065e-05, |
| "loss": 1.47, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 1.9998384421347525e-05, |
| "loss": 1.5568, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 1.999812631920916e-05, |
| "loss": 1.5439, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 1.999784910086745e-05, |
| "loss": 1.6501, |
| "step": 86 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 1.9997552766852434e-05, |
| "loss": 1.6142, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 1.9997237317730688e-05, |
| "loss": 1.5881, |
| "step": 88 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 1.9996902754105356e-05, |
| "loss": 1.5508, |
| "step": 89 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 1.999654907661611e-05, |
| "loss": 1.5169, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 1.9996176285939174e-05, |
| "loss": 1.5327, |
| "step": 91 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 1.9995784382787317e-05, |
| "loss": 1.4774, |
| "step": 92 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 1.9995373367909846e-05, |
| "loss": 1.5419, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 1.9994943242092616e-05, |
| "loss": 1.6472, |
| "step": 94 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 1.9994494006158018e-05, |
| "loss": 1.6383, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 1.9994025660964983e-05, |
| "loss": 1.5829, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.12, |
| "learning_rate": 1.9993538207408975e-05, |
| "loss": 1.5911, |
| "step": 97 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 1.9993031646421994e-05, |
| "loss": 1.4514, |
| "step": 98 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 1.9992505978972575e-05, |
| "loss": 1.6075, |
| "step": 99 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 1.9991961206065783e-05, |
| "loss": 1.5791, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 1.999139732874321e-05, |
| "loss": 1.6027, |
| "step": 101 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 1.999081434808298e-05, |
| "loss": 1.6114, |
| "step": 102 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 1.9990212265199738e-05, |
| "loss": 1.5659, |
| "step": 103 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 1.9989591081244654e-05, |
| "loss": 1.5509, |
| "step": 104 |
| }, |
| { |
| "epoch": 0.13, |
| "learning_rate": 1.998895079740542e-05, |
| "loss": 1.6459, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 1.9988291414906243e-05, |
| "loss": 1.5128, |
| "step": 106 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 1.9987612935007847e-05, |
| "loss": 1.6003, |
| "step": 107 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 1.9986915359007475e-05, |
| "loss": 1.6368, |
| "step": 108 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 1.9986198688238874e-05, |
| "loss": 1.5721, |
| "step": 109 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 1.9985462924072305e-05, |
| "loss": 1.4756, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 1.9984708067914533e-05, |
| "loss": 1.521, |
| "step": 111 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 1.9983934121208826e-05, |
| "loss": 1.5336, |
| "step": 112 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 1.9983141085434952e-05, |
| "loss": 1.5823, |
| "step": 113 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 1.9982328962109183e-05, |
| "loss": 1.4413, |
| "step": 114 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 1.998149775278428e-05, |
| "loss": 1.4935, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 1.9980647459049494e-05, |
| "loss": 1.5255, |
| "step": 116 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 1.9979778082530572e-05, |
| "loss": 1.561, |
| "step": 117 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 1.997888962488974e-05, |
| "loss": 1.5925, |
| "step": 118 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 1.9977982087825714e-05, |
| "loss": 1.6209, |
| "step": 119 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 1.997705547307368e-05, |
| "loss": 1.5829, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.15, |
| "learning_rate": 1.9976109782405314e-05, |
| "loss": 1.5361, |
| "step": 121 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 1.9975145017628745e-05, |
| "loss": 1.528, |
| "step": 122 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 1.9974161180588593e-05, |
| "loss": 1.5733, |
| "step": 123 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 1.9973158273165925e-05, |
| "loss": 1.5336, |
| "step": 124 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 1.9972136297278285e-05, |
| "loss": 1.5187, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 1.997109525487967e-05, |
| "loss": 1.5704, |
| "step": 126 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 1.9970035147960524e-05, |
| "loss": 1.6741, |
| "step": 127 |
| }, |
| { |
| "epoch": 0.16, |
| "learning_rate": 1.9968955978547753e-05, |
| "loss": 1.5884, |
| "step": 128 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 1.9967857748704705e-05, |
| "loss": 1.5621, |
| "step": 129 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 1.9966740460531172e-05, |
| "loss": 1.4549, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 1.996560411616338e-05, |
| "loss": 1.5865, |
| "step": 131 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 1.9964448717774008e-05, |
| "loss": 1.5352, |
| "step": 132 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 1.9963274267572143e-05, |
| "loss": 1.4776, |
| "step": 133 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 1.9962080767803303e-05, |
| "loss": 1.5916, |
| "step": 134 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 1.996086822074945e-05, |
| "loss": 1.423, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 1.9959636628728938e-05, |
| "loss": 1.5618, |
| "step": 136 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 1.9958385994096543e-05, |
| "loss": 1.4755, |
| "step": 137 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 1.9957116319243458e-05, |
| "loss": 1.4957, |
| "step": 138 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 1.995582760659727e-05, |
| "loss": 1.5618, |
| "step": 139 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 1.995451985862197e-05, |
| "loss": 1.4713, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 1.995319307781794e-05, |
| "loss": 1.5717, |
| "step": 141 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 1.995184726672197e-05, |
| "loss": 1.5397, |
| "step": 142 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 1.995048242790721e-05, |
| "loss": 1.5383, |
| "step": 143 |
| }, |
| { |
| "epoch": 0.18, |
| "learning_rate": 1.9949098563983206e-05, |
| "loss": 1.5889, |
| "step": 144 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 1.9947695677595877e-05, |
| "loss": 1.4919, |
| "step": 145 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 1.994627377142751e-05, |
| "loss": 1.4836, |
| "step": 146 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 1.9944832848196757e-05, |
| "loss": 1.4114, |
| "step": 147 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 1.994337291065864e-05, |
| "loss": 1.5828, |
| "step": 148 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 1.994189396160452e-05, |
| "loss": 1.5641, |
| "step": 149 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 1.9940396003862123e-05, |
| "loss": 1.5425, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 1.9938879040295508e-05, |
| "loss": 1.5539, |
| "step": 151 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 1.9937343073805078e-05, |
| "loss": 1.5581, |
| "step": 152 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 1.9935788107327565e-05, |
| "loss": 1.6151, |
| "step": 153 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 1.9934214143836038e-05, |
| "loss": 1.581, |
| "step": 154 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 1.9932621186339873e-05, |
| "loss": 1.4626, |
| "step": 155 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 1.993100923788478e-05, |
| "loss": 1.5291, |
| "step": 156 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 1.992937830155276e-05, |
| "loss": 1.6632, |
| "step": 157 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 1.9927728380462133e-05, |
| "loss": 1.5383, |
| "step": 158 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 1.992605947776752e-05, |
| "loss": 1.5726, |
| "step": 159 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 1.9924371596659816e-05, |
| "loss": 1.5397, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.21, |
| "learning_rate": 1.992266474036622e-05, |
| "loss": 1.5604, |
| "step": 161 |
| }, |
| { |
| "epoch": 0.21, |
| "learning_rate": 1.99209389121502e-05, |
| "loss": 1.6912, |
| "step": 162 |
| }, |
| { |
| "epoch": 0.21, |
| "learning_rate": 1.9919194115311505e-05, |
| "loss": 1.6031, |
| "step": 163 |
| }, |
| { |
| "epoch": 0.21, |
| "learning_rate": 1.9917430353186155e-05, |
| "loss": 1.4782, |
| "step": 164 |
| }, |
| { |
| "epoch": 0.21, |
| "learning_rate": 1.991564762914641e-05, |
| "loss": 1.5983, |
| "step": 165 |
| }, |
| { |
| "epoch": 0.21, |
| "learning_rate": 1.991384594660082e-05, |
| "loss": 1.5054, |
| "step": 166 |
| }, |
| { |
| "epoch": 0.21, |
| "learning_rate": 1.9912025308994146e-05, |
| "loss": 1.5919, |
| "step": 167 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 1.991018571980742e-05, |
| "loss": 1.5814, |
| "step": 168 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 1.990832718255789e-05, |
| "loss": 1.6701, |
| "step": 169 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 1.9906449700799045e-05, |
| "loss": 1.5032, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 1.9904553278120585e-05, |
| "loss": 1.4681, |
| "step": 171 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 1.9902637918148434e-05, |
| "loss": 1.6125, |
| "step": 172 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 1.9900703624544713e-05, |
| "loss": 1.5088, |
| "step": 173 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 1.9898750401007755e-05, |
| "loss": 1.527, |
| "step": 174 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 1.989677825127208e-05, |
| "loss": 1.554, |
| "step": 175 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 1.9894787179108395e-05, |
| "loss": 1.5321, |
| "step": 176 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 1.989277718832359e-05, |
| "loss": 1.584, |
| "step": 177 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 1.9890748282760723e-05, |
| "loss": 1.6019, |
| "step": 178 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 1.988870046629902e-05, |
| "loss": 1.5916, |
| "step": 179 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 1.988663374285386e-05, |
| "loss": 1.5527, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 1.9884548116376767e-05, |
| "loss": 1.5735, |
| "step": 181 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 1.9882443590855427e-05, |
| "loss": 1.6428, |
| "step": 182 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 1.988032017031364e-05, |
| "loss": 1.5624, |
| "step": 183 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 1.9878177858811346e-05, |
| "loss": 1.5176, |
| "step": 184 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 1.9876016660444594e-05, |
| "loss": 1.4849, |
| "step": 185 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 1.987383657934555e-05, |
| "loss": 1.5434, |
| "step": 186 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 1.987163761968248e-05, |
| "loss": 1.5246, |
| "step": 187 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 1.986941978565975e-05, |
| "loss": 1.346, |
| "step": 188 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 1.9867183081517812e-05, |
| "loss": 1.604, |
| "step": 189 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 1.98649275115332e-05, |
| "loss": 1.4752, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.24, |
| "learning_rate": 1.9862653080018508e-05, |
| "loss": 1.5684, |
| "step": 191 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 1.9860359791322403e-05, |
| "loss": 1.6076, |
| "step": 192 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 1.9858047649829607e-05, |
| "loss": 1.5748, |
| "step": 193 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 1.985571665996088e-05, |
| "loss": 1.5715, |
| "step": 194 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 1.985336682617303e-05, |
| "loss": 1.6314, |
| "step": 195 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 1.985099815295888e-05, |
| "loss": 1.5235, |
| "step": 196 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 1.9848610644847296e-05, |
| "loss": 1.4343, |
| "step": 197 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 1.984620430640313e-05, |
| "loss": 1.5752, |
| "step": 198 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 1.9843779142227258e-05, |
| "loss": 1.5915, |
| "step": 199 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 1.9841335156956535e-05, |
| "loss": 1.6355, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 1.9838872355263813e-05, |
| "loss": 1.5881, |
| "step": 201 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 1.9836390741857914e-05, |
| "loss": 1.4578, |
| "step": 202 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 1.9833890321483636e-05, |
| "loss": 1.6122, |
| "step": 203 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 1.983137109892172e-05, |
| "loss": 1.4533, |
| "step": 204 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 1.982883307898887e-05, |
| "loss": 1.6115, |
| "step": 205 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 1.9826276266537726e-05, |
| "loss": 1.5985, |
| "step": 206 |
| }, |
| { |
| "epoch": 0.26, |
| "learning_rate": 1.9823700666456854e-05, |
| "loss": 1.5907, |
| "step": 207 |
| }, |
| { |
| "epoch": 0.27, |
| "learning_rate": 1.9821106283670753e-05, |
| "loss": 1.575, |
| "step": 208 |
| }, |
| { |
| "epoch": 0.27, |
| "learning_rate": 1.9818493123139825e-05, |
| "loss": 1.6302, |
| "step": 209 |
| }, |
| { |
| "epoch": 0.27, |
| "learning_rate": 1.9815861189860373e-05, |
| "loss": 1.5568, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.27, |
| "learning_rate": 1.9813210488864604e-05, |
| "loss": 1.6496, |
| "step": 211 |
| }, |
| { |
| "epoch": 0.27, |
| "learning_rate": 1.9810541025220594e-05, |
| "loss": 1.526, |
| "step": 212 |
| }, |
| { |
| "epoch": 0.27, |
| "learning_rate": 1.9807852804032306e-05, |
| "loss": 1.5699, |
| "step": 213 |
| }, |
| { |
| "epoch": 0.27, |
| "learning_rate": 1.980514583043956e-05, |
| "loss": 1.4829, |
| "step": 214 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 1.980242010961803e-05, |
| "loss": 1.5604, |
| "step": 215 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 1.979967564677924e-05, |
| "loss": 1.4396, |
| "step": 216 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 1.9796912447170536e-05, |
| "loss": 1.5015, |
| "step": 217 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 1.9794130516075104e-05, |
| "loss": 1.563, |
| "step": 218 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 1.979132985881193e-05, |
| "loss": 1.602, |
| "step": 219 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 1.9788510480735822e-05, |
| "loss": 1.6708, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 1.9785672387237357e-05, |
| "loss": 1.443, |
| "step": 221 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 1.9782815583742917e-05, |
| "loss": 1.6391, |
| "step": 222 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 1.977994007571465e-05, |
| "loss": 1.4925, |
| "step": 223 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 1.977704586865046e-05, |
| "loss": 1.5083, |
| "step": 224 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 1.9774132968084013e-05, |
| "loss": 1.5564, |
| "step": 225 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 1.977120137958471e-05, |
| "loss": 1.5835, |
| "step": 226 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 1.976825110875769e-05, |
| "loss": 1.4686, |
| "step": 227 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 1.9765282161243803e-05, |
| "loss": 1.4561, |
| "step": 228 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 1.9762294542719612e-05, |
| "loss": 1.5031, |
| "step": 229 |
| }, |
| { |
| "epoch": 0.29, |
| "learning_rate": 1.9759288258897385e-05, |
| "loss": 1.5902, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.3, |
| "learning_rate": 1.975626331552507e-05, |
| "loss": 1.6597, |
| "step": 231 |
| }, |
| { |
| "epoch": 0.3, |
| "learning_rate": 1.9753219718386293e-05, |
| "loss": 1.508, |
| "step": 232 |
| }, |
| { |
| "epoch": 0.3, |
| "learning_rate": 1.9750157473300346e-05, |
| "loss": 1.6212, |
| "step": 233 |
| }, |
| { |
| "epoch": 0.3, |
| "learning_rate": 1.9747076586122175e-05, |
| "loss": 1.5397, |
| "step": 234 |
| }, |
| { |
| "epoch": 0.3, |
| "learning_rate": 1.9743977062742373e-05, |
| "loss": 1.6058, |
| "step": 235 |
| }, |
| { |
| "epoch": 0.3, |
| "learning_rate": 1.9740858909087165e-05, |
| "loss": 1.5216, |
| "step": 236 |
| }, |
| { |
| "epoch": 0.3, |
| "learning_rate": 1.9737722131118385e-05, |
| "loss": 1.62, |
| "step": 237 |
| }, |
| { |
| "epoch": 0.3, |
| "learning_rate": 1.9734566734833492e-05, |
| "loss": 1.5419, |
| "step": 238 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 1.9731392726265538e-05, |
| "loss": 1.5845, |
| "step": 239 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 1.9728200111483155e-05, |
| "loss": 1.5045, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 1.9724988896590553e-05, |
| "loss": 1.5712, |
| "step": 241 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 1.972175908772751e-05, |
| "loss": 1.5249, |
| "step": 242 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 1.9718510691069352e-05, |
| "loss": 1.5461, |
| "step": 243 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 1.9715243712826938e-05, |
| "loss": 1.52, |
| "step": 244 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 1.9711958159246665e-05, |
| "loss": 1.4509, |
| "step": 245 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 1.970865403661044e-05, |
| "loss": 1.4433, |
| "step": 246 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 1.9705331351235673e-05, |
| "loss": 1.555, |
| "step": 247 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 1.9701990109475273e-05, |
| "loss": 1.5605, |
| "step": 248 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 1.9698630317717617e-05, |
| "loss": 1.6769, |
| "step": 249 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 1.9695251982386553e-05, |
| "loss": 1.6464, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 1.9691855109941394e-05, |
| "loss": 1.4653, |
| "step": 251 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 1.968843970687688e-05, |
| "loss": 1.4217, |
| "step": 252 |
| }, |
| { |
| "epoch": 0.32, |
| "learning_rate": 1.9685005779723192e-05, |
| "loss": 1.5234, |
| "step": 253 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 1.9681553335045925e-05, |
| "loss": 1.5572, |
| "step": 254 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 1.967808237944608e-05, |
| "loss": 1.4501, |
| "step": 255 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 1.9674592919560046e-05, |
| "loss": 1.5247, |
| "step": 256 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 1.9671084962059596e-05, |
| "loss": 1.5897, |
| "step": 257 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 1.9667558513651875e-05, |
| "loss": 1.5331, |
| "step": 258 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 1.966401358107937e-05, |
| "loss": 1.4277, |
| "step": 259 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 1.9660450171119923e-05, |
| "loss": 1.6509, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 1.965686829058669e-05, |
| "loss": 1.469, |
| "step": 261 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 1.965326794632816e-05, |
| "loss": 1.5032, |
| "step": 262 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 1.96496491452281e-05, |
| "loss": 1.4701, |
| "step": 263 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 1.9646011894205596e-05, |
| "loss": 1.6634, |
| "step": 264 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 1.9642356200214977e-05, |
| "loss": 1.568, |
| "step": 265 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 1.963868207024587e-05, |
| "loss": 1.6312, |
| "step": 266 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 1.9634989511323118e-05, |
| "loss": 1.6529, |
| "step": 267 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 1.9631278530506817e-05, |
| "loss": 1.5365, |
| "step": 268 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 1.9627549134892293e-05, |
| "loss": 1.3881, |
| "step": 269 |
| }, |
| { |
| "epoch": 0.35, |
| "learning_rate": 1.962380133161006e-05, |
| "loss": 1.5303, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.35, |
| "learning_rate": 1.962003512782584e-05, |
| "loss": 1.6535, |
| "step": 271 |
| }, |
| { |
| "epoch": 0.35, |
| "learning_rate": 1.961625053074054e-05, |
| "loss": 1.5175, |
| "step": 272 |
| }, |
| { |
| "epoch": 0.35, |
| "learning_rate": 1.9612447547590216e-05, |
| "loss": 1.6222, |
| "step": 273 |
| }, |
| { |
| "epoch": 0.35, |
| "learning_rate": 1.96086261856461e-05, |
| "loss": 1.5488, |
| "step": 274 |
| }, |
| { |
| "epoch": 0.35, |
| "learning_rate": 1.9604786452214554e-05, |
| "loss": 1.5521, |
| "step": 275 |
| }, |
| { |
| "epoch": 0.35, |
| "learning_rate": 1.9600928354637063e-05, |
| "loss": 1.6231, |
| "step": 276 |
| }, |
| { |
| "epoch": 0.35, |
| "learning_rate": 1.959705190029023e-05, |
| "loss": 1.6081, |
| "step": 277 |
| }, |
| { |
| "epoch": 0.36, |
| "learning_rate": 1.9593157096585747e-05, |
| "loss": 1.5731, |
| "step": 278 |
| }, |
| { |
| "epoch": 0.36, |
| "learning_rate": 1.95892439509704e-05, |
| "loss": 1.5925, |
| "step": 279 |
| }, |
| { |
| "epoch": 0.36, |
| "learning_rate": 1.9585312470926043e-05, |
| "loss": 1.4548, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.36, |
| "learning_rate": 1.9581362663969572e-05, |
| "loss": 1.5281, |
| "step": 281 |
| }, |
| { |
| "epoch": 0.36, |
| "learning_rate": 1.957739453765294e-05, |
| "loss": 1.5626, |
| "step": 282 |
| }, |
| { |
| "epoch": 0.36, |
| "learning_rate": 1.957340809956312e-05, |
| "loss": 1.646, |
| "step": 283 |
| }, |
| { |
| "epoch": 0.36, |
| "learning_rate": 1.956940335732209e-05, |
| "loss": 1.4231, |
| "step": 284 |
| }, |
| { |
| "epoch": 0.36, |
| "learning_rate": 1.9565380318586838e-05, |
| "loss": 1.5245, |
| "step": 285 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 1.9561338991049323e-05, |
| "loss": 1.4708, |
| "step": 286 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 1.9557279382436483e-05, |
| "loss": 1.6048, |
| "step": 287 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 1.9553201500510197e-05, |
| "loss": 1.5894, |
| "step": 288 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 1.954910535306729e-05, |
| "loss": 1.4496, |
| "step": 289 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 1.9544990947939504e-05, |
| "loss": 1.4983, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 1.95408582929935e-05, |
| "loss": 1.4941, |
| "step": 291 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 1.9536707396130826e-05, |
| "loss": 1.5229, |
| "step": 292 |
| }, |
| { |
| "epoch": 0.38, |
| "learning_rate": 1.95325382652879e-05, |
| "loss": 1.5914, |
| "step": 293 |
| }, |
| { |
| "epoch": 0.38, |
| "learning_rate": 1.952835090843602e-05, |
| "loss": 1.4926, |
| "step": 294 |
| }, |
| { |
| "epoch": 0.38, |
| "learning_rate": 1.9524145333581315e-05, |
| "loss": 1.5251, |
| "step": 295 |
| }, |
| { |
| "epoch": 0.38, |
| "learning_rate": 1.951992154876476e-05, |
| "loss": 1.6215, |
| "step": 296 |
| }, |
| { |
| "epoch": 0.38, |
| "learning_rate": 1.951567956206214e-05, |
| "loss": 1.5768, |
| "step": 297 |
| }, |
| { |
| "epoch": 0.38, |
| "learning_rate": 1.9511419381584033e-05, |
| "loss": 1.5086, |
| "step": 298 |
| }, |
| { |
| "epoch": 0.38, |
| "learning_rate": 1.9507141015475828e-05, |
| "loss": 1.5019, |
| "step": 299 |
| }, |
| { |
| "epoch": 0.38, |
| "learning_rate": 1.950284447191766e-05, |
| "loss": 1.5381, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.39, |
| "learning_rate": 1.949852975912443e-05, |
| "loss": 1.5158, |
| "step": 301 |
| }, |
| { |
| "epoch": 0.39, |
| "learning_rate": 1.949419688534578e-05, |
| "loss": 1.507, |
| "step": 302 |
| }, |
| { |
| "epoch": 0.39, |
| "learning_rate": 1.9489845858866066e-05, |
| "loss": 1.6007, |
| "step": 303 |
| }, |
| { |
| "epoch": 0.39, |
| "learning_rate": 1.948547668800436e-05, |
| "loss": 1.5318, |
| "step": 304 |
| }, |
| { |
| "epoch": 0.39, |
| "learning_rate": 1.9481089381114427e-05, |
| "loss": 1.544, |
| "step": 305 |
| }, |
| { |
| "epoch": 0.39, |
| "learning_rate": 1.94766839465847e-05, |
| "loss": 1.6778, |
| "step": 306 |
| }, |
| { |
| "epoch": 0.39, |
| "learning_rate": 1.9472260392838272e-05, |
| "loss": 1.4475, |
| "step": 307 |
| }, |
| { |
| "epoch": 0.39, |
| "learning_rate": 1.9467818728332887e-05, |
| "loss": 1.4946, |
| "step": 308 |
| }, |
| { |
| "epoch": 0.4, |
| "learning_rate": 1.946335896156091e-05, |
| "loss": 1.4903, |
| "step": 309 |
| }, |
| { |
| "epoch": 0.4, |
| "learning_rate": 1.945888110104933e-05, |
| "loss": 1.5962, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.4, |
| "learning_rate": 1.9454385155359704e-05, |
| "loss": 1.4955, |
| "step": 311 |
| }, |
| { |
| "epoch": 0.4, |
| "learning_rate": 1.9449871133088197e-05, |
| "loss": 1.4937, |
| "step": 312 |
| }, |
| { |
| "epoch": 0.4, |
| "learning_rate": 1.9445339042865513e-05, |
| "loss": 1.5271, |
| "step": 313 |
| }, |
| { |
| "epoch": 0.4, |
| "learning_rate": 1.9440788893356917e-05, |
| "loss": 1.4787, |
| "step": 314 |
| }, |
| { |
| "epoch": 0.4, |
| "learning_rate": 1.9436220693262196e-05, |
| "loss": 1.5266, |
| "step": 315 |
| }, |
| { |
| "epoch": 0.4, |
| "learning_rate": 1.9431634451315656e-05, |
| "loss": 1.59, |
| "step": 316 |
| }, |
| { |
| "epoch": 0.41, |
| "learning_rate": 1.9427030176286084e-05, |
| "loss": 1.5541, |
| "step": 317 |
| }, |
| { |
| "epoch": 0.41, |
| "learning_rate": 1.942240787697676e-05, |
| "loss": 1.5428, |
| "step": 318 |
| }, |
| { |
| "epoch": 0.41, |
| "learning_rate": 1.9417767562225422e-05, |
| "loss": 1.5705, |
| "step": 319 |
| }, |
| { |
| "epoch": 0.41, |
| "learning_rate": 1.941310924090425e-05, |
| "loss": 1.5075, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.41, |
| "learning_rate": 1.9408432921919858e-05, |
| "loss": 1.6058, |
| "step": 321 |
| }, |
| { |
| "epoch": 0.41, |
| "learning_rate": 1.9403738614213266e-05, |
| "loss": 1.5554, |
| "step": 322 |
| }, |
| { |
| "epoch": 0.41, |
| "learning_rate": 1.9399026326759886e-05, |
| "loss": 1.426, |
| "step": 323 |
| }, |
| { |
| "epoch": 0.41, |
| "learning_rate": 1.9394296068569517e-05, |
| "loss": 1.5018, |
| "step": 324 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 1.938954784868631e-05, |
| "loss": 1.479, |
| "step": 325 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 1.9384781676188756e-05, |
| "loss": 1.6239, |
| "step": 326 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 1.9379997560189677e-05, |
| "loss": 1.4323, |
| "step": 327 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 1.93751955098362e-05, |
| "loss": 1.5905, |
| "step": 328 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 1.9370375534309743e-05, |
| "loss": 1.559, |
| "step": 329 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 1.9365537642825996e-05, |
| "loss": 1.5472, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 1.9360681844634903e-05, |
| "loss": 1.4352, |
| "step": 331 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 1.9355808149020645e-05, |
| "loss": 1.5759, |
| "step": 332 |
| }, |
| { |
| "epoch": 0.43, |
| "learning_rate": 1.935091656530163e-05, |
| "loss": 1.5138, |
| "step": 333 |
| }, |
| { |
| "epoch": 0.43, |
| "learning_rate": 1.934600710283045e-05, |
| "loss": 1.4506, |
| "step": 334 |
| }, |
| { |
| "epoch": 0.43, |
| "learning_rate": 1.93410797709939e-05, |
| "loss": 1.5263, |
| "step": 335 |
| }, |
| { |
| "epoch": 0.43, |
| "learning_rate": 1.9336134579212933e-05, |
| "loss": 1.447, |
| "step": 336 |
| }, |
| { |
| "epoch": 0.43, |
| "learning_rate": 1.933117153694265e-05, |
| "loss": 1.4765, |
| "step": 337 |
| }, |
| { |
| "epoch": 0.43, |
| "learning_rate": 1.9326190653672277e-05, |
| "loss": 1.5694, |
| "step": 338 |
| }, |
| { |
| "epoch": 0.43, |
| "learning_rate": 1.9321191938925156e-05, |
| "loss": 1.5056, |
| "step": 339 |
| }, |
| { |
| "epoch": 0.44, |
| "learning_rate": 1.9316175402258733e-05, |
| "loss": 1.53, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.44, |
| "learning_rate": 1.9311141053264502e-05, |
| "loss": 1.5962, |
| "step": 341 |
| }, |
| { |
| "epoch": 0.44, |
| "learning_rate": 1.9306088901568047e-05, |
| "loss": 1.6619, |
| "step": 342 |
| }, |
| { |
| "epoch": 0.44, |
| "learning_rate": 1.9301018956828966e-05, |
| "loss": 1.5669, |
| "step": 343 |
| }, |
| { |
| "epoch": 0.44, |
| "learning_rate": 1.9295931228740882e-05, |
| "loss": 1.5196, |
| "step": 344 |
| }, |
| { |
| "epoch": 0.44, |
| "learning_rate": 1.9290825727031425e-05, |
| "loss": 1.5164, |
| "step": 345 |
| }, |
| { |
| "epoch": 0.44, |
| "learning_rate": 1.9285702461462204e-05, |
| "loss": 1.5818, |
| "step": 346 |
| }, |
| { |
| "epoch": 0.44, |
| "learning_rate": 1.9280561441828794e-05, |
| "loss": 1.5524, |
| "step": 347 |
| }, |
| { |
| "epoch": 0.45, |
| "learning_rate": 1.9275402677960716e-05, |
| "loss": 1.5758, |
| "step": 348 |
| }, |
| { |
| "epoch": 0.45, |
| "learning_rate": 1.9270226179721416e-05, |
| "loss": 1.4895, |
| "step": 349 |
| }, |
| { |
| "epoch": 0.45, |
| "learning_rate": 1.9265031957008243e-05, |
| "loss": 1.5604, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.45, |
| "learning_rate": 1.9259820019752445e-05, |
| "loss": 1.6659, |
| "step": 351 |
| }, |
| { |
| "epoch": 0.45, |
| "learning_rate": 1.9254590377919128e-05, |
| "loss": 1.5113, |
| "step": 352 |
| }, |
| { |
| "epoch": 0.45, |
| "learning_rate": 1.9249343041507264e-05, |
| "loss": 1.492, |
| "step": 353 |
| }, |
| { |
| "epoch": 0.45, |
| "learning_rate": 1.924407802054964e-05, |
| "loss": 1.5277, |
| "step": 354 |
| }, |
| { |
| "epoch": 0.45, |
| "learning_rate": 1.9238795325112867e-05, |
| "loss": 1.5835, |
| "step": 355 |
| }, |
| { |
| "epoch": 0.46, |
| "learning_rate": 1.923349496529735e-05, |
| "loss": 1.5611, |
| "step": 356 |
| }, |
| { |
| "epoch": 0.46, |
| "learning_rate": 1.9228176951237252e-05, |
| "loss": 1.6418, |
| "step": 357 |
| }, |
| { |
| "epoch": 0.46, |
| "learning_rate": 1.922284129310051e-05, |
| "loss": 1.512, |
| "step": 358 |
| }, |
| { |
| "epoch": 0.46, |
| "learning_rate": 1.9217488001088784e-05, |
| "loss": 1.6668, |
| "step": 359 |
| }, |
| { |
| "epoch": 0.46, |
| "learning_rate": 1.921211708543746e-05, |
| "loss": 1.4548, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.46, |
| "learning_rate": 1.9206728556415604e-05, |
| "loss": 1.5987, |
| "step": 361 |
| }, |
| { |
| "epoch": 0.46, |
| "learning_rate": 1.9201322424325982e-05, |
| "loss": 1.5106, |
| "step": 362 |
| }, |
| { |
| "epoch": 0.46, |
| "learning_rate": 1.919589869950499e-05, |
| "loss": 1.6269, |
| "step": 363 |
| }, |
| { |
| "epoch": 0.47, |
| "learning_rate": 1.9190457392322677e-05, |
| "loss": 1.5116, |
| "step": 364 |
| }, |
| { |
| "epoch": 0.47, |
| "learning_rate": 1.9184998513182705e-05, |
| "loss": 1.5542, |
| "step": 365 |
| }, |
| { |
| "epoch": 0.47, |
| "learning_rate": 1.917952207252234e-05, |
| "loss": 1.6317, |
| "step": 366 |
| }, |
| { |
| "epoch": 0.47, |
| "learning_rate": 1.9174028080812415e-05, |
| "loss": 1.5089, |
| "step": 367 |
| }, |
| { |
| "epoch": 0.47, |
| "learning_rate": 1.9168516548557322e-05, |
| "loss": 1.4765, |
| "step": 368 |
| }, |
| { |
| "epoch": 0.47, |
| "learning_rate": 1.9162987486295e-05, |
| "loss": 1.5436, |
| "step": 369 |
| }, |
| { |
| "epoch": 0.47, |
| "learning_rate": 1.9157440904596886e-05, |
| "loss": 1.536, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.47, |
| "learning_rate": 1.9151876814067932e-05, |
| "loss": 1.6647, |
| "step": 371 |
| }, |
| { |
| "epoch": 0.48, |
| "learning_rate": 1.9146295225346566e-05, |
| "loss": 1.4496, |
| "step": 372 |
| }, |
| { |
| "epoch": 0.48, |
| "learning_rate": 1.9140696149104657e-05, |
| "loss": 1.5858, |
| "step": 373 |
| }, |
| { |
| "epoch": 0.48, |
| "learning_rate": 1.9135079596047522e-05, |
| "loss": 1.4961, |
| "step": 374 |
| }, |
| { |
| "epoch": 0.48, |
| "learning_rate": 1.9129445576913886e-05, |
| "loss": 1.5313, |
| "step": 375 |
| }, |
| { |
| "epoch": 0.48, |
| "learning_rate": 1.9123794102475884e-05, |
| "loss": 1.6396, |
| "step": 376 |
| }, |
| { |
| "epoch": 0.48, |
| "learning_rate": 1.9118125183539003e-05, |
| "loss": 1.5159, |
| "step": 377 |
| }, |
| { |
| "epoch": 0.48, |
| "learning_rate": 1.91124388309421e-05, |
| "loss": 1.5542, |
| "step": 378 |
| }, |
| { |
| "epoch": 0.49, |
| "learning_rate": 1.910673505555736e-05, |
| "loss": 1.5676, |
| "step": 379 |
| }, |
| { |
| "epoch": 0.49, |
| "learning_rate": 1.9101013868290274e-05, |
| "loss": 1.4897, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.49, |
| "learning_rate": 1.909527528007963e-05, |
| "loss": 1.5595, |
| "step": 381 |
| }, |
| { |
| "epoch": 0.49, |
| "learning_rate": 1.9089519301897494e-05, |
| "loss": 1.528, |
| "step": 382 |
| }, |
| { |
| "epoch": 0.49, |
| "learning_rate": 1.9083745944749163e-05, |
| "loss": 1.4923, |
| "step": 383 |
| }, |
| { |
| "epoch": 0.49, |
| "learning_rate": 1.907795521967318e-05, |
| "loss": 1.4445, |
| "step": 384 |
| }, |
| { |
| "epoch": 0.49, |
| "learning_rate": 1.907214713774128e-05, |
| "loss": 1.3598, |
| "step": 385 |
| }, |
| { |
| "epoch": 0.49, |
| "learning_rate": 1.90663217100584e-05, |
| "loss": 1.5494, |
| "step": 386 |
| }, |
| { |
| "epoch": 0.5, |
| "learning_rate": 1.9060478947762625e-05, |
| "loss": 1.5066, |
| "step": 387 |
| }, |
| { |
| "epoch": 0.5, |
| "learning_rate": 1.9054618862025198e-05, |
| "loss": 1.5562, |
| "step": 388 |
| }, |
| { |
| "epoch": 0.5, |
| "learning_rate": 1.9048741464050468e-05, |
| "loss": 1.5286, |
| "step": 389 |
| }, |
| { |
| "epoch": 0.5, |
| "learning_rate": 1.9042846765075905e-05, |
| "loss": 1.5854, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.5, |
| "learning_rate": 1.903693477637204e-05, |
| "loss": 1.6033, |
| "step": 391 |
| }, |
| { |
| "epoch": 0.5, |
| "learning_rate": 1.903100550924247e-05, |
| "loss": 1.5679, |
| "step": 392 |
| }, |
| { |
| "epoch": 0.5, |
| "learning_rate": 1.902505897502383e-05, |
| "loss": 1.5367, |
| "step": 393 |
| }, |
| { |
| "epoch": 0.5, |
| "learning_rate": 1.901909518508576e-05, |
| "loss": 1.5428, |
| "step": 394 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 1.9013114150830904e-05, |
| "loss": 1.5304, |
| "step": 395 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 1.9007115883694872e-05, |
| "loss": 1.6188, |
| "step": 396 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 1.9001100395146216e-05, |
| "loss": 1.5219, |
| "step": 397 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 1.899506769668643e-05, |
| "loss": 1.5169, |
| "step": 398 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 1.8989017799849896e-05, |
| "loss": 1.3971, |
| "step": 399 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 1.8982950716203888e-05, |
| "loss": 1.5518, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 1.8976866457348547e-05, |
| "loss": 1.4723, |
| "step": 401 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 1.897076503491684e-05, |
| "loss": 1.4152, |
| "step": 402 |
| }, |
| { |
| "epoch": 0.52, |
| "learning_rate": 1.8964646460574555e-05, |
| "loss": 1.5079, |
| "step": 403 |
| }, |
| { |
| "epoch": 0.52, |
| "learning_rate": 1.8958510746020276e-05, |
| "loss": 1.5701, |
| "step": 404 |
| }, |
| { |
| "epoch": 0.52, |
| "learning_rate": 1.8952357902985363e-05, |
| "loss": 1.6675, |
| "step": 405 |
| }, |
| { |
| "epoch": 0.52, |
| "learning_rate": 1.894618794323391e-05, |
| "loss": 1.5312, |
| "step": 406 |
| }, |
| { |
| "epoch": 0.52, |
| "learning_rate": 1.8940000878562758e-05, |
| "loss": 1.503, |
| "step": 407 |
| }, |
| { |
| "epoch": 0.52, |
| "learning_rate": 1.8933796720801437e-05, |
| "loss": 1.5588, |
| "step": 408 |
| }, |
| { |
| "epoch": 0.52, |
| "learning_rate": 1.8927575481812168e-05, |
| "loss": 1.6306, |
| "step": 409 |
| }, |
| { |
| "epoch": 0.52, |
| "learning_rate": 1.892133717348982e-05, |
| "loss": 1.5523, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.53, |
| "learning_rate": 1.8915081807761907e-05, |
| "loss": 1.489, |
| "step": 411 |
| }, |
| { |
| "epoch": 0.53, |
| "learning_rate": 1.8908809396588557e-05, |
| "loss": 1.4722, |
| "step": 412 |
| }, |
| { |
| "epoch": 0.53, |
| "learning_rate": 1.8902519951962483e-05, |
| "loss": 1.6297, |
| "step": 413 |
| }, |
| { |
| "epoch": 0.53, |
| "learning_rate": 1.889621348590897e-05, |
| "loss": 1.4199, |
| "step": 414 |
| }, |
| { |
| "epoch": 0.53, |
| "learning_rate": 1.8889890010485847e-05, |
| "loss": 1.5319, |
| "step": 415 |
| }, |
| { |
| "epoch": 0.53, |
| "learning_rate": 1.888354953778346e-05, |
| "loss": 1.5826, |
| "step": 416 |
| }, |
| { |
| "epoch": 0.53, |
| "learning_rate": 1.887719207992466e-05, |
| "loss": 1.5537, |
| "step": 417 |
| }, |
| { |
| "epoch": 0.54, |
| "learning_rate": 1.887081764906477e-05, |
| "loss": 1.6173, |
| "step": 418 |
| }, |
| { |
| "epoch": 0.54, |
| "learning_rate": 1.8864426257391567e-05, |
| "loss": 1.5516, |
| "step": 419 |
| }, |
| { |
| "epoch": 0.54, |
| "learning_rate": 1.885801791712525e-05, |
| "loss": 1.6347, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.54, |
| "learning_rate": 1.8851592640518433e-05, |
| "loss": 1.4851, |
| "step": 421 |
| }, |
| { |
| "epoch": 0.54, |
| "learning_rate": 1.8845150439856102e-05, |
| "loss": 1.3995, |
| "step": 422 |
| }, |
| { |
| "epoch": 0.54, |
| "learning_rate": 1.883869132745561e-05, |
| "loss": 1.5242, |
| "step": 423 |
| }, |
| { |
| "epoch": 0.54, |
| "learning_rate": 1.8832215315666645e-05, |
| "loss": 1.4844, |
| "step": 424 |
| }, |
| { |
| "epoch": 0.54, |
| "learning_rate": 1.8825722416871197e-05, |
| "loss": 1.6318, |
| "step": 425 |
| }, |
| { |
| "epoch": 0.55, |
| "learning_rate": 1.881921264348355e-05, |
| "loss": 1.5529, |
| "step": 426 |
| }, |
| { |
| "epoch": 0.55, |
| "learning_rate": 1.8812686007950257e-05, |
| "loss": 1.571, |
| "step": 427 |
| }, |
| { |
| "epoch": 0.55, |
| "learning_rate": 1.8806142522750098e-05, |
| "loss": 1.481, |
| "step": 428 |
| }, |
| { |
| "epoch": 0.55, |
| "learning_rate": 1.879958220039408e-05, |
| "loss": 1.5949, |
| "step": 429 |
| }, |
| { |
| "epoch": 0.55, |
| "learning_rate": 1.87930050534254e-05, |
| "loss": 1.5699, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.55, |
| "learning_rate": 1.878641109441942e-05, |
| "loss": 1.5897, |
| "step": 431 |
| }, |
| { |
| "epoch": 0.55, |
| "learning_rate": 1.8779800335983648e-05, |
| "loss": 1.4738, |
| "step": 432 |
| }, |
| { |
| "epoch": 0.55, |
| "learning_rate": 1.8773172790757713e-05, |
| "loss": 1.5033, |
| "step": 433 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 1.876652847141334e-05, |
| "loss": 1.4432, |
| "step": 434 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 1.8759867390654326e-05, |
| "loss": 1.4978, |
| "step": 435 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 1.8753189561216514e-05, |
| "loss": 1.6507, |
| "step": 436 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 1.874649499586777e-05, |
| "loss": 1.4505, |
| "step": 437 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 1.8739783707407965e-05, |
| "loss": 1.5597, |
| "step": 438 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 1.8733055708668928e-05, |
| "loss": 1.5849, |
| "step": 439 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 1.872631101251446e-05, |
| "loss": 1.4635, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 1.8719549631840266e-05, |
| "loss": 1.5478, |
| "step": 441 |
| }, |
| { |
| "epoch": 0.57, |
| "learning_rate": 1.8712771579573968e-05, |
| "loss": 1.5115, |
| "step": 442 |
| }, |
| { |
| "epoch": 0.57, |
| "learning_rate": 1.870597686867505e-05, |
| "loss": 1.6138, |
| "step": 443 |
| }, |
| { |
| "epoch": 0.57, |
| "learning_rate": 1.869916551213486e-05, |
| "loss": 1.441, |
| "step": 444 |
| }, |
| { |
| "epoch": 0.57, |
| "learning_rate": 1.869233752297656e-05, |
| "loss": 1.4458, |
| "step": 445 |
| }, |
| { |
| "epoch": 0.57, |
| "learning_rate": 1.868549291425513e-05, |
| "loss": 1.4579, |
| "step": 446 |
| }, |
| { |
| "epoch": 0.57, |
| "learning_rate": 1.86786316990573e-05, |
| "loss": 1.4856, |
| "step": 447 |
| }, |
| { |
| "epoch": 0.57, |
| "learning_rate": 1.867175389050158e-05, |
| "loss": 1.6457, |
| "step": 448 |
| }, |
| { |
| "epoch": 0.57, |
| "learning_rate": 1.8664859501738183e-05, |
| "loss": 1.5344, |
| "step": 449 |
| }, |
| { |
| "epoch": 0.58, |
| "learning_rate": 1.8657948545949036e-05, |
| "loss": 1.4316, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.58, |
| "learning_rate": 1.865102103634774e-05, |
| "loss": 1.6006, |
| "step": 451 |
| }, |
| { |
| "epoch": 0.58, |
| "learning_rate": 1.8644076986179543e-05, |
| "loss": 1.5681, |
| "step": 452 |
| }, |
| { |
| "epoch": 0.58, |
| "learning_rate": 1.8637116408721324e-05, |
| "loss": 1.4692, |
| "step": 453 |
| }, |
| { |
| "epoch": 0.58, |
| "learning_rate": 1.8630139317281554e-05, |
| "loss": 1.4087, |
| "step": 454 |
| }, |
| { |
| "epoch": 0.58, |
| "learning_rate": 1.862314572520028e-05, |
| "loss": 1.637, |
| "step": 455 |
| }, |
| { |
| "epoch": 0.58, |
| "learning_rate": 1.8616135645849106e-05, |
| "loss": 1.5485, |
| "step": 456 |
| }, |
| { |
| "epoch": 0.58, |
| "learning_rate": 1.860910909263115e-05, |
| "loss": 1.3215, |
| "step": 457 |
| }, |
| { |
| "epoch": 0.59, |
| "learning_rate": 1.860206607898103e-05, |
| "loss": 1.6119, |
| "step": 458 |
| }, |
| { |
| "epoch": 0.59, |
| "learning_rate": 1.8595006618364843e-05, |
| "loss": 1.582, |
| "step": 459 |
| }, |
| { |
| "epoch": 0.59, |
| "learning_rate": 1.8587930724280124e-05, |
| "loss": 1.5267, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.59, |
| "learning_rate": 1.8580838410255825e-05, |
| "loss": 1.5518, |
| "step": 461 |
| }, |
| { |
| "epoch": 0.59, |
| "learning_rate": 1.8573729689852307e-05, |
| "loss": 1.4138, |
| "step": 462 |
| }, |
| { |
| "epoch": 0.59, |
| "learning_rate": 1.8566604576661288e-05, |
| "loss": 1.654, |
| "step": 463 |
| }, |
| { |
| "epoch": 0.59, |
| "learning_rate": 1.855946308430583e-05, |
| "loss": 1.6676, |
| "step": 464 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 1.8552305226440315e-05, |
| "loss": 1.5533, |
| "step": 465 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 1.8545131016750413e-05, |
| "loss": 1.6206, |
| "step": 466 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 1.853794046895306e-05, |
| "loss": 1.5267, |
| "step": 467 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 1.853073359679643e-05, |
| "loss": 1.4732, |
| "step": 468 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 1.8523510414059903e-05, |
| "loss": 1.5234, |
| "step": 469 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 1.851627093455406e-05, |
| "loss": 1.5687, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 1.850901517212062e-05, |
| "loss": 1.4919, |
| "step": 471 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 1.8501743140632457e-05, |
| "loss": 1.4328, |
| "step": 472 |
| }, |
| { |
| "epoch": 0.61, |
| "learning_rate": 1.8494454853993527e-05, |
| "loss": 1.5762, |
| "step": 473 |
| }, |
| { |
| "epoch": 0.61, |
| "learning_rate": 1.8487150326138884e-05, |
| "loss": 1.5325, |
| "step": 474 |
| }, |
| { |
| "epoch": 0.61, |
| "learning_rate": 1.847982957103463e-05, |
| "loss": 1.5702, |
| "step": 475 |
| }, |
| { |
| "epoch": 0.61, |
| "learning_rate": 1.847249260267789e-05, |
| "loss": 1.5487, |
| "step": 476 |
| }, |
| { |
| "epoch": 0.61, |
| "learning_rate": 1.8465139435096787e-05, |
| "loss": 1.5621, |
| "step": 477 |
| }, |
| { |
| "epoch": 0.61, |
| "learning_rate": 1.8457770082350426e-05, |
| "loss": 1.6269, |
| "step": 478 |
| }, |
| { |
| "epoch": 0.61, |
| "learning_rate": 1.8450384558528848e-05, |
| "loss": 1.5245, |
| "step": 479 |
| }, |
| { |
| "epoch": 0.61, |
| "learning_rate": 1.8442982877753016e-05, |
| "loss": 1.4432, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 1.8435565054174792e-05, |
| "loss": 1.6035, |
| "step": 481 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 1.8428131101976884e-05, |
| "loss": 1.5814, |
| "step": 482 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 1.8420681035372858e-05, |
| "loss": 1.6586, |
| "step": 483 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 1.841321486860708e-05, |
| "loss": 1.4699, |
| "step": 484 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 1.84057326159547e-05, |
| "loss": 1.5394, |
| "step": 485 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 1.8398234291721622e-05, |
| "loss": 1.5886, |
| "step": 486 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 1.8390719910244487e-05, |
| "loss": 1.5667, |
| "step": 487 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 1.8383189485890632e-05, |
| "loss": 1.5693, |
| "step": 488 |
| }, |
| { |
| "epoch": 0.63, |
| "learning_rate": 1.8375643033058058e-05, |
| "loss": 1.4013, |
| "step": 489 |
| }, |
| { |
| "epoch": 0.63, |
| "learning_rate": 1.836808056617543e-05, |
| "loss": 1.4577, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.63, |
| "learning_rate": 1.8360502099702016e-05, |
| "loss": 1.4651, |
| "step": 491 |
| }, |
| { |
| "epoch": 0.63, |
| "learning_rate": 1.8352907648127688e-05, |
| "loss": 1.5245, |
| "step": 492 |
| }, |
| { |
| "epoch": 0.63, |
| "learning_rate": 1.8345297225972873e-05, |
| "loss": 1.4243, |
| "step": 493 |
| }, |
| { |
| "epoch": 0.63, |
| "learning_rate": 1.8337670847788535e-05, |
| "loss": 1.5857, |
| "step": 494 |
| }, |
| { |
| "epoch": 0.63, |
| "learning_rate": 1.8330028528156138e-05, |
| "loss": 1.5347, |
| "step": 495 |
| }, |
| { |
| "epoch": 0.63, |
| "learning_rate": 1.8322370281687644e-05, |
| "loss": 1.6043, |
| "step": 496 |
| }, |
| { |
| "epoch": 0.64, |
| "learning_rate": 1.8314696123025456e-05, |
| "loss": 1.5089, |
| "step": 497 |
| }, |
| { |
| "epoch": 0.64, |
| "learning_rate": 1.830700606684239e-05, |
| "loss": 1.4643, |
| "step": 498 |
| }, |
| { |
| "epoch": 0.64, |
| "learning_rate": 1.8299300127841683e-05, |
| "loss": 1.6145, |
| "step": 499 |
| }, |
| { |
| "epoch": 0.64, |
| "learning_rate": 1.8291578320756913e-05, |
| "loss": 1.4843, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.64, |
| "learning_rate": 1.8283840660352017e-05, |
| "loss": 1.6487, |
| "step": 501 |
| }, |
| { |
| "epoch": 0.64, |
| "learning_rate": 1.8276087161421234e-05, |
| "loss": 1.5171, |
| "step": 502 |
| }, |
| { |
| "epoch": 0.64, |
| "learning_rate": 1.826831783878909e-05, |
| "loss": 1.5883, |
| "step": 503 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 1.826053270731036e-05, |
| "loss": 1.5624, |
| "step": 504 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 1.8252731781870046e-05, |
| "loss": 1.5277, |
| "step": 505 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 1.8244915077383358e-05, |
| "loss": 1.4787, |
| "step": 506 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 1.8237082608795664e-05, |
| "loss": 1.5277, |
| "step": 507 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 1.8229234391082472e-05, |
| "loss": 1.5259, |
| "step": 508 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 1.8221370439249412e-05, |
| "loss": 1.628, |
| "step": 509 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 1.8213490768332193e-05, |
| "loss": 1.5003, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 1.820559539339657e-05, |
| "loss": 1.5083, |
| "step": 511 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 1.819768432953833e-05, |
| "loss": 1.4897, |
| "step": 512 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 1.818975759188327e-05, |
| "loss": 1.5501, |
| "step": 513 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 1.818181519558713e-05, |
| "loss": 1.5783, |
| "step": 514 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 1.817385715583561e-05, |
| "loss": 1.4158, |
| "step": 515 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 1.81658834878443e-05, |
| "loss": 1.5979, |
| "step": 516 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 1.8157894206858698e-05, |
| "loss": 1.5257, |
| "step": 517 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 1.8149889328154123e-05, |
| "loss": 1.4823, |
| "step": 518 |
| }, |
| { |
| "epoch": 0.66, |
| "learning_rate": 1.8141868867035745e-05, |
| "loss": 1.6418, |
| "step": 519 |
| }, |
| { |
| "epoch": 0.67, |
| "learning_rate": 1.8133832838838503e-05, |
| "loss": 1.5138, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.67, |
| "learning_rate": 1.8125781258927116e-05, |
| "loss": 1.5885, |
| "step": 521 |
| }, |
| { |
| "epoch": 0.67, |
| "learning_rate": 1.811771414269603e-05, |
| "loss": 1.412, |
| "step": 522 |
| }, |
| { |
| "epoch": 0.67, |
| "learning_rate": 1.8109631505569404e-05, |
| "loss": 1.5144, |
| "step": 523 |
| }, |
| { |
| "epoch": 0.67, |
| "learning_rate": 1.810153336300106e-05, |
| "loss": 1.6172, |
| "step": 524 |
| }, |
| { |
| "epoch": 0.67, |
| "learning_rate": 1.809341973047448e-05, |
| "loss": 1.6128, |
| "step": 525 |
| }, |
| { |
| "epoch": 0.67, |
| "learning_rate": 1.808529062350275e-05, |
| "loss": 1.4841, |
| "step": 526 |
| }, |
| { |
| "epoch": 0.67, |
| "learning_rate": 1.8077146057628547e-05, |
| "loss": 1.543, |
| "step": 527 |
| }, |
| { |
| "epoch": 0.68, |
| "learning_rate": 1.806898604842411e-05, |
| "loss": 1.5263, |
| "step": 528 |
| }, |
| { |
| "epoch": 0.68, |
| "learning_rate": 1.8060810611491203e-05, |
| "loss": 1.4654, |
| "step": 529 |
| }, |
| { |
| "epoch": 0.68, |
| "learning_rate": 1.8052619762461084e-05, |
| "loss": 1.59, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.68, |
| "learning_rate": 1.8044413516994483e-05, |
| "loss": 1.4533, |
| "step": 531 |
| }, |
| { |
| "epoch": 0.68, |
| "learning_rate": 1.8036191890781563e-05, |
| "loss": 1.5751, |
| "step": 532 |
| }, |
| { |
| "epoch": 0.68, |
| "learning_rate": 1.8027954899541894e-05, |
| "loss": 1.5516, |
| "step": 533 |
| }, |
| { |
| "epoch": 0.68, |
| "learning_rate": 1.8019702559024437e-05, |
| "loss": 1.546, |
| "step": 534 |
| }, |
| { |
| "epoch": 0.68, |
| "learning_rate": 1.8011434885007482e-05, |
| "loss": 1.5633, |
| "step": 535 |
| }, |
| { |
| "epoch": 0.69, |
| "learning_rate": 1.8003151893298646e-05, |
| "loss": 1.5256, |
| "step": 536 |
| }, |
| { |
| "epoch": 0.69, |
| "learning_rate": 1.7994853599734835e-05, |
| "loss": 1.5775, |
| "step": 537 |
| }, |
| { |
| "epoch": 0.69, |
| "learning_rate": 1.7986540020182207e-05, |
| "loss": 1.4661, |
| "step": 538 |
| }, |
| { |
| "epoch": 0.69, |
| "learning_rate": 1.7978211170536145e-05, |
| "loss": 1.4995, |
| "step": 539 |
| }, |
| { |
| "epoch": 0.69, |
| "learning_rate": 1.7969867066721234e-05, |
| "loss": 1.6578, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.69, |
| "learning_rate": 1.796150772469122e-05, |
| "loss": 1.4593, |
| "step": 541 |
| }, |
| { |
| "epoch": 0.69, |
| "learning_rate": 1.795313316042899e-05, |
| "loss": 1.5147, |
| "step": 542 |
| }, |
| { |
| "epoch": 0.7, |
| "learning_rate": 1.7944743389946524e-05, |
| "loss": 1.6462, |
| "step": 543 |
| }, |
| { |
| "epoch": 0.7, |
| "learning_rate": 1.793633842928489e-05, |
| "loss": 1.4628, |
| "step": 544 |
| }, |
| { |
| "epoch": 0.7, |
| "learning_rate": 1.792791829451419e-05, |
| "loss": 1.3819, |
| "step": 545 |
| }, |
| { |
| "epoch": 0.7, |
| "learning_rate": 1.791948300173354e-05, |
| "loss": 1.5975, |
| "step": 546 |
| }, |
| { |
| "epoch": 0.7, |
| "learning_rate": 1.7911032567071044e-05, |
| "loss": 1.5558, |
| "step": 547 |
| }, |
| { |
| "epoch": 0.7, |
| "learning_rate": 1.7902567006683747e-05, |
| "loss": 1.6226, |
| "step": 548 |
| }, |
| { |
| "epoch": 0.7, |
| "learning_rate": 1.7894086336757627e-05, |
| "loss": 1.5697, |
| "step": 549 |
| }, |
| { |
| "epoch": 0.7, |
| "learning_rate": 1.788559057350753e-05, |
| "loss": 1.4851, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 1.7877079733177185e-05, |
| "loss": 1.4857, |
| "step": 551 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 1.7868553832039128e-05, |
| "loss": 1.4548, |
| "step": 552 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 1.786001288639471e-05, |
| "loss": 1.5372, |
| "step": 553 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 1.7851456912574022e-05, |
| "loss": 1.667, |
| "step": 554 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 1.7842885926935917e-05, |
| "loss": 1.5669, |
| "step": 555 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 1.7834299945867926e-05, |
| "loss": 1.5722, |
| "step": 556 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 1.7825698985786264e-05, |
| "loss": 1.4489, |
| "step": 557 |
| }, |
| { |
| "epoch": 0.71, |
| "learning_rate": 1.7817083063135782e-05, |
| "loss": 1.4824, |
| "step": 558 |
| }, |
| { |
| "epoch": 0.72, |
| "learning_rate": 1.780845219438994e-05, |
| "loss": 1.5446, |
| "step": 559 |
| }, |
| { |
| "epoch": 0.72, |
| "learning_rate": 1.779980639605077e-05, |
| "loss": 1.5276, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.72, |
| "learning_rate": 1.7791145684648864e-05, |
| "loss": 1.4949, |
| "step": 561 |
| }, |
| { |
| "epoch": 0.72, |
| "learning_rate": 1.778247007674331e-05, |
| "loss": 1.4785, |
| "step": 562 |
| }, |
| { |
| "epoch": 0.72, |
| "learning_rate": 1.777377958892168e-05, |
| "loss": 1.6022, |
| "step": 563 |
| }, |
| { |
| "epoch": 0.72, |
| "learning_rate": 1.776507423780001e-05, |
| "loss": 1.5496, |
| "step": 564 |
| }, |
| { |
| "epoch": 0.72, |
| "learning_rate": 1.7756354040022736e-05, |
| "loss": 1.5412, |
| "step": 565 |
| }, |
| { |
| "epoch": 0.72, |
| "learning_rate": 1.7747619012262697e-05, |
| "loss": 1.5029, |
| "step": 566 |
| }, |
| { |
| "epoch": 0.73, |
| "learning_rate": 1.773886917122107e-05, |
| "loss": 1.5501, |
| "step": 567 |
| }, |
| { |
| "epoch": 0.73, |
| "learning_rate": 1.773010453362737e-05, |
| "loss": 1.4715, |
| "step": 568 |
| }, |
| { |
| "epoch": 0.73, |
| "learning_rate": 1.77213251162394e-05, |
| "loss": 1.4653, |
| "step": 569 |
| }, |
| { |
| "epoch": 0.73, |
| "learning_rate": 1.7712530935843206e-05, |
| "loss": 1.4231, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.73, |
| "learning_rate": 1.7703722009253084e-05, |
| "loss": 1.6637, |
| "step": 571 |
| }, |
| { |
| "epoch": 0.73, |
| "learning_rate": 1.769489835331151e-05, |
| "loss": 1.5361, |
| "step": 572 |
| }, |
| { |
| "epoch": 0.73, |
| "learning_rate": 1.7686059984889123e-05, |
| "loss": 1.4445, |
| "step": 573 |
| }, |
| { |
| "epoch": 0.73, |
| "learning_rate": 1.767720692088469e-05, |
| "loss": 1.6327, |
| "step": 574 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 1.766833917822509e-05, |
| "loss": 1.5256, |
| "step": 575 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 1.7659456773865247e-05, |
| "loss": 1.4394, |
| "step": 576 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 1.7650559724788137e-05, |
| "loss": 1.5231, |
| "step": 577 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 1.764164804800472e-05, |
| "loss": 1.5967, |
| "step": 578 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 1.7632721760553935e-05, |
| "loss": 1.5865, |
| "step": 579 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 1.762378087950265e-05, |
| "loss": 1.4517, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 1.7614825421945642e-05, |
| "loss": 1.5339, |
| "step": 581 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 1.7605855405005548e-05, |
| "loss": 1.5612, |
| "step": 582 |
| }, |
| { |
| "epoch": 0.75, |
| "learning_rate": 1.759687084583285e-05, |
| "loss": 1.6679, |
| "step": 583 |
| }, |
| { |
| "epoch": 0.75, |
| "learning_rate": 1.7587871761605838e-05, |
| "loss": 1.5264, |
| "step": 584 |
| }, |
| { |
| "epoch": 0.75, |
| "learning_rate": 1.7578858169530565e-05, |
| "loss": 1.5529, |
| "step": 585 |
| }, |
| { |
| "epoch": 0.75, |
| "learning_rate": 1.756983008684082e-05, |
| "loss": 1.5614, |
| "step": 586 |
| }, |
| { |
| "epoch": 0.75, |
| "learning_rate": 1.756078753079811e-05, |
| "loss": 1.5096, |
| "step": 587 |
| }, |
| { |
| "epoch": 0.75, |
| "learning_rate": 1.7551730518691612e-05, |
| "loss": 1.525, |
| "step": 588 |
| }, |
| { |
| "epoch": 0.75, |
| "learning_rate": 1.7542659067838132e-05, |
| "loss": 1.4647, |
| "step": 589 |
| }, |
| { |
| "epoch": 0.76, |
| "learning_rate": 1.7533573195582093e-05, |
| "loss": 1.5689, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.76, |
| "learning_rate": 1.7524472919295488e-05, |
| "loss": 1.5574, |
| "step": 591 |
| }, |
| { |
| "epoch": 0.76, |
| "learning_rate": 1.751535825637785e-05, |
| "loss": 1.4986, |
| "step": 592 |
| }, |
| { |
| "epoch": 0.76, |
| "learning_rate": 1.7506229224256216e-05, |
| "loss": 1.4711, |
| "step": 593 |
| }, |
| { |
| "epoch": 0.76, |
| "learning_rate": 1.7497085840385105e-05, |
| "loss": 1.4352, |
| "step": 594 |
| }, |
| { |
| "epoch": 0.76, |
| "learning_rate": 1.7487928122246467e-05, |
| "loss": 1.5839, |
| "step": 595 |
| }, |
| { |
| "epoch": 0.76, |
| "learning_rate": 1.7478756087349665e-05, |
| "loss": 1.3915, |
| "step": 596 |
| }, |
| { |
| "epoch": 0.76, |
| "learning_rate": 1.7469569753231436e-05, |
| "loss": 1.4873, |
| "step": 597 |
| }, |
| { |
| "epoch": 0.77, |
| "learning_rate": 1.7460369137455848e-05, |
| "loss": 1.5048, |
| "step": 598 |
| }, |
| { |
| "epoch": 0.77, |
| "learning_rate": 1.7451154257614287e-05, |
| "loss": 1.6288, |
| "step": 599 |
| }, |
| { |
| "epoch": 0.77, |
| "learning_rate": 1.7441925131325403e-05, |
| "loss": 1.3306, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.77, |
| "learning_rate": 1.7432681776235085e-05, |
| "loss": 1.557, |
| "step": 601 |
| }, |
| { |
| "epoch": 0.77, |
| "learning_rate": 1.7423424210016438e-05, |
| "loss": 1.4688, |
| "step": 602 |
| }, |
| { |
| "epoch": 0.77, |
| "learning_rate": 1.741415245036972e-05, |
| "loss": 1.4829, |
| "step": 603 |
| }, |
| { |
| "epoch": 0.77, |
| "learning_rate": 1.7404866515022344e-05, |
| "loss": 1.6249, |
| "step": 604 |
| }, |
| { |
| "epoch": 0.77, |
| "learning_rate": 1.7395566421728818e-05, |
| "loss": 1.5846, |
| "step": 605 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 1.738625218827072e-05, |
| "loss": 1.5239, |
| "step": 606 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 1.7376923832456665e-05, |
| "loss": 1.4548, |
| "step": 607 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 1.736758137212227e-05, |
| "loss": 1.5579, |
| "step": 608 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 1.735822482513012e-05, |
| "loss": 1.5518, |
| "step": 609 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 1.7348854209369733e-05, |
| "loss": 1.4994, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 1.7339469542757524e-05, |
| "loss": 1.5803, |
| "step": 611 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 1.7330070843236772e-05, |
| "loss": 1.4356, |
| "step": 612 |
| }, |
| { |
| "epoch": 0.78, |
| "learning_rate": 1.7320658128777594e-05, |
| "loss": 1.5938, |
| "step": 613 |
| }, |
| { |
| "epoch": 0.79, |
| "learning_rate": 1.7311231417376894e-05, |
| "loss": 1.4163, |
| "step": 614 |
| }, |
| { |
| "epoch": 0.79, |
| "learning_rate": 1.7301790727058344e-05, |
| "loss": 1.5661, |
| "step": 615 |
| }, |
| { |
| "epoch": 0.79, |
| "learning_rate": 1.729233607587234e-05, |
| "loss": 1.5186, |
| "step": 616 |
| }, |
| { |
| "epoch": 0.79, |
| "learning_rate": 1.7282867481895983e-05, |
| "loss": 1.4679, |
| "step": 617 |
| }, |
| { |
| "epoch": 0.79, |
| "learning_rate": 1.7273384963233005e-05, |
| "loss": 1.5399, |
| "step": 618 |
| }, |
| { |
| "epoch": 0.79, |
| "learning_rate": 1.726388853801379e-05, |
| "loss": 1.5167, |
| "step": 619 |
| }, |
| { |
| "epoch": 0.79, |
| "learning_rate": 1.72543782243953e-05, |
| "loss": 1.3608, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.79, |
| "learning_rate": 1.7244854040561042e-05, |
| "loss": 1.6059, |
| "step": 621 |
| }, |
| { |
| "epoch": 0.8, |
| "learning_rate": 1.7235316004721058e-05, |
| "loss": 1.4435, |
| "step": 622 |
| }, |
| { |
| "epoch": 0.8, |
| "learning_rate": 1.7225764135111867e-05, |
| "loss": 1.5231, |
| "step": 623 |
| }, |
| { |
| "epoch": 0.8, |
| "learning_rate": 1.7216198449996442e-05, |
| "loss": 1.4482, |
| "step": 624 |
| }, |
| { |
| "epoch": 0.8, |
| "learning_rate": 1.7206618967664162e-05, |
| "loss": 1.5029, |
| "step": 625 |
| }, |
| { |
| "epoch": 0.8, |
| "learning_rate": 1.7197025706430797e-05, |
| "loss": 1.4924, |
| "step": 626 |
| }, |
| { |
| "epoch": 0.8, |
| "learning_rate": 1.7187418684638456e-05, |
| "loss": 1.3754, |
| "step": 627 |
| }, |
| { |
| "epoch": 0.8, |
| "learning_rate": 1.7177797920655555e-05, |
| "loss": 1.6074, |
| "step": 628 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 1.7168163432876796e-05, |
| "loss": 1.4987, |
| "step": 629 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 1.7158515239723107e-05, |
| "loss": 1.523, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 1.7148853359641627e-05, |
| "loss": 1.6024, |
| "step": 631 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 1.7139177811105666e-05, |
| "loss": 1.5471, |
| "step": 632 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 1.7129488612614668e-05, |
| "loss": 1.4727, |
| "step": 633 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 1.711978578269416e-05, |
| "loss": 1.5078, |
| "step": 634 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 1.7110069339895764e-05, |
| "loss": 1.5186, |
| "step": 635 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 1.71003393027971e-05, |
| "loss": 1.6478, |
| "step": 636 |
| }, |
| { |
| "epoch": 0.82, |
| "learning_rate": 1.709059569000179e-05, |
| "loss": 1.5051, |
| "step": 637 |
| }, |
| { |
| "epoch": 0.82, |
| "learning_rate": 1.7080838520139417e-05, |
| "loss": 1.3942, |
| "step": 638 |
| }, |
| { |
| "epoch": 0.82, |
| "learning_rate": 1.7071067811865477e-05, |
| "loss": 1.5837, |
| "step": 639 |
| }, |
| { |
| "epoch": 0.82, |
| "learning_rate": 1.7061283583861358e-05, |
| "loss": 1.5204, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.82, |
| "learning_rate": 1.705148585483429e-05, |
| "loss": 1.5019, |
| "step": 641 |
| }, |
| { |
| "epoch": 0.82, |
| "learning_rate": 1.7041674643517322e-05, |
| "loss": 1.519, |
| "step": 642 |
| }, |
| { |
| "epoch": 0.82, |
| "learning_rate": 1.7031849968669287e-05, |
| "loss": 1.565, |
| "step": 643 |
| }, |
| { |
| "epoch": 0.82, |
| "learning_rate": 1.7022011849074745e-05, |
| "loss": 1.436, |
| "step": 644 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 1.701216030354397e-05, |
| "loss": 1.4744, |
| "step": 645 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 1.700229535091291e-05, |
| "loss": 1.4501, |
| "step": 646 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 1.6992417010043144e-05, |
| "loss": 1.4366, |
| "step": 647 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 1.698252529982184e-05, |
| "loss": 1.4416, |
| "step": 648 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 1.6972620239161747e-05, |
| "loss": 1.4952, |
| "step": 649 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 1.696270184700112e-05, |
| "loss": 1.5683, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 1.695277014230372e-05, |
| "loss": 1.6548, |
| "step": 651 |
| }, |
| { |
| "epoch": 0.83, |
| "learning_rate": 1.6942825144058744e-05, |
| "loss": 1.5788, |
| "step": 652 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 1.6932866871280823e-05, |
| "loss": 1.4559, |
| "step": 653 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 1.6922895343009963e-05, |
| "loss": 1.5352, |
| "step": 654 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 1.6912910578311503e-05, |
| "loss": 1.5393, |
| "step": 655 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 1.6902912596276107e-05, |
| "loss": 1.4919, |
| "step": 656 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 1.6892901416019702e-05, |
| "loss": 1.5079, |
| "step": 657 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 1.6882877056683446e-05, |
| "loss": 1.5547, |
| "step": 658 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 1.6872839537433706e-05, |
| "loss": 1.5188, |
| "step": 659 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 1.6862788877461994e-05, |
| "loss": 1.4864, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.85, |
| "learning_rate": 1.6852725095984964e-05, |
| "loss": 1.5786, |
| "step": 661 |
| }, |
| { |
| "epoch": 0.85, |
| "learning_rate": 1.684264821224435e-05, |
| "loss": 1.5396, |
| "step": 662 |
| }, |
| { |
| "epoch": 0.85, |
| "learning_rate": 1.6832558245506937e-05, |
| "loss": 1.479, |
| "step": 663 |
| }, |
| { |
| "epoch": 0.85, |
| "learning_rate": 1.6822455215064522e-05, |
| "loss": 1.5636, |
| "step": 664 |
| }, |
| { |
| "epoch": 0.85, |
| "learning_rate": 1.681233914023389e-05, |
| "loss": 1.4777, |
| "step": 665 |
| }, |
| { |
| "epoch": 0.85, |
| "learning_rate": 1.6802210040356753e-05, |
| "loss": 1.6042, |
| "step": 666 |
| }, |
| { |
| "epoch": 0.85, |
| "learning_rate": 1.6792067934799734e-05, |
| "loss": 1.444, |
| "step": 667 |
| }, |
| { |
| "epoch": 0.86, |
| "learning_rate": 1.6781912842954323e-05, |
| "loss": 1.6779, |
| "step": 668 |
| }, |
| { |
| "epoch": 0.86, |
| "learning_rate": 1.677174478423684e-05, |
| "loss": 1.5744, |
| "step": 669 |
| }, |
| { |
| "epoch": 0.86, |
| "learning_rate": 1.676156377808839e-05, |
| "loss": 1.5028, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.86, |
| "learning_rate": 1.6751369843974842e-05, |
| "loss": 1.5694, |
| "step": 671 |
| }, |
| { |
| "epoch": 0.86, |
| "learning_rate": 1.6741163001386783e-05, |
| "loss": 1.5982, |
| "step": 672 |
| }, |
| { |
| "epoch": 0.86, |
| "learning_rate": 1.673094326983947e-05, |
| "loss": 1.4068, |
| "step": 673 |
| }, |
| { |
| "epoch": 0.86, |
| "learning_rate": 1.6720710668872816e-05, |
| "loss": 1.5212, |
| "step": 674 |
| }, |
| { |
| "epoch": 0.86, |
| "learning_rate": 1.6710465218051332e-05, |
| "loss": 1.572, |
| "step": 675 |
| }, |
| { |
| "epoch": 0.87, |
| "learning_rate": 1.6700206936964102e-05, |
| "loss": 1.6098, |
| "step": 676 |
| }, |
| { |
| "epoch": 0.87, |
| "learning_rate": 1.6689935845224742e-05, |
| "loss": 1.5454, |
| "step": 677 |
| }, |
| { |
| "epoch": 0.87, |
| "learning_rate": 1.6679651962471356e-05, |
| "loss": 1.5341, |
| "step": 678 |
| }, |
| { |
| "epoch": 0.87, |
| "learning_rate": 1.666935530836651e-05, |
| "loss": 1.5899, |
| "step": 679 |
| }, |
| { |
| "epoch": 0.87, |
| "learning_rate": 1.665904590259718e-05, |
| "loss": 1.5433, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.87, |
| "learning_rate": 1.6648723764874737e-05, |
| "loss": 1.5102, |
| "step": 681 |
| }, |
| { |
| "epoch": 0.87, |
| "learning_rate": 1.6638388914934886e-05, |
| "loss": 1.4966, |
| "step": 682 |
| }, |
| { |
| "epoch": 0.87, |
| "learning_rate": 1.6628041372537633e-05, |
| "loss": 1.5101, |
| "step": 683 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 1.661768115746726e-05, |
| "loss": 1.494, |
| "step": 684 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 1.660730828953228e-05, |
| "loss": 1.5223, |
| "step": 685 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 1.659692278856539e-05, |
| "loss": 1.4483, |
| "step": 686 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 1.658652467442345e-05, |
| "loss": 1.5636, |
| "step": 687 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 1.6576113966987423e-05, |
| "loss": 1.4911, |
| "step": 688 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 1.6565690686162365e-05, |
| "loss": 1.5708, |
| "step": 689 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 1.6555254851877367e-05, |
| "loss": 1.5029, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 1.6544806484085517e-05, |
| "loss": 1.5355, |
| "step": 691 |
| }, |
| { |
| "epoch": 0.89, |
| "learning_rate": 1.653434560276387e-05, |
| "loss": 1.6601, |
| "step": 692 |
| }, |
| { |
| "epoch": 0.89, |
| "learning_rate": 1.652387222791341e-05, |
| "loss": 1.5048, |
| "step": 693 |
| }, |
| { |
| "epoch": 0.89, |
| "learning_rate": 1.6513386379559006e-05, |
| "loss": 1.4535, |
| "step": 694 |
| }, |
| { |
| "epoch": 0.89, |
| "learning_rate": 1.650288807774937e-05, |
| "loss": 1.4837, |
| "step": 695 |
| }, |
| { |
| "epoch": 0.89, |
| "learning_rate": 1.6492377342557036e-05, |
| "loss": 1.6034, |
| "step": 696 |
| }, |
| { |
| "epoch": 0.89, |
| "learning_rate": 1.64818541940783e-05, |
| "loss": 1.546, |
| "step": 697 |
| }, |
| { |
| "epoch": 0.89, |
| "learning_rate": 1.6471318652433198e-05, |
| "loss": 1.5467, |
| "step": 698 |
| }, |
| { |
| "epoch": 0.89, |
| "learning_rate": 1.6460770737765465e-05, |
| "loss": 1.4843, |
| "step": 699 |
| }, |
| { |
| "epoch": 0.9, |
| "learning_rate": 1.645021047024248e-05, |
| "loss": 1.4456, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.9, |
| "learning_rate": 1.643963787005525e-05, |
| "loss": 1.4996, |
| "step": 701 |
| }, |
| { |
| "epoch": 0.9, |
| "learning_rate": 1.6429052957418363e-05, |
| "loss": 1.5743, |
| "step": 702 |
| }, |
| { |
| "epoch": 0.9, |
| "learning_rate": 1.6418455752569945e-05, |
| "loss": 1.5112, |
| "step": 703 |
| }, |
| { |
| "epoch": 0.9, |
| "learning_rate": 1.6407846275771625e-05, |
| "loss": 1.4653, |
| "step": 704 |
| }, |
| { |
| "epoch": 0.9, |
| "learning_rate": 1.639722454730849e-05, |
| "loss": 1.4576, |
| "step": 705 |
| }, |
| { |
| "epoch": 0.9, |
| "learning_rate": 1.638659058748906e-05, |
| "loss": 1.472, |
| "step": 706 |
| }, |
| { |
| "epoch": 0.9, |
| "learning_rate": 1.637594441664524e-05, |
| "loss": 1.5423, |
| "step": 707 |
| }, |
| { |
| "epoch": 0.91, |
| "learning_rate": 1.636528605513229e-05, |
| "loss": 1.5335, |
| "step": 708 |
| }, |
| { |
| "epoch": 0.91, |
| "learning_rate": 1.635461552332875e-05, |
| "loss": 1.5117, |
| "step": 709 |
| }, |
| { |
| "epoch": 0.91, |
| "learning_rate": 1.6343932841636455e-05, |
| "loss": 1.4179, |
| "step": 710 |
| }, |
| { |
| "epoch": 0.91, |
| "learning_rate": 1.6333238030480473e-05, |
| "loss": 1.448, |
| "step": 711 |
| }, |
| { |
| "epoch": 0.91, |
| "learning_rate": 1.632253111030904e-05, |
| "loss": 1.5192, |
| "step": 712 |
| }, |
| { |
| "epoch": 0.91, |
| "learning_rate": 1.631181210159357e-05, |
| "loss": 1.4766, |
| "step": 713 |
| }, |
| { |
| "epoch": 0.91, |
| "learning_rate": 1.630108102482857e-05, |
| "loss": 1.5815, |
| "step": 714 |
| }, |
| { |
| "epoch": 0.92, |
| "learning_rate": 1.6290337900531633e-05, |
| "loss": 1.4862, |
| "step": 715 |
| }, |
| { |
| "epoch": 0.92, |
| "learning_rate": 1.627958274924338e-05, |
| "loss": 1.5528, |
| "step": 716 |
| }, |
| { |
| "epoch": 0.92, |
| "learning_rate": 1.626881559152743e-05, |
| "loss": 1.6059, |
| "step": 717 |
| }, |
| { |
| "epoch": 0.92, |
| "learning_rate": 1.6258036447970363e-05, |
| "loss": 1.5849, |
| "step": 718 |
| }, |
| { |
| "epoch": 0.92, |
| "learning_rate": 1.624724533918166e-05, |
| "loss": 1.5834, |
| "step": 719 |
| }, |
| { |
| "epoch": 0.92, |
| "learning_rate": 1.62364422857937e-05, |
| "loss": 1.5903, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.92, |
| "learning_rate": 1.6225627308461688e-05, |
| "loss": 1.5857, |
| "step": 721 |
| }, |
| { |
| "epoch": 0.92, |
| "learning_rate": 1.6214800427863626e-05, |
| "loss": 1.4953, |
| "step": 722 |
| }, |
| { |
| "epoch": 0.93, |
| "learning_rate": 1.6203961664700275e-05, |
| "loss": 1.5037, |
| "step": 723 |
| }, |
| { |
| "epoch": 0.93, |
| "learning_rate": 1.6193111039695124e-05, |
| "loss": 1.4149, |
| "step": 724 |
| }, |
| { |
| "epoch": 0.93, |
| "learning_rate": 1.6182248573594334e-05, |
| "loss": 1.5717, |
| "step": 725 |
| }, |
| { |
| "epoch": 0.93, |
| "learning_rate": 1.6171374287166708e-05, |
| "loss": 1.5662, |
| "step": 726 |
| }, |
| { |
| "epoch": 0.93, |
| "learning_rate": 1.6160488201203643e-05, |
| "loss": 1.507, |
| "step": 727 |
| }, |
| { |
| "epoch": 0.93, |
| "learning_rate": 1.614959033651911e-05, |
| "loss": 1.5662, |
| "step": 728 |
| }, |
| { |
| "epoch": 0.93, |
| "learning_rate": 1.613868071394959e-05, |
| "loss": 1.5028, |
| "step": 729 |
| }, |
| { |
| "epoch": 0.93, |
| "learning_rate": 1.6127759354354044e-05, |
| "loss": 1.504, |
| "step": 730 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 1.6116826278613882e-05, |
| "loss": 1.4723, |
| "step": 731 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 1.610588150763291e-05, |
| "loss": 1.4376, |
| "step": 732 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 1.6094925062337294e-05, |
| "loss": 1.4708, |
| "step": 733 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 1.6083956963675524e-05, |
| "loss": 1.55, |
| "step": 734 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 1.607297723261837e-05, |
| "loss": 1.4513, |
| "step": 735 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 1.606198589015884e-05, |
| "loss": 1.5062, |
| "step": 736 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 1.6050982957312148e-05, |
| "loss": 1.4572, |
| "step": 737 |
| }, |
| { |
| "epoch": 0.94, |
| "learning_rate": 1.6039968455115666e-05, |
| "loss": 1.542, |
| "step": 738 |
| }, |
| { |
| "epoch": 0.95, |
| "learning_rate": 1.6028942404628886e-05, |
| "loss": 1.5446, |
| "step": 739 |
| }, |
| { |
| "epoch": 0.95, |
| "learning_rate": 1.6017904826933376e-05, |
| "loss": 1.5324, |
| "step": 740 |
| }, |
| { |
| "epoch": 0.95, |
| "learning_rate": 1.6006855743132756e-05, |
| "loss": 1.5279, |
| "step": 741 |
| }, |
| { |
| "epoch": 0.95, |
| "learning_rate": 1.599579517435263e-05, |
| "loss": 1.419, |
| "step": 742 |
| }, |
| { |
| "epoch": 0.95, |
| "learning_rate": 1.5984723141740578e-05, |
| "loss": 1.5599, |
| "step": 743 |
| }, |
| { |
| "epoch": 0.95, |
| "learning_rate": 1.5973639666466078e-05, |
| "loss": 1.5176, |
| "step": 744 |
| }, |
| { |
| "epoch": 0.95, |
| "learning_rate": 1.59625447697205e-05, |
| "loss": 1.6002, |
| "step": 745 |
| }, |
| { |
| "epoch": 0.95, |
| "learning_rate": 1.5951438472717055e-05, |
| "loss": 1.4306, |
| "step": 746 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 1.5940320796690742e-05, |
| "loss": 1.4968, |
| "step": 747 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 1.5929191762898315e-05, |
| "loss": 1.5079, |
| "step": 748 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 1.5918051392618255e-05, |
| "loss": 1.5674, |
| "step": 749 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 1.590689970715071e-05, |
| "loss": 1.5115, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 1.5895736727817457e-05, |
| "loss": 1.5265, |
| "step": 751 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 1.588456247596188e-05, |
| "loss": 1.5439, |
| "step": 752 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 1.5873376972948905e-05, |
| "loss": 1.4767, |
| "step": 753 |
| }, |
| { |
| "epoch": 0.97, |
| "learning_rate": 1.586218024016498e-05, |
| "loss": 1.5562, |
| "step": 754 |
| }, |
| { |
| "epoch": 0.97, |
| "learning_rate": 1.5850972299018014e-05, |
| "loss": 1.6845, |
| "step": 755 |
| }, |
| { |
| "epoch": 0.97, |
| "learning_rate": 1.583975317093735e-05, |
| "loss": 1.6212, |
| "step": 756 |
| }, |
| { |
| "epoch": 0.97, |
| "learning_rate": 1.5828522877373728e-05, |
| "loss": 1.5755, |
| "step": 757 |
| }, |
| { |
| "epoch": 0.97, |
| "learning_rate": 1.5817281439799215e-05, |
| "loss": 1.548, |
| "step": 758 |
| }, |
| { |
| "epoch": 0.97, |
| "learning_rate": 1.580602887970721e-05, |
| "loss": 1.451, |
| "step": 759 |
| }, |
| { |
| "epoch": 0.97, |
| "learning_rate": 1.579476521861236e-05, |
| "loss": 1.5423, |
| "step": 760 |
| }, |
| { |
| "epoch": 0.97, |
| "learning_rate": 1.578349047805055e-05, |
| "loss": 1.5677, |
| "step": 761 |
| }, |
| { |
| "epoch": 0.98, |
| "learning_rate": 1.5772204679578835e-05, |
| "loss": 1.5563, |
| "step": 762 |
| }, |
| { |
| "epoch": 0.98, |
| "learning_rate": 1.5760907844775428e-05, |
| "loss": 1.5435, |
| "step": 763 |
| }, |
| { |
| "epoch": 0.98, |
| "learning_rate": 1.574959999523963e-05, |
| "loss": 1.4779, |
| "step": 764 |
| }, |
| { |
| "epoch": 0.98, |
| "learning_rate": 1.573828115259181e-05, |
| "loss": 1.5168, |
| "step": 765 |
| }, |
| { |
| "epoch": 0.98, |
| "learning_rate": 1.572695133847335e-05, |
| "loss": 1.5292, |
| "step": 766 |
| }, |
| { |
| "epoch": 0.98, |
| "learning_rate": 1.5715610574546612e-05, |
| "loss": 1.4631, |
| "step": 767 |
| }, |
| { |
| "epoch": 0.98, |
| "learning_rate": 1.5704258882494893e-05, |
| "loss": 1.4746, |
| "step": 768 |
| }, |
| { |
| "epoch": 0.98, |
| "learning_rate": 1.569289628402239e-05, |
| "loss": 1.4808, |
| "step": 769 |
| }, |
| { |
| "epoch": 0.99, |
| "learning_rate": 1.5681522800854147e-05, |
| "loss": 1.5122, |
| "step": 770 |
| }, |
| { |
| "epoch": 0.99, |
| "learning_rate": 1.5670138454736012e-05, |
| "loss": 1.4242, |
| "step": 771 |
| }, |
| { |
| "epoch": 0.99, |
| "learning_rate": 1.565874326743462e-05, |
| "loss": 1.5265, |
| "step": 772 |
| }, |
| { |
| "epoch": 0.99, |
| "learning_rate": 1.564733726073732e-05, |
| "loss": 1.4896, |
| "step": 773 |
| }, |
| { |
| "epoch": 0.99, |
| "learning_rate": 1.5635920456452152e-05, |
| "loss": 1.6303, |
| "step": 774 |
| }, |
| { |
| "epoch": 0.99, |
| "learning_rate": 1.562449287640781e-05, |
| "loss": 1.5172, |
| "step": 775 |
| }, |
| { |
| "epoch": 0.99, |
| "learning_rate": 1.561305454245357e-05, |
| "loss": 1.507, |
| "step": 776 |
| }, |
| { |
| "epoch": 0.99, |
| "learning_rate": 1.560160547645929e-05, |
| "loss": 1.6261, |
| "step": 777 |
| }, |
| { |
| "epoch": 1.0, |
| "learning_rate": 1.5590145700315333e-05, |
| "loss": 1.5126, |
| "step": 778 |
| }, |
| { |
| "epoch": 1.0, |
| "learning_rate": 1.557867523593255e-05, |
| "loss": 1.5811, |
| "step": 779 |
| }, |
| { |
| "epoch": 1.0, |
| "learning_rate": 1.556719410524222e-05, |
| "loss": 1.4222, |
| "step": 780 |
| }, |
| { |
| "epoch": 1.0, |
| "learning_rate": 1.5555702330196024e-05, |
| "loss": 1.5278, |
| "step": 781 |
| }, |
| { |
| "epoch": 1.0, |
| "learning_rate": 1.5544199932765983e-05, |
| "loss": 1.3436, |
| "step": 782 |
| }, |
| { |
| "epoch": 1.0, |
| "learning_rate": 1.553268693494444e-05, |
| "loss": 1.2307, |
| "step": 783 |
| }, |
| { |
| "epoch": 1.0, |
| "learning_rate": 1.5521163358743998e-05, |
| "loss": 1.17, |
| "step": 784 |
| }, |
| { |
| "epoch": 1.0, |
| "learning_rate": 1.550962922619749e-05, |
| "loss": 1.225, |
| "step": 785 |
| }, |
| { |
| "epoch": 1.01, |
| "learning_rate": 1.5498084559357922e-05, |
| "loss": 1.2589, |
| "step": 786 |
| }, |
| { |
| "epoch": 1.01, |
| "learning_rate": 1.548652938029846e-05, |
| "loss": 1.2029, |
| "step": 787 |
| }, |
| { |
| "epoch": 1.01, |
| "learning_rate": 1.5474963711112354e-05, |
| "loss": 1.3469, |
| "step": 788 |
| }, |
| { |
| "epoch": 1.01, |
| "learning_rate": 1.5463387573912916e-05, |
| "loss": 1.1938, |
| "step": 789 |
| }, |
| { |
| "epoch": 1.01, |
| "learning_rate": 1.5451800990833473e-05, |
| "loss": 1.2913, |
| "step": 790 |
| }, |
| { |
| "epoch": 1.01, |
| "learning_rate": 1.5440203984027323e-05, |
| "loss": 1.3701, |
| "step": 791 |
| }, |
| { |
| "epoch": 1.01, |
| "learning_rate": 1.5428596575667697e-05, |
| "loss": 1.1922, |
| "step": 792 |
| }, |
| { |
| "epoch": 1.02, |
| "learning_rate": 1.5416978787947708e-05, |
| "loss": 1.2417, |
| "step": 793 |
| }, |
| { |
| "epoch": 1.02, |
| "learning_rate": 1.5405350643080315e-05, |
| "loss": 1.1946, |
| "step": 794 |
| }, |
| { |
| "epoch": 1.02, |
| "learning_rate": 1.539371216329829e-05, |
| "loss": 1.1704, |
| "step": 795 |
| }, |
| { |
| "epoch": 1.02, |
| "learning_rate": 1.5382063370854154e-05, |
| "loss": 1.3232, |
| "step": 796 |
| }, |
| { |
| "epoch": 1.02, |
| "learning_rate": 1.5370404288020142e-05, |
| "loss": 1.333, |
| "step": 797 |
| }, |
| { |
| "epoch": 1.02, |
| "learning_rate": 1.5358734937088174e-05, |
| "loss": 1.134, |
| "step": 798 |
| }, |
| { |
| "epoch": 1.02, |
| "learning_rate": 1.5347055340369806e-05, |
| "loss": 1.1303, |
| "step": 799 |
| }, |
| { |
| "epoch": 1.02, |
| "learning_rate": 1.533536552019617e-05, |
| "loss": 1.1818, |
| "step": 800 |
| }, |
| { |
| "epoch": 1.03, |
| "learning_rate": 1.532366549891795e-05, |
| "loss": 1.3051, |
| "step": 801 |
| }, |
| { |
| "epoch": 1.03, |
| "learning_rate": 1.5311955298905342e-05, |
| "loss": 1.2986, |
| "step": 802 |
| }, |
| { |
| "epoch": 1.03, |
| "learning_rate": 1.530023494254799e-05, |
| "loss": 1.2276, |
| "step": 803 |
| }, |
| { |
| "epoch": 1.03, |
| "learning_rate": 1.5288504452254975e-05, |
| "loss": 1.2336, |
| "step": 804 |
| }, |
| { |
| "epoch": 1.03, |
| "learning_rate": 1.5276763850454735e-05, |
| "loss": 1.1609, |
| "step": 805 |
| }, |
| { |
| "epoch": 1.03, |
| "learning_rate": 1.526501315959505e-05, |
| "loss": 1.3192, |
| "step": 806 |
| }, |
| { |
| "epoch": 1.03, |
| "learning_rate": 1.5253252402142989e-05, |
| "loss": 1.2879, |
| "step": 807 |
| }, |
| { |
| "epoch": 1.03, |
| "learning_rate": 1.5241481600584874e-05, |
| "loss": 1.3112, |
| "step": 808 |
| }, |
| { |
| "epoch": 1.04, |
| "learning_rate": 1.5229700777426217e-05, |
| "loss": 1.163, |
| "step": 809 |
| }, |
| { |
| "epoch": 1.04, |
| "learning_rate": 1.5217909955191706e-05, |
| "loss": 1.1871, |
| "step": 810 |
| }, |
| { |
| "epoch": 1.04, |
| "learning_rate": 1.5206109156425137e-05, |
| "loss": 1.1491, |
| "step": 811 |
| }, |
| { |
| "epoch": 1.04, |
| "learning_rate": 1.5194298403689386e-05, |
| "loss": 1.1259, |
| "step": 812 |
| }, |
| { |
| "epoch": 1.04, |
| "learning_rate": 1.518247771956636e-05, |
| "loss": 1.1368, |
| "step": 813 |
| }, |
| { |
| "epoch": 1.04, |
| "learning_rate": 1.517064712665695e-05, |
| "loss": 1.1642, |
| "step": 814 |
| }, |
| { |
| "epoch": 1.04, |
| "learning_rate": 1.5158806647581e-05, |
| "loss": 1.2282, |
| "step": 815 |
| }, |
| { |
| "epoch": 1.04, |
| "learning_rate": 1.5146956304977254e-05, |
| "loss": 1.1852, |
| "step": 816 |
| }, |
| { |
| "epoch": 1.05, |
| "learning_rate": 1.5135096121503308e-05, |
| "loss": 1.2004, |
| "step": 817 |
| }, |
| { |
| "epoch": 1.05, |
| "learning_rate": 1.5123226119835581e-05, |
| "loss": 1.1955, |
| "step": 818 |
| }, |
| { |
| "epoch": 1.05, |
| "learning_rate": 1.5111346322669262e-05, |
| "loss": 1.117, |
| "step": 819 |
| }, |
| { |
| "epoch": 1.05, |
| "learning_rate": 1.5099456752718266e-05, |
| "loss": 1.2484, |
| "step": 820 |
| }, |
| { |
| "epoch": 1.05, |
| "learning_rate": 1.5087557432715206e-05, |
| "loss": 1.3413, |
| "step": 821 |
| }, |
| { |
| "epoch": 1.05, |
| "learning_rate": 1.5075648385411314e-05, |
| "loss": 1.1343, |
| "step": 822 |
| }, |
| { |
| "epoch": 1.05, |
| "learning_rate": 1.506372963357644e-05, |
| "loss": 1.2481, |
| "step": 823 |
| }, |
| { |
| "epoch": 1.05, |
| "learning_rate": 1.5051801199998983e-05, |
| "loss": 1.2712, |
| "step": 824 |
| }, |
| { |
| "epoch": 1.06, |
| "learning_rate": 1.503986310748585e-05, |
| "loss": 1.1932, |
| "step": 825 |
| }, |
| { |
| "epoch": 1.06, |
| "learning_rate": 1.502791537886242e-05, |
| "loss": 1.2357, |
| "step": 826 |
| }, |
| { |
| "epoch": 1.06, |
| "learning_rate": 1.5015958036972488e-05, |
| "loss": 1.2222, |
| "step": 827 |
| }, |
| { |
| "epoch": 1.06, |
| "learning_rate": 1.5003991104678245e-05, |
| "loss": 1.2228, |
| "step": 828 |
| }, |
| { |
| "epoch": 1.06, |
| "learning_rate": 1.4992014604860203e-05, |
| "loss": 1.2384, |
| "step": 829 |
| }, |
| { |
| "epoch": 1.06, |
| "learning_rate": 1.4980028560417176e-05, |
| "loss": 1.2331, |
| "step": 830 |
| }, |
| { |
| "epoch": 1.06, |
| "learning_rate": 1.4968032994266224e-05, |
| "loss": 1.2519, |
| "step": 831 |
| }, |
| { |
| "epoch": 1.06, |
| "learning_rate": 1.4956027929342611e-05, |
| "loss": 1.2266, |
| "step": 832 |
| }, |
| { |
| "epoch": 1.07, |
| "learning_rate": 1.4944013388599763e-05, |
| "loss": 1.1775, |
| "step": 833 |
| }, |
| { |
| "epoch": 1.07, |
| "learning_rate": 1.493198939500923e-05, |
| "loss": 1.1914, |
| "step": 834 |
| }, |
| { |
| "epoch": 1.07, |
| "learning_rate": 1.4919955971560622e-05, |
| "loss": 1.1442, |
| "step": 835 |
| }, |
| { |
| "epoch": 1.07, |
| "learning_rate": 1.4907913141261591e-05, |
| "loss": 1.1764, |
| "step": 836 |
| }, |
| { |
| "epoch": 1.07, |
| "learning_rate": 1.4895860927137773e-05, |
| "loss": 1.2792, |
| "step": 837 |
| }, |
| { |
| "epoch": 1.07, |
| "learning_rate": 1.4883799352232743e-05, |
| "loss": 1.2952, |
| "step": 838 |
| }, |
| { |
| "epoch": 1.07, |
| "learning_rate": 1.4871728439607967e-05, |
| "loss": 1.2613, |
| "step": 839 |
| }, |
| { |
| "epoch": 1.08, |
| "learning_rate": 1.4859648212342778e-05, |
| "loss": 1.3075, |
| "step": 840 |
| }, |
| { |
| "epoch": 1.08, |
| "learning_rate": 1.4847558693534309e-05, |
| "loss": 1.1774, |
| "step": 841 |
| }, |
| { |
| "epoch": 1.08, |
| "learning_rate": 1.4835459906297462e-05, |
| "loss": 1.1291, |
| "step": 842 |
| }, |
| { |
| "epoch": 1.08, |
| "learning_rate": 1.4823351873764862e-05, |
| "loss": 1.2469, |
| "step": 843 |
| }, |
| { |
| "epoch": 1.08, |
| "learning_rate": 1.4811234619086802e-05, |
| "loss": 1.0765, |
| "step": 844 |
| }, |
| { |
| "epoch": 1.08, |
| "learning_rate": 1.4799108165431215e-05, |
| "loss": 1.1946, |
| "step": 845 |
| }, |
| { |
| "epoch": 1.08, |
| "learning_rate": 1.4786972535983624e-05, |
| "loss": 1.173, |
| "step": 846 |
| }, |
| { |
| "epoch": 1.08, |
| "learning_rate": 1.4774827753947088e-05, |
| "loss": 1.2652, |
| "step": 847 |
| }, |
| { |
| "epoch": 1.09, |
| "learning_rate": 1.4762673842542175e-05, |
| "loss": 1.1688, |
| "step": 848 |
| }, |
| { |
| "epoch": 1.09, |
| "learning_rate": 1.4750510825006898e-05, |
| "loss": 1.2725, |
| "step": 849 |
| }, |
| { |
| "epoch": 1.09, |
| "learning_rate": 1.4738338724596691e-05, |
| "loss": 1.1952, |
| "step": 850 |
| }, |
| { |
| "epoch": 1.09, |
| "learning_rate": 1.4726157564584345e-05, |
| "loss": 1.1522, |
| "step": 851 |
| }, |
| { |
| "epoch": 1.09, |
| "learning_rate": 1.4713967368259981e-05, |
| "loss": 1.178, |
| "step": 852 |
| }, |
| { |
| "epoch": 1.09, |
| "learning_rate": 1.4701768158930987e-05, |
| "loss": 1.2104, |
| "step": 853 |
| }, |
| { |
| "epoch": 1.09, |
| "learning_rate": 1.4689559959921995e-05, |
| "loss": 1.3163, |
| "step": 854 |
| }, |
| { |
| "epoch": 1.09, |
| "learning_rate": 1.4677342794574819e-05, |
| "loss": 1.1721, |
| "step": 855 |
| }, |
| { |
| "epoch": 1.1, |
| "learning_rate": 1.4665116686248418e-05, |
| "loss": 1.1627, |
| "step": 856 |
| }, |
| { |
| "epoch": 1.1, |
| "learning_rate": 1.465288165831885e-05, |
| "loss": 1.1635, |
| "step": 857 |
| }, |
| { |
| "epoch": 1.1, |
| "learning_rate": 1.4640637734179222e-05, |
| "loss": 1.1497, |
| "step": 858 |
| }, |
| { |
| "epoch": 1.1, |
| "learning_rate": 1.4628384937239659e-05, |
| "loss": 1.1312, |
| "step": 859 |
| }, |
| { |
| "epoch": 1.1, |
| "learning_rate": 1.4616123290927244e-05, |
| "loss": 1.2433, |
| "step": 860 |
| }, |
| { |
| "epoch": 1.1, |
| "learning_rate": 1.4603852818685985e-05, |
| "loss": 1.092, |
| "step": 861 |
| }, |
| { |
| "epoch": 1.1, |
| "learning_rate": 1.4591573543976758e-05, |
| "loss": 1.3162, |
| "step": 862 |
| }, |
| { |
| "epoch": 1.1, |
| "learning_rate": 1.4579285490277275e-05, |
| "loss": 1.1653, |
| "step": 863 |
| }, |
| { |
| "epoch": 1.11, |
| "learning_rate": 1.456698868108203e-05, |
| "loss": 1.2219, |
| "step": 864 |
| }, |
| { |
| "epoch": 1.11, |
| "learning_rate": 1.4554683139902262e-05, |
| "loss": 1.206, |
| "step": 865 |
| }, |
| { |
| "epoch": 1.11, |
| "learning_rate": 1.4542368890265901e-05, |
| "loss": 1.4149, |
| "step": 866 |
| }, |
| { |
| "epoch": 1.11, |
| "learning_rate": 1.4530045955717528e-05, |
| "loss": 1.2155, |
| "step": 867 |
| }, |
| { |
| "epoch": 1.11, |
| "learning_rate": 1.4517714359818329e-05, |
| "loss": 1.198, |
| "step": 868 |
| }, |
| { |
| "epoch": 1.11, |
| "learning_rate": 1.4505374126146053e-05, |
| "loss": 1.2839, |
| "step": 869 |
| }, |
| { |
| "epoch": 1.11, |
| "learning_rate": 1.449302527829496e-05, |
| "loss": 1.139, |
| "step": 870 |
| }, |
| { |
| "epoch": 1.11, |
| "learning_rate": 1.4480667839875786e-05, |
| "loss": 1.2797, |
| "step": 871 |
| }, |
| { |
| "epoch": 1.12, |
| "learning_rate": 1.4468301834515685e-05, |
| "loss": 1.1736, |
| "step": 872 |
| }, |
| { |
| "epoch": 1.12, |
| "learning_rate": 1.4455927285858196e-05, |
| "loss": 1.1984, |
| "step": 873 |
| }, |
| { |
| "epoch": 1.12, |
| "learning_rate": 1.444354421756319e-05, |
| "loss": 1.2049, |
| "step": 874 |
| }, |
| { |
| "epoch": 1.12, |
| "learning_rate": 1.4431152653306833e-05, |
| "loss": 1.2572, |
| "step": 875 |
| }, |
| { |
| "epoch": 1.12, |
| "learning_rate": 1.4418752616781522e-05, |
| "loss": 1.2538, |
| "step": 876 |
| }, |
| { |
| "epoch": 1.12, |
| "learning_rate": 1.4406344131695868e-05, |
| "loss": 1.2, |
| "step": 877 |
| }, |
| { |
| "epoch": 1.12, |
| "learning_rate": 1.4393927221774627e-05, |
| "loss": 1.1266, |
| "step": 878 |
| }, |
| { |
| "epoch": 1.13, |
| "learning_rate": 1.4381501910758662e-05, |
| "loss": 1.2158, |
| "step": 879 |
| }, |
| { |
| "epoch": 1.13, |
| "learning_rate": 1.4369068222404904e-05, |
| "loss": 1.2406, |
| "step": 880 |
| }, |
| { |
| "epoch": 1.13, |
| "learning_rate": 1.4356626180486299e-05, |
| "loss": 1.0863, |
| "step": 881 |
| }, |
| { |
| "epoch": 1.13, |
| "learning_rate": 1.4344175808791765e-05, |
| "loss": 1.2011, |
| "step": 882 |
| }, |
| { |
| "epoch": 1.13, |
| "learning_rate": 1.4331717131126142e-05, |
| "loss": 1.1927, |
| "step": 883 |
| }, |
| { |
| "epoch": 1.13, |
| "learning_rate": 1.4319250171310164e-05, |
| "loss": 1.2026, |
| "step": 884 |
| }, |
| { |
| "epoch": 1.13, |
| "learning_rate": 1.430677495318038e-05, |
| "loss": 1.203, |
| "step": 885 |
| }, |
| { |
| "epoch": 1.13, |
| "learning_rate": 1.4294291500589145e-05, |
| "loss": 1.2063, |
| "step": 886 |
| }, |
| { |
| "epoch": 1.14, |
| "learning_rate": 1.4281799837404553e-05, |
| "loss": 1.0684, |
| "step": 887 |
| }, |
| { |
| "epoch": 1.14, |
| "learning_rate": 1.4269299987510398e-05, |
| "loss": 1.3337, |
| "step": 888 |
| }, |
| { |
| "epoch": 1.14, |
| "learning_rate": 1.4256791974806122e-05, |
| "loss": 1.1662, |
| "step": 889 |
| }, |
| { |
| "epoch": 1.14, |
| "learning_rate": 1.4244275823206776e-05, |
| "loss": 1.2638, |
| "step": 890 |
| }, |
| { |
| "epoch": 1.14, |
| "learning_rate": 1.4231751556642981e-05, |
| "loss": 1.2312, |
| "step": 891 |
| }, |
| { |
| "epoch": 1.14, |
| "learning_rate": 1.421921919906086e-05, |
| "loss": 1.2843, |
| "step": 892 |
| }, |
| { |
| "epoch": 1.14, |
| "learning_rate": 1.4206678774422016e-05, |
| "loss": 1.2139, |
| "step": 893 |
| }, |
| { |
| "epoch": 1.14, |
| "learning_rate": 1.4194130306703467e-05, |
| "loss": 1.192, |
| "step": 894 |
| }, |
| { |
| "epoch": 1.15, |
| "learning_rate": 1.4181573819897618e-05, |
| "loss": 1.1594, |
| "step": 895 |
| }, |
| { |
| "epoch": 1.15, |
| "learning_rate": 1.4169009338012205e-05, |
| "loss": 1.2109, |
| "step": 896 |
| }, |
| { |
| "epoch": 1.15, |
| "learning_rate": 1.4156436885070243e-05, |
| "loss": 1.2549, |
| "step": 897 |
| }, |
| { |
| "epoch": 1.15, |
| "learning_rate": 1.4143856485109999e-05, |
| "loss": 1.2495, |
| "step": 898 |
| }, |
| { |
| "epoch": 1.15, |
| "learning_rate": 1.4131268162184921e-05, |
| "loss": 1.1652, |
| "step": 899 |
| }, |
| { |
| "epoch": 1.15, |
| "learning_rate": 1.4118671940363618e-05, |
| "loss": 1.1898, |
| "step": 900 |
| }, |
| { |
| "epoch": 1.15, |
| "learning_rate": 1.4106067843729797e-05, |
| "loss": 1.1882, |
| "step": 901 |
| }, |
| { |
| "epoch": 1.15, |
| "learning_rate": 1.4093455896382215e-05, |
| "loss": 1.3092, |
| "step": 902 |
| }, |
| { |
| "epoch": 1.16, |
| "learning_rate": 1.408083612243465e-05, |
| "loss": 1.2652, |
| "step": 903 |
| }, |
| { |
| "epoch": 1.16, |
| "learning_rate": 1.4068208546015838e-05, |
| "loss": 1.1824, |
| "step": 904 |
| }, |
| { |
| "epoch": 1.16, |
| "learning_rate": 1.4055573191269436e-05, |
| "loss": 1.2462, |
| "step": 905 |
| }, |
| { |
| "epoch": 1.16, |
| "learning_rate": 1.4042930082353969e-05, |
| "loss": 1.3578, |
| "step": 906 |
| }, |
| { |
| "epoch": 1.16, |
| "learning_rate": 1.4030279243442793e-05, |
| "loss": 1.178, |
| "step": 907 |
| }, |
| { |
| "epoch": 1.16, |
| "learning_rate": 1.4017620698724037e-05, |
| "loss": 1.2513, |
| "step": 908 |
| }, |
| { |
| "epoch": 1.16, |
| "learning_rate": 1.4004954472400574e-05, |
| "loss": 1.239, |
| "step": 909 |
| }, |
| { |
| "epoch": 1.16, |
| "learning_rate": 1.399228058868995e-05, |
| "loss": 1.1456, |
| "step": 910 |
| }, |
| { |
| "epoch": 1.17, |
| "learning_rate": 1.3979599071824363e-05, |
| "loss": 1.2961, |
| "step": 911 |
| }, |
| { |
| "epoch": 1.17, |
| "learning_rate": 1.3966909946050602e-05, |
| "loss": 1.1437, |
| "step": 912 |
| }, |
| { |
| "epoch": 1.17, |
| "learning_rate": 1.3954213235630005e-05, |
| "loss": 1.2119, |
| "step": 913 |
| }, |
| { |
| "epoch": 1.17, |
| "learning_rate": 1.3941508964838407e-05, |
| "loss": 1.3639, |
| "step": 914 |
| }, |
| { |
| "epoch": 1.17, |
| "learning_rate": 1.3928797157966098e-05, |
| "loss": 1.1819, |
| "step": 915 |
| }, |
| { |
| "epoch": 1.17, |
| "learning_rate": 1.391607783931779e-05, |
| "loss": 1.2616, |
| "step": 916 |
| }, |
| { |
| "epoch": 1.17, |
| "learning_rate": 1.3903351033212544e-05, |
| "loss": 1.2256, |
| "step": 917 |
| }, |
| { |
| "epoch": 1.18, |
| "learning_rate": 1.3890616763983738e-05, |
| "loss": 1.3166, |
| "step": 918 |
| }, |
| { |
| "epoch": 1.18, |
| "learning_rate": 1.3877875055979021e-05, |
| "loss": 1.2222, |
| "step": 919 |
| }, |
| { |
| "epoch": 1.18, |
| "learning_rate": 1.3865125933560272e-05, |
| "loss": 1.2539, |
| "step": 920 |
| }, |
| { |
| "epoch": 1.18, |
| "learning_rate": 1.3852369421103538e-05, |
| "loss": 1.2659, |
| "step": 921 |
| }, |
| { |
| "epoch": 1.18, |
| "learning_rate": 1.3839605542998991e-05, |
| "loss": 1.2749, |
| "step": 922 |
| }, |
| { |
| "epoch": 1.18, |
| "learning_rate": 1.3826834323650899e-05, |
| "loss": 1.1186, |
| "step": 923 |
| }, |
| { |
| "epoch": 1.18, |
| "learning_rate": 1.3814055787477558e-05, |
| "loss": 1.2641, |
| "step": 924 |
| }, |
| { |
| "epoch": 1.18, |
| "learning_rate": 1.3801269958911256e-05, |
| "loss": 1.0724, |
| "step": 925 |
| }, |
| { |
| "epoch": 1.19, |
| "learning_rate": 1.3788476862398226e-05, |
| "loss": 1.2218, |
| "step": 926 |
| }, |
| { |
| "epoch": 1.19, |
| "learning_rate": 1.3775676522398587e-05, |
| "loss": 1.241, |
| "step": 927 |
| }, |
| { |
| "epoch": 1.19, |
| "learning_rate": 1.376286896338632e-05, |
| "loss": 1.2912, |
| "step": 928 |
| }, |
| { |
| "epoch": 1.19, |
| "learning_rate": 1.3750054209849204e-05, |
| "loss": 1.2464, |
| "step": 929 |
| }, |
| { |
| "epoch": 1.19, |
| "learning_rate": 1.3737232286288773e-05, |
| "loss": 1.2726, |
| "step": 930 |
| }, |
| { |
| "epoch": 1.19, |
| "learning_rate": 1.3724403217220268e-05, |
| "loss": 1.1272, |
| "step": 931 |
| }, |
| { |
| "epoch": 1.19, |
| "learning_rate": 1.3711567027172594e-05, |
| "loss": 1.2658, |
| "step": 932 |
| }, |
| { |
| "epoch": 1.19, |
| "learning_rate": 1.3698723740688275e-05, |
| "loss": 1.2059, |
| "step": 933 |
| }, |
| { |
| "epoch": 1.2, |
| "learning_rate": 1.3685873382323398e-05, |
| "loss": 1.2691, |
| "step": 934 |
| }, |
| { |
| "epoch": 1.2, |
| "learning_rate": 1.367301597664757e-05, |
| "loss": 1.1836, |
| "step": 935 |
| }, |
| { |
| "epoch": 1.2, |
| "learning_rate": 1.3660151548243879e-05, |
| "loss": 1.2218, |
| "step": 936 |
| }, |
| { |
| "epoch": 1.2, |
| "learning_rate": 1.3647280121708835e-05, |
| "loss": 1.2973, |
| "step": 937 |
| }, |
| { |
| "epoch": 1.2, |
| "learning_rate": 1.363440172165233e-05, |
| "loss": 1.1976, |
| "step": 938 |
| }, |
| { |
| "epoch": 1.2, |
| "learning_rate": 1.3621516372697587e-05, |
| "loss": 1.1687, |
| "step": 939 |
| }, |
| { |
| "epoch": 1.2, |
| "learning_rate": 1.360862409948112e-05, |
| "loss": 1.2191, |
| "step": 940 |
| }, |
| { |
| "epoch": 1.2, |
| "learning_rate": 1.3595724926652679e-05, |
| "loss": 1.076, |
| "step": 941 |
| }, |
| { |
| "epoch": 1.21, |
| "learning_rate": 1.3582818878875209e-05, |
| "loss": 1.2148, |
| "step": 942 |
| }, |
| { |
| "epoch": 1.21, |
| "learning_rate": 1.3569905980824789e-05, |
| "loss": 1.1488, |
| "step": 943 |
| }, |
| { |
| "epoch": 1.21, |
| "learning_rate": 1.3556986257190608e-05, |
| "loss": 1.1193, |
| "step": 944 |
| }, |
| { |
| "epoch": 1.21, |
| "learning_rate": 1.35440597326749e-05, |
| "loss": 1.338, |
| "step": 945 |
| }, |
| { |
| "epoch": 1.21, |
| "learning_rate": 1.3531126431992906e-05, |
| "loss": 1.2717, |
| "step": 946 |
| }, |
| { |
| "epoch": 1.21, |
| "learning_rate": 1.3518186379872815e-05, |
| "loss": 1.2307, |
| "step": 947 |
| }, |
| { |
| "epoch": 1.21, |
| "learning_rate": 1.3505239601055734e-05, |
| "loss": 1.296, |
| "step": 948 |
| }, |
| { |
| "epoch": 1.21, |
| "learning_rate": 1.3492286120295624e-05, |
| "loss": 1.3322, |
| "step": 949 |
| }, |
| { |
| "epoch": 1.22, |
| "learning_rate": 1.3479325962359263e-05, |
| "loss": 1.1685, |
| "step": 950 |
| }, |
| { |
| "epoch": 1.22, |
| "learning_rate": 1.3466359152026197e-05, |
| "loss": 1.0792, |
| "step": 951 |
| }, |
| { |
| "epoch": 1.22, |
| "learning_rate": 1.3453385714088686e-05, |
| "loss": 1.3691, |
| "step": 952 |
| }, |
| { |
| "epoch": 1.22, |
| "learning_rate": 1.3440405673351668e-05, |
| "loss": 1.2432, |
| "step": 953 |
| }, |
| { |
| "epoch": 1.22, |
| "learning_rate": 1.3427419054632705e-05, |
| "loss": 1.2322, |
| "step": 954 |
| }, |
| { |
| "epoch": 1.22, |
| "learning_rate": 1.341442588276193e-05, |
| "loss": 1.2585, |
| "step": 955 |
| }, |
| { |
| "epoch": 1.22, |
| "learning_rate": 1.3401426182582009e-05, |
| "loss": 1.2795, |
| "step": 956 |
| }, |
| { |
| "epoch": 1.22, |
| "learning_rate": 1.3388419978948092e-05, |
| "loss": 1.222, |
| "step": 957 |
| }, |
| { |
| "epoch": 1.23, |
| "learning_rate": 1.3375407296727762e-05, |
| "loss": 1.0831, |
| "step": 958 |
| }, |
| { |
| "epoch": 1.23, |
| "learning_rate": 1.3362388160800992e-05, |
| "loss": 1.1849, |
| "step": 959 |
| }, |
| { |
| "epoch": 1.23, |
| "learning_rate": 1.3349362596060084e-05, |
| "loss": 1.2971, |
| "step": 960 |
| }, |
| { |
| "epoch": 1.23, |
| "learning_rate": 1.3336330627409642e-05, |
| "loss": 1.1618, |
| "step": 961 |
| }, |
| { |
| "epoch": 1.23, |
| "learning_rate": 1.332329227976651e-05, |
| "loss": 1.205, |
| "step": 962 |
| }, |
| { |
| "epoch": 1.23, |
| "learning_rate": 1.3310247578059734e-05, |
| "loss": 1.2194, |
| "step": 963 |
| }, |
| { |
| "epoch": 1.23, |
| "learning_rate": 1.3297196547230497e-05, |
| "loss": 1.223, |
| "step": 964 |
| }, |
| { |
| "epoch": 1.24, |
| "learning_rate": 1.3284139212232095e-05, |
| "loss": 1.3378, |
| "step": 965 |
| }, |
| { |
| "epoch": 1.24, |
| "learning_rate": 1.327107559802987e-05, |
| "loss": 1.2353, |
| "step": 966 |
| }, |
| { |
| "epoch": 1.24, |
| "learning_rate": 1.3258005729601178e-05, |
| "loss": 1.1581, |
| "step": 967 |
| }, |
| { |
| "epoch": 1.24, |
| "learning_rate": 1.3244929631935322e-05, |
| "loss": 1.2425, |
| "step": 968 |
| }, |
| { |
| "epoch": 1.24, |
| "learning_rate": 1.3231847330033521e-05, |
| "loss": 1.1883, |
| "step": 969 |
| }, |
| { |
| "epoch": 1.24, |
| "learning_rate": 1.3218758848908857e-05, |
| "loss": 1.1119, |
| "step": 970 |
| }, |
| { |
| "epoch": 1.24, |
| "learning_rate": 1.3205664213586227e-05, |
| "loss": 1.2546, |
| "step": 971 |
| }, |
| { |
| "epoch": 1.24, |
| "learning_rate": 1.3192563449102291e-05, |
| "loss": 1.0959, |
| "step": 972 |
| }, |
| { |
| "epoch": 1.25, |
| "learning_rate": 1.3179456580505432e-05, |
| "loss": 1.2063, |
| "step": 973 |
| }, |
| { |
| "epoch": 1.25, |
| "learning_rate": 1.31663436328557e-05, |
| "loss": 1.1055, |
| "step": 974 |
| }, |
| { |
| "epoch": 1.25, |
| "learning_rate": 1.3153224631224772e-05, |
| "loss": 1.2694, |
| "step": 975 |
| }, |
| { |
| "epoch": 1.25, |
| "learning_rate": 1.3140099600695899e-05, |
| "loss": 1.0975, |
| "step": 976 |
| }, |
| { |
| "epoch": 1.25, |
| "learning_rate": 1.3126968566363853e-05, |
| "loss": 1.2222, |
| "step": 977 |
| }, |
| { |
| "epoch": 1.25, |
| "learning_rate": 1.3113831553334894e-05, |
| "loss": 1.286, |
| "step": 978 |
| }, |
| { |
| "epoch": 1.25, |
| "learning_rate": 1.310068858672671e-05, |
| "loss": 1.2431, |
| "step": 979 |
| }, |
| { |
| "epoch": 1.25, |
| "learning_rate": 1.3087539691668373e-05, |
| "loss": 1.2218, |
| "step": 980 |
| }, |
| { |
| "epoch": 1.26, |
| "learning_rate": 1.3074384893300285e-05, |
| "loss": 1.1672, |
| "step": 981 |
| }, |
| { |
| "epoch": 1.26, |
| "learning_rate": 1.3061224216774137e-05, |
| "loss": 1.252, |
| "step": 982 |
| }, |
| { |
| "epoch": 1.26, |
| "learning_rate": 1.3048057687252866e-05, |
| "loss": 1.0989, |
| "step": 983 |
| }, |
| { |
| "epoch": 1.26, |
| "learning_rate": 1.3034885329910593e-05, |
| "loss": 1.3368, |
| "step": 984 |
| }, |
| { |
| "epoch": 1.26, |
| "learning_rate": 1.3021707169932585e-05, |
| "loss": 1.2396, |
| "step": 985 |
| }, |
| { |
| "epoch": 1.26, |
| "learning_rate": 1.3008523232515193e-05, |
| "loss": 1.1901, |
| "step": 986 |
| }, |
| { |
| "epoch": 1.26, |
| "learning_rate": 1.2995333542865831e-05, |
| "loss": 1.1528, |
| "step": 987 |
| }, |
| { |
| "epoch": 1.26, |
| "learning_rate": 1.2982138126202901e-05, |
| "loss": 1.2131, |
| "step": 988 |
| }, |
| { |
| "epoch": 1.27, |
| "learning_rate": 1.2968937007755756e-05, |
| "loss": 1.1468, |
| "step": 989 |
| }, |
| { |
| "epoch": 1.27, |
| "learning_rate": 1.2955730212764655e-05, |
| "loss": 1.2623, |
| "step": 990 |
| }, |
| { |
| "epoch": 1.27, |
| "learning_rate": 1.2942517766480702e-05, |
| "loss": 1.2391, |
| "step": 991 |
| }, |
| { |
| "epoch": 1.27, |
| "learning_rate": 1.2929299694165813e-05, |
| "loss": 1.2141, |
| "step": 992 |
| }, |
| { |
| "epoch": 1.27, |
| "learning_rate": 1.2916076021092661e-05, |
| "loss": 1.1282, |
| "step": 993 |
| }, |
| { |
| "epoch": 1.27, |
| "learning_rate": 1.2902846772544625e-05, |
| "loss": 1.2265, |
| "step": 994 |
| }, |
| { |
| "epoch": 1.27, |
| "learning_rate": 1.2889611973815746e-05, |
| "loss": 1.2512, |
| "step": 995 |
| }, |
| { |
| "epoch": 1.27, |
| "learning_rate": 1.2876371650210671e-05, |
| "loss": 1.3452, |
| "step": 996 |
| }, |
| { |
| "epoch": 1.28, |
| "learning_rate": 1.2863125827044619e-05, |
| "loss": 1.2332, |
| "step": 997 |
| }, |
| { |
| "epoch": 1.28, |
| "learning_rate": 1.2849874529643318e-05, |
| "loss": 1.3424, |
| "step": 998 |
| }, |
| { |
| "epoch": 1.28, |
| "learning_rate": 1.2836617783342968e-05, |
| "loss": 1.1914, |
| "step": 999 |
| }, |
| { |
| "epoch": 1.28, |
| "learning_rate": 1.2823355613490182e-05, |
| "loss": 1.2177, |
| "step": 1000 |
| }, |
| { |
| "epoch": 1.28, |
| "learning_rate": 1.2810088045441944e-05, |
| "loss": 1.1754, |
| "step": 1001 |
| }, |
| { |
| "epoch": 1.28, |
| "learning_rate": 1.2796815104565559e-05, |
| "loss": 1.1826, |
| "step": 1002 |
| }, |
| { |
| "epoch": 1.28, |
| "learning_rate": 1.278353681623861e-05, |
| "loss": 1.2538, |
| "step": 1003 |
| }, |
| { |
| "epoch": 1.29, |
| "learning_rate": 1.2770253205848895e-05, |
| "loss": 1.1912, |
| "step": 1004 |
| }, |
| { |
| "epoch": 1.29, |
| "learning_rate": 1.2756964298794396e-05, |
| "loss": 1.0092, |
| "step": 1005 |
| }, |
| { |
| "epoch": 1.29, |
| "learning_rate": 1.2743670120483216e-05, |
| "loss": 1.1961, |
| "step": 1006 |
| }, |
| { |
| "epoch": 1.29, |
| "learning_rate": 1.2730370696333538e-05, |
| "loss": 1.204, |
| "step": 1007 |
| }, |
| { |
| "epoch": 1.29, |
| "learning_rate": 1.271706605177358e-05, |
| "loss": 1.3658, |
| "step": 1008 |
| }, |
| { |
| "epoch": 1.29, |
| "learning_rate": 1.270375621224154e-05, |
| "loss": 1.2451, |
| "step": 1009 |
| }, |
| { |
| "epoch": 1.29, |
| "learning_rate": 1.2690441203185537e-05, |
| "loss": 1.2666, |
| "step": 1010 |
| }, |
| { |
| "epoch": 1.29, |
| "learning_rate": 1.2677121050063588e-05, |
| "loss": 1.0864, |
| "step": 1011 |
| }, |
| { |
| "epoch": 1.3, |
| "learning_rate": 1.2663795778343543e-05, |
| "loss": 1.1692, |
| "step": 1012 |
| }, |
| { |
| "epoch": 1.3, |
| "learning_rate": 1.2650465413503033e-05, |
| "loss": 1.3123, |
| "step": 1013 |
| }, |
| { |
| "epoch": 1.3, |
| "learning_rate": 1.2637129981029428e-05, |
| "loss": 1.2077, |
| "step": 1014 |
| }, |
| { |
| "epoch": 1.3, |
| "learning_rate": 1.2623789506419792e-05, |
| "loss": 1.1327, |
| "step": 1015 |
| }, |
| { |
| "epoch": 1.3, |
| "learning_rate": 1.2610444015180825e-05, |
| "loss": 1.2426, |
| "step": 1016 |
| }, |
| { |
| "epoch": 1.3, |
| "learning_rate": 1.2597093532828826e-05, |
| "loss": 1.285, |
| "step": 1017 |
| }, |
| { |
| "epoch": 1.3, |
| "learning_rate": 1.2583738084889623e-05, |
| "loss": 1.2837, |
| "step": 1018 |
| }, |
| { |
| "epoch": 1.3, |
| "learning_rate": 1.2570377696898552e-05, |
| "loss": 1.1604, |
| "step": 1019 |
| }, |
| { |
| "epoch": 1.31, |
| "learning_rate": 1.2557012394400387e-05, |
| "loss": 1.1691, |
| "step": 1020 |
| }, |
| { |
| "epoch": 1.31, |
| "learning_rate": 1.2543642202949297e-05, |
| "loss": 1.246, |
| "step": 1021 |
| }, |
| { |
| "epoch": 1.31, |
| "learning_rate": 1.253026714810881e-05, |
| "loss": 1.1309, |
| "step": 1022 |
| }, |
| { |
| "epoch": 1.31, |
| "learning_rate": 1.2516887255451735e-05, |
| "loss": 1.2366, |
| "step": 1023 |
| }, |
| { |
| "epoch": 1.31, |
| "learning_rate": 1.2503502550560145e-05, |
| "loss": 1.3223, |
| "step": 1024 |
| }, |
| { |
| "epoch": 1.31, |
| "learning_rate": 1.2490113059025308e-05, |
| "loss": 1.2855, |
| "step": 1025 |
| }, |
| { |
| "epoch": 1.31, |
| "learning_rate": 1.2476718806447646e-05, |
| "loss": 1.2117, |
| "step": 1026 |
| }, |
| { |
| "epoch": 1.31, |
| "learning_rate": 1.2463319818436681e-05, |
| "loss": 1.2173, |
| "step": 1027 |
| }, |
| { |
| "epoch": 1.32, |
| "learning_rate": 1.2449916120610991e-05, |
| "loss": 1.2821, |
| "step": 1028 |
| }, |
| { |
| "epoch": 1.32, |
| "learning_rate": 1.2436507738598162e-05, |
| "loss": 1.178, |
| "step": 1029 |
| }, |
| { |
| "epoch": 1.32, |
| "learning_rate": 1.2423094698034732e-05, |
| "loss": 1.2223, |
| "step": 1030 |
| }, |
| { |
| "epoch": 1.32, |
| "learning_rate": 1.2409677024566145e-05, |
| "loss": 1.2041, |
| "step": 1031 |
| }, |
| { |
| "epoch": 1.32, |
| "learning_rate": 1.2396254743846707e-05, |
| "loss": 1.2204, |
| "step": 1032 |
| }, |
| { |
| "epoch": 1.32, |
| "learning_rate": 1.2382827881539531e-05, |
| "loss": 1.2247, |
| "step": 1033 |
| }, |
| { |
| "epoch": 1.32, |
| "learning_rate": 1.2369396463316491e-05, |
| "loss": 1.2221, |
| "step": 1034 |
| }, |
| { |
| "epoch": 1.32, |
| "learning_rate": 1.2355960514858171e-05, |
| "loss": 1.1776, |
| "step": 1035 |
| }, |
| { |
| "epoch": 1.33, |
| "learning_rate": 1.2342520061853812e-05, |
| "loss": 1.2304, |
| "step": 1036 |
| }, |
| { |
| "epoch": 1.33, |
| "learning_rate": 1.232907513000128e-05, |
| "loss": 1.1772, |
| "step": 1037 |
| }, |
| { |
| "epoch": 1.33, |
| "learning_rate": 1.2315625745006991e-05, |
| "loss": 1.1897, |
| "step": 1038 |
| }, |
| { |
| "epoch": 1.33, |
| "learning_rate": 1.2302171932585885e-05, |
| "loss": 1.226, |
| "step": 1039 |
| }, |
| { |
| "epoch": 1.33, |
| "learning_rate": 1.2288713718461356e-05, |
| "loss": 1.1921, |
| "step": 1040 |
| }, |
| { |
| "epoch": 1.33, |
| "learning_rate": 1.227525112836523e-05, |
| "loss": 1.2414, |
| "step": 1041 |
| }, |
| { |
| "epoch": 1.33, |
| "learning_rate": 1.2261784188037682e-05, |
| "loss": 1.2773, |
| "step": 1042 |
| }, |
| { |
| "epoch": 1.34, |
| "learning_rate": 1.2248312923227224e-05, |
| "loss": 1.3523, |
| "step": 1043 |
| }, |
| { |
| "epoch": 1.34, |
| "learning_rate": 1.2234837359690616e-05, |
| "loss": 1.1346, |
| "step": 1044 |
| }, |
| { |
| "epoch": 1.34, |
| "learning_rate": 1.2221357523192851e-05, |
| "loss": 1.2644, |
| "step": 1045 |
| }, |
| { |
| "epoch": 1.34, |
| "learning_rate": 1.2207873439507087e-05, |
| "loss": 1.209, |
| "step": 1046 |
| }, |
| { |
| "epoch": 1.34, |
| "learning_rate": 1.2194385134414608e-05, |
| "loss": 1.2497, |
| "step": 1047 |
| }, |
| { |
| "epoch": 1.34, |
| "learning_rate": 1.2180892633704756e-05, |
| "loss": 1.2583, |
| "step": 1048 |
| }, |
| { |
| "epoch": 1.34, |
| "learning_rate": 1.2167395963174906e-05, |
| "loss": 1.2783, |
| "step": 1049 |
| }, |
| { |
| "epoch": 1.34, |
| "learning_rate": 1.2153895148630406e-05, |
| "loss": 1.1737, |
| "step": 1050 |
| }, |
| { |
| "epoch": 1.35, |
| "learning_rate": 1.2140390215884523e-05, |
| "loss": 1.2858, |
| "step": 1051 |
| }, |
| { |
| "epoch": 1.35, |
| "learning_rate": 1.2126881190758397e-05, |
| "loss": 1.2904, |
| "step": 1052 |
| }, |
| { |
| "epoch": 1.35, |
| "learning_rate": 1.2113368099080997e-05, |
| "loss": 1.1233, |
| "step": 1053 |
| }, |
| { |
| "epoch": 1.35, |
| "learning_rate": 1.2099850966689064e-05, |
| "loss": 1.2289, |
| "step": 1054 |
| }, |
| { |
| "epoch": 1.35, |
| "learning_rate": 1.2086329819427065e-05, |
| "loss": 1.1652, |
| "step": 1055 |
| }, |
| { |
| "epoch": 1.35, |
| "learning_rate": 1.2072804683147146e-05, |
| "loss": 1.288, |
| "step": 1056 |
| }, |
| { |
| "epoch": 1.35, |
| "learning_rate": 1.2059275583709074e-05, |
| "loss": 1.2332, |
| "step": 1057 |
| }, |
| { |
| "epoch": 1.35, |
| "learning_rate": 1.20457425469802e-05, |
| "loss": 1.2306, |
| "step": 1058 |
| }, |
| { |
| "epoch": 1.36, |
| "learning_rate": 1.20322055988354e-05, |
| "loss": 1.2533, |
| "step": 1059 |
| }, |
| { |
| "epoch": 1.36, |
| "learning_rate": 1.2018664765157029e-05, |
| "loss": 1.1684, |
| "step": 1060 |
| }, |
| { |
| "epoch": 1.36, |
| "learning_rate": 1.200512007183487e-05, |
| "loss": 1.1907, |
| "step": 1061 |
| }, |
| { |
| "epoch": 1.36, |
| "learning_rate": 1.1991571544766085e-05, |
| "loss": 1.2096, |
| "step": 1062 |
| }, |
| { |
| "epoch": 1.36, |
| "learning_rate": 1.1978019209855174e-05, |
| "loss": 1.1476, |
| "step": 1063 |
| }, |
| { |
| "epoch": 1.36, |
| "learning_rate": 1.1964463093013905e-05, |
| "loss": 1.2318, |
| "step": 1064 |
| }, |
| { |
| "epoch": 1.36, |
| "learning_rate": 1.1950903220161286e-05, |
| "loss": 1.1828, |
| "step": 1065 |
| }, |
| { |
| "epoch": 1.36, |
| "learning_rate": 1.1937339617223498e-05, |
| "loss": 1.1491, |
| "step": 1066 |
| }, |
| { |
| "epoch": 1.37, |
| "learning_rate": 1.1923772310133866e-05, |
| "loss": 1.2285, |
| "step": 1067 |
| }, |
| { |
| "epoch": 1.37, |
| "learning_rate": 1.191020132483279e-05, |
| "loss": 1.2351, |
| "step": 1068 |
| }, |
| { |
| "epoch": 1.37, |
| "learning_rate": 1.1896626687267698e-05, |
| "loss": 1.1737, |
| "step": 1069 |
| }, |
| { |
| "epoch": 1.37, |
| "learning_rate": 1.188304842339301e-05, |
| "loss": 1.2429, |
| "step": 1070 |
| }, |
| { |
| "epoch": 1.37, |
| "learning_rate": 1.1869466559170073e-05, |
| "loss": 1.1993, |
| "step": 1071 |
| }, |
| { |
| "epoch": 1.37, |
| "learning_rate": 1.1855881120567125e-05, |
| "loss": 1.2313, |
| "step": 1072 |
| }, |
| { |
| "epoch": 1.37, |
| "learning_rate": 1.184229213355923e-05, |
| "loss": 1.1466, |
| "step": 1073 |
| }, |
| { |
| "epoch": 1.37, |
| "learning_rate": 1.1828699624128241e-05, |
| "loss": 1.1341, |
| "step": 1074 |
| }, |
| { |
| "epoch": 1.38, |
| "learning_rate": 1.1815103618262746e-05, |
| "loss": 1.1827, |
| "step": 1075 |
| }, |
| { |
| "epoch": 1.38, |
| "learning_rate": 1.1801504141958016e-05, |
| "loss": 1.3454, |
| "step": 1076 |
| }, |
| { |
| "epoch": 1.38, |
| "learning_rate": 1.1787901221215956e-05, |
| "loss": 1.1905, |
| "step": 1077 |
| }, |
| { |
| "epoch": 1.38, |
| "learning_rate": 1.1774294882045063e-05, |
| "loss": 1.2414, |
| "step": 1078 |
| }, |
| { |
| "epoch": 1.38, |
| "learning_rate": 1.1760685150460363e-05, |
| "loss": 1.1697, |
| "step": 1079 |
| }, |
| { |
| "epoch": 1.38, |
| "learning_rate": 1.1747072052483374e-05, |
| "loss": 1.2075, |
| "step": 1080 |
| }, |
| { |
| "epoch": 1.38, |
| "learning_rate": 1.1733455614142047e-05, |
| "loss": 1.2436, |
| "step": 1081 |
| }, |
| { |
| "epoch": 1.38, |
| "learning_rate": 1.1719835861470717e-05, |
| "loss": 1.1449, |
| "step": 1082 |
| }, |
| { |
| "epoch": 1.39, |
| "learning_rate": 1.1706212820510061e-05, |
| "loss": 1.1728, |
| "step": 1083 |
| }, |
| { |
| "epoch": 1.39, |
| "learning_rate": 1.1692586517307047e-05, |
| "loss": 1.2989, |
| "step": 1084 |
| }, |
| { |
| "epoch": 1.39, |
| "learning_rate": 1.1678956977914868e-05, |
| "loss": 1.194, |
| "step": 1085 |
| }, |
| { |
| "epoch": 1.39, |
| "learning_rate": 1.1665324228392914e-05, |
| "loss": 1.1926, |
| "step": 1086 |
| }, |
| { |
| "epoch": 1.39, |
| "learning_rate": 1.1651688294806706e-05, |
| "loss": 1.2826, |
| "step": 1087 |
| }, |
| { |
| "epoch": 1.39, |
| "learning_rate": 1.1638049203227868e-05, |
| "loss": 1.2815, |
| "step": 1088 |
| }, |
| { |
| "epoch": 1.39, |
| "learning_rate": 1.1624406979734038e-05, |
| "loss": 1.4368, |
| "step": 1089 |
| }, |
| { |
| "epoch": 1.4, |
| "learning_rate": 1.1610761650408862e-05, |
| "loss": 1.292, |
| "step": 1090 |
| }, |
| { |
| "epoch": 1.4, |
| "learning_rate": 1.1597113241341915e-05, |
| "loss": 1.2566, |
| "step": 1091 |
| }, |
| { |
| "epoch": 1.4, |
| "learning_rate": 1.1583461778628664e-05, |
| "loss": 1.2278, |
| "step": 1092 |
| }, |
| { |
| "epoch": 1.4, |
| "learning_rate": 1.1569807288370417e-05, |
| "loss": 1.1719, |
| "step": 1093 |
| }, |
| { |
| "epoch": 1.4, |
| "learning_rate": 1.1556149796674259e-05, |
| "loss": 1.2424, |
| "step": 1094 |
| }, |
| { |
| "epoch": 1.4, |
| "learning_rate": 1.1542489329653024e-05, |
| "loss": 1.2071, |
| "step": 1095 |
| }, |
| { |
| "epoch": 1.4, |
| "learning_rate": 1.1528825913425237e-05, |
| "loss": 1.2317, |
| "step": 1096 |
| }, |
| { |
| "epoch": 1.4, |
| "learning_rate": 1.1515159574115053e-05, |
| "loss": 1.1678, |
| "step": 1097 |
| }, |
| { |
| "epoch": 1.41, |
| "learning_rate": 1.1501490337852221e-05, |
| "loss": 1.1983, |
| "step": 1098 |
| }, |
| { |
| "epoch": 1.41, |
| "learning_rate": 1.1487818230772025e-05, |
| "loss": 1.2461, |
| "step": 1099 |
| }, |
| { |
| "epoch": 1.41, |
| "learning_rate": 1.1474143279015247e-05, |
| "loss": 1.2219, |
| "step": 1100 |
| }, |
| { |
| "epoch": 1.41, |
| "learning_rate": 1.1460465508728097e-05, |
| "loss": 1.2425, |
| "step": 1101 |
| }, |
| { |
| "epoch": 1.41, |
| "learning_rate": 1.1446784946062181e-05, |
| "loss": 1.165, |
| "step": 1102 |
| }, |
| { |
| "epoch": 1.41, |
| "learning_rate": 1.143310161717444e-05, |
| "loss": 1.1056, |
| "step": 1103 |
| }, |
| { |
| "epoch": 1.41, |
| "learning_rate": 1.1419415548227109e-05, |
| "loss": 1.2139, |
| "step": 1104 |
| }, |
| { |
| "epoch": 1.41, |
| "learning_rate": 1.1405726765387654e-05, |
| "loss": 1.2269, |
| "step": 1105 |
| }, |
| { |
| "epoch": 1.42, |
| "learning_rate": 1.1392035294828738e-05, |
| "loss": 1.1764, |
| "step": 1106 |
| }, |
| { |
| "epoch": 1.42, |
| "learning_rate": 1.1378341162728158e-05, |
| "loss": 1.1935, |
| "step": 1107 |
| }, |
| { |
| "epoch": 1.42, |
| "learning_rate": 1.13646443952688e-05, |
| "loss": 1.2995, |
| "step": 1108 |
| }, |
| { |
| "epoch": 1.42, |
| "learning_rate": 1.1350945018638596e-05, |
| "loss": 1.2678, |
| "step": 1109 |
| }, |
| { |
| "epoch": 1.42, |
| "learning_rate": 1.1337243059030451e-05, |
| "loss": 1.259, |
| "step": 1110 |
| }, |
| { |
| "epoch": 1.42, |
| "learning_rate": 1.1323538542642227e-05, |
| "loss": 1.2291, |
| "step": 1111 |
| }, |
| { |
| "epoch": 1.42, |
| "learning_rate": 1.130983149567666e-05, |
| "loss": 1.2805, |
| "step": 1112 |
| }, |
| { |
| "epoch": 1.42, |
| "learning_rate": 1.1296121944341332e-05, |
| "loss": 1.2107, |
| "step": 1113 |
| }, |
| { |
| "epoch": 1.43, |
| "learning_rate": 1.1282409914848616e-05, |
| "loss": 1.1662, |
| "step": 1114 |
| }, |
| { |
| "epoch": 1.43, |
| "learning_rate": 1.1268695433415608e-05, |
| "loss": 1.2584, |
| "step": 1115 |
| }, |
| { |
| "epoch": 1.43, |
| "learning_rate": 1.1254978526264113e-05, |
| "loss": 1.2842, |
| "step": 1116 |
| }, |
| { |
| "epoch": 1.43, |
| "learning_rate": 1.1241259219620559e-05, |
| "loss": 1.2067, |
| "step": 1117 |
| }, |
| { |
| "epoch": 1.43, |
| "learning_rate": 1.1227537539715969e-05, |
| "loss": 1.2046, |
| "step": 1118 |
| }, |
| { |
| "epoch": 1.43, |
| "learning_rate": 1.1213813512785897e-05, |
| "loss": 1.2397, |
| "step": 1119 |
| }, |
| { |
| "epoch": 1.43, |
| "learning_rate": 1.1200087165070393e-05, |
| "loss": 1.2248, |
| "step": 1120 |
| }, |
| { |
| "epoch": 1.43, |
| "learning_rate": 1.1186358522813935e-05, |
| "loss": 1.1961, |
| "step": 1121 |
| }, |
| { |
| "epoch": 1.44, |
| "learning_rate": 1.1172627612265397e-05, |
| "loss": 1.2, |
| "step": 1122 |
| }, |
| { |
| "epoch": 1.44, |
| "learning_rate": 1.1158894459677983e-05, |
| "loss": 1.2396, |
| "step": 1123 |
| }, |
| { |
| "epoch": 1.44, |
| "learning_rate": 1.1145159091309189e-05, |
| "loss": 1.1822, |
| "step": 1124 |
| }, |
| { |
| "epoch": 1.44, |
| "learning_rate": 1.113142153342074e-05, |
| "loss": 1.2142, |
| "step": 1125 |
| }, |
| { |
| "epoch": 1.44, |
| "learning_rate": 1.111768181227856e-05, |
| "loss": 1.1318, |
| "step": 1126 |
| }, |
| { |
| "epoch": 1.44, |
| "learning_rate": 1.11039399541527e-05, |
| "loss": 1.2041, |
| "step": 1127 |
| }, |
| { |
| "epoch": 1.44, |
| "learning_rate": 1.1090195985317294e-05, |
| "loss": 1.2663, |
| "step": 1128 |
| }, |
| { |
| "epoch": 1.45, |
| "learning_rate": 1.1076449932050517e-05, |
| "loss": 1.266, |
| "step": 1129 |
| }, |
| { |
| "epoch": 1.45, |
| "learning_rate": 1.106270182063453e-05, |
| "loss": 1.1773, |
| "step": 1130 |
| }, |
| { |
| "epoch": 1.45, |
| "learning_rate": 1.1048951677355426e-05, |
| "loss": 1.1955, |
| "step": 1131 |
| }, |
| { |
| "epoch": 1.45, |
| "learning_rate": 1.1035199528503188e-05, |
| "loss": 1.0673, |
| "step": 1132 |
| }, |
| { |
| "epoch": 1.45, |
| "learning_rate": 1.1021445400371626e-05, |
| "loss": 1.2931, |
| "step": 1133 |
| }, |
| { |
| "epoch": 1.45, |
| "learning_rate": 1.1007689319258339e-05, |
| "loss": 1.1637, |
| "step": 1134 |
| }, |
| { |
| "epoch": 1.45, |
| "learning_rate": 1.099393131146466e-05, |
| "loss": 1.2179, |
| "step": 1135 |
| }, |
| { |
| "epoch": 1.45, |
| "learning_rate": 1.098017140329561e-05, |
| "loss": 1.2468, |
| "step": 1136 |
| }, |
| { |
| "epoch": 1.46, |
| "learning_rate": 1.0966409621059827e-05, |
| "loss": 1.1762, |
| "step": 1137 |
| }, |
| { |
| "epoch": 1.46, |
| "learning_rate": 1.0952645991069555e-05, |
| "loss": 1.2549, |
| "step": 1138 |
| }, |
| { |
| "epoch": 1.46, |
| "learning_rate": 1.0938880539640555e-05, |
| "loss": 1.2583, |
| "step": 1139 |
| }, |
| { |
| "epoch": 1.46, |
| "learning_rate": 1.0925113293092076e-05, |
| "loss": 1.1927, |
| "step": 1140 |
| }, |
| { |
| "epoch": 1.46, |
| "learning_rate": 1.0911344277746798e-05, |
| "loss": 1.101, |
| "step": 1141 |
| }, |
| { |
| "epoch": 1.46, |
| "learning_rate": 1.0897573519930785e-05, |
| "loss": 1.2897, |
| "step": 1142 |
| }, |
| { |
| "epoch": 1.46, |
| "learning_rate": 1.0883801045973425e-05, |
| "loss": 1.2225, |
| "step": 1143 |
| }, |
| { |
| "epoch": 1.46, |
| "learning_rate": 1.08700268822074e-05, |
| "loss": 1.2594, |
| "step": 1144 |
| }, |
| { |
| "epoch": 1.47, |
| "learning_rate": 1.0856251054968615e-05, |
| "loss": 1.1848, |
| "step": 1145 |
| }, |
| { |
| "epoch": 1.47, |
| "learning_rate": 1.0842473590596158e-05, |
| "loss": 1.2013, |
| "step": 1146 |
| }, |
| { |
| "epoch": 1.47, |
| "learning_rate": 1.0828694515432239e-05, |
| "loss": 1.2891, |
| "step": 1147 |
| }, |
| { |
| "epoch": 1.47, |
| "learning_rate": 1.081491385582216e-05, |
| "loss": 1.3359, |
| "step": 1148 |
| }, |
| { |
| "epoch": 1.47, |
| "learning_rate": 1.0801131638114247e-05, |
| "loss": 1.1543, |
| "step": 1149 |
| }, |
| { |
| "epoch": 1.47, |
| "learning_rate": 1.07873478886598e-05, |
| "loss": 1.2348, |
| "step": 1150 |
| }, |
| { |
| "epoch": 1.47, |
| "learning_rate": 1.0773562633813061e-05, |
| "loss": 1.2465, |
| "step": 1151 |
| }, |
| { |
| "epoch": 1.47, |
| "learning_rate": 1.0759775899931135e-05, |
| "loss": 1.128, |
| "step": 1152 |
| }, |
| { |
| "epoch": 1.48, |
| "learning_rate": 1.0745987713373962e-05, |
| "loss": 1.2658, |
| "step": 1153 |
| }, |
| { |
| "epoch": 1.48, |
| "learning_rate": 1.0732198100504266e-05, |
| "loss": 1.0994, |
| "step": 1154 |
| }, |
| { |
| "epoch": 1.48, |
| "learning_rate": 1.0718407087687486e-05, |
| "loss": 1.1552, |
| "step": 1155 |
| }, |
| { |
| "epoch": 1.48, |
| "learning_rate": 1.070461470129174e-05, |
| "loss": 1.2499, |
| "step": 1156 |
| }, |
| { |
| "epoch": 1.48, |
| "learning_rate": 1.0690820967687782e-05, |
| "loss": 1.1218, |
| "step": 1157 |
| }, |
| { |
| "epoch": 1.48, |
| "learning_rate": 1.0677025913248933e-05, |
| "loss": 1.1318, |
| "step": 1158 |
| }, |
| { |
| "epoch": 1.48, |
| "learning_rate": 1.066322956435104e-05, |
| "loss": 1.145, |
| "step": 1159 |
| }, |
| { |
| "epoch": 1.48, |
| "learning_rate": 1.0649431947372428e-05, |
| "loss": 1.2118, |
| "step": 1160 |
| }, |
| { |
| "epoch": 1.49, |
| "learning_rate": 1.0635633088693844e-05, |
| "loss": 1.1378, |
| "step": 1161 |
| }, |
| { |
| "epoch": 1.49, |
| "learning_rate": 1.0621833014698413e-05, |
| "loss": 1.1867, |
| "step": 1162 |
| }, |
| { |
| "epoch": 1.49, |
| "learning_rate": 1.0608031751771577e-05, |
| "loss": 1.2299, |
| "step": 1163 |
| }, |
| { |
| "epoch": 1.49, |
| "learning_rate": 1.0594229326301058e-05, |
| "loss": 1.3025, |
| "step": 1164 |
| }, |
| { |
| "epoch": 1.49, |
| "learning_rate": 1.0580425764676796e-05, |
| "loss": 1.1908, |
| "step": 1165 |
| }, |
| { |
| "epoch": 1.49, |
| "learning_rate": 1.0566621093290905e-05, |
| "loss": 1.2278, |
| "step": 1166 |
| }, |
| { |
| "epoch": 1.49, |
| "learning_rate": 1.0552815338537621e-05, |
| "loss": 1.1709, |
| "step": 1167 |
| }, |
| { |
| "epoch": 1.5, |
| "learning_rate": 1.0539008526813255e-05, |
| "loss": 1.1814, |
| "step": 1168 |
| }, |
| { |
| "epoch": 1.5, |
| "learning_rate": 1.0525200684516131e-05, |
| "loss": 1.1933, |
| "step": 1169 |
| }, |
| { |
| "epoch": 1.5, |
| "learning_rate": 1.0511391838046547e-05, |
| "loss": 1.2449, |
| "step": 1170 |
| }, |
| { |
| "epoch": 1.5, |
| "learning_rate": 1.0497582013806723e-05, |
| "loss": 1.1893, |
| "step": 1171 |
| }, |
| { |
| "epoch": 1.5, |
| "learning_rate": 1.0483771238200752e-05, |
| "loss": 1.1745, |
| "step": 1172 |
| }, |
| { |
| "epoch": 1.5, |
| "learning_rate": 1.0469959537634535e-05, |
| "loss": 1.1528, |
| "step": 1173 |
| }, |
| { |
| "epoch": 1.5, |
| "learning_rate": 1.045614693851575e-05, |
| "loss": 1.1539, |
| "step": 1174 |
| }, |
| { |
| "epoch": 1.5, |
| "learning_rate": 1.044233346725379e-05, |
| "loss": 1.1539, |
| "step": 1175 |
| }, |
| { |
| "epoch": 1.51, |
| "learning_rate": 1.0428519150259717e-05, |
| "loss": 1.2377, |
| "step": 1176 |
| }, |
| { |
| "epoch": 1.51, |
| "learning_rate": 1.0414704013946206e-05, |
| "loss": 1.2317, |
| "step": 1177 |
| }, |
| { |
| "epoch": 1.51, |
| "learning_rate": 1.0400888084727506e-05, |
| "loss": 1.217, |
| "step": 1178 |
| }, |
| { |
| "epoch": 1.51, |
| "learning_rate": 1.0387071389019371e-05, |
| "loss": 1.177, |
| "step": 1179 |
| }, |
| { |
| "epoch": 1.51, |
| "learning_rate": 1.037325395323904e-05, |
| "loss": 1.3033, |
| "step": 1180 |
| }, |
| { |
| "epoch": 1.51, |
| "learning_rate": 1.035943580380514e-05, |
| "loss": 1.2733, |
| "step": 1181 |
| }, |
| { |
| "epoch": 1.51, |
| "learning_rate": 1.0345616967137686e-05, |
| "loss": 1.2211, |
| "step": 1182 |
| }, |
| { |
| "epoch": 1.51, |
| "learning_rate": 1.0331797469657992e-05, |
| "loss": 1.3157, |
| "step": 1183 |
| }, |
| { |
| "epoch": 1.52, |
| "learning_rate": 1.0317977337788646e-05, |
| "loss": 1.1992, |
| "step": 1184 |
| }, |
| { |
| "epoch": 1.52, |
| "learning_rate": 1.0304156597953443e-05, |
| "loss": 1.1105, |
| "step": 1185 |
| }, |
| { |
| "epoch": 1.52, |
| "learning_rate": 1.0290335276577336e-05, |
| "loss": 1.2747, |
| "step": 1186 |
| }, |
| { |
| "epoch": 1.52, |
| "learning_rate": 1.0276513400086403e-05, |
| "loss": 1.3201, |
| "step": 1187 |
| }, |
| { |
| "epoch": 1.52, |
| "learning_rate": 1.0262690994907771e-05, |
| "loss": 1.2312, |
| "step": 1188 |
| }, |
| { |
| "epoch": 1.52, |
| "learning_rate": 1.0248868087469586e-05, |
| "loss": 1.2129, |
| "step": 1189 |
| }, |
| { |
| "epoch": 1.52, |
| "learning_rate": 1.0235044704200947e-05, |
| "loss": 1.1454, |
| "step": 1190 |
| }, |
| { |
| "epoch": 1.52, |
| "learning_rate": 1.022122087153187e-05, |
| "loss": 1.3206, |
| "step": 1191 |
| }, |
| { |
| "epoch": 1.53, |
| "learning_rate": 1.0207396615893223e-05, |
| "loss": 1.3062, |
| "step": 1192 |
| }, |
| { |
| "epoch": 1.53, |
| "learning_rate": 1.0193571963716695e-05, |
| "loss": 1.2413, |
| "step": 1193 |
| }, |
| { |
| "epoch": 1.53, |
| "learning_rate": 1.0179746941434714e-05, |
| "loss": 1.2914, |
| "step": 1194 |
| }, |
| { |
| "epoch": 1.53, |
| "learning_rate": 1.0165921575480433e-05, |
| "loss": 1.1856, |
| "step": 1195 |
| }, |
| { |
| "epoch": 1.53, |
| "learning_rate": 1.0152095892287653e-05, |
| "loss": 1.2448, |
| "step": 1196 |
| }, |
| { |
| "epoch": 1.53, |
| "learning_rate": 1.0138269918290788e-05, |
| "loss": 1.2546, |
| "step": 1197 |
| }, |
| { |
| "epoch": 1.53, |
| "learning_rate": 1.0124443679924799e-05, |
| "loss": 1.2695, |
| "step": 1198 |
| }, |
| { |
| "epoch": 1.53, |
| "learning_rate": 1.011061720362516e-05, |
| "loss": 1.2566, |
| "step": 1199 |
| }, |
| { |
| "epoch": 1.54, |
| "learning_rate": 1.0096790515827795e-05, |
| "loss": 1.2004, |
| "step": 1200 |
| }, |
| { |
| "epoch": 1.54, |
| "learning_rate": 1.0082963642969041e-05, |
| "loss": 1.2348, |
| "step": 1201 |
| }, |
| { |
| "epoch": 1.54, |
| "learning_rate": 1.0069136611485577e-05, |
| "loss": 1.2215, |
| "step": 1202 |
| }, |
| { |
| "epoch": 1.54, |
| "learning_rate": 1.0055309447814394e-05, |
| "loss": 1.1867, |
| "step": 1203 |
| }, |
| { |
| "epoch": 1.54, |
| "learning_rate": 1.004148217839273e-05, |
| "loss": 1.0689, |
| "step": 1204 |
| }, |
| { |
| "epoch": 1.54, |
| "learning_rate": 1.0027654829658031e-05, |
| "loss": 1.1713, |
| "step": 1205 |
| }, |
| { |
| "epoch": 1.54, |
| "learning_rate": 1.001382742804789e-05, |
| "loss": 1.1697, |
| "step": 1206 |
| }, |
| { |
| "epoch": 1.54, |
| "learning_rate": 1e-05, |
| "loss": 1.2225, |
| "step": 1207 |
| }, |
| { |
| "epoch": 1.55, |
| "learning_rate": 9.986172571952116e-06, |
| "loss": 1.2603, |
| "step": 1208 |
| }, |
| { |
| "epoch": 1.55, |
| "learning_rate": 9.972345170341972e-06, |
| "loss": 1.3214, |
| "step": 1209 |
| }, |
| { |
| "epoch": 1.55, |
| "learning_rate": 9.958517821607272e-06, |
| "loss": 1.1838, |
| "step": 1210 |
| }, |
| { |
| "epoch": 1.55, |
| "learning_rate": 9.944690552185607e-06, |
| "loss": 1.2249, |
| "step": 1211 |
| }, |
| { |
| "epoch": 1.55, |
| "learning_rate": 9.930863388514425e-06, |
| "loss": 1.2893, |
| "step": 1212 |
| }, |
| { |
| "epoch": 1.55, |
| "learning_rate": 9.917036357030964e-06, |
| "loss": 1.3325, |
| "step": 1213 |
| }, |
| { |
| "epoch": 1.55, |
| "learning_rate": 9.903209484172205e-06, |
| "loss": 1.21, |
| "step": 1214 |
| }, |
| { |
| "epoch": 1.56, |
| "learning_rate": 9.889382796374844e-06, |
| "loss": 1.2518, |
| "step": 1215 |
| }, |
| { |
| "epoch": 1.56, |
| "learning_rate": 9.875556320075206e-06, |
| "loss": 1.2567, |
| "step": 1216 |
| }, |
| { |
| "epoch": 1.56, |
| "learning_rate": 9.861730081709216e-06, |
| "loss": 1.1867, |
| "step": 1217 |
| }, |
| { |
| "epoch": 1.56, |
| "learning_rate": 9.84790410771235e-06, |
| "loss": 1.2231, |
| "step": 1218 |
| }, |
| { |
| "epoch": 1.56, |
| "learning_rate": 9.834078424519568e-06, |
| "loss": 1.1213, |
| "step": 1219 |
| }, |
| { |
| "epoch": 1.56, |
| "learning_rate": 9.820253058565289e-06, |
| "loss": 1.2608, |
| "step": 1220 |
| }, |
| { |
| "epoch": 1.56, |
| "learning_rate": 9.806428036283312e-06, |
| "loss": 1.1948, |
| "step": 1221 |
| }, |
| { |
| "epoch": 1.56, |
| "learning_rate": 9.792603384106777e-06, |
| "loss": 1.2356, |
| "step": 1222 |
| }, |
| { |
| "epoch": 1.57, |
| "learning_rate": 9.778779128468133e-06, |
| "loss": 1.2408, |
| "step": 1223 |
| }, |
| { |
| "epoch": 1.57, |
| "learning_rate": 9.764955295799056e-06, |
| "loss": 1.1777, |
| "step": 1224 |
| }, |
| { |
| "epoch": 1.57, |
| "learning_rate": 9.751131912530416e-06, |
| "loss": 1.3538, |
| "step": 1225 |
| }, |
| { |
| "epoch": 1.57, |
| "learning_rate": 9.737309005092232e-06, |
| "loss": 1.1906, |
| "step": 1226 |
| }, |
| { |
| "epoch": 1.57, |
| "learning_rate": 9.723486599913598e-06, |
| "loss": 1.2888, |
| "step": 1227 |
| }, |
| { |
| "epoch": 1.57, |
| "learning_rate": 9.709664723422666e-06, |
| "loss": 1.2535, |
| "step": 1228 |
| }, |
| { |
| "epoch": 1.57, |
| "learning_rate": 9.695843402046562e-06, |
| "loss": 1.2193, |
| "step": 1229 |
| }, |
| { |
| "epoch": 1.57, |
| "learning_rate": 9.682022662211356e-06, |
| "loss": 1.3029, |
| "step": 1230 |
| }, |
| { |
| "epoch": 1.58, |
| "learning_rate": 9.66820253034201e-06, |
| "loss": 1.3119, |
| "step": 1231 |
| }, |
| { |
| "epoch": 1.58, |
| "learning_rate": 9.654383032862316e-06, |
| "loss": 1.177, |
| "step": 1232 |
| }, |
| { |
| "epoch": 1.58, |
| "learning_rate": 9.640564196194862e-06, |
| "loss": 1.2104, |
| "step": 1233 |
| }, |
| { |
| "epoch": 1.58, |
| "learning_rate": 9.626746046760967e-06, |
| "loss": 1.2007, |
| "step": 1234 |
| }, |
| { |
| "epoch": 1.58, |
| "learning_rate": 9.612928610980627e-06, |
| "loss": 1.1755, |
| "step": 1235 |
| }, |
| { |
| "epoch": 1.58, |
| "learning_rate": 9.5991119152725e-06, |
| "loss": 1.1419, |
| "step": 1236 |
| }, |
| { |
| "epoch": 1.58, |
| "learning_rate": 9.5852959860538e-06, |
| "loss": 1.2028, |
| "step": 1237 |
| }, |
| { |
| "epoch": 1.58, |
| "learning_rate": 9.571480849740287e-06, |
| "loss": 1.2802, |
| "step": 1238 |
| }, |
| { |
| "epoch": 1.59, |
| "learning_rate": 9.557666532746214e-06, |
| "loss": 1.3392, |
| "step": 1239 |
| }, |
| { |
| "epoch": 1.59, |
| "learning_rate": 9.543853061484251e-06, |
| "loss": 1.2309, |
| "step": 1240 |
| }, |
| { |
| "epoch": 1.59, |
| "learning_rate": 9.530040462365467e-06, |
| "loss": 1.1705, |
| "step": 1241 |
| }, |
| { |
| "epoch": 1.59, |
| "learning_rate": 9.516228761799251e-06, |
| "loss": 1.2026, |
| "step": 1242 |
| }, |
| { |
| "epoch": 1.59, |
| "learning_rate": 9.502417986193275e-06, |
| "loss": 1.2524, |
| "step": 1243 |
| }, |
| { |
| "epoch": 1.59, |
| "learning_rate": 9.488608161953455e-06, |
| "loss": 1.1734, |
| "step": 1244 |
| }, |
| { |
| "epoch": 1.59, |
| "learning_rate": 9.474799315483874e-06, |
| "loss": 1.159, |
| "step": 1245 |
| }, |
| { |
| "epoch": 1.59, |
| "learning_rate": 9.460991473186748e-06, |
| "loss": 1.2606, |
| "step": 1246 |
| }, |
| { |
| "epoch": 1.6, |
| "learning_rate": 9.447184661462382e-06, |
| "loss": 1.2185, |
| "step": 1247 |
| }, |
| { |
| "epoch": 1.6, |
| "learning_rate": 9.433378906709097e-06, |
| "loss": 1.1551, |
| "step": 1248 |
| }, |
| { |
| "epoch": 1.6, |
| "learning_rate": 9.419574235323208e-06, |
| "loss": 1.1854, |
| "step": 1249 |
| }, |
| { |
| "epoch": 1.6, |
| "learning_rate": 9.405770673698947e-06, |
| "loss": 1.2538, |
| "step": 1250 |
| }, |
| { |
| "epoch": 1.6, |
| "learning_rate": 9.391968248228425e-06, |
| "loss": 1.1945, |
| "step": 1251 |
| }, |
| { |
| "epoch": 1.6, |
| "learning_rate": 9.37816698530159e-06, |
| "loss": 1.227, |
| "step": 1252 |
| }, |
| { |
| "epoch": 1.6, |
| "learning_rate": 9.364366911306158e-06, |
| "loss": 1.2195, |
| "step": 1253 |
| }, |
| { |
| "epoch": 1.61, |
| "learning_rate": 9.350568052627574e-06, |
| "loss": 1.1223, |
| "step": 1254 |
| }, |
| { |
| "epoch": 1.61, |
| "learning_rate": 9.336770435648963e-06, |
| "loss": 1.1254, |
| "step": 1255 |
| }, |
| { |
| "epoch": 1.61, |
| "learning_rate": 9.322974086751069e-06, |
| "loss": 1.2231, |
| "step": 1256 |
| }, |
| { |
| "epoch": 1.61, |
| "learning_rate": 9.309179032312221e-06, |
| "loss": 1.1632, |
| "step": 1257 |
| }, |
| { |
| "epoch": 1.61, |
| "learning_rate": 9.29538529870826e-06, |
| "loss": 1.2553, |
| "step": 1258 |
| }, |
| { |
| "epoch": 1.61, |
| "learning_rate": 9.281592912312519e-06, |
| "loss": 1.1913, |
| "step": 1259 |
| }, |
| { |
| "epoch": 1.61, |
| "learning_rate": 9.26780189949574e-06, |
| "loss": 1.2201, |
| "step": 1260 |
| }, |
| { |
| "epoch": 1.61, |
| "learning_rate": 9.254012286626037e-06, |
| "loss": 1.2545, |
| "step": 1261 |
| }, |
| { |
| "epoch": 1.62, |
| "learning_rate": 9.240224100068868e-06, |
| "loss": 1.1376, |
| "step": 1262 |
| }, |
| { |
| "epoch": 1.62, |
| "learning_rate": 9.226437366186942e-06, |
| "loss": 1.2654, |
| "step": 1263 |
| }, |
| { |
| "epoch": 1.62, |
| "learning_rate": 9.2126521113402e-06, |
| "loss": 1.2355, |
| "step": 1264 |
| }, |
| { |
| "epoch": 1.62, |
| "learning_rate": 9.198868361885758e-06, |
| "loss": 1.2238, |
| "step": 1265 |
| }, |
| { |
| "epoch": 1.62, |
| "learning_rate": 9.185086144177843e-06, |
| "loss": 1.3105, |
| "step": 1266 |
| }, |
| { |
| "epoch": 1.62, |
| "learning_rate": 9.171305484567763e-06, |
| "loss": 1.2534, |
| "step": 1267 |
| }, |
| { |
| "epoch": 1.62, |
| "learning_rate": 9.157526409403845e-06, |
| "loss": 1.2002, |
| "step": 1268 |
| }, |
| { |
| "epoch": 1.62, |
| "learning_rate": 9.143748945031386e-06, |
| "loss": 1.1885, |
| "step": 1269 |
| }, |
| { |
| "epoch": 1.63, |
| "learning_rate": 9.129973117792601e-06, |
| "loss": 1.1954, |
| "step": 1270 |
| }, |
| { |
| "epoch": 1.63, |
| "learning_rate": 9.116198954026577e-06, |
| "loss": 1.1995, |
| "step": 1271 |
| }, |
| { |
| "epoch": 1.63, |
| "learning_rate": 9.10242648006922e-06, |
| "loss": 1.103, |
| "step": 1272 |
| }, |
| { |
| "epoch": 1.63, |
| "learning_rate": 9.088655722253205e-06, |
| "loss": 1.3276, |
| "step": 1273 |
| }, |
| { |
| "epoch": 1.63, |
| "learning_rate": 9.074886706907928e-06, |
| "loss": 1.1884, |
| "step": 1274 |
| }, |
| { |
| "epoch": 1.63, |
| "learning_rate": 9.061119460359448e-06, |
| "loss": 1.1646, |
| "step": 1275 |
| }, |
| { |
| "epoch": 1.63, |
| "learning_rate": 9.047354008930448e-06, |
| "loss": 1.2076, |
| "step": 1276 |
| }, |
| { |
| "epoch": 1.63, |
| "learning_rate": 9.033590378940174e-06, |
| "loss": 1.3016, |
| "step": 1277 |
| }, |
| { |
| "epoch": 1.64, |
| "learning_rate": 9.019828596704394e-06, |
| "loss": 1.2364, |
| "step": 1278 |
| }, |
| { |
| "epoch": 1.64, |
| "learning_rate": 9.006068688535342e-06, |
| "loss": 1.1564, |
| "step": 1279 |
| }, |
| { |
| "epoch": 1.64, |
| "learning_rate": 8.992310680741663e-06, |
| "loss": 1.2053, |
| "step": 1280 |
| }, |
| { |
| "epoch": 1.64, |
| "learning_rate": 8.978554599628377e-06, |
| "loss": 1.2804, |
| "step": 1281 |
| }, |
| { |
| "epoch": 1.64, |
| "learning_rate": 8.964800471496815e-06, |
| "loss": 1.1789, |
| "step": 1282 |
| }, |
| { |
| "epoch": 1.64, |
| "learning_rate": 8.951048322644576e-06, |
| "loss": 1.2137, |
| "step": 1283 |
| }, |
| { |
| "epoch": 1.64, |
| "learning_rate": 8.937298179365474e-06, |
| "loss": 1.2667, |
| "step": 1284 |
| }, |
| { |
| "epoch": 1.64, |
| "learning_rate": 8.923550067949486e-06, |
| "loss": 1.2012, |
| "step": 1285 |
| }, |
| { |
| "epoch": 1.65, |
| "learning_rate": 8.90980401468271e-06, |
| "loss": 1.1657, |
| "step": 1286 |
| }, |
| { |
| "epoch": 1.65, |
| "learning_rate": 8.896060045847305e-06, |
| "loss": 1.2268, |
| "step": 1287 |
| }, |
| { |
| "epoch": 1.65, |
| "learning_rate": 8.882318187721441e-06, |
| "loss": 1.1706, |
| "step": 1288 |
| }, |
| { |
| "epoch": 1.65, |
| "learning_rate": 8.868578466579261e-06, |
| "loss": 1.286, |
| "step": 1289 |
| }, |
| { |
| "epoch": 1.65, |
| "learning_rate": 8.854840908690815e-06, |
| "loss": 1.1666, |
| "step": 1290 |
| }, |
| { |
| "epoch": 1.65, |
| "learning_rate": 8.841105540322018e-06, |
| "loss": 1.1818, |
| "step": 1291 |
| }, |
| { |
| "epoch": 1.65, |
| "learning_rate": 8.827372387734607e-06, |
| "loss": 1.1979, |
| "step": 1292 |
| }, |
| { |
| "epoch": 1.66, |
| "learning_rate": 8.813641477186069e-06, |
| "loss": 1.3155, |
| "step": 1293 |
| }, |
| { |
| "epoch": 1.66, |
| "learning_rate": 8.799912834929612e-06, |
| "loss": 1.2488, |
| "step": 1294 |
| }, |
| { |
| "epoch": 1.66, |
| "learning_rate": 8.786186487214108e-06, |
| "loss": 1.2747, |
| "step": 1295 |
| }, |
| { |
| "epoch": 1.66, |
| "learning_rate": 8.772462460284033e-06, |
| "loss": 1.1985, |
| "step": 1296 |
| }, |
| { |
| "epoch": 1.66, |
| "learning_rate": 8.758740780379443e-06, |
| "loss": 1.1771, |
| "step": 1297 |
| }, |
| { |
| "epoch": 1.66, |
| "learning_rate": 8.745021473735889e-06, |
| "loss": 1.254, |
| "step": 1298 |
| }, |
| { |
| "epoch": 1.66, |
| "learning_rate": 8.731304566584394e-06, |
| "loss": 1.2877, |
| "step": 1299 |
| }, |
| { |
| "epoch": 1.66, |
| "learning_rate": 8.71759008515139e-06, |
| "loss": 1.2556, |
| "step": 1300 |
| }, |
| { |
| "epoch": 1.67, |
| "learning_rate": 8.703878055658668e-06, |
| "loss": 1.2291, |
| "step": 1301 |
| }, |
| { |
| "epoch": 1.67, |
| "learning_rate": 8.690168504323343e-06, |
| "loss": 1.2848, |
| "step": 1302 |
| }, |
| { |
| "epoch": 1.67, |
| "learning_rate": 8.676461457357777e-06, |
| "loss": 1.2288, |
| "step": 1303 |
| }, |
| { |
| "epoch": 1.67, |
| "learning_rate": 8.66275694096955e-06, |
| "loss": 1.2396, |
| "step": 1304 |
| }, |
| { |
| "epoch": 1.67, |
| "learning_rate": 8.64905498136141e-06, |
| "loss": 1.2925, |
| "step": 1305 |
| }, |
| { |
| "epoch": 1.67, |
| "learning_rate": 8.6353556047312e-06, |
| "loss": 1.137, |
| "step": 1306 |
| }, |
| { |
| "epoch": 1.67, |
| "learning_rate": 8.621658837271844e-06, |
| "loss": 1.1228, |
| "step": 1307 |
| }, |
| { |
| "epoch": 1.67, |
| "learning_rate": 8.607964705171267e-06, |
| "loss": 1.2767, |
| "step": 1308 |
| }, |
| { |
| "epoch": 1.68, |
| "learning_rate": 8.594273234612347e-06, |
| "loss": 1.089, |
| "step": 1309 |
| }, |
| { |
| "epoch": 1.68, |
| "learning_rate": 8.580584451772895e-06, |
| "loss": 1.2522, |
| "step": 1310 |
| }, |
| { |
| "epoch": 1.68, |
| "learning_rate": 8.566898382825558e-06, |
| "loss": 1.2113, |
| "step": 1311 |
| }, |
| { |
| "epoch": 1.68, |
| "learning_rate": 8.55321505393782e-06, |
| "loss": 1.1131, |
| "step": 1312 |
| }, |
| { |
| "epoch": 1.68, |
| "learning_rate": 8.539534491271906e-06, |
| "loss": 1.2038, |
| "step": 1313 |
| }, |
| { |
| "epoch": 1.68, |
| "learning_rate": 8.525856720984755e-06, |
| "loss": 1.151, |
| "step": 1314 |
| }, |
| { |
| "epoch": 1.68, |
| "learning_rate": 8.512181769227976e-06, |
| "loss": 1.1602, |
| "step": 1315 |
| }, |
| { |
| "epoch": 1.68, |
| "learning_rate": 8.498509662147784e-06, |
| "loss": 1.1346, |
| "step": 1316 |
| }, |
| { |
| "epoch": 1.69, |
| "learning_rate": 8.484840425884949e-06, |
| "loss": 1.3731, |
| "step": 1317 |
| }, |
| { |
| "epoch": 1.69, |
| "learning_rate": 8.471174086574767e-06, |
| "loss": 1.2035, |
| "step": 1318 |
| }, |
| { |
| "epoch": 1.69, |
| "learning_rate": 8.457510670346976e-06, |
| "loss": 1.2039, |
| "step": 1319 |
| }, |
| { |
| "epoch": 1.69, |
| "learning_rate": 8.443850203325745e-06, |
| "loss": 1.0988, |
| "step": 1320 |
| }, |
| { |
| "epoch": 1.69, |
| "learning_rate": 8.430192711629588e-06, |
| "loss": 1.3004, |
| "step": 1321 |
| }, |
| { |
| "epoch": 1.69, |
| "learning_rate": 8.416538221371336e-06, |
| "loss": 1.2406, |
| "step": 1322 |
| }, |
| { |
| "epoch": 1.69, |
| "learning_rate": 8.402886758658087e-06, |
| "loss": 1.1473, |
| "step": 1323 |
| }, |
| { |
| "epoch": 1.69, |
| "learning_rate": 8.389238349591143e-06, |
| "loss": 1.2939, |
| "step": 1324 |
| }, |
| { |
| "epoch": 1.7, |
| "learning_rate": 8.375593020265964e-06, |
| "loss": 1.1418, |
| "step": 1325 |
| }, |
| { |
| "epoch": 1.7, |
| "learning_rate": 8.361950796772139e-06, |
| "loss": 1.1144, |
| "step": 1326 |
| }, |
| { |
| "epoch": 1.7, |
| "learning_rate": 8.348311705193294e-06, |
| "loss": 1.2204, |
| "step": 1327 |
| }, |
| { |
| "epoch": 1.7, |
| "learning_rate": 8.334675771607091e-06, |
| "loss": 1.2422, |
| "step": 1328 |
| }, |
| { |
| "epoch": 1.7, |
| "learning_rate": 8.321043022085139e-06, |
| "loss": 1.2705, |
| "step": 1329 |
| }, |
| { |
| "epoch": 1.7, |
| "learning_rate": 8.307413482692955e-06, |
| "loss": 1.2326, |
| "step": 1330 |
| }, |
| { |
| "epoch": 1.7, |
| "learning_rate": 8.29378717948994e-06, |
| "loss": 1.1773, |
| "step": 1331 |
| }, |
| { |
| "epoch": 1.7, |
| "learning_rate": 8.280164138529285e-06, |
| "loss": 1.201, |
| "step": 1332 |
| }, |
| { |
| "epoch": 1.71, |
| "learning_rate": 8.266544385857958e-06, |
| "loss": 1.1861, |
| "step": 1333 |
| }, |
| { |
| "epoch": 1.71, |
| "learning_rate": 8.252927947516631e-06, |
| "loss": 1.0718, |
| "step": 1334 |
| }, |
| { |
| "epoch": 1.71, |
| "learning_rate": 8.239314849539639e-06, |
| "loss": 1.3154, |
| "step": 1335 |
| }, |
| { |
| "epoch": 1.71, |
| "learning_rate": 8.22570511795494e-06, |
| "loss": 1.2147, |
| "step": 1336 |
| }, |
| { |
| "epoch": 1.71, |
| "learning_rate": 8.212098778784049e-06, |
| "loss": 1.2228, |
| "step": 1337 |
| }, |
| { |
| "epoch": 1.71, |
| "learning_rate": 8.198495858041988e-06, |
| "loss": 1.2397, |
| "step": 1338 |
| }, |
| { |
| "epoch": 1.71, |
| "learning_rate": 8.184896381737257e-06, |
| "loss": 1.1285, |
| "step": 1339 |
| }, |
| { |
| "epoch": 1.72, |
| "learning_rate": 8.171300375871759e-06, |
| "loss": 1.0844, |
| "step": 1340 |
| }, |
| { |
| "epoch": 1.72, |
| "learning_rate": 8.157707866440772e-06, |
| "loss": 1.169, |
| "step": 1341 |
| }, |
| { |
| "epoch": 1.72, |
| "learning_rate": 8.144118879432879e-06, |
| "loss": 1.1869, |
| "step": 1342 |
| }, |
| { |
| "epoch": 1.72, |
| "learning_rate": 8.130533440829927e-06, |
| "loss": 1.3084, |
| "step": 1343 |
| }, |
| { |
| "epoch": 1.72, |
| "learning_rate": 8.116951576606994e-06, |
| "loss": 1.2427, |
| "step": 1344 |
| }, |
| { |
| "epoch": 1.72, |
| "learning_rate": 8.103373312732308e-06, |
| "loss": 1.2171, |
| "step": 1345 |
| }, |
| { |
| "epoch": 1.72, |
| "learning_rate": 8.089798675167214e-06, |
| "loss": 1.2308, |
| "step": 1346 |
| }, |
| { |
| "epoch": 1.72, |
| "learning_rate": 8.076227689866138e-06, |
| "loss": 1.3029, |
| "step": 1347 |
| }, |
| { |
| "epoch": 1.73, |
| "learning_rate": 8.062660382776503e-06, |
| "loss": 1.1341, |
| "step": 1348 |
| }, |
| { |
| "epoch": 1.73, |
| "learning_rate": 8.04909677983872e-06, |
| "loss": 1.1793, |
| "step": 1349 |
| }, |
| { |
| "epoch": 1.73, |
| "learning_rate": 8.0355369069861e-06, |
| "loss": 1.2853, |
| "step": 1350 |
| }, |
| { |
| "epoch": 1.73, |
| "learning_rate": 8.021980790144828e-06, |
| "loss": 1.2242, |
| "step": 1351 |
| }, |
| { |
| "epoch": 1.73, |
| "learning_rate": 8.008428455233916e-06, |
| "loss": 1.2111, |
| "step": 1352 |
| }, |
| { |
| "epoch": 1.73, |
| "learning_rate": 7.994879928165132e-06, |
| "loss": 1.2698, |
| "step": 1353 |
| }, |
| { |
| "epoch": 1.73, |
| "learning_rate": 7.981335234842975e-06, |
| "loss": 1.2268, |
| "step": 1354 |
| }, |
| { |
| "epoch": 1.73, |
| "learning_rate": 7.967794401164604e-06, |
| "loss": 1.2708, |
| "step": 1355 |
| }, |
| { |
| "epoch": 1.74, |
| "learning_rate": 7.954257453019802e-06, |
| "loss": 1.2289, |
| "step": 1356 |
| }, |
| { |
| "epoch": 1.74, |
| "learning_rate": 7.940724416290931e-06, |
| "loss": 1.2533, |
| "step": 1357 |
| }, |
| { |
| "epoch": 1.74, |
| "learning_rate": 7.927195316852861e-06, |
| "loss": 1.184, |
| "step": 1358 |
| }, |
| { |
| "epoch": 1.74, |
| "learning_rate": 7.913670180572936e-06, |
| "loss": 1.2034, |
| "step": 1359 |
| }, |
| { |
| "epoch": 1.74, |
| "learning_rate": 7.90014903331094e-06, |
| "loss": 1.1699, |
| "step": 1360 |
| }, |
| { |
| "epoch": 1.74, |
| "learning_rate": 7.886631900919005e-06, |
| "loss": 1.1744, |
| "step": 1361 |
| }, |
| { |
| "epoch": 1.74, |
| "learning_rate": 7.873118809241605e-06, |
| "loss": 1.2295, |
| "step": 1362 |
| }, |
| { |
| "epoch": 1.74, |
| "learning_rate": 7.85960978411548e-06, |
| "loss": 1.1816, |
| "step": 1363 |
| }, |
| { |
| "epoch": 1.75, |
| "learning_rate": 7.846104851369595e-06, |
| "loss": 1.196, |
| "step": 1364 |
| }, |
| { |
| "epoch": 1.75, |
| "learning_rate": 7.832604036825096e-06, |
| "loss": 1.2683, |
| "step": 1365 |
| }, |
| { |
| "epoch": 1.75, |
| "learning_rate": 7.81910736629525e-06, |
| "loss": 1.2391, |
| "step": 1366 |
| }, |
| { |
| "epoch": 1.75, |
| "learning_rate": 7.805614865585397e-06, |
| "loss": 1.2124, |
| "step": 1367 |
| }, |
| { |
| "epoch": 1.75, |
| "learning_rate": 7.792126560492915e-06, |
| "loss": 1.2382, |
| "step": 1368 |
| }, |
| { |
| "epoch": 1.75, |
| "learning_rate": 7.778642476807149e-06, |
| "loss": 1.1709, |
| "step": 1369 |
| }, |
| { |
| "epoch": 1.75, |
| "learning_rate": 7.765162640309387e-06, |
| "loss": 1.3215, |
| "step": 1370 |
| }, |
| { |
| "epoch": 1.75, |
| "learning_rate": 7.751687076772781e-06, |
| "loss": 1.1953, |
| "step": 1371 |
| }, |
| { |
| "epoch": 1.76, |
| "learning_rate": 7.738215811962318e-06, |
| "loss": 1.1826, |
| "step": 1372 |
| }, |
| { |
| "epoch": 1.76, |
| "learning_rate": 7.724748871634776e-06, |
| "loss": 1.1949, |
| "step": 1373 |
| }, |
| { |
| "epoch": 1.76, |
| "learning_rate": 7.711286281538644e-06, |
| "loss": 1.2285, |
| "step": 1374 |
| }, |
| { |
| "epoch": 1.76, |
| "learning_rate": 7.69782806741412e-06, |
| "loss": 1.1769, |
| "step": 1375 |
| }, |
| { |
| "epoch": 1.76, |
| "learning_rate": 7.684374254993012e-06, |
| "loss": 1.171, |
| "step": 1376 |
| }, |
| { |
| "epoch": 1.76, |
| "learning_rate": 7.670924869998721e-06, |
| "loss": 1.298, |
| "step": 1377 |
| }, |
| { |
| "epoch": 1.76, |
| "learning_rate": 7.65747993814619e-06, |
| "loss": 1.1573, |
| "step": 1378 |
| }, |
| { |
| "epoch": 1.77, |
| "learning_rate": 7.644039485141834e-06, |
| "loss": 1.203, |
| "step": 1379 |
| }, |
| { |
| "epoch": 1.77, |
| "learning_rate": 7.63060353668351e-06, |
| "loss": 1.2373, |
| "step": 1380 |
| }, |
| { |
| "epoch": 1.77, |
| "learning_rate": 7.617172118460472e-06, |
| "loss": 1.1766, |
| "step": 1381 |
| }, |
| { |
| "epoch": 1.77, |
| "learning_rate": 7.603745256153293e-06, |
| "loss": 1.1084, |
| "step": 1382 |
| }, |
| { |
| "epoch": 1.77, |
| "learning_rate": 7.590322975433857e-06, |
| "loss": 1.1462, |
| "step": 1383 |
| }, |
| { |
| "epoch": 1.77, |
| "learning_rate": 7.576905301965273e-06, |
| "loss": 1.1686, |
| "step": 1384 |
| }, |
| { |
| "epoch": 1.77, |
| "learning_rate": 7.5634922614018395e-06, |
| "loss": 1.2987, |
| "step": 1385 |
| }, |
| { |
| "epoch": 1.77, |
| "learning_rate": 7.550083879389012e-06, |
| "loss": 1.1828, |
| "step": 1386 |
| }, |
| { |
| "epoch": 1.78, |
| "learning_rate": 7.536680181563324e-06, |
| "loss": 1.2234, |
| "step": 1387 |
| }, |
| { |
| "epoch": 1.78, |
| "learning_rate": 7.5232811935523564e-06, |
| "loss": 1.2498, |
| "step": 1388 |
| }, |
| { |
| "epoch": 1.78, |
| "learning_rate": 7.5098869409746956e-06, |
| "loss": 1.1995, |
| "step": 1389 |
| }, |
| { |
| "epoch": 1.78, |
| "learning_rate": 7.496497449439857e-06, |
| "loss": 1.2482, |
| "step": 1390 |
| }, |
| { |
| "epoch": 1.78, |
| "learning_rate": 7.4831127445482675e-06, |
| "loss": 1.1674, |
| "step": 1391 |
| }, |
| { |
| "epoch": 1.78, |
| "learning_rate": 7.469732851891196e-06, |
| "loss": 1.3185, |
| "step": 1392 |
| }, |
| { |
| "epoch": 1.78, |
| "learning_rate": 7.456357797050702e-06, |
| "loss": 1.2152, |
| "step": 1393 |
| }, |
| { |
| "epoch": 1.78, |
| "learning_rate": 7.442987605599617e-06, |
| "loss": 1.2486, |
| "step": 1394 |
| }, |
| { |
| "epoch": 1.79, |
| "learning_rate": 7.429622303101452e-06, |
| "loss": 1.1224, |
| "step": 1395 |
| }, |
| { |
| "epoch": 1.79, |
| "learning_rate": 7.416261915110379e-06, |
| "loss": 1.1414, |
| "step": 1396 |
| }, |
| { |
| "epoch": 1.79, |
| "learning_rate": 7.402906467171178e-06, |
| "loss": 1.2516, |
| "step": 1397 |
| }, |
| { |
| "epoch": 1.79, |
| "learning_rate": 7.389555984819175e-06, |
| "loss": 1.2951, |
| "step": 1398 |
| }, |
| { |
| "epoch": 1.79, |
| "learning_rate": 7.376210493580211e-06, |
| "loss": 1.1454, |
| "step": 1399 |
| }, |
| { |
| "epoch": 1.79, |
| "learning_rate": 7.362870018970576e-06, |
| "loss": 1.2325, |
| "step": 1400 |
| }, |
| { |
| "epoch": 1.79, |
| "learning_rate": 7.349534586496972e-06, |
| "loss": 1.1991, |
| "step": 1401 |
| }, |
| { |
| "epoch": 1.79, |
| "learning_rate": 7.336204221656461e-06, |
| "loss": 1.2179, |
| "step": 1402 |
| }, |
| { |
| "epoch": 1.8, |
| "learning_rate": 7.3228789499364114e-06, |
| "loss": 1.3545, |
| "step": 1403 |
| }, |
| { |
| "epoch": 1.8, |
| "learning_rate": 7.309558796814466e-06, |
| "loss": 1.1229, |
| "step": 1404 |
| }, |
| { |
| "epoch": 1.8, |
| "learning_rate": 7.296243787758463e-06, |
| "loss": 1.1341, |
| "step": 1405 |
| }, |
| { |
| "epoch": 1.8, |
| "learning_rate": 7.282933948226418e-06, |
| "loss": 1.2491, |
| "step": 1406 |
| }, |
| { |
| "epoch": 1.8, |
| "learning_rate": 7.269629303666463e-06, |
| "loss": 1.2477, |
| "step": 1407 |
| }, |
| { |
| "epoch": 1.8, |
| "learning_rate": 7.256329879516788e-06, |
| "loss": 1.2735, |
| "step": 1408 |
| }, |
| { |
| "epoch": 1.8, |
| "learning_rate": 7.243035701205608e-06, |
| "loss": 1.1586, |
| "step": 1409 |
| }, |
| { |
| "epoch": 1.8, |
| "learning_rate": 7.229746794151107e-06, |
| "loss": 1.2793, |
| "step": 1410 |
| }, |
| { |
| "epoch": 1.81, |
| "learning_rate": 7.216463183761392e-06, |
| "loss": 1.1497, |
| "step": 1411 |
| }, |
| { |
| "epoch": 1.81, |
| "learning_rate": 7.203184895434443e-06, |
| "loss": 1.1339, |
| "step": 1412 |
| }, |
| { |
| "epoch": 1.81, |
| "learning_rate": 7.1899119545580595e-06, |
| "loss": 1.17, |
| "step": 1413 |
| }, |
| { |
| "epoch": 1.81, |
| "learning_rate": 7.176644386509821e-06, |
| "loss": 1.2709, |
| "step": 1414 |
| }, |
| { |
| "epoch": 1.81, |
| "learning_rate": 7.163382216657033e-06, |
| "loss": 1.1636, |
| "step": 1415 |
| }, |
| { |
| "epoch": 1.81, |
| "learning_rate": 7.1501254703566845e-06, |
| "loss": 1.2577, |
| "step": 1416 |
| }, |
| { |
| "epoch": 1.81, |
| "learning_rate": 7.136874172955385e-06, |
| "loss": 1.008, |
| "step": 1417 |
| }, |
| { |
| "epoch": 1.82, |
| "learning_rate": 7.123628349789332e-06, |
| "loss": 1.2189, |
| "step": 1418 |
| }, |
| { |
| "epoch": 1.82, |
| "learning_rate": 7.110388026184259e-06, |
| "loss": 1.1411, |
| "step": 1419 |
| }, |
| { |
| "epoch": 1.82, |
| "learning_rate": 7.097153227455379e-06, |
| "loss": 1.1516, |
| "step": 1420 |
| }, |
| { |
| "epoch": 1.82, |
| "learning_rate": 7.083923978907341e-06, |
| "loss": 1.2314, |
| "step": 1421 |
| }, |
| { |
| "epoch": 1.82, |
| "learning_rate": 7.070700305834189e-06, |
| "loss": 1.204, |
| "step": 1422 |
| }, |
| { |
| "epoch": 1.82, |
| "learning_rate": 7.0574822335193015e-06, |
| "loss": 1.2731, |
| "step": 1423 |
| }, |
| { |
| "epoch": 1.82, |
| "learning_rate": 7.044269787235348e-06, |
| "loss": 1.3234, |
| "step": 1424 |
| }, |
| { |
| "epoch": 1.82, |
| "learning_rate": 7.031062992244246e-06, |
| "loss": 1.1738, |
| "step": 1425 |
| }, |
| { |
| "epoch": 1.83, |
| "learning_rate": 7.017861873797102e-06, |
| "loss": 1.1253, |
| "step": 1426 |
| }, |
| { |
| "epoch": 1.83, |
| "learning_rate": 7.0046664571341715e-06, |
| "loss": 1.1937, |
| "step": 1427 |
| }, |
| { |
| "epoch": 1.83, |
| "learning_rate": 6.991476767484809e-06, |
| "loss": 1.1622, |
| "step": 1428 |
| }, |
| { |
| "epoch": 1.83, |
| "learning_rate": 6.9782928300674215e-06, |
| "loss": 1.1891, |
| "step": 1429 |
| }, |
| { |
| "epoch": 1.83, |
| "learning_rate": 6.965114670089409e-06, |
| "loss": 1.0855, |
| "step": 1430 |
| }, |
| { |
| "epoch": 1.83, |
| "learning_rate": 6.951942312747135e-06, |
| "loss": 1.0969, |
| "step": 1431 |
| }, |
| { |
| "epoch": 1.83, |
| "learning_rate": 6.938775783225864e-06, |
| "loss": 1.2792, |
| "step": 1432 |
| }, |
| { |
| "epoch": 1.83, |
| "learning_rate": 6.925615106699718e-06, |
| "loss": 1.1715, |
| "step": 1433 |
| }, |
| { |
| "epoch": 1.84, |
| "learning_rate": 6.912460308331631e-06, |
| "loss": 1.2868, |
| "step": 1434 |
| }, |
| { |
| "epoch": 1.84, |
| "learning_rate": 6.899311413273292e-06, |
| "loss": 1.2059, |
| "step": 1435 |
| }, |
| { |
| "epoch": 1.84, |
| "learning_rate": 6.886168446665107e-06, |
| "loss": 1.1905, |
| "step": 1436 |
| }, |
| { |
| "epoch": 1.84, |
| "learning_rate": 6.873031433636151e-06, |
| "loss": 1.2055, |
| "step": 1437 |
| }, |
| { |
| "epoch": 1.84, |
| "learning_rate": 6.859900399304104e-06, |
| "loss": 1.2846, |
| "step": 1438 |
| }, |
| { |
| "epoch": 1.84, |
| "learning_rate": 6.846775368775231e-06, |
| "loss": 1.159, |
| "step": 1439 |
| }, |
| { |
| "epoch": 1.84, |
| "learning_rate": 6.8336563671443015e-06, |
| "loss": 1.2263, |
| "step": 1440 |
| }, |
| { |
| "epoch": 1.84, |
| "learning_rate": 6.82054341949457e-06, |
| "loss": 1.2224, |
| "step": 1441 |
| }, |
| { |
| "epoch": 1.85, |
| "learning_rate": 6.807436550897713e-06, |
| "loss": 1.2751, |
| "step": 1442 |
| }, |
| { |
| "epoch": 1.85, |
| "learning_rate": 6.794335786413775e-06, |
| "loss": 1.2331, |
| "step": 1443 |
| }, |
| { |
| "epoch": 1.85, |
| "learning_rate": 6.781241151091146e-06, |
| "loss": 1.2263, |
| "step": 1444 |
| }, |
| { |
| "epoch": 1.85, |
| "learning_rate": 6.768152669966482e-06, |
| "loss": 1.1878, |
| "step": 1445 |
| }, |
| { |
| "epoch": 1.85, |
| "learning_rate": 6.755070368064682e-06, |
| "loss": 1.2407, |
| "step": 1446 |
| }, |
| { |
| "epoch": 1.85, |
| "learning_rate": 6.741994270398826e-06, |
| "loss": 1.2185, |
| "step": 1447 |
| }, |
| { |
| "epoch": 1.85, |
| "learning_rate": 6.72892440197013e-06, |
| "loss": 1.2299, |
| "step": 1448 |
| }, |
| { |
| "epoch": 1.85, |
| "learning_rate": 6.715860787767908e-06, |
| "loss": 1.2441, |
| "step": 1449 |
| }, |
| { |
| "epoch": 1.86, |
| "learning_rate": 6.702803452769507e-06, |
| "loss": 1.167, |
| "step": 1450 |
| }, |
| { |
| "epoch": 1.86, |
| "learning_rate": 6.68975242194027e-06, |
| "loss": 1.2689, |
| "step": 1451 |
| }, |
| { |
| "epoch": 1.86, |
| "learning_rate": 6.676707720233493e-06, |
| "loss": 1.2297, |
| "step": 1452 |
| }, |
| { |
| "epoch": 1.86, |
| "learning_rate": 6.663669372590359e-06, |
| "loss": 1.2276, |
| "step": 1453 |
| }, |
| { |
| "epoch": 1.86, |
| "learning_rate": 6.650637403939919e-06, |
| "loss": 1.0347, |
| "step": 1454 |
| }, |
| { |
| "epoch": 1.86, |
| "learning_rate": 6.637611839199013e-06, |
| "loss": 1.245, |
| "step": 1455 |
| }, |
| { |
| "epoch": 1.86, |
| "learning_rate": 6.624592703272237e-06, |
| "loss": 1.1784, |
| "step": 1456 |
| }, |
| { |
| "epoch": 1.86, |
| "learning_rate": 6.61158002105191e-06, |
| "loss": 1.2064, |
| "step": 1457 |
| }, |
| { |
| "epoch": 1.87, |
| "learning_rate": 6.598573817417995e-06, |
| "loss": 1.2198, |
| "step": 1458 |
| }, |
| { |
| "epoch": 1.87, |
| "learning_rate": 6.585574117238074e-06, |
| "loss": 1.1274, |
| "step": 1459 |
| }, |
| { |
| "epoch": 1.87, |
| "learning_rate": 6.572580945367299e-06, |
| "loss": 1.2477, |
| "step": 1460 |
| }, |
| { |
| "epoch": 1.87, |
| "learning_rate": 6.559594326648333e-06, |
| "loss": 1.2273, |
| "step": 1461 |
| }, |
| { |
| "epoch": 1.87, |
| "learning_rate": 6.546614285911317e-06, |
| "loss": 1.2612, |
| "step": 1462 |
| }, |
| { |
| "epoch": 1.87, |
| "learning_rate": 6.533640847973809e-06, |
| "loss": 1.2091, |
| "step": 1463 |
| }, |
| { |
| "epoch": 1.87, |
| "learning_rate": 6.52067403764074e-06, |
| "loss": 1.2417, |
| "step": 1464 |
| }, |
| { |
| "epoch": 1.88, |
| "learning_rate": 6.50771387970438e-06, |
| "loss": 1.2327, |
| "step": 1465 |
| }, |
| { |
| "epoch": 1.88, |
| "learning_rate": 6.494760398944271e-06, |
| "loss": 1.2424, |
| "step": 1466 |
| }, |
| { |
| "epoch": 1.88, |
| "learning_rate": 6.4818136201271864e-06, |
| "loss": 1.2398, |
| "step": 1467 |
| }, |
| { |
| "epoch": 1.88, |
| "learning_rate": 6.468873568007099e-06, |
| "loss": 1.2524, |
| "step": 1468 |
| }, |
| { |
| "epoch": 1.88, |
| "learning_rate": 6.4559402673251005e-06, |
| "loss": 1.132, |
| "step": 1469 |
| }, |
| { |
| "epoch": 1.88, |
| "learning_rate": 6.443013742809397e-06, |
| "loss": 1.1357, |
| "step": 1470 |
| }, |
| { |
| "epoch": 1.88, |
| "learning_rate": 6.430094019175217e-06, |
| "loss": 1.1602, |
| "step": 1471 |
| }, |
| { |
| "epoch": 1.88, |
| "learning_rate": 6.417181121124796e-06, |
| "loss": 1.248, |
| "step": 1472 |
| }, |
| { |
| "epoch": 1.89, |
| "learning_rate": 6.404275073347324e-06, |
| "loss": 1.1438, |
| "step": 1473 |
| }, |
| { |
| "epoch": 1.89, |
| "learning_rate": 6.391375900518879e-06, |
| "loss": 1.1996, |
| "step": 1474 |
| }, |
| { |
| "epoch": 1.89, |
| "learning_rate": 6.378483627302415e-06, |
| "loss": 1.1685, |
| "step": 1475 |
| }, |
| { |
| "epoch": 1.89, |
| "learning_rate": 6.365598278347675e-06, |
| "loss": 1.274, |
| "step": 1476 |
| }, |
| { |
| "epoch": 1.89, |
| "learning_rate": 6.352719878291167e-06, |
| "loss": 1.3252, |
| "step": 1477 |
| }, |
| { |
| "epoch": 1.89, |
| "learning_rate": 6.3398484517561255e-06, |
| "loss": 1.1939, |
| "step": 1478 |
| }, |
| { |
| "epoch": 1.89, |
| "learning_rate": 6.326984023352435e-06, |
| "loss": 1.2356, |
| "step": 1479 |
| }, |
| { |
| "epoch": 1.89, |
| "learning_rate": 6.314126617676606e-06, |
| "loss": 1.1645, |
| "step": 1480 |
| }, |
| { |
| "epoch": 1.9, |
| "learning_rate": 6.301276259311728e-06, |
| "loss": 1.2177, |
| "step": 1481 |
| }, |
| { |
| "epoch": 1.9, |
| "learning_rate": 6.288432972827406e-06, |
| "loss": 1.1975, |
| "step": 1482 |
| }, |
| { |
| "epoch": 1.9, |
| "learning_rate": 6.2755967827797346e-06, |
| "loss": 1.2273, |
| "step": 1483 |
| }, |
| { |
| "epoch": 1.9, |
| "learning_rate": 6.262767713711231e-06, |
| "loss": 1.1871, |
| "step": 1484 |
| }, |
| { |
| "epoch": 1.9, |
| "learning_rate": 6.249945790150798e-06, |
| "loss": 1.2117, |
| "step": 1485 |
| }, |
| { |
| "epoch": 1.9, |
| "learning_rate": 6.237131036613682e-06, |
| "loss": 1.1942, |
| "step": 1486 |
| }, |
| { |
| "epoch": 1.9, |
| "learning_rate": 6.224323477601417e-06, |
| "loss": 1.1107, |
| "step": 1487 |
| }, |
| { |
| "epoch": 1.9, |
| "learning_rate": 6.211523137601778e-06, |
| "loss": 1.1699, |
| "step": 1488 |
| }, |
| { |
| "epoch": 1.91, |
| "learning_rate": 6.198730041088747e-06, |
| "loss": 1.1788, |
| "step": 1489 |
| }, |
| { |
| "epoch": 1.91, |
| "learning_rate": 6.1859442125224425e-06, |
| "loss": 1.1894, |
| "step": 1490 |
| }, |
| { |
| "epoch": 1.91, |
| "learning_rate": 6.173165676349103e-06, |
| "loss": 1.188, |
| "step": 1491 |
| }, |
| { |
| "epoch": 1.91, |
| "learning_rate": 6.160394457001013e-06, |
| "loss": 1.1823, |
| "step": 1492 |
| }, |
| { |
| "epoch": 1.91, |
| "learning_rate": 6.147630578896467e-06, |
| "loss": 1.1767, |
| "step": 1493 |
| }, |
| { |
| "epoch": 1.91, |
| "learning_rate": 6.134874066439731e-06, |
| "loss": 1.151, |
| "step": 1494 |
| }, |
| { |
| "epoch": 1.91, |
| "learning_rate": 6.122124944020978e-06, |
| "loss": 1.1289, |
| "step": 1495 |
| }, |
| { |
| "epoch": 1.91, |
| "learning_rate": 6.109383236016266e-06, |
| "loss": 1.2195, |
| "step": 1496 |
| }, |
| { |
| "epoch": 1.92, |
| "learning_rate": 6.0966489667874604e-06, |
| "loss": 1.2644, |
| "step": 1497 |
| }, |
| { |
| "epoch": 1.92, |
| "learning_rate": 6.08392216068221e-06, |
| "loss": 1.1951, |
| "step": 1498 |
| }, |
| { |
| "epoch": 1.92, |
| "learning_rate": 6.071202842033903e-06, |
| "loss": 1.1681, |
| "step": 1499 |
| }, |
| { |
| "epoch": 1.92, |
| "learning_rate": 6.0584910351616e-06, |
| "loss": 1.0822, |
| "step": 1500 |
| }, |
| { |
| "epoch": 1.92, |
| "learning_rate": 6.045786764369999e-06, |
| "loss": 1.205, |
| "step": 1501 |
| }, |
| { |
| "epoch": 1.92, |
| "learning_rate": 6.033090053949401e-06, |
| "loss": 1.1279, |
| "step": 1502 |
| }, |
| { |
| "epoch": 1.92, |
| "learning_rate": 6.020400928175637e-06, |
| "loss": 1.2276, |
| "step": 1503 |
| }, |
| { |
| "epoch": 1.93, |
| "learning_rate": 6.007719411310053e-06, |
| "loss": 1.2384, |
| "step": 1504 |
| }, |
| { |
| "epoch": 1.93, |
| "learning_rate": 5.995045527599432e-06, |
| "loss": 1.1605, |
| "step": 1505 |
| }, |
| { |
| "epoch": 1.93, |
| "learning_rate": 5.982379301275965e-06, |
| "loss": 1.1511, |
| "step": 1506 |
| }, |
| { |
| "epoch": 1.93, |
| "learning_rate": 5.969720756557211e-06, |
| "loss": 1.1394, |
| "step": 1507 |
| }, |
| { |
| "epoch": 1.93, |
| "learning_rate": 5.9570699176460356e-06, |
| "loss": 1.2302, |
| "step": 1508 |
| }, |
| { |
| "epoch": 1.93, |
| "learning_rate": 5.944426808730567e-06, |
| "loss": 1.2512, |
| "step": 1509 |
| }, |
| { |
| "epoch": 1.93, |
| "learning_rate": 5.931791453984165e-06, |
| "loss": 1.309, |
| "step": 1510 |
| }, |
| { |
| "epoch": 1.93, |
| "learning_rate": 5.919163877565351e-06, |
| "loss": 1.1683, |
| "step": 1511 |
| }, |
| { |
| "epoch": 1.94, |
| "learning_rate": 5.906544103617787e-06, |
| "loss": 1.1527, |
| "step": 1512 |
| }, |
| { |
| "epoch": 1.94, |
| "learning_rate": 5.8939321562702075e-06, |
| "loss": 1.1849, |
| "step": 1513 |
| }, |
| { |
| "epoch": 1.94, |
| "learning_rate": 5.881328059636382e-06, |
| "loss": 1.261, |
| "step": 1514 |
| }, |
| { |
| "epoch": 1.94, |
| "learning_rate": 5.868731837815082e-06, |
| "loss": 1.1192, |
| "step": 1515 |
| }, |
| { |
| "epoch": 1.94, |
| "learning_rate": 5.856143514890003e-06, |
| "loss": 1.2567, |
| "step": 1516 |
| }, |
| { |
| "epoch": 1.94, |
| "learning_rate": 5.843563114929758e-06, |
| "loss": 1.1921, |
| "step": 1517 |
| }, |
| { |
| "epoch": 1.94, |
| "learning_rate": 5.830990661987797e-06, |
| "loss": 1.3111, |
| "step": 1518 |
| }, |
| { |
| "epoch": 1.94, |
| "learning_rate": 5.818426180102382e-06, |
| "loss": 1.2476, |
| "step": 1519 |
| }, |
| { |
| "epoch": 1.95, |
| "learning_rate": 5.805869693296534e-06, |
| "loss": 1.2868, |
| "step": 1520 |
| }, |
| { |
| "epoch": 1.95, |
| "learning_rate": 5.79332122557799e-06, |
| "loss": 1.2284, |
| "step": 1521 |
| }, |
| { |
| "epoch": 1.95, |
| "learning_rate": 5.78078080093914e-06, |
| "loss": 1.2468, |
| "step": 1522 |
| }, |
| { |
| "epoch": 1.95, |
| "learning_rate": 5.768248443357023e-06, |
| "loss": 1.2817, |
| "step": 1523 |
| }, |
| { |
| "epoch": 1.95, |
| "learning_rate": 5.75572417679322e-06, |
| "loss": 1.2287, |
| "step": 1524 |
| }, |
| { |
| "epoch": 1.95, |
| "learning_rate": 5.74320802519388e-06, |
| "loss": 1.2079, |
| "step": 1525 |
| }, |
| { |
| "epoch": 1.95, |
| "learning_rate": 5.730700012489605e-06, |
| "loss": 1.1556, |
| "step": 1526 |
| }, |
| { |
| "epoch": 1.95, |
| "learning_rate": 5.718200162595449e-06, |
| "loss": 1.1299, |
| "step": 1527 |
| }, |
| { |
| "epoch": 1.96, |
| "learning_rate": 5.705708499410856e-06, |
| "loss": 1.1807, |
| "step": 1528 |
| }, |
| { |
| "epoch": 1.96, |
| "learning_rate": 5.693225046819626e-06, |
| "loss": 1.2614, |
| "step": 1529 |
| }, |
| { |
| "epoch": 1.96, |
| "learning_rate": 5.680749828689838e-06, |
| "loss": 1.212, |
| "step": 1530 |
| }, |
| { |
| "epoch": 1.96, |
| "learning_rate": 5.66828286887386e-06, |
| "loss": 1.1789, |
| "step": 1531 |
| }, |
| { |
| "epoch": 1.96, |
| "learning_rate": 5.655824191208235e-06, |
| "loss": 1.2098, |
| "step": 1532 |
| }, |
| { |
| "epoch": 1.96, |
| "learning_rate": 5.6433738195137035e-06, |
| "loss": 1.1266, |
| "step": 1533 |
| }, |
| { |
| "epoch": 1.96, |
| "learning_rate": 5.630931777595099e-06, |
| "loss": 1.2683, |
| "step": 1534 |
| }, |
| { |
| "epoch": 1.96, |
| "learning_rate": 5.61849808924134e-06, |
| "loss": 1.1881, |
| "step": 1535 |
| }, |
| { |
| "epoch": 1.97, |
| "learning_rate": 5.606072778225376e-06, |
| "loss": 1.2769, |
| "step": 1536 |
| }, |
| { |
| "epoch": 1.97, |
| "learning_rate": 5.593655868304137e-06, |
| "loss": 1.1944, |
| "step": 1537 |
| }, |
| { |
| "epoch": 1.97, |
| "learning_rate": 5.581247383218477e-06, |
| "loss": 1.1937, |
| "step": 1538 |
| }, |
| { |
| "epoch": 1.97, |
| "learning_rate": 5.568847346693172e-06, |
| "loss": 1.2951, |
| "step": 1539 |
| }, |
| { |
| "epoch": 1.97, |
| "learning_rate": 5.556455782436808e-06, |
| "loss": 1.1568, |
| "step": 1540 |
| }, |
| { |
| "epoch": 1.97, |
| "learning_rate": 5.544072714141806e-06, |
| "loss": 1.1523, |
| "step": 1541 |
| }, |
| { |
| "epoch": 1.97, |
| "learning_rate": 5.531698165484318e-06, |
| "loss": 1.1155, |
| "step": 1542 |
| }, |
| { |
| "epoch": 1.98, |
| "learning_rate": 5.519332160124215e-06, |
| "loss": 1.2658, |
| "step": 1543 |
| }, |
| { |
| "epoch": 1.98, |
| "learning_rate": 5.506974721705041e-06, |
| "loss": 1.1159, |
| "step": 1544 |
| }, |
| { |
| "epoch": 1.98, |
| "learning_rate": 5.494625873853948e-06, |
| "loss": 1.1216, |
| "step": 1545 |
| }, |
| { |
| "epoch": 1.98, |
| "learning_rate": 5.48228564018167e-06, |
| "loss": 1.2539, |
| "step": 1546 |
| }, |
| { |
| "epoch": 1.98, |
| "learning_rate": 5.469954044282475e-06, |
| "loss": 1.1125, |
| "step": 1547 |
| }, |
| { |
| "epoch": 1.98, |
| "learning_rate": 5.457631109734098e-06, |
| "loss": 1.2033, |
| "step": 1548 |
| }, |
| { |
| "epoch": 1.98, |
| "learning_rate": 5.44531686009774e-06, |
| "loss": 1.1368, |
| "step": 1549 |
| }, |
| { |
| "epoch": 1.98, |
| "learning_rate": 5.433011318917971e-06, |
| "loss": 1.1438, |
| "step": 1550 |
| }, |
| { |
| "epoch": 1.99, |
| "learning_rate": 5.420714509722729e-06, |
| "loss": 1.1716, |
| "step": 1551 |
| }, |
| { |
| "epoch": 1.99, |
| "learning_rate": 5.408426456023246e-06, |
| "loss": 1.1518, |
| "step": 1552 |
| }, |
| { |
| "epoch": 1.99, |
| "learning_rate": 5.396147181314018e-06, |
| "loss": 1.1704, |
| "step": 1553 |
| }, |
| { |
| "epoch": 1.99, |
| "learning_rate": 5.383876709072757e-06, |
| "loss": 1.2655, |
| "step": 1554 |
| }, |
| { |
| "epoch": 1.99, |
| "learning_rate": 5.371615062760346e-06, |
| "loss": 1.1582, |
| "step": 1555 |
| }, |
| { |
| "epoch": 1.99, |
| "learning_rate": 5.359362265820778e-06, |
| "loss": 1.1072, |
| "step": 1556 |
| }, |
| { |
| "epoch": 1.99, |
| "learning_rate": 5.347118341681155e-06, |
| "loss": 1.2682, |
| "step": 1557 |
| }, |
| { |
| "epoch": 1.99, |
| "learning_rate": 5.3348833137515845e-06, |
| "loss": 1.2154, |
| "step": 1558 |
| }, |
| { |
| "epoch": 2.0, |
| "learning_rate": 5.322657205425184e-06, |
| "loss": 1.17, |
| "step": 1559 |
| }, |
| { |
| "epoch": 2.0, |
| "learning_rate": 5.310440040078008e-06, |
| "loss": 1.1694, |
| "step": 1560 |
| }, |
| { |
| "epoch": 2.0, |
| "learning_rate": 5.298231841069017e-06, |
| "loss": 1.2991, |
| "step": 1561 |
| }, |
| { |
| "epoch": 2.0, |
| "learning_rate": 5.286032631740023e-06, |
| "loss": 1.2575, |
| "step": 1562 |
| }, |
| { |
| "epoch": 2.0, |
| "learning_rate": 5.273842435415661e-06, |
| "loss": 1.0553, |
| "step": 1563 |
| }, |
| { |
| "epoch": 2.0, |
| "learning_rate": 5.2616612754033115e-06, |
| "loss": 0.9876, |
| "step": 1564 |
| }, |
| { |
| "epoch": 2.0, |
| "learning_rate": 5.249489174993106e-06, |
| "loss": 1.0595, |
| "step": 1565 |
| }, |
| { |
| "epoch": 2.0, |
| "learning_rate": 5.237326157457825e-06, |
| "loss": 0.8973, |
| "step": 1566 |
| }, |
| { |
| "epoch": 2.01, |
| "learning_rate": 5.225172246052914e-06, |
| "loss": 0.9984, |
| "step": 1567 |
| }, |
| { |
| "epoch": 2.01, |
| "learning_rate": 5.2130274640163795e-06, |
| "loss": 0.9015, |
| "step": 1568 |
| }, |
| { |
| "epoch": 2.01, |
| "learning_rate": 5.200891834568786e-06, |
| "loss": 0.9749, |
| "step": 1569 |
| }, |
| { |
| "epoch": 2.01, |
| "learning_rate": 5.188765380913201e-06, |
| "loss": 1.0185, |
| "step": 1570 |
| }, |
| { |
| "epoch": 2.01, |
| "learning_rate": 5.176648126235144e-06, |
| "loss": 0.8835, |
| "step": 1571 |
| }, |
| { |
| "epoch": 2.01, |
| "learning_rate": 5.164540093702537e-06, |
| "loss": 1.0687, |
| "step": 1572 |
| }, |
| { |
| "epoch": 2.01, |
| "learning_rate": 5.152441306465694e-06, |
| "loss": 0.9187, |
| "step": 1573 |
| }, |
| { |
| "epoch": 2.01, |
| "learning_rate": 5.140351787657223e-06, |
| "loss": 1.0196, |
| "step": 1574 |
| }, |
| { |
| "epoch": 2.02, |
| "learning_rate": 5.128271560392037e-06, |
| "loss": 0.867, |
| "step": 1575 |
| }, |
| { |
| "epoch": 2.02, |
| "learning_rate": 5.1162006477672625e-06, |
| "loss": 1.0205, |
| "step": 1576 |
| }, |
| { |
| "epoch": 2.02, |
| "learning_rate": 5.104139072862229e-06, |
| "loss": 0.9017, |
| "step": 1577 |
| }, |
| { |
| "epoch": 2.02, |
| "learning_rate": 5.09208685873841e-06, |
| "loss": 0.9108, |
| "step": 1578 |
| }, |
| { |
| "epoch": 2.02, |
| "learning_rate": 5.080044028439384e-06, |
| "loss": 0.9202, |
| "step": 1579 |
| }, |
| { |
| "epoch": 2.02, |
| "learning_rate": 5.068010604990774e-06, |
| "loss": 1.0001, |
| "step": 1580 |
| }, |
| { |
| "epoch": 2.02, |
| "learning_rate": 5.055986611400241e-06, |
| "loss": 1.0204, |
| "step": 1581 |
| }, |
| { |
| "epoch": 2.02, |
| "learning_rate": 5.04397207065739e-06, |
| "loss": 0.9883, |
| "step": 1582 |
| }, |
| { |
| "epoch": 2.03, |
| "learning_rate": 5.03196700573378e-06, |
| "loss": 0.9101, |
| "step": 1583 |
| }, |
| { |
| "epoch": 2.03, |
| "learning_rate": 5.019971439582827e-06, |
| "loss": 0.9032, |
| "step": 1584 |
| }, |
| { |
| "epoch": 2.03, |
| "learning_rate": 5.007985395139798e-06, |
| "loss": 0.9103, |
| "step": 1585 |
| }, |
| { |
| "epoch": 2.03, |
| "learning_rate": 4.996008895321758e-06, |
| "loss": 0.9687, |
| "step": 1586 |
| }, |
| { |
| "epoch": 2.03, |
| "learning_rate": 4.984041963027512e-06, |
| "loss": 0.8934, |
| "step": 1587 |
| }, |
| { |
| "epoch": 2.03, |
| "learning_rate": 4.972084621137584e-06, |
| "loss": 1.0048, |
| "step": 1588 |
| }, |
| { |
| "epoch": 2.03, |
| "learning_rate": 4.9601368925141555e-06, |
| "loss": 0.9536, |
| "step": 1589 |
| }, |
| { |
| "epoch": 2.04, |
| "learning_rate": 4.948198800001018e-06, |
| "loss": 0.9386, |
| "step": 1590 |
| }, |
| { |
| "epoch": 2.04, |
| "learning_rate": 4.936270366423563e-06, |
| "loss": 0.9867, |
| "step": 1591 |
| }, |
| { |
| "epoch": 2.04, |
| "learning_rate": 4.92435161458869e-06, |
| "loss": 1.0543, |
| "step": 1592 |
| }, |
| { |
| "epoch": 2.04, |
| "learning_rate": 4.9124425672847984e-06, |
| "loss": 0.8894, |
| "step": 1593 |
| }, |
| { |
| "epoch": 2.04, |
| "learning_rate": 4.900543247281735e-06, |
| "loss": 0.9253, |
| "step": 1594 |
| }, |
| { |
| "epoch": 2.04, |
| "learning_rate": 4.888653677330742e-06, |
| "loss": 0.9791, |
| "step": 1595 |
| }, |
| { |
| "epoch": 2.04, |
| "learning_rate": 4.8767738801644215e-06, |
| "loss": 0.9468, |
| "step": 1596 |
| }, |
| { |
| "epoch": 2.04, |
| "learning_rate": 4.864903878496698e-06, |
| "loss": 0.9756, |
| "step": 1597 |
| }, |
| { |
| "epoch": 2.05, |
| "learning_rate": 4.8530436950227485e-06, |
| "loss": 0.9662, |
| "step": 1598 |
| }, |
| { |
| "epoch": 2.05, |
| "learning_rate": 4.841193352419003e-06, |
| "loss": 0.8578, |
| "step": 1599 |
| }, |
| { |
| "epoch": 2.05, |
| "learning_rate": 4.829352873343054e-06, |
| "loss": 0.9664, |
| "step": 1600 |
| }, |
| { |
| "epoch": 2.05, |
| "learning_rate": 4.817522280433644e-06, |
| "loss": 0.8967, |
| "step": 1601 |
| }, |
| { |
| "epoch": 2.05, |
| "learning_rate": 4.805701596310617e-06, |
| "loss": 0.8151, |
| "step": 1602 |
| }, |
| { |
| "epoch": 2.05, |
| "learning_rate": 4.793890843574865e-06, |
| "loss": 0.9862, |
| "step": 1603 |
| }, |
| { |
| "epoch": 2.05, |
| "learning_rate": 4.782090044808297e-06, |
| "loss": 0.8758, |
| "step": 1604 |
| }, |
| { |
| "epoch": 2.05, |
| "learning_rate": 4.770299222573788e-06, |
| "loss": 0.9145, |
| "step": 1605 |
| }, |
| { |
| "epoch": 2.06, |
| "learning_rate": 4.758518399415128e-06, |
| "loss": 0.9274, |
| "step": 1606 |
| }, |
| { |
| "epoch": 2.06, |
| "learning_rate": 4.746747597857014e-06, |
| "loss": 0.9774, |
| "step": 1607 |
| }, |
| { |
| "epoch": 2.06, |
| "learning_rate": 4.734986840404955e-06, |
| "loss": 0.8476, |
| "step": 1608 |
| }, |
| { |
| "epoch": 2.06, |
| "learning_rate": 4.72323614954527e-06, |
| "loss": 0.8496, |
| "step": 1609 |
| }, |
| { |
| "epoch": 2.06, |
| "learning_rate": 4.71149554774503e-06, |
| "loss": 1.0089, |
| "step": 1610 |
| }, |
| { |
| "epoch": 2.06, |
| "learning_rate": 4.699765057452012e-06, |
| "loss": 0.945, |
| "step": 1611 |
| }, |
| { |
| "epoch": 2.06, |
| "learning_rate": 4.688044701094662e-06, |
| "loss": 0.9744, |
| "step": 1612 |
| }, |
| { |
| "epoch": 2.06, |
| "learning_rate": 4.676334501082053e-06, |
| "loss": 0.9844, |
| "step": 1613 |
| }, |
| { |
| "epoch": 2.07, |
| "learning_rate": 4.664634479803834e-06, |
| "loss": 0.9222, |
| "step": 1614 |
| }, |
| { |
| "epoch": 2.07, |
| "learning_rate": 4.6529446596301996e-06, |
| "loss": 0.9032, |
| "step": 1615 |
| }, |
| { |
| "epoch": 2.07, |
| "learning_rate": 4.641265062911825e-06, |
| "loss": 0.8921, |
| "step": 1616 |
| }, |
| { |
| "epoch": 2.07, |
| "learning_rate": 4.629595711979863e-06, |
| "loss": 0.9486, |
| "step": 1617 |
| }, |
| { |
| "epoch": 2.07, |
| "learning_rate": 4.617936629145852e-06, |
| "loss": 0.9193, |
| "step": 1618 |
| }, |
| { |
| "epoch": 2.07, |
| "learning_rate": 4.606287836701712e-06, |
| "loss": 0.8869, |
| "step": 1619 |
| }, |
| { |
| "epoch": 2.07, |
| "learning_rate": 4.5946493569196846e-06, |
| "loss": 0.8432, |
| "step": 1620 |
| }, |
| { |
| "epoch": 2.07, |
| "learning_rate": 4.583021212052294e-06, |
| "loss": 1.0055, |
| "step": 1621 |
| }, |
| { |
| "epoch": 2.08, |
| "learning_rate": 4.571403424332304e-06, |
| "loss": 0.8568, |
| "step": 1622 |
| }, |
| { |
| "epoch": 2.08, |
| "learning_rate": 4.559796015972677e-06, |
| "loss": 0.9654, |
| "step": 1623 |
| }, |
| { |
| "epoch": 2.08, |
| "learning_rate": 4.5481990091665276e-06, |
| "loss": 0.9807, |
| "step": 1624 |
| }, |
| { |
| "epoch": 2.08, |
| "learning_rate": 4.536612426087088e-06, |
| "loss": 0.7968, |
| "step": 1625 |
| }, |
| { |
| "epoch": 2.08, |
| "learning_rate": 4.52503628888765e-06, |
| "loss": 0.9867, |
| "step": 1626 |
| }, |
| { |
| "epoch": 2.08, |
| "learning_rate": 4.5134706197015435e-06, |
| "loss": 0.8893, |
| "step": 1627 |
| }, |
| { |
| "epoch": 2.08, |
| "learning_rate": 4.50191544064208e-06, |
| "loss": 0.9086, |
| "step": 1628 |
| }, |
| { |
| "epoch": 2.09, |
| "learning_rate": 4.490370773802515e-06, |
| "loss": 1.0271, |
| "step": 1629 |
| }, |
| { |
| "epoch": 2.09, |
| "learning_rate": 4.4788366412560035e-06, |
| "loss": 0.8963, |
| "step": 1630 |
| }, |
| { |
| "epoch": 2.09, |
| "learning_rate": 4.467313065055561e-06, |
| "loss": 0.8981, |
| "step": 1631 |
| }, |
| { |
| "epoch": 2.09, |
| "learning_rate": 4.455800067234017e-06, |
| "loss": 0.9212, |
| "step": 1632 |
| }, |
| { |
| "epoch": 2.09, |
| "learning_rate": 4.444297669803981e-06, |
| "loss": 1.0019, |
| "step": 1633 |
| }, |
| { |
| "epoch": 2.09, |
| "learning_rate": 4.4328058947577825e-06, |
| "loss": 0.9532, |
| "step": 1634 |
| }, |
| { |
| "epoch": 2.09, |
| "learning_rate": 4.421324764067454e-06, |
| "loss": 0.8828, |
| "step": 1635 |
| }, |
| { |
| "epoch": 2.09, |
| "learning_rate": 4.4098542996846705e-06, |
| "loss": 0.8679, |
| "step": 1636 |
| }, |
| { |
| "epoch": 2.1, |
| "learning_rate": 4.398394523540714e-06, |
| "loss": 0.8877, |
| "step": 1637 |
| }, |
| { |
| "epoch": 2.1, |
| "learning_rate": 4.386945457546433e-06, |
| "loss": 0.998, |
| "step": 1638 |
| }, |
| { |
| "epoch": 2.1, |
| "learning_rate": 4.375507123592194e-06, |
| "loss": 0.9224, |
| "step": 1639 |
| }, |
| { |
| "epoch": 2.1, |
| "learning_rate": 4.364079543547848e-06, |
| "loss": 0.9539, |
| "step": 1640 |
| }, |
| { |
| "epoch": 2.1, |
| "learning_rate": 4.352662739262682e-06, |
| "loss": 0.9783, |
| "step": 1641 |
| }, |
| { |
| "epoch": 2.1, |
| "learning_rate": 4.341256732565386e-06, |
| "loss": 0.9187, |
| "step": 1642 |
| }, |
| { |
| "epoch": 2.1, |
| "learning_rate": 4.329861545263992e-06, |
| "loss": 0.9102, |
| "step": 1643 |
| }, |
| { |
| "epoch": 2.1, |
| "learning_rate": 4.3184771991458584e-06, |
| "loss": 0.9146, |
| "step": 1644 |
| }, |
| { |
| "epoch": 2.11, |
| "learning_rate": 4.307103715977612e-06, |
| "loss": 0.8811, |
| "step": 1645 |
| }, |
| { |
| "epoch": 2.11, |
| "learning_rate": 4.295741117505108e-06, |
| "loss": 1.0502, |
| "step": 1646 |
| }, |
| { |
| "epoch": 2.11, |
| "learning_rate": 4.2843894254533906e-06, |
| "loss": 0.9476, |
| "step": 1647 |
| }, |
| { |
| "epoch": 2.11, |
| "learning_rate": 4.273048661526653e-06, |
| "loss": 0.8427, |
| "step": 1648 |
| }, |
| { |
| "epoch": 2.11, |
| "learning_rate": 4.261718847408192e-06, |
| "loss": 0.8971, |
| "step": 1649 |
| }, |
| { |
| "epoch": 2.11, |
| "learning_rate": 4.2504000047603734e-06, |
| "loss": 0.9371, |
| "step": 1650 |
| }, |
| { |
| "epoch": 2.11, |
| "learning_rate": 4.239092155224572e-06, |
| "loss": 0.8408, |
| "step": 1651 |
| }, |
| { |
| "epoch": 2.11, |
| "learning_rate": 4.227795320421167e-06, |
| "loss": 0.9018, |
| "step": 1652 |
| }, |
| { |
| "epoch": 2.12, |
| "learning_rate": 4.216509521949455e-06, |
| "loss": 0.8654, |
| "step": 1653 |
| }, |
| { |
| "epoch": 2.12, |
| "learning_rate": 4.205234781387643e-06, |
| "loss": 0.8716, |
| "step": 1654 |
| }, |
| { |
| "epoch": 2.12, |
| "learning_rate": 4.193971120292793e-06, |
| "loss": 0.8943, |
| "step": 1655 |
| }, |
| { |
| "epoch": 2.12, |
| "learning_rate": 4.182718560200787e-06, |
| "loss": 0.9488, |
| "step": 1656 |
| }, |
| { |
| "epoch": 2.12, |
| "learning_rate": 4.171477122626276e-06, |
| "loss": 0.8276, |
| "step": 1657 |
| }, |
| { |
| "epoch": 2.12, |
| "learning_rate": 4.160246829062648e-06, |
| "loss": 0.957, |
| "step": 1658 |
| }, |
| { |
| "epoch": 2.12, |
| "learning_rate": 4.1490277009819845e-06, |
| "loss": 0.8498, |
| "step": 1659 |
| }, |
| { |
| "epoch": 2.12, |
| "learning_rate": 4.137819759835023e-06, |
| "loss": 0.9623, |
| "step": 1660 |
| }, |
| { |
| "epoch": 2.13, |
| "learning_rate": 4.126623027051092e-06, |
| "loss": 0.8857, |
| "step": 1661 |
| }, |
| { |
| "epoch": 2.13, |
| "learning_rate": 4.115437524038124e-06, |
| "loss": 0.9144, |
| "step": 1662 |
| }, |
| { |
| "epoch": 2.13, |
| "learning_rate": 4.104263272182546e-06, |
| "loss": 0.9597, |
| "step": 1663 |
| }, |
| { |
| "epoch": 2.13, |
| "learning_rate": 4.093100292849295e-06, |
| "loss": 0.9234, |
| "step": 1664 |
| }, |
| { |
| "epoch": 2.13, |
| "learning_rate": 4.081948607381746e-06, |
| "loss": 0.9882, |
| "step": 1665 |
| }, |
| { |
| "epoch": 2.13, |
| "learning_rate": 4.070808237101686e-06, |
| "loss": 0.8833, |
| "step": 1666 |
| }, |
| { |
| "epoch": 2.13, |
| "learning_rate": 4.0596792033092605e-06, |
| "loss": 0.8744, |
| "step": 1667 |
| }, |
| { |
| "epoch": 2.14, |
| "learning_rate": 4.048561527282948e-06, |
| "loss": 0.9768, |
| "step": 1668 |
| }, |
| { |
| "epoch": 2.14, |
| "learning_rate": 4.037455230279499e-06, |
| "loss": 0.9219, |
| "step": 1669 |
| }, |
| { |
| "epoch": 2.14, |
| "learning_rate": 4.026360333533927e-06, |
| "loss": 1.0341, |
| "step": 1670 |
| }, |
| { |
| "epoch": 2.14, |
| "learning_rate": 4.015276858259427e-06, |
| "loss": 0.9072, |
| "step": 1671 |
| }, |
| { |
| "epoch": 2.14, |
| "learning_rate": 4.004204825647371e-06, |
| "loss": 0.9757, |
| "step": 1672 |
| }, |
| { |
| "epoch": 2.14, |
| "learning_rate": 3.993144256867246e-06, |
| "loss": 0.8854, |
| "step": 1673 |
| }, |
| { |
| "epoch": 2.14, |
| "learning_rate": 3.982095173066624e-06, |
| "loss": 0.7917, |
| "step": 1674 |
| }, |
| { |
| "epoch": 2.14, |
| "learning_rate": 3.971057595371116e-06, |
| "loss": 0.8593, |
| "step": 1675 |
| }, |
| { |
| "epoch": 2.15, |
| "learning_rate": 3.960031544884338e-06, |
| "loss": 0.9831, |
| "step": 1676 |
| }, |
| { |
| "epoch": 2.15, |
| "learning_rate": 3.949017042687851e-06, |
| "loss": 0.7972, |
| "step": 1677 |
| }, |
| { |
| "epoch": 2.15, |
| "learning_rate": 3.938014109841163e-06, |
| "loss": 0.9441, |
| "step": 1678 |
| }, |
| { |
| "epoch": 2.15, |
| "learning_rate": 3.927022767381634e-06, |
| "loss": 0.8867, |
| "step": 1679 |
| }, |
| { |
| "epoch": 2.15, |
| "learning_rate": 3.91604303632448e-06, |
| "loss": 0.9397, |
| "step": 1680 |
| }, |
| { |
| "epoch": 2.15, |
| "learning_rate": 3.9050749376627085e-06, |
| "loss": 0.9325, |
| "step": 1681 |
| }, |
| { |
| "epoch": 2.15, |
| "learning_rate": 3.8941184923670925e-06, |
| "loss": 1.0011, |
| "step": 1682 |
| }, |
| { |
| "epoch": 2.15, |
| "learning_rate": 3.88317372138612e-06, |
| "loss": 0.938, |
| "step": 1683 |
| }, |
| { |
| "epoch": 2.16, |
| "learning_rate": 3.872240645645962e-06, |
| "loss": 0.9411, |
| "step": 1684 |
| }, |
| { |
| "epoch": 2.16, |
| "learning_rate": 3.861319286050413e-06, |
| "loss": 1.0486, |
| "step": 1685 |
| }, |
| { |
| "epoch": 2.16, |
| "learning_rate": 3.850409663480894e-06, |
| "loss": 0.9033, |
| "step": 1686 |
| }, |
| { |
| "epoch": 2.16, |
| "learning_rate": 3.839511798796357e-06, |
| "loss": 1.0295, |
| "step": 1687 |
| }, |
| { |
| "epoch": 2.16, |
| "learning_rate": 3.828625712833297e-06, |
| "loss": 0.9372, |
| "step": 1688 |
| }, |
| { |
| "epoch": 2.16, |
| "learning_rate": 3.817751426405669e-06, |
| "loss": 0.819, |
| "step": 1689 |
| }, |
| { |
| "epoch": 2.16, |
| "learning_rate": 3.8068889603048773e-06, |
| "loss": 0.939, |
| "step": 1690 |
| }, |
| { |
| "epoch": 2.16, |
| "learning_rate": 3.7960383352997264e-06, |
| "loss": 0.9189, |
| "step": 1691 |
| }, |
| { |
| "epoch": 2.17, |
| "learning_rate": 3.785199572136381e-06, |
| "loss": 0.9362, |
| "step": 1692 |
| }, |
| { |
| "epoch": 2.17, |
| "learning_rate": 3.774372691538314e-06, |
| "loss": 1.021, |
| "step": 1693 |
| }, |
| { |
| "epoch": 2.17, |
| "learning_rate": 3.763557714206303e-06, |
| "loss": 0.963, |
| "step": 1694 |
| }, |
| { |
| "epoch": 2.17, |
| "learning_rate": 3.7527546608183386e-06, |
| "loss": 0.9682, |
| "step": 1695 |
| }, |
| { |
| "epoch": 2.17, |
| "learning_rate": 3.7419635520296426e-06, |
| "loss": 0.8836, |
| "step": 1696 |
| }, |
| { |
| "epoch": 2.17, |
| "learning_rate": 3.7311844084725725e-06, |
| "loss": 0.9708, |
| "step": 1697 |
| }, |
| { |
| "epoch": 2.17, |
| "learning_rate": 3.720417250756623e-06, |
| "loss": 0.9227, |
| "step": 1698 |
| }, |
| { |
| "epoch": 2.17, |
| "learning_rate": 3.7096620994683695e-06, |
| "loss": 0.8323, |
| "step": 1699 |
| }, |
| { |
| "epoch": 2.18, |
| "learning_rate": 3.698918975171435e-06, |
| "loss": 0.96, |
| "step": 1700 |
| }, |
| { |
| "epoch": 2.18, |
| "learning_rate": 3.6881878984064323e-06, |
| "loss": 0.9208, |
| "step": 1701 |
| }, |
| { |
| "epoch": 2.18, |
| "learning_rate": 3.6774688896909625e-06, |
| "loss": 0.9071, |
| "step": 1702 |
| }, |
| { |
| "epoch": 2.18, |
| "learning_rate": 3.6667619695195287e-06, |
| "loss": 1.03, |
| "step": 1703 |
| }, |
| { |
| "epoch": 2.18, |
| "learning_rate": 3.6560671583635467e-06, |
| "loss": 0.9356, |
| "step": 1704 |
| }, |
| { |
| "epoch": 2.18, |
| "learning_rate": 3.6453844766712553e-06, |
| "loss": 1.012, |
| "step": 1705 |
| }, |
| { |
| "epoch": 2.18, |
| "learning_rate": 3.634713944867716e-06, |
| "loss": 0.8723, |
| "step": 1706 |
| }, |
| { |
| "epoch": 2.18, |
| "learning_rate": 3.6240555833547587e-06, |
| "loss": 0.9082, |
| "step": 1707 |
| }, |
| { |
| "epoch": 2.19, |
| "learning_rate": 3.6134094125109407e-06, |
| "loss": 0.9916, |
| "step": 1708 |
| }, |
| { |
| "epoch": 2.19, |
| "learning_rate": 3.6027754526915126e-06, |
| "loss": 0.9049, |
| "step": 1709 |
| }, |
| { |
| "epoch": 2.19, |
| "learning_rate": 3.592153724228382e-06, |
| "loss": 1.0006, |
| "step": 1710 |
| }, |
| { |
| "epoch": 2.19, |
| "learning_rate": 3.5815442474300564e-06, |
| "loss": 0.9074, |
| "step": 1711 |
| }, |
| { |
| "epoch": 2.19, |
| "learning_rate": 3.5709470425816407e-06, |
| "loss": 0.9461, |
| "step": 1712 |
| }, |
| { |
| "epoch": 2.19, |
| "learning_rate": 3.5603621299447534e-06, |
| "loss": 0.8634, |
| "step": 1713 |
| }, |
| { |
| "epoch": 2.19, |
| "learning_rate": 3.5497895297575248e-06, |
| "loss": 0.8921, |
| "step": 1714 |
| }, |
| { |
| "epoch": 2.2, |
| "learning_rate": 3.539229262234539e-06, |
| "loss": 0.9433, |
| "step": 1715 |
| }, |
| { |
| "epoch": 2.2, |
| "learning_rate": 3.5286813475668027e-06, |
| "loss": 0.8936, |
| "step": 1716 |
| }, |
| { |
| "epoch": 2.2, |
| "learning_rate": 3.518145805921701e-06, |
| "loss": 0.7956, |
| "step": 1717 |
| }, |
| { |
| "epoch": 2.2, |
| "learning_rate": 3.5076226574429694e-06, |
| "loss": 0.9654, |
| "step": 1718 |
| }, |
| { |
| "epoch": 2.2, |
| "learning_rate": 3.49711192225063e-06, |
| "loss": 0.7656, |
| "step": 1719 |
| }, |
| { |
| "epoch": 2.2, |
| "learning_rate": 3.4866136204409985e-06, |
| "loss": 0.7926, |
| "step": 1720 |
| }, |
| { |
| "epoch": 2.2, |
| "learning_rate": 3.4761277720865918e-06, |
| "loss": 0.9608, |
| "step": 1721 |
| }, |
| { |
| "epoch": 2.2, |
| "learning_rate": 3.4656543972361322e-06, |
| "loss": 0.9616, |
| "step": 1722 |
| }, |
| { |
| "epoch": 2.21, |
| "learning_rate": 3.4551935159144854e-06, |
| "loss": 0.9001, |
| "step": 1723 |
| }, |
| { |
| "epoch": 2.21, |
| "learning_rate": 3.4447451481226357e-06, |
| "loss": 0.9394, |
| "step": 1724 |
| }, |
| { |
| "epoch": 2.21, |
| "learning_rate": 3.434309313837636e-06, |
| "loss": 0.812, |
| "step": 1725 |
| }, |
| { |
| "epoch": 2.21, |
| "learning_rate": 3.4238860330125822e-06, |
| "loss": 0.9025, |
| "step": 1726 |
| }, |
| { |
| "epoch": 2.21, |
| "learning_rate": 3.413475325576554e-06, |
| "loss": 0.9188, |
| "step": 1727 |
| }, |
| { |
| "epoch": 2.21, |
| "learning_rate": 3.403077211434613e-06, |
| "loss": 0.8714, |
| "step": 1728 |
| }, |
| { |
| "epoch": 2.21, |
| "learning_rate": 3.3926917104677205e-06, |
| "loss": 0.9143, |
| "step": 1729 |
| }, |
| { |
| "epoch": 2.21, |
| "learning_rate": 3.3823188425327424e-06, |
| "loss": 0.7811, |
| "step": 1730 |
| }, |
| { |
| "epoch": 2.22, |
| "learning_rate": 3.3719586274623707e-06, |
| "loss": 0.9018, |
| "step": 1731 |
| }, |
| { |
| "epoch": 2.22, |
| "learning_rate": 3.3616110850651174e-06, |
| "loss": 0.8254, |
| "step": 1732 |
| }, |
| { |
| "epoch": 2.22, |
| "learning_rate": 3.3512762351252638e-06, |
| "loss": 0.9799, |
| "step": 1733 |
| }, |
| { |
| "epoch": 2.22, |
| "learning_rate": 3.340954097402823e-06, |
| "loss": 0.8388, |
| "step": 1734 |
| }, |
| { |
| "epoch": 2.22, |
| "learning_rate": 3.330644691633492e-06, |
| "loss": 0.9305, |
| "step": 1735 |
| }, |
| { |
| "epoch": 2.22, |
| "learning_rate": 3.320348037528648e-06, |
| "loss": 0.9721, |
| "step": 1736 |
| }, |
| { |
| "epoch": 2.22, |
| "learning_rate": 3.3100641547752576e-06, |
| "loss": 0.9311, |
| "step": 1737 |
| }, |
| { |
| "epoch": 2.22, |
| "learning_rate": 3.2997930630359e-06, |
| "loss": 0.9306, |
| "step": 1738 |
| }, |
| { |
| "epoch": 2.23, |
| "learning_rate": 3.28953478194867e-06, |
| "loss": 0.959, |
| "step": 1739 |
| }, |
| { |
| "epoch": 2.23, |
| "learning_rate": 3.2792893311271867e-06, |
| "loss": 0.8128, |
| "step": 1740 |
| }, |
| { |
| "epoch": 2.23, |
| "learning_rate": 3.2690567301605324e-06, |
| "loss": 0.8764, |
| "step": 1741 |
| }, |
| { |
| "epoch": 2.23, |
| "learning_rate": 3.258836998613224e-06, |
| "loss": 0.9671, |
| "step": 1742 |
| }, |
| { |
| "epoch": 2.23, |
| "learning_rate": 3.248630156025159e-06, |
| "loss": 0.9977, |
| "step": 1743 |
| }, |
| { |
| "epoch": 2.23, |
| "learning_rate": 3.238436221911614e-06, |
| "loss": 0.8843, |
| "step": 1744 |
| }, |
| { |
| "epoch": 2.23, |
| "learning_rate": 3.228255215763162e-06, |
| "loss": 0.9281, |
| "step": 1745 |
| }, |
| { |
| "epoch": 2.23, |
| "learning_rate": 3.2180871570456797e-06, |
| "loss": 0.925, |
| "step": 1746 |
| }, |
| { |
| "epoch": 2.24, |
| "learning_rate": 3.2079320652002686e-06, |
| "loss": 1.0351, |
| "step": 1747 |
| }, |
| { |
| "epoch": 2.24, |
| "learning_rate": 3.1977899596432506e-06, |
| "loss": 0.9145, |
| "step": 1748 |
| }, |
| { |
| "epoch": 2.24, |
| "learning_rate": 3.187660859766113e-06, |
| "loss": 0.9336, |
| "step": 1749 |
| }, |
| { |
| "epoch": 2.24, |
| "learning_rate": 3.177544784935479e-06, |
| "loss": 0.9329, |
| "step": 1750 |
| }, |
| { |
| "epoch": 2.24, |
| "learning_rate": 3.167441754493066e-06, |
| "loss": 0.9723, |
| "step": 1751 |
| }, |
| { |
| "epoch": 2.24, |
| "learning_rate": 3.1573517877556548e-06, |
| "loss": 0.8157, |
| "step": 1752 |
| }, |
| { |
| "epoch": 2.24, |
| "learning_rate": 3.1472749040150365e-06, |
| "loss": 0.9577, |
| "step": 1753 |
| }, |
| { |
| "epoch": 2.25, |
| "learning_rate": 3.1372111225380096e-06, |
| "loss": 0.9359, |
| "step": 1754 |
| }, |
| { |
| "epoch": 2.25, |
| "learning_rate": 3.127160462566301e-06, |
| "loss": 1.0164, |
| "step": 1755 |
| }, |
| { |
| "epoch": 2.25, |
| "learning_rate": 3.1171229433165575e-06, |
| "loss": 0.937, |
| "step": 1756 |
| }, |
| { |
| "epoch": 2.25, |
| "learning_rate": 3.107098583980301e-06, |
| "loss": 0.9567, |
| "step": 1757 |
| }, |
| { |
| "epoch": 2.25, |
| "learning_rate": 3.0970874037238952e-06, |
| "loss": 0.9211, |
| "step": 1758 |
| }, |
| { |
| "epoch": 2.25, |
| "learning_rate": 3.0870894216884994e-06, |
| "loss": 0.9142, |
| "step": 1759 |
| }, |
| { |
| "epoch": 2.25, |
| "learning_rate": 3.077104656990041e-06, |
| "loss": 0.9803, |
| "step": 1760 |
| }, |
| { |
| "epoch": 2.25, |
| "learning_rate": 3.0671331287191773e-06, |
| "loss": 0.9646, |
| "step": 1761 |
| }, |
| { |
| "epoch": 2.26, |
| "learning_rate": 3.0571748559412595e-06, |
| "loss": 0.9349, |
| "step": 1762 |
| }, |
| { |
| "epoch": 2.26, |
| "learning_rate": 3.0472298576962857e-06, |
| "loss": 0.9774, |
| "step": 1763 |
| }, |
| { |
| "epoch": 2.26, |
| "learning_rate": 3.037298152998882e-06, |
| "loss": 0.9041, |
| "step": 1764 |
| }, |
| { |
| "epoch": 2.26, |
| "learning_rate": 3.027379760838256e-06, |
| "loss": 0.8421, |
| "step": 1765 |
| }, |
| { |
| "epoch": 2.26, |
| "learning_rate": 3.017474700178161e-06, |
| "loss": 0.8946, |
| "step": 1766 |
| }, |
| { |
| "epoch": 2.26, |
| "learning_rate": 3.00758298995686e-06, |
| "loss": 0.8961, |
| "step": 1767 |
| }, |
| { |
| "epoch": 2.26, |
| "learning_rate": 2.997704649087091e-06, |
| "loss": 0.7834, |
| "step": 1768 |
| }, |
| { |
| "epoch": 2.26, |
| "learning_rate": 2.9878396964560307e-06, |
| "loss": 0.8679, |
| "step": 1769 |
| }, |
| { |
| "epoch": 2.27, |
| "learning_rate": 2.9779881509252574e-06, |
| "loss": 0.9619, |
| "step": 1770 |
| }, |
| { |
| "epoch": 2.27, |
| "learning_rate": 2.9681500313307176e-06, |
| "loss": 0.9902, |
| "step": 1771 |
| }, |
| { |
| "epoch": 2.27, |
| "learning_rate": 2.95832535648268e-06, |
| "loss": 0.9478, |
| "step": 1772 |
| }, |
| { |
| "epoch": 2.27, |
| "learning_rate": 2.948514145165714e-06, |
| "loss": 0.9337, |
| "step": 1773 |
| }, |
| { |
| "epoch": 2.27, |
| "learning_rate": 2.938716416138647e-06, |
| "loss": 0.9225, |
| "step": 1774 |
| }, |
| { |
| "epoch": 2.27, |
| "learning_rate": 2.9289321881345257e-06, |
| "loss": 1.0805, |
| "step": 1775 |
| }, |
| { |
| "epoch": 2.27, |
| "learning_rate": 2.919161479860585e-06, |
| "loss": 0.8907, |
| "step": 1776 |
| }, |
| { |
| "epoch": 2.27, |
| "learning_rate": 2.9094043099982104e-06, |
| "loss": 0.9301, |
| "step": 1777 |
| }, |
| { |
| "epoch": 2.28, |
| "learning_rate": 2.899660697202901e-06, |
| "loss": 0.9534, |
| "step": 1778 |
| }, |
| { |
| "epoch": 2.28, |
| "learning_rate": 2.8899306601042353e-06, |
| "loss": 0.9699, |
| "step": 1779 |
| }, |
| { |
| "epoch": 2.28, |
| "learning_rate": 2.88021421730584e-06, |
| "loss": 0.9823, |
| "step": 1780 |
| }, |
| { |
| "epoch": 2.28, |
| "learning_rate": 2.8705113873853384e-06, |
| "loss": 0.863, |
| "step": 1781 |
| }, |
| { |
| "epoch": 2.28, |
| "learning_rate": 2.860822188894338e-06, |
| "loss": 0.9131, |
| "step": 1782 |
| }, |
| { |
| "epoch": 2.28, |
| "learning_rate": 2.8511466403583766e-06, |
| "loss": 0.8497, |
| "step": 1783 |
| }, |
| { |
| "epoch": 2.28, |
| "learning_rate": 2.8414847602768968e-06, |
| "loss": 0.9432, |
| "step": 1784 |
| }, |
| { |
| "epoch": 2.28, |
| "learning_rate": 2.831836567123207e-06, |
| "loss": 0.8666, |
| "step": 1785 |
| }, |
| { |
| "epoch": 2.29, |
| "learning_rate": 2.822202079344446e-06, |
| "loss": 0.9243, |
| "step": 1786 |
| }, |
| { |
| "epoch": 2.29, |
| "learning_rate": 2.8125813153615455e-06, |
| "loss": 0.9677, |
| "step": 1787 |
| }, |
| { |
| "epoch": 2.29, |
| "learning_rate": 2.802974293569204e-06, |
| "loss": 0.8537, |
| "step": 1788 |
| }, |
| { |
| "epoch": 2.29, |
| "learning_rate": 2.7933810323358414e-06, |
| "loss": 0.9206, |
| "step": 1789 |
| }, |
| { |
| "epoch": 2.29, |
| "learning_rate": 2.783801550003562e-06, |
| "loss": 0.9126, |
| "step": 1790 |
| }, |
| { |
| "epoch": 2.29, |
| "learning_rate": 2.774235864888135e-06, |
| "loss": 0.9411, |
| "step": 1791 |
| }, |
| { |
| "epoch": 2.29, |
| "learning_rate": 2.764683995278944e-06, |
| "loss": 0.9471, |
| "step": 1792 |
| }, |
| { |
| "epoch": 2.3, |
| "learning_rate": 2.7551459594389605e-06, |
| "loss": 1.0012, |
| "step": 1793 |
| }, |
| { |
| "epoch": 2.3, |
| "learning_rate": 2.7456217756047044e-06, |
| "loss": 0.9193, |
| "step": 1794 |
| }, |
| { |
| "epoch": 2.3, |
| "learning_rate": 2.7361114619862105e-06, |
| "loss": 1.0385, |
| "step": 1795 |
| }, |
| { |
| "epoch": 2.3, |
| "learning_rate": 2.7266150367669953e-06, |
| "loss": 0.9454, |
| "step": 1796 |
| }, |
| { |
| "epoch": 2.3, |
| "learning_rate": 2.717132518104023e-06, |
| "loss": 0.9185, |
| "step": 1797 |
| }, |
| { |
| "epoch": 2.3, |
| "learning_rate": 2.707663924127657e-06, |
| "loss": 0.8957, |
| "step": 1798 |
| }, |
| { |
| "epoch": 2.3, |
| "learning_rate": 2.698209272941659e-06, |
| "loss": 1.0563, |
| "step": 1799 |
| }, |
| { |
| "epoch": 2.3, |
| "learning_rate": 2.68876858262311e-06, |
| "loss": 0.9441, |
| "step": 1800 |
| }, |
| { |
| "epoch": 2.31, |
| "learning_rate": 2.679341871222411e-06, |
| "loss": 0.8075, |
| "step": 1801 |
| }, |
| { |
| "epoch": 2.31, |
| "learning_rate": 2.66992915676323e-06, |
| "loss": 0.9214, |
| "step": 1802 |
| }, |
| { |
| "epoch": 2.31, |
| "learning_rate": 2.6605304572424793e-06, |
| "loss": 0.9836, |
| "step": 1803 |
| }, |
| { |
| "epoch": 2.31, |
| "learning_rate": 2.6511457906302685e-06, |
| "loss": 0.9149, |
| "step": 1804 |
| }, |
| { |
| "epoch": 2.31, |
| "learning_rate": 2.641775174869882e-06, |
| "loss": 0.9025, |
| "step": 1805 |
| }, |
| { |
| "epoch": 2.31, |
| "learning_rate": 2.6324186278777285e-06, |
| "loss": 0.866, |
| "step": 1806 |
| }, |
| { |
| "epoch": 2.31, |
| "learning_rate": 2.623076167543337e-06, |
| "loss": 0.9417, |
| "step": 1807 |
| }, |
| { |
| "epoch": 2.31, |
| "learning_rate": 2.6137478117292792e-06, |
| "loss": 1.0292, |
| "step": 1808 |
| }, |
| { |
| "epoch": 2.32, |
| "learning_rate": 2.604433578271185e-06, |
| "loss": 0.9518, |
| "step": 1809 |
| }, |
| { |
| "epoch": 2.32, |
| "learning_rate": 2.5951334849776576e-06, |
| "loss": 0.8617, |
| "step": 1810 |
| }, |
| { |
| "epoch": 2.32, |
| "learning_rate": 2.5858475496302813e-06, |
| "loss": 0.893, |
| "step": 1811 |
| }, |
| { |
| "epoch": 2.32, |
| "learning_rate": 2.5765757899835664e-06, |
| "loss": 0.9568, |
| "step": 1812 |
| }, |
| { |
| "epoch": 2.32, |
| "learning_rate": 2.5673182237649187e-06, |
| "loss": 0.9372, |
| "step": 1813 |
| }, |
| { |
| "epoch": 2.32, |
| "learning_rate": 2.5580748686745995e-06, |
| "loss": 0.9749, |
| "step": 1814 |
| }, |
| { |
| "epoch": 2.32, |
| "learning_rate": 2.548845742385717e-06, |
| "loss": 1.076, |
| "step": 1815 |
| }, |
| { |
| "epoch": 2.32, |
| "learning_rate": 2.539630862544151e-06, |
| "loss": 0.9242, |
| "step": 1816 |
| }, |
| { |
| "epoch": 2.33, |
| "learning_rate": 2.530430246768567e-06, |
| "loss": 0.9107, |
| "step": 1817 |
| }, |
| { |
| "epoch": 2.33, |
| "learning_rate": 2.5212439126503364e-06, |
| "loss": 0.9187, |
| "step": 1818 |
| }, |
| { |
| "epoch": 2.33, |
| "learning_rate": 2.5120718777535345e-06, |
| "loss": 0.8949, |
| "step": 1819 |
| }, |
| { |
| "epoch": 2.33, |
| "learning_rate": 2.5029141596148985e-06, |
| "loss": 1.0752, |
| "step": 1820 |
| }, |
| { |
| "epoch": 2.33, |
| "learning_rate": 2.493770775743789e-06, |
| "loss": 0.9732, |
| "step": 1821 |
| }, |
| { |
| "epoch": 2.33, |
| "learning_rate": 2.4846417436221536e-06, |
| "loss": 0.9328, |
| "step": 1822 |
| }, |
| { |
| "epoch": 2.33, |
| "learning_rate": 2.4755270807045174e-06, |
| "loss": 0.9301, |
| "step": 1823 |
| }, |
| { |
| "epoch": 2.33, |
| "learning_rate": 2.4664268044179075e-06, |
| "loss": 0.9307, |
| "step": 1824 |
| }, |
| { |
| "epoch": 2.34, |
| "learning_rate": 2.45734093216187e-06, |
| "loss": 0.9487, |
| "step": 1825 |
| }, |
| { |
| "epoch": 2.34, |
| "learning_rate": 2.4482694813083908e-06, |
| "loss": 0.9483, |
| "step": 1826 |
| }, |
| { |
| "epoch": 2.34, |
| "learning_rate": 2.439212469201889e-06, |
| "loss": 0.8474, |
| "step": 1827 |
| }, |
| { |
| "epoch": 2.34, |
| "learning_rate": 2.4301699131591817e-06, |
| "loss": 0.8025, |
| "step": 1828 |
| }, |
| { |
| "epoch": 2.34, |
| "learning_rate": 2.421141830469439e-06, |
| "loss": 0.9041, |
| "step": 1829 |
| }, |
| { |
| "epoch": 2.34, |
| "learning_rate": 2.4121282383941637e-06, |
| "loss": 0.9092, |
| "step": 1830 |
| }, |
| { |
| "epoch": 2.34, |
| "learning_rate": 2.403129154167153e-06, |
| "loss": 1.012, |
| "step": 1831 |
| }, |
| { |
| "epoch": 2.34, |
| "learning_rate": 2.3941445949944542e-06, |
| "loss": 0.9348, |
| "step": 1832 |
| }, |
| { |
| "epoch": 2.35, |
| "learning_rate": 2.3851745780543634e-06, |
| "loss": 0.9443, |
| "step": 1833 |
| }, |
| { |
| "epoch": 2.35, |
| "learning_rate": 2.376219120497353e-06, |
| "loss": 0.9222, |
| "step": 1834 |
| }, |
| { |
| "epoch": 2.35, |
| "learning_rate": 2.367278239446068e-06, |
| "loss": 0.9394, |
| "step": 1835 |
| }, |
| { |
| "epoch": 2.35, |
| "learning_rate": 2.3583519519952815e-06, |
| "loss": 0.8816, |
| "step": 1836 |
| }, |
| { |
| "epoch": 2.35, |
| "learning_rate": 2.349440275211865e-06, |
| "loss": 0.9472, |
| "step": 1837 |
| }, |
| { |
| "epoch": 2.35, |
| "learning_rate": 2.3405432261347525e-06, |
| "loss": 0.9385, |
| "step": 1838 |
| }, |
| { |
| "epoch": 2.35, |
| "learning_rate": 2.331660821774915e-06, |
| "loss": 1.0436, |
| "step": 1839 |
| }, |
| { |
| "epoch": 2.36, |
| "learning_rate": 2.3227930791153096e-06, |
| "loss": 0.8765, |
| "step": 1840 |
| }, |
| { |
| "epoch": 2.36, |
| "learning_rate": 2.3139400151108825e-06, |
| "loss": 0.9332, |
| "step": 1841 |
| }, |
| { |
| "epoch": 2.36, |
| "learning_rate": 2.3051016466884946e-06, |
| "loss": 0.8792, |
| "step": 1842 |
| }, |
| { |
| "epoch": 2.36, |
| "learning_rate": 2.296277990746918e-06, |
| "loss": 0.869, |
| "step": 1843 |
| }, |
| { |
| "epoch": 2.36, |
| "learning_rate": 2.2874690641567952e-06, |
| "loss": 0.9428, |
| "step": 1844 |
| }, |
| { |
| "epoch": 2.36, |
| "learning_rate": 2.278674883760604e-06, |
| "loss": 0.9838, |
| "step": 1845 |
| }, |
| { |
| "epoch": 2.36, |
| "learning_rate": 2.26989546637263e-06, |
| "loss": 0.8958, |
| "step": 1846 |
| }, |
| { |
| "epoch": 2.36, |
| "learning_rate": 2.2611308287789347e-06, |
| "loss": 0.9466, |
| "step": 1847 |
| }, |
| { |
| "epoch": 2.37, |
| "learning_rate": 2.252380987737307e-06, |
| "loss": 0.9677, |
| "step": 1848 |
| }, |
| { |
| "epoch": 2.37, |
| "learning_rate": 2.2436459599772676e-06, |
| "loss": 0.9, |
| "step": 1849 |
| }, |
| { |
| "epoch": 2.37, |
| "learning_rate": 2.234925762199992e-06, |
| "loss": 0.8713, |
| "step": 1850 |
| }, |
| { |
| "epoch": 2.37, |
| "learning_rate": 2.2262204110783213e-06, |
| "loss": 1.0113, |
| "step": 1851 |
| }, |
| { |
| "epoch": 2.37, |
| "learning_rate": 2.217529923256694e-06, |
| "loss": 0.9633, |
| "step": 1852 |
| }, |
| { |
| "epoch": 2.37, |
| "learning_rate": 2.2088543153511368e-06, |
| "loss": 0.9615, |
| "step": 1853 |
| }, |
| { |
| "epoch": 2.37, |
| "learning_rate": 2.200193603949229e-06, |
| "loss": 0.8904, |
| "step": 1854 |
| }, |
| { |
| "epoch": 2.37, |
| "learning_rate": 2.191547805610066e-06, |
| "loss": 0.9729, |
| "step": 1855 |
| }, |
| { |
| "epoch": 2.38, |
| "learning_rate": 2.1829169368642213e-06, |
| "loss": 0.9741, |
| "step": 1856 |
| }, |
| { |
| "epoch": 2.38, |
| "learning_rate": 2.174301014213741e-06, |
| "loss": 0.9176, |
| "step": 1857 |
| }, |
| { |
| "epoch": 2.38, |
| "learning_rate": 2.165700054132076e-06, |
| "loss": 0.9284, |
| "step": 1858 |
| }, |
| { |
| "epoch": 2.38, |
| "learning_rate": 2.157114073064087e-06, |
| "loss": 0.9563, |
| "step": 1859 |
| }, |
| { |
| "epoch": 2.38, |
| "learning_rate": 2.148543087425979e-06, |
| "loss": 1.036, |
| "step": 1860 |
| }, |
| { |
| "epoch": 2.38, |
| "learning_rate": 2.139987113605294e-06, |
| "loss": 0.9852, |
| "step": 1861 |
| }, |
| { |
| "epoch": 2.38, |
| "learning_rate": 2.1314461679608724e-06, |
| "loss": 0.935, |
| "step": 1862 |
| }, |
| { |
| "epoch": 2.38, |
| "learning_rate": 2.1229202668228197e-06, |
| "loss": 0.9547, |
| "step": 1863 |
| }, |
| { |
| "epoch": 2.39, |
| "learning_rate": 2.114409426492471e-06, |
| "loss": 0.9068, |
| "step": 1864 |
| }, |
| { |
| "epoch": 2.39, |
| "learning_rate": 2.105913663242378e-06, |
| "loss": 0.9174, |
| "step": 1865 |
| }, |
| { |
| "epoch": 2.39, |
| "learning_rate": 2.0974329933162505e-06, |
| "loss": 1.0306, |
| "step": 1866 |
| }, |
| { |
| "epoch": 2.39, |
| "learning_rate": 2.088967432928958e-06, |
| "loss": 0.8501, |
| "step": 1867 |
| }, |
| { |
| "epoch": 2.39, |
| "learning_rate": 2.0805169982664607e-06, |
| "loss": 0.7826, |
| "step": 1868 |
| }, |
| { |
| "epoch": 2.39, |
| "learning_rate": 2.0720817054858124e-06, |
| "loss": 0.9202, |
| "step": 1869 |
| }, |
| { |
| "epoch": 2.39, |
| "learning_rate": 2.063661570715113e-06, |
| "loss": 0.9672, |
| "step": 1870 |
| }, |
| { |
| "epoch": 2.39, |
| "learning_rate": 2.0552566100534787e-06, |
| "loss": 1.0096, |
| "step": 1871 |
| }, |
| { |
| "epoch": 2.4, |
| "learning_rate": 2.0468668395710145e-06, |
| "loss": 0.9538, |
| "step": 1872 |
| }, |
| { |
| "epoch": 2.4, |
| "learning_rate": 2.0384922753087843e-06, |
| "loss": 0.9075, |
| "step": 1873 |
| }, |
| { |
| "epoch": 2.4, |
| "learning_rate": 2.0301329332787676e-06, |
| "loss": 0.8199, |
| "step": 1874 |
| }, |
| { |
| "epoch": 2.4, |
| "learning_rate": 2.021788829463859e-06, |
| "loss": 0.9543, |
| "step": 1875 |
| }, |
| { |
| "epoch": 2.4, |
| "learning_rate": 2.013459979817797e-06, |
| "loss": 0.93, |
| "step": 1876 |
| }, |
| { |
| "epoch": 2.4, |
| "learning_rate": 2.0051464002651665e-06, |
| "loss": 0.8649, |
| "step": 1877 |
| }, |
| { |
| "epoch": 2.4, |
| "learning_rate": 1.9968481067013546e-06, |
| "loss": 0.9399, |
| "step": 1878 |
| }, |
| { |
| "epoch": 2.41, |
| "learning_rate": 1.988565114992519e-06, |
| "loss": 0.9868, |
| "step": 1879 |
| }, |
| { |
| "epoch": 2.41, |
| "learning_rate": 1.980297440975565e-06, |
| "loss": 0.8749, |
| "step": 1880 |
| }, |
| { |
| "epoch": 2.41, |
| "learning_rate": 1.9720451004581077e-06, |
| "loss": 0.8347, |
| "step": 1881 |
| }, |
| { |
| "epoch": 2.41, |
| "learning_rate": 1.9638081092184403e-06, |
| "loss": 0.8397, |
| "step": 1882 |
| }, |
| { |
| "epoch": 2.41, |
| "learning_rate": 1.9555864830055215e-06, |
| "loss": 1.0395, |
| "step": 1883 |
| }, |
| { |
| "epoch": 2.41, |
| "learning_rate": 1.9473802375389196e-06, |
| "loss": 0.883, |
| "step": 1884 |
| }, |
| { |
| "epoch": 2.41, |
| "learning_rate": 1.939189388508801e-06, |
| "loss": 0.9367, |
| "step": 1885 |
| }, |
| { |
| "epoch": 2.41, |
| "learning_rate": 1.931013951575893e-06, |
| "loss": 0.9554, |
| "step": 1886 |
| }, |
| { |
| "epoch": 2.42, |
| "learning_rate": 1.9228539423714564e-06, |
| "loss": 0.8722, |
| "step": 1887 |
| }, |
| { |
| "epoch": 2.42, |
| "learning_rate": 1.9147093764972547e-06, |
| "loss": 0.8608, |
| "step": 1888 |
| }, |
| { |
| "epoch": 2.42, |
| "learning_rate": 1.9065802695255265e-06, |
| "loss": 0.9613, |
| "step": 1889 |
| }, |
| { |
| "epoch": 2.42, |
| "learning_rate": 1.8984666369989403e-06, |
| "loss": 0.9606, |
| "step": 1890 |
| }, |
| { |
| "epoch": 2.42, |
| "learning_rate": 1.8903684944306e-06, |
| "loss": 0.8746, |
| "step": 1891 |
| }, |
| { |
| "epoch": 2.42, |
| "learning_rate": 1.8822858573039715e-06, |
| "loss": 0.8865, |
| "step": 1892 |
| }, |
| { |
| "epoch": 2.42, |
| "learning_rate": 1.8742187410728862e-06, |
| "loss": 0.9769, |
| "step": 1893 |
| }, |
| { |
| "epoch": 2.42, |
| "learning_rate": 1.8661671611614995e-06, |
| "loss": 0.9306, |
| "step": 1894 |
| }, |
| { |
| "epoch": 2.43, |
| "learning_rate": 1.8581311329642592e-06, |
| "loss": 0.9716, |
| "step": 1895 |
| }, |
| { |
| "epoch": 2.43, |
| "learning_rate": 1.8501106718458773e-06, |
| "loss": 0.9364, |
| "step": 1896 |
| }, |
| { |
| "epoch": 2.43, |
| "learning_rate": 1.8421057931413055e-06, |
| "loss": 0.9583, |
| "step": 1897 |
| }, |
| { |
| "epoch": 2.43, |
| "learning_rate": 1.8341165121556992e-06, |
| "loss": 0.8756, |
| "step": 1898 |
| }, |
| { |
| "epoch": 2.43, |
| "learning_rate": 1.8261428441643957e-06, |
| "loss": 0.9299, |
| "step": 1899 |
| }, |
| { |
| "epoch": 2.43, |
| "learning_rate": 1.8181848044128703e-06, |
| "loss": 1.0125, |
| "step": 1900 |
| }, |
| { |
| "epoch": 2.43, |
| "learning_rate": 1.8102424081167336e-06, |
| "loss": 0.9913, |
| "step": 1901 |
| }, |
| { |
| "epoch": 2.43, |
| "learning_rate": 1.8023156704616696e-06, |
| "loss": 0.9831, |
| "step": 1902 |
| }, |
| { |
| "epoch": 2.44, |
| "learning_rate": 1.7944046066034338e-06, |
| "loss": 0.788, |
| "step": 1903 |
| }, |
| { |
| "epoch": 2.44, |
| "learning_rate": 1.7865092316678112e-06, |
| "loss": 0.8973, |
| "step": 1904 |
| }, |
| { |
| "epoch": 2.44, |
| "learning_rate": 1.7786295607505888e-06, |
| "loss": 0.9919, |
| "step": 1905 |
| }, |
| { |
| "epoch": 2.44, |
| "learning_rate": 1.7707656089175286e-06, |
| "loss": 0.9962, |
| "step": 1906 |
| }, |
| { |
| "epoch": 2.44, |
| "learning_rate": 1.7629173912043397e-06, |
| "loss": 0.8648, |
| "step": 1907 |
| }, |
| { |
| "epoch": 2.44, |
| "learning_rate": 1.7550849226166432e-06, |
| "loss": 1.0581, |
| "step": 1908 |
| }, |
| { |
| "epoch": 2.44, |
| "learning_rate": 1.7472682181299572e-06, |
| "loss": 0.8977, |
| "step": 1909 |
| }, |
| { |
| "epoch": 2.44, |
| "learning_rate": 1.7394672926896449e-06, |
| "loss": 0.8958, |
| "step": 1910 |
| }, |
| { |
| "epoch": 2.45, |
| "learning_rate": 1.7316821612109136e-06, |
| "loss": 0.865, |
| "step": 1911 |
| }, |
| { |
| "epoch": 2.45, |
| "learning_rate": 1.7239128385787674e-06, |
| "loss": 0.9205, |
| "step": 1912 |
| }, |
| { |
| "epoch": 2.45, |
| "learning_rate": 1.7161593396479848e-06, |
| "loss": 0.9056, |
| "step": 1913 |
| }, |
| { |
| "epoch": 2.45, |
| "learning_rate": 1.7084216792430885e-06, |
| "loss": 0.9238, |
| "step": 1914 |
| }, |
| { |
| "epoch": 2.45, |
| "learning_rate": 1.70069987215832e-06, |
| "loss": 0.9981, |
| "step": 1915 |
| }, |
| { |
| "epoch": 2.45, |
| "learning_rate": 1.6929939331576096e-06, |
| "loss": 0.8721, |
| "step": 1916 |
| }, |
| { |
| "epoch": 2.45, |
| "learning_rate": 1.6853038769745466e-06, |
| "loss": 0.8501, |
| "step": 1917 |
| }, |
| { |
| "epoch": 2.46, |
| "learning_rate": 1.677629718312358e-06, |
| "loss": 0.8986, |
| "step": 1918 |
| }, |
| { |
| "epoch": 2.46, |
| "learning_rate": 1.6699714718438643e-06, |
| "loss": 0.7869, |
| "step": 1919 |
| }, |
| { |
| "epoch": 2.46, |
| "learning_rate": 1.6623291522114705e-06, |
| "loss": 0.9695, |
| "step": 1920 |
| }, |
| { |
| "epoch": 2.46, |
| "learning_rate": 1.6547027740271304e-06, |
| "loss": 0.9308, |
| "step": 1921 |
| }, |
| { |
| "epoch": 2.46, |
| "learning_rate": 1.6470923518723137e-06, |
| "loss": 0.9173, |
| "step": 1922 |
| }, |
| { |
| "epoch": 2.46, |
| "learning_rate": 1.6394979002979849e-06, |
| "loss": 0.929, |
| "step": 1923 |
| }, |
| { |
| "epoch": 2.46, |
| "learning_rate": 1.6319194338245725e-06, |
| "loss": 0.9998, |
| "step": 1924 |
| }, |
| { |
| "epoch": 2.46, |
| "learning_rate": 1.6243569669419434e-06, |
| "loss": 0.9173, |
| "step": 1925 |
| }, |
| { |
| "epoch": 2.47, |
| "learning_rate": 1.6168105141093737e-06, |
| "loss": 0.9882, |
| "step": 1926 |
| }, |
| { |
| "epoch": 2.47, |
| "learning_rate": 1.609280089755515e-06, |
| "loss": 0.9423, |
| "step": 1927 |
| }, |
| { |
| "epoch": 2.47, |
| "learning_rate": 1.6017657082783788e-06, |
| "loss": 0.8875, |
| "step": 1928 |
| }, |
| { |
| "epoch": 2.47, |
| "learning_rate": 1.5942673840453038e-06, |
| "loss": 1.0213, |
| "step": 1929 |
| }, |
| { |
| "epoch": 2.47, |
| "learning_rate": 1.5867851313929229e-06, |
| "loss": 0.9758, |
| "step": 1930 |
| }, |
| { |
| "epoch": 2.47, |
| "learning_rate": 1.5793189646271434e-06, |
| "loss": 0.9056, |
| "step": 1931 |
| }, |
| { |
| "epoch": 2.47, |
| "learning_rate": 1.5718688980231168e-06, |
| "loss": 0.8455, |
| "step": 1932 |
| }, |
| { |
| "epoch": 2.47, |
| "learning_rate": 1.5644349458252117e-06, |
| "loss": 0.9708, |
| "step": 1933 |
| }, |
| { |
| "epoch": 2.48, |
| "learning_rate": 1.5570171222469854e-06, |
| "loss": 0.9892, |
| "step": 1934 |
| }, |
| { |
| "epoch": 2.48, |
| "learning_rate": 1.549615441471153e-06, |
| "loss": 0.9308, |
| "step": 1935 |
| }, |
| { |
| "epoch": 2.48, |
| "learning_rate": 1.5422299176495782e-06, |
| "loss": 0.9043, |
| "step": 1936 |
| }, |
| { |
| "epoch": 2.48, |
| "learning_rate": 1.5348605649032166e-06, |
| "loss": 0.9824, |
| "step": 1937 |
| }, |
| { |
| "epoch": 2.48, |
| "learning_rate": 1.5275073973221154e-06, |
| "loss": 0.9638, |
| "step": 1938 |
| }, |
| { |
| "epoch": 2.48, |
| "learning_rate": 1.520170428965374e-06, |
| "loss": 0.9408, |
| "step": 1939 |
| }, |
| { |
| "epoch": 2.48, |
| "learning_rate": 1.5128496738611187e-06, |
| "loss": 0.9569, |
| "step": 1940 |
| }, |
| { |
| "epoch": 2.48, |
| "learning_rate": 1.5055451460064752e-06, |
| "loss": 1.0012, |
| "step": 1941 |
| }, |
| { |
| "epoch": 2.49, |
| "learning_rate": 1.4982568593675462e-06, |
| "loss": 0.979, |
| "step": 1942 |
| }, |
| { |
| "epoch": 2.49, |
| "learning_rate": 1.490984827879378e-06, |
| "loss": 0.9667, |
| "step": 1943 |
| }, |
| { |
| "epoch": 2.49, |
| "learning_rate": 1.4837290654459425e-06, |
| "loss": 0.9506, |
| "step": 1944 |
| }, |
| { |
| "epoch": 2.49, |
| "learning_rate": 1.4764895859400952e-06, |
| "loss": 0.879, |
| "step": 1945 |
| }, |
| { |
| "epoch": 2.49, |
| "learning_rate": 1.469266403203573e-06, |
| "loss": 0.8935, |
| "step": 1946 |
| }, |
| { |
| "epoch": 2.49, |
| "learning_rate": 1.4620595310469421e-06, |
| "loss": 0.8769, |
| "step": 1947 |
| }, |
| { |
| "epoch": 2.49, |
| "learning_rate": 1.45486898324959e-06, |
| "loss": 0.8919, |
| "step": 1948 |
| }, |
| { |
| "epoch": 2.49, |
| "learning_rate": 1.447694773559688e-06, |
| "loss": 0.9093, |
| "step": 1949 |
| }, |
| { |
| "epoch": 2.5, |
| "learning_rate": 1.4405369156941717e-06, |
| "loss": 0.9248, |
| "step": 1950 |
| }, |
| { |
| "epoch": 2.5, |
| "learning_rate": 1.4333954233387138e-06, |
| "loss": 0.8488, |
| "step": 1951 |
| }, |
| { |
| "epoch": 2.5, |
| "learning_rate": 1.426270310147695e-06, |
| "loss": 0.9435, |
| "step": 1952 |
| }, |
| { |
| "epoch": 2.5, |
| "learning_rate": 1.4191615897441746e-06, |
| "loss": 0.8837, |
| "step": 1953 |
| }, |
| { |
| "epoch": 2.5, |
| "learning_rate": 1.4120692757198806e-06, |
| "loss": 1.0227, |
| "step": 1954 |
| }, |
| { |
| "epoch": 2.5, |
| "learning_rate": 1.404993381635159e-06, |
| "loss": 0.8377, |
| "step": 1955 |
| }, |
| { |
| "epoch": 2.5, |
| "learning_rate": 1.3979339210189703e-06, |
| "loss": 0.9342, |
| "step": 1956 |
| }, |
| { |
| "epoch": 2.5, |
| "learning_rate": 1.3908909073688526e-06, |
| "loss": 1.0163, |
| "step": 1957 |
| }, |
| { |
| "epoch": 2.51, |
| "learning_rate": 1.3838643541508967e-06, |
| "loss": 0.897, |
| "step": 1958 |
| }, |
| { |
| "epoch": 2.51, |
| "learning_rate": 1.3768542747997215e-06, |
| "loss": 0.9731, |
| "step": 1959 |
| }, |
| { |
| "epoch": 2.51, |
| "learning_rate": 1.3698606827184513e-06, |
| "loss": 0.931, |
| "step": 1960 |
| }, |
| { |
| "epoch": 2.51, |
| "learning_rate": 1.3628835912786774e-06, |
| "loss": 0.9389, |
| "step": 1961 |
| }, |
| { |
| "epoch": 2.51, |
| "learning_rate": 1.3559230138204583e-06, |
| "loss": 0.9478, |
| "step": 1962 |
| }, |
| { |
| "epoch": 2.51, |
| "learning_rate": 1.3489789636522621e-06, |
| "loss": 0.8997, |
| "step": 1963 |
| }, |
| { |
| "epoch": 2.51, |
| "learning_rate": 1.3420514540509655e-06, |
| "loss": 0.933, |
| "step": 1964 |
| }, |
| { |
| "epoch": 2.52, |
| "learning_rate": 1.3351404982618199e-06, |
| "loss": 0.9028, |
| "step": 1965 |
| }, |
| { |
| "epoch": 2.52, |
| "learning_rate": 1.3282461094984233e-06, |
| "loss": 0.9937, |
| "step": 1966 |
| }, |
| { |
| "epoch": 2.52, |
| "learning_rate": 1.3213683009427004e-06, |
| "loss": 0.9106, |
| "step": 1967 |
| }, |
| { |
| "epoch": 2.52, |
| "learning_rate": 1.314507085744875e-06, |
| "loss": 0.892, |
| "step": 1968 |
| }, |
| { |
| "epoch": 2.52, |
| "learning_rate": 1.307662477023438e-06, |
| "loss": 0.8452, |
| "step": 1969 |
| }, |
| { |
| "epoch": 2.52, |
| "learning_rate": 1.3008344878651435e-06, |
| "loss": 0.8714, |
| "step": 1970 |
| }, |
| { |
| "epoch": 2.52, |
| "learning_rate": 1.2940231313249508e-06, |
| "loss": 0.9794, |
| "step": 1971 |
| }, |
| { |
| "epoch": 2.52, |
| "learning_rate": 1.2872284204260365e-06, |
| "loss": 0.8886, |
| "step": 1972 |
| }, |
| { |
| "epoch": 2.53, |
| "learning_rate": 1.2804503681597369e-06, |
| "loss": 0.9441, |
| "step": 1973 |
| }, |
| { |
| "epoch": 2.53, |
| "learning_rate": 1.273688987485544e-06, |
| "loss": 0.9442, |
| "step": 1974 |
| }, |
| { |
| "epoch": 2.53, |
| "learning_rate": 1.2669442913310725e-06, |
| "loss": 0.8971, |
| "step": 1975 |
| }, |
| { |
| "epoch": 2.53, |
| "learning_rate": 1.2602162925920403e-06, |
| "loss": 0.9381, |
| "step": 1976 |
| }, |
| { |
| "epoch": 2.53, |
| "learning_rate": 1.2535050041322293e-06, |
| "loss": 0.8865, |
| "step": 1977 |
| }, |
| { |
| "epoch": 2.53, |
| "learning_rate": 1.246810438783488e-06, |
| "loss": 0.9203, |
| "step": 1978 |
| }, |
| { |
| "epoch": 2.53, |
| "learning_rate": 1.2401326093456733e-06, |
| "loss": 0.9632, |
| "step": 1979 |
| }, |
| { |
| "epoch": 2.53, |
| "learning_rate": 1.2334715285866616e-06, |
| "loss": 1.0848, |
| "step": 1980 |
| }, |
| { |
| "epoch": 2.54, |
| "learning_rate": 1.2268272092422895e-06, |
| "loss": 0.9116, |
| "step": 1981 |
| }, |
| { |
| "epoch": 2.54, |
| "learning_rate": 1.220199664016355e-06, |
| "loss": 0.9392, |
| "step": 1982 |
| }, |
| { |
| "epoch": 2.54, |
| "learning_rate": 1.213588905580584e-06, |
| "loss": 0.9275, |
| "step": 1983 |
| }, |
| { |
| "epoch": 2.54, |
| "learning_rate": 1.2069949465746044e-06, |
| "loss": 0.8655, |
| "step": 1984 |
| }, |
| { |
| "epoch": 2.54, |
| "learning_rate": 1.2004177996059208e-06, |
| "loss": 0.8951, |
| "step": 1985 |
| }, |
| { |
| "epoch": 2.54, |
| "learning_rate": 1.1938574772499056e-06, |
| "loss": 0.9523, |
| "step": 1986 |
| }, |
| { |
| "epoch": 2.54, |
| "learning_rate": 1.1873139920497445e-06, |
| "loss": 0.9707, |
| "step": 1987 |
| }, |
| { |
| "epoch": 2.54, |
| "learning_rate": 1.1807873565164507e-06, |
| "loss": 0.9084, |
| "step": 1988 |
| }, |
| { |
| "epoch": 2.55, |
| "learning_rate": 1.1742775831288056e-06, |
| "loss": 0.8355, |
| "step": 1989 |
| }, |
| { |
| "epoch": 2.55, |
| "learning_rate": 1.1677846843333574e-06, |
| "loss": 1.0318, |
| "step": 1990 |
| }, |
| { |
| "epoch": 2.55, |
| "learning_rate": 1.161308672544389e-06, |
| "loss": 0.9273, |
| "step": 1991 |
| }, |
| { |
| "epoch": 2.55, |
| "learning_rate": 1.1548495601438981e-06, |
| "loss": 0.9532, |
| "step": 1992 |
| }, |
| { |
| "epoch": 2.55, |
| "learning_rate": 1.1484073594815691e-06, |
| "loss": 0.8998, |
| "step": 1993 |
| }, |
| { |
| "epoch": 2.55, |
| "learning_rate": 1.141982082874752e-06, |
| "loss": 0.8319, |
| "step": 1994 |
| }, |
| { |
| "epoch": 2.55, |
| "learning_rate": 1.1355737426084335e-06, |
| "loss": 0.911, |
| "step": 1995 |
| }, |
| { |
| "epoch": 2.55, |
| "learning_rate": 1.1291823509352295e-06, |
| "loss": 0.9495, |
| "step": 1996 |
| }, |
| { |
| "epoch": 2.56, |
| "learning_rate": 1.1228079200753394e-06, |
| "loss": 0.9539, |
| "step": 1997 |
| }, |
| { |
| "epoch": 2.56, |
| "learning_rate": 1.1164504622165407e-06, |
| "loss": 0.9018, |
| "step": 1998 |
| }, |
| { |
| "epoch": 2.56, |
| "learning_rate": 1.1101099895141542e-06, |
| "loss": 0.877, |
| "step": 1999 |
| }, |
| { |
| "epoch": 2.56, |
| "learning_rate": 1.103786514091031e-06, |
| "loss": 0.9637, |
| "step": 2000 |
| }, |
| { |
| "epoch": 2.56, |
| "learning_rate": 1.097480048037518e-06, |
| "loss": 0.9737, |
| "step": 2001 |
| }, |
| { |
| "epoch": 2.56, |
| "learning_rate": 1.0911906034114472e-06, |
| "loss": 0.9838, |
| "step": 2002 |
| }, |
| { |
| "epoch": 2.56, |
| "learning_rate": 1.0849181922380947e-06, |
| "loss": 0.8776, |
| "step": 2003 |
| }, |
| { |
| "epoch": 2.57, |
| "learning_rate": 1.078662826510185e-06, |
| "loss": 0.9595, |
| "step": 2004 |
| }, |
| { |
| "epoch": 2.57, |
| "learning_rate": 1.0724245181878369e-06, |
| "loss": 0.907, |
| "step": 2005 |
| }, |
| { |
| "epoch": 2.57, |
| "learning_rate": 1.0662032791985644e-06, |
| "loss": 0.825, |
| "step": 2006 |
| }, |
| { |
| "epoch": 2.57, |
| "learning_rate": 1.059999121437244e-06, |
| "loss": 0.9094, |
| "step": 2007 |
| }, |
| { |
| "epoch": 2.57, |
| "learning_rate": 1.0538120567660904e-06, |
| "loss": 0.976, |
| "step": 2008 |
| }, |
| { |
| "epoch": 2.57, |
| "learning_rate": 1.0476420970146406e-06, |
| "loss": 0.8653, |
| "step": 2009 |
| }, |
| { |
| "epoch": 2.57, |
| "learning_rate": 1.0414892539797272e-06, |
| "loss": 0.9359, |
| "step": 2010 |
| }, |
| { |
| "epoch": 2.57, |
| "learning_rate": 1.0353535394254477e-06, |
| "loss": 0.9514, |
| "step": 2011 |
| }, |
| { |
| "epoch": 2.58, |
| "learning_rate": 1.0292349650831656e-06, |
| "loss": 0.9776, |
| "step": 2012 |
| }, |
| { |
| "epoch": 2.58, |
| "learning_rate": 1.0231335426514555e-06, |
| "loss": 0.8467, |
| "step": 2013 |
| }, |
| { |
| "epoch": 2.58, |
| "learning_rate": 1.0170492837961133e-06, |
| "loss": 0.9453, |
| "step": 2014 |
| }, |
| { |
| "epoch": 2.58, |
| "learning_rate": 1.010982200150108e-06, |
| "loss": 0.935, |
| "step": 2015 |
| }, |
| { |
| "epoch": 2.58, |
| "learning_rate": 1.0049323033135738e-06, |
| "loss": 0.9227, |
| "step": 2016 |
| }, |
| { |
| "epoch": 2.58, |
| "learning_rate": 9.988996048537848e-07, |
| "loss": 0.8756, |
| "step": 2017 |
| }, |
| { |
| "epoch": 2.58, |
| "learning_rate": 9.928841163051318e-07, |
| "loss": 0.9422, |
| "step": 2018 |
| }, |
| { |
| "epoch": 2.58, |
| "learning_rate": 9.868858491690947e-07, |
| "loss": 0.9629, |
| "step": 2019 |
| }, |
| { |
| "epoch": 2.59, |
| "learning_rate": 9.809048149142409e-07, |
| "loss": 0.9206, |
| "step": 2020 |
| }, |
| { |
| "epoch": 2.59, |
| "learning_rate": 9.749410249761715e-07, |
| "loss": 0.9085, |
| "step": 2021 |
| }, |
| { |
| "epoch": 2.59, |
| "learning_rate": 9.68994490757531e-07, |
| "loss": 0.9067, |
| "step": 2022 |
| }, |
| { |
| "epoch": 2.59, |
| "learning_rate": 9.630652236279626e-07, |
| "loss": 0.9401, |
| "step": 2023 |
| }, |
| { |
| "epoch": 2.59, |
| "learning_rate": 9.571532349240976e-07, |
| "loss": 0.9138, |
| "step": 2024 |
| }, |
| { |
| "epoch": 2.59, |
| "learning_rate": 9.512585359495319e-07, |
| "loss": 1.0039, |
| "step": 2025 |
| }, |
| { |
| "epoch": 2.59, |
| "learning_rate": 9.453811379748068e-07, |
| "loss": 0.9856, |
| "step": 2026 |
| }, |
| { |
| "epoch": 2.59, |
| "learning_rate": 9.395210522373755e-07, |
| "loss": 1.0115, |
| "step": 2027 |
| }, |
| { |
| "epoch": 2.6, |
| "learning_rate": 9.336782899416031e-07, |
| "loss": 1.0045, |
| "step": 2028 |
| }, |
| { |
| "epoch": 2.6, |
| "learning_rate": 9.278528622587191e-07, |
| "loss": 0.9497, |
| "step": 2029 |
| }, |
| { |
| "epoch": 2.6, |
| "learning_rate": 9.220447803268228e-07, |
| "loss": 0.867, |
| "step": 2030 |
| }, |
| { |
| "epoch": 2.6, |
| "learning_rate": 9.162540552508392e-07, |
| "loss": 0.9456, |
| "step": 2031 |
| }, |
| { |
| "epoch": 2.6, |
| "learning_rate": 9.104806981025094e-07, |
| "loss": 0.96, |
| "step": 2032 |
| }, |
| { |
| "epoch": 2.6, |
| "learning_rate": 9.047247199203713e-07, |
| "loss": 0.8925, |
| "step": 2033 |
| }, |
| { |
| "epoch": 2.6, |
| "learning_rate": 8.989861317097304e-07, |
| "loss": 0.8728, |
| "step": 2034 |
| }, |
| { |
| "epoch": 2.6, |
| "learning_rate": 8.932649444426445e-07, |
| "loss": 0.9253, |
| "step": 2035 |
| }, |
| { |
| "epoch": 2.61, |
| "learning_rate": 8.875611690579033e-07, |
| "loss": 0.8838, |
| "step": 2036 |
| }, |
| { |
| "epoch": 2.61, |
| "learning_rate": 8.81874816460998e-07, |
| "loss": 0.9694, |
| "step": 2037 |
| }, |
| { |
| "epoch": 2.61, |
| "learning_rate": 8.762058975241194e-07, |
| "loss": 0.9279, |
| "step": 2038 |
| }, |
| { |
| "epoch": 2.61, |
| "learning_rate": 8.705544230861141e-07, |
| "loss": 1.0201, |
| "step": 2039 |
| }, |
| { |
| "epoch": 2.61, |
| "learning_rate": 8.649204039524816e-07, |
| "loss": 0.7873, |
| "step": 2040 |
| }, |
| { |
| "epoch": 2.61, |
| "learning_rate": 8.593038508953466e-07, |
| "loss": 0.9852, |
| "step": 2041 |
| }, |
| { |
| "epoch": 2.61, |
| "learning_rate": 8.537047746534377e-07, |
| "loss": 0.9373, |
| "step": 2042 |
| }, |
| { |
| "epoch": 2.62, |
| "learning_rate": 8.481231859320682e-07, |
| "loss": 0.9618, |
| "step": 2043 |
| }, |
| { |
| "epoch": 2.62, |
| "learning_rate": 8.425590954031171e-07, |
| "loss": 1.0313, |
| "step": 2044 |
| }, |
| { |
| "epoch": 2.62, |
| "learning_rate": 8.370125137050056e-07, |
| "loss": 0.9545, |
| "step": 2045 |
| }, |
| { |
| "epoch": 2.62, |
| "learning_rate": 8.314834514426806e-07, |
| "loss": 0.9007, |
| "step": 2046 |
| }, |
| { |
| "epoch": 2.62, |
| "learning_rate": 8.259719191875892e-07, |
| "loss": 0.9018, |
| "step": 2047 |
| }, |
| { |
| "epoch": 2.62, |
| "learning_rate": 8.20477927477662e-07, |
| "loss": 1.0339, |
| "step": 2048 |
| }, |
| { |
| "epoch": 2.62, |
| "learning_rate": 8.150014868172951e-07, |
| "loss": 1.0561, |
| "step": 2049 |
| }, |
| { |
| "epoch": 2.62, |
| "learning_rate": 8.095426076773261e-07, |
| "loss": 0.9442, |
| "step": 2050 |
| }, |
| { |
| "epoch": 2.63, |
| "learning_rate": 8.041013004950138e-07, |
| "loss": 0.9251, |
| "step": 2051 |
| }, |
| { |
| "epoch": 2.63, |
| "learning_rate": 7.986775756740206e-07, |
| "loss": 0.918, |
| "step": 2052 |
| }, |
| { |
| "epoch": 2.63, |
| "learning_rate": 7.932714435843947e-07, |
| "loss": 0.8165, |
| "step": 2053 |
| }, |
| { |
| "epoch": 2.63, |
| "learning_rate": 7.878829145625422e-07, |
| "loss": 0.8565, |
| "step": 2054 |
| }, |
| { |
| "epoch": 2.63, |
| "learning_rate": 7.825119989112173e-07, |
| "loss": 0.925, |
| "step": 2055 |
| }, |
| { |
| "epoch": 2.63, |
| "learning_rate": 7.771587068994935e-07, |
| "loss": 1.0064, |
| "step": 2056 |
| }, |
| { |
| "epoch": 2.63, |
| "learning_rate": 7.71823048762751e-07, |
| "loss": 0.947, |
| "step": 2057 |
| }, |
| { |
| "epoch": 2.63, |
| "learning_rate": 7.665050347026548e-07, |
| "loss": 1.0526, |
| "step": 2058 |
| }, |
| { |
| "epoch": 2.64, |
| "learning_rate": 7.612046748871327e-07, |
| "loss": 0.9703, |
| "step": 2059 |
| }, |
| { |
| "epoch": 2.64, |
| "learning_rate": 7.559219794503603e-07, |
| "loss": 0.9854, |
| "step": 2060 |
| }, |
| { |
| "epoch": 2.64, |
| "learning_rate": 7.50656958492737e-07, |
| "loss": 0.9629, |
| "step": 2061 |
| }, |
| { |
| "epoch": 2.64, |
| "learning_rate": 7.454096220808715e-07, |
| "loss": 0.9814, |
| "step": 2062 |
| }, |
| { |
| "epoch": 2.64, |
| "learning_rate": 7.401799802475574e-07, |
| "loss": 0.982, |
| "step": 2063 |
| }, |
| { |
| "epoch": 2.64, |
| "learning_rate": 7.349680429917594e-07, |
| "loss": 0.8733, |
| "step": 2064 |
| }, |
| { |
| "epoch": 2.64, |
| "learning_rate": 7.297738202785875e-07, |
| "loss": 0.86, |
| "step": 2065 |
| }, |
| { |
| "epoch": 2.64, |
| "learning_rate": 7.245973220392854e-07, |
| "loss": 0.8781, |
| "step": 2066 |
| }, |
| { |
| "epoch": 2.65, |
| "learning_rate": 7.194385581712071e-07, |
| "loss": 0.9554, |
| "step": 2067 |
| }, |
| { |
| "epoch": 2.65, |
| "learning_rate": 7.142975385377981e-07, |
| "loss": 0.9102, |
| "step": 2068 |
| }, |
| { |
| "epoch": 2.65, |
| "learning_rate": 7.09174272968578e-07, |
| "loss": 0.9443, |
| "step": 2069 |
| }, |
| { |
| "epoch": 2.65, |
| "learning_rate": 7.040687712591221e-07, |
| "loss": 0.975, |
| "step": 2070 |
| }, |
| { |
| "epoch": 2.65, |
| "learning_rate": 6.989810431710375e-07, |
| "loss": 1.0177, |
| "step": 2071 |
| }, |
| { |
| "epoch": 2.65, |
| "learning_rate": 6.939110984319541e-07, |
| "loss": 0.9926, |
| "step": 2072 |
| }, |
| { |
| "epoch": 2.65, |
| "learning_rate": 6.888589467354978e-07, |
| "loss": 0.9268, |
| "step": 2073 |
| }, |
| { |
| "epoch": 2.65, |
| "learning_rate": 6.838245977412717e-07, |
| "loss": 0.9057, |
| "step": 2074 |
| }, |
| { |
| "epoch": 2.66, |
| "learning_rate": 6.788080610748437e-07, |
| "loss": 0.9604, |
| "step": 2075 |
| }, |
| { |
| "epoch": 2.66, |
| "learning_rate": 6.738093463277262e-07, |
| "loss": 0.954, |
| "step": 2076 |
| }, |
| { |
| "epoch": 2.66, |
| "learning_rate": 6.688284630573538e-07, |
| "loss": 0.9154, |
| "step": 2077 |
| }, |
| { |
| "epoch": 2.66, |
| "learning_rate": 6.638654207870688e-07, |
| "loss": 0.894, |
| "step": 2078 |
| }, |
| { |
| "epoch": 2.66, |
| "learning_rate": 6.589202290061015e-07, |
| "loss": 0.9209, |
| "step": 2079 |
| }, |
| { |
| "epoch": 2.66, |
| "learning_rate": 6.539928971695531e-07, |
| "loss": 0.9184, |
| "step": 2080 |
| }, |
| { |
| "epoch": 2.66, |
| "learning_rate": 6.490834346983765e-07, |
| "loss": 0.8968, |
| "step": 2081 |
| }, |
| { |
| "epoch": 2.66, |
| "learning_rate": 6.441918509793565e-07, |
| "loss": 0.9123, |
| "step": 2082 |
| }, |
| { |
| "epoch": 2.67, |
| "learning_rate": 6.393181553651007e-07, |
| "loss": 0.9991, |
| "step": 2083 |
| }, |
| { |
| "epoch": 2.67, |
| "learning_rate": 6.344623571740083e-07, |
| "loss": 0.8547, |
| "step": 2084 |
| }, |
| { |
| "epoch": 2.67, |
| "learning_rate": 6.296244656902595e-07, |
| "loss": 0.8819, |
| "step": 2085 |
| }, |
| { |
| "epoch": 2.67, |
| "learning_rate": 6.248044901638028e-07, |
| "loss": 0.8958, |
| "step": 2086 |
| }, |
| { |
| "epoch": 2.67, |
| "learning_rate": 6.200024398103255e-07, |
| "loss": 0.9298, |
| "step": 2087 |
| }, |
| { |
| "epoch": 2.67, |
| "learning_rate": 6.152183238112464e-07, |
| "loss": 0.91, |
| "step": 2088 |
| }, |
| { |
| "epoch": 2.67, |
| "learning_rate": 6.104521513136941e-07, |
| "loss": 0.964, |
| "step": 2089 |
| }, |
| { |
| "epoch": 2.68, |
| "learning_rate": 6.057039314304824e-07, |
| "loss": 0.9043, |
| "step": 2090 |
| }, |
| { |
| "epoch": 2.68, |
| "learning_rate": 6.00973673240115e-07, |
| "loss": 0.89, |
| "step": 2091 |
| }, |
| { |
| "epoch": 2.68, |
| "learning_rate": 5.96261385786735e-07, |
| "loss": 0.892, |
| "step": 2092 |
| }, |
| { |
| "epoch": 2.68, |
| "learning_rate": 5.915670780801441e-07, |
| "loss": 0.8748, |
| "step": 2093 |
| }, |
| { |
| "epoch": 2.68, |
| "learning_rate": 5.868907590957507e-07, |
| "loss": 0.8928, |
| "step": 2094 |
| }, |
| { |
| "epoch": 2.68, |
| "learning_rate": 5.822324377745792e-07, |
| "loss": 0.9251, |
| "step": 2095 |
| }, |
| { |
| "epoch": 2.68, |
| "learning_rate": 5.775921230232407e-07, |
| "loss": 0.8978, |
| "step": 2096 |
| }, |
| { |
| "epoch": 2.68, |
| "learning_rate": 5.729698237139191e-07, |
| "loss": 0.8754, |
| "step": 2097 |
| }, |
| { |
| "epoch": 2.69, |
| "learning_rate": 5.683655486843465e-07, |
| "loss": 1.0626, |
| "step": 2098 |
| }, |
| { |
| "epoch": 2.69, |
| "learning_rate": 5.63779306737805e-07, |
| "loss": 1.0134, |
| "step": 2099 |
| }, |
| { |
| "epoch": 2.69, |
| "learning_rate": 5.592111066430828e-07, |
| "loss": 0.9338, |
| "step": 2100 |
| }, |
| { |
| "epoch": 2.69, |
| "learning_rate": 5.546609571344897e-07, |
| "loss": 0.8544, |
| "step": 2101 |
| }, |
| { |
| "epoch": 2.69, |
| "learning_rate": 5.501288669118077e-07, |
| "loss": 1.0095, |
| "step": 2102 |
| }, |
| { |
| "epoch": 2.69, |
| "learning_rate": 5.456148446402976e-07, |
| "loss": 0.9461, |
| "step": 2103 |
| }, |
| { |
| "epoch": 2.69, |
| "learning_rate": 5.411188989506733e-07, |
| "loss": 0.906, |
| "step": 2104 |
| }, |
| { |
| "epoch": 2.69, |
| "learning_rate": 5.366410384390897e-07, |
| "loss": 0.8949, |
| "step": 2105 |
| }, |
| { |
| "epoch": 2.7, |
| "learning_rate": 5.321812716671137e-07, |
| "loss": 0.8961, |
| "step": 2106 |
| }, |
| { |
| "epoch": 2.7, |
| "learning_rate": 5.277396071617313e-07, |
| "loss": 0.9842, |
| "step": 2107 |
| }, |
| { |
| "epoch": 2.7, |
| "learning_rate": 5.233160534153037e-07, |
| "loss": 0.8478, |
| "step": 2108 |
| }, |
| { |
| "epoch": 2.7, |
| "learning_rate": 5.189106188855753e-07, |
| "loss": 0.9286, |
| "step": 2109 |
| }, |
| { |
| "epoch": 2.7, |
| "learning_rate": 5.145233119956394e-07, |
| "loss": 0.9315, |
| "step": 2110 |
| }, |
| { |
| "epoch": 2.7, |
| "learning_rate": 5.101541411339351e-07, |
| "loss": 0.9396, |
| "step": 2111 |
| }, |
| { |
| "epoch": 2.7, |
| "learning_rate": 5.058031146542219e-07, |
| "loss": 0.9953, |
| "step": 2112 |
| }, |
| { |
| "epoch": 2.7, |
| "learning_rate": 5.0147024087557e-07, |
| "loss": 1.0198, |
| "step": 2113 |
| }, |
| { |
| "epoch": 2.71, |
| "learning_rate": 4.971555280823415e-07, |
| "loss": 1.0282, |
| "step": 2114 |
| }, |
| { |
| "epoch": 2.71, |
| "learning_rate": 4.92858984524175e-07, |
| "loss": 0.9356, |
| "step": 2115 |
| }, |
| { |
| "epoch": 2.71, |
| "learning_rate": 4.885806184159669e-07, |
| "loss": 0.8674, |
| "step": 2116 |
| }, |
| { |
| "epoch": 2.71, |
| "learning_rate": 4.843204379378663e-07, |
| "loss": 0.9352, |
| "step": 2117 |
| }, |
| { |
| "epoch": 2.71, |
| "learning_rate": 4.80078451235243e-07, |
| "loss": 0.9148, |
| "step": 2118 |
| }, |
| { |
| "epoch": 2.71, |
| "learning_rate": 4.7585466641868696e-07, |
| "loss": 0.9017, |
| "step": 2119 |
| }, |
| { |
| "epoch": 2.71, |
| "learning_rate": 4.7164909156398265e-07, |
| "loss": 0.9265, |
| "step": 2120 |
| }, |
| { |
| "epoch": 2.71, |
| "learning_rate": 4.674617347121013e-07, |
| "loss": 0.9299, |
| "step": 2121 |
| }, |
| { |
| "epoch": 2.72, |
| "learning_rate": 4.6329260386917654e-07, |
| "loss": 0.9358, |
| "step": 2122 |
| }, |
| { |
| "epoch": 2.72, |
| "learning_rate": 4.5914170700650184e-07, |
| "loss": 0.9032, |
| "step": 2123 |
| }, |
| { |
| "epoch": 2.72, |
| "learning_rate": 4.5500905206049663e-07, |
| "loss": 0.9696, |
| "step": 2124 |
| }, |
| { |
| "epoch": 2.72, |
| "learning_rate": 4.5089464693271467e-07, |
| "loss": 0.9912, |
| "step": 2125 |
| }, |
| { |
| "epoch": 2.72, |
| "learning_rate": 4.467984994898067e-07, |
| "loss": 0.8969, |
| "step": 2126 |
| }, |
| { |
| "epoch": 2.72, |
| "learning_rate": 4.427206175635201e-07, |
| "loss": 0.9538, |
| "step": 2127 |
| }, |
| { |
| "epoch": 2.72, |
| "learning_rate": 4.386610089506771e-07, |
| "loss": 0.9487, |
| "step": 2128 |
| }, |
| { |
| "epoch": 2.73, |
| "learning_rate": 4.3461968141316427e-07, |
| "loss": 0.9271, |
| "step": 2129 |
| }, |
| { |
| "epoch": 2.73, |
| "learning_rate": 4.305966426779118e-07, |
| "loss": 0.9953, |
| "step": 2130 |
| }, |
| { |
| "epoch": 2.73, |
| "learning_rate": 4.265919004368846e-07, |
| "loss": 0.962, |
| "step": 2131 |
| }, |
| { |
| "epoch": 2.73, |
| "learning_rate": 4.22605462347061e-07, |
| "loss": 0.932, |
| "step": 2132 |
| }, |
| { |
| "epoch": 2.73, |
| "learning_rate": 4.186373360304297e-07, |
| "loss": 1.0048, |
| "step": 2133 |
| }, |
| { |
| "epoch": 2.73, |
| "learning_rate": 4.1468752907395826e-07, |
| "loss": 0.7749, |
| "step": 2134 |
| }, |
| { |
| "epoch": 2.73, |
| "learning_rate": 4.107560490295992e-07, |
| "loss": 0.9691, |
| "step": 2135 |
| }, |
| { |
| "epoch": 2.73, |
| "learning_rate": 4.068429034142529e-07, |
| "loss": 0.8954, |
| "step": 2136 |
| }, |
| { |
| "epoch": 2.74, |
| "learning_rate": 4.0294809970977235e-07, |
| "loss": 0.8931, |
| "step": 2137 |
| }, |
| { |
| "epoch": 2.74, |
| "learning_rate": 3.990716453629373e-07, |
| "loss": 0.8449, |
| "step": 2138 |
| }, |
| { |
| "epoch": 2.74, |
| "learning_rate": 3.9521354778544794e-07, |
| "loss": 0.9297, |
| "step": 2139 |
| }, |
| { |
| "epoch": 2.74, |
| "learning_rate": 3.913738143538992e-07, |
| "loss": 0.9894, |
| "step": 2140 |
| }, |
| { |
| "epoch": 2.74, |
| "learning_rate": 3.8755245240978623e-07, |
| "loss": 0.8579, |
| "step": 2141 |
| }, |
| { |
| "epoch": 2.74, |
| "learning_rate": 3.8374946925946457e-07, |
| "loss": 0.8972, |
| "step": 2142 |
| }, |
| { |
| "epoch": 2.74, |
| "learning_rate": 3.799648721741622e-07, |
| "loss": 0.9176, |
| "step": 2143 |
| }, |
| { |
| "epoch": 2.74, |
| "learning_rate": 3.7619866838994324e-07, |
| "loss": 0.8867, |
| "step": 2144 |
| }, |
| { |
| "epoch": 2.75, |
| "learning_rate": 3.7245086510770965e-07, |
| "loss": 0.9033, |
| "step": 2145 |
| }, |
| { |
| "epoch": 2.75, |
| "learning_rate": 3.687214694931829e-07, |
| "loss": 1.0276, |
| "step": 2146 |
| }, |
| { |
| "epoch": 2.75, |
| "learning_rate": 3.650104886768868e-07, |
| "loss": 0.8234, |
| "step": 2147 |
| }, |
| { |
| "epoch": 2.75, |
| "learning_rate": 3.6131792975413473e-07, |
| "loss": 0.914, |
| "step": 2148 |
| }, |
| { |
| "epoch": 2.75, |
| "learning_rate": 3.576437997850235e-07, |
| "loss": 0.9492, |
| "step": 2149 |
| }, |
| { |
| "epoch": 2.75, |
| "learning_rate": 3.5398810579440834e-07, |
| "loss": 0.9243, |
| "step": 2150 |
| }, |
| { |
| "epoch": 2.75, |
| "learning_rate": 3.5035085477190143e-07, |
| "loss": 0.963, |
| "step": 2151 |
| }, |
| { |
| "epoch": 2.75, |
| "learning_rate": 3.4673205367184437e-07, |
| "loss": 0.8881, |
| "step": 2152 |
| }, |
| { |
| "epoch": 2.76, |
| "learning_rate": 3.431317094133102e-07, |
| "loss": 0.9065, |
| "step": 2153 |
| }, |
| { |
| "epoch": 2.76, |
| "learning_rate": 3.3954982888007915e-07, |
| "loss": 0.9524, |
| "step": 2154 |
| }, |
| { |
| "epoch": 2.76, |
| "learning_rate": 3.3598641892063075e-07, |
| "loss": 1.0529, |
| "step": 2155 |
| }, |
| { |
| "epoch": 2.76, |
| "learning_rate": 3.3244148634812733e-07, |
| "loss": 1.0048, |
| "step": 2156 |
| }, |
| { |
| "epoch": 2.76, |
| "learning_rate": 3.28915037940406e-07, |
| "loss": 1.022, |
| "step": 2157 |
| }, |
| { |
| "epoch": 2.76, |
| "learning_rate": 3.2540708043995674e-07, |
| "loss": 0.9923, |
| "step": 2158 |
| }, |
| { |
| "epoch": 2.76, |
| "learning_rate": 3.219176205539243e-07, |
| "loss": 0.939, |
| "step": 2159 |
| }, |
| { |
| "epoch": 2.76, |
| "learning_rate": 3.184466649540763e-07, |
| "loss": 0.8588, |
| "step": 2160 |
| }, |
| { |
| "epoch": 2.77, |
| "learning_rate": 3.149942202768086e-07, |
| "loss": 0.9248, |
| "step": 2161 |
| }, |
| { |
| "epoch": 2.77, |
| "learning_rate": 3.1156029312312097e-07, |
| "loss": 0.8605, |
| "step": 2162 |
| }, |
| { |
| "epoch": 2.77, |
| "learning_rate": 3.08144890058607e-07, |
| "loss": 0.82, |
| "step": 2163 |
| }, |
| { |
| "epoch": 2.77, |
| "learning_rate": 3.0474801761344654e-07, |
| "loss": 1.0657, |
| "step": 2164 |
| }, |
| { |
| "epoch": 2.77, |
| "learning_rate": 3.013696822823864e-07, |
| "loss": 0.9589, |
| "step": 2165 |
| }, |
| { |
| "epoch": 2.77, |
| "learning_rate": 2.9800989052472863e-07, |
| "loss": 0.8637, |
| "step": 2166 |
| }, |
| { |
| "epoch": 2.77, |
| "learning_rate": 2.94668648764328e-07, |
| "loss": 0.8438, |
| "step": 2167 |
| }, |
| { |
| "epoch": 2.78, |
| "learning_rate": 2.913459633895621e-07, |
| "loss": 0.899, |
| "step": 2168 |
| }, |
| { |
| "epoch": 2.78, |
| "learning_rate": 2.8804184075333695e-07, |
| "loss": 0.8585, |
| "step": 2169 |
| }, |
| { |
| "epoch": 2.78, |
| "learning_rate": 2.847562871730647e-07, |
| "loss": 1.0156, |
| "step": 2170 |
| }, |
| { |
| "epoch": 2.78, |
| "learning_rate": 2.814893089306514e-07, |
| "loss": 0.868, |
| "step": 2171 |
| }, |
| { |
| "epoch": 2.78, |
| "learning_rate": 2.782409122724916e-07, |
| "loss": 0.9976, |
| "step": 2172 |
| }, |
| { |
| "epoch": 2.78, |
| "learning_rate": 2.750111034094494e-07, |
| "loss": 0.9533, |
| "step": 2173 |
| }, |
| { |
| "epoch": 2.78, |
| "learning_rate": 2.717998885168471e-07, |
| "loss": 0.9652, |
| "step": 2174 |
| }, |
| { |
| "epoch": 2.78, |
| "learning_rate": 2.6860727373446473e-07, |
| "loss": 1.0484, |
| "step": 2175 |
| }, |
| { |
| "epoch": 2.79, |
| "learning_rate": 2.6543326516650815e-07, |
| "loss": 0.9585, |
| "step": 2176 |
| }, |
| { |
| "epoch": 2.79, |
| "learning_rate": 2.6227786888161634e-07, |
| "loss": 0.9765, |
| "step": 2177 |
| }, |
| { |
| "epoch": 2.79, |
| "learning_rate": 2.591410909128389e-07, |
| "loss": 0.9476, |
| "step": 2178 |
| }, |
| { |
| "epoch": 2.79, |
| "learning_rate": 2.560229372576284e-07, |
| "loss": 0.8876, |
| "step": 2179 |
| }, |
| { |
| "epoch": 2.79, |
| "learning_rate": 2.5292341387782695e-07, |
| "loss": 0.9232, |
| "step": 2180 |
| }, |
| { |
| "epoch": 2.79, |
| "learning_rate": 2.4984252669965734e-07, |
| "loss": 0.8714, |
| "step": 2181 |
| }, |
| { |
| "epoch": 2.79, |
| "learning_rate": 2.4678028161370994e-07, |
| "loss": 0.9322, |
| "step": 2182 |
| }, |
| { |
| "epoch": 2.79, |
| "learning_rate": 2.4373668447493225e-07, |
| "loss": 0.9322, |
| "step": 2183 |
| }, |
| { |
| "epoch": 2.8, |
| "learning_rate": 2.40711741102615e-07, |
| "loss": 1.0387, |
| "step": 2184 |
| }, |
| { |
| "epoch": 2.8, |
| "learning_rate": 2.377054572803883e-07, |
| "loss": 1.0648, |
| "step": 2185 |
| }, |
| { |
| "epoch": 2.8, |
| "learning_rate": 2.347178387561999e-07, |
| "loss": 0.9778, |
| "step": 2186 |
| }, |
| { |
| "epoch": 2.8, |
| "learning_rate": 2.3174889124231158e-07, |
| "loss": 1.0032, |
| "step": 2187 |
| }, |
| { |
| "epoch": 2.8, |
| "learning_rate": 2.2879862041529034e-07, |
| "loss": 0.9426, |
| "step": 2188 |
| }, |
| { |
| "epoch": 2.8, |
| "learning_rate": 2.2586703191598836e-07, |
| "loss": 0.8759, |
| "step": 2189 |
| }, |
| { |
| "epoch": 2.8, |
| "learning_rate": 2.229541313495409e-07, |
| "loss": 0.8073, |
| "step": 2190 |
| }, |
| { |
| "epoch": 2.8, |
| "learning_rate": 2.2005992428535184e-07, |
| "loss": 0.9205, |
| "step": 2191 |
| }, |
| { |
| "epoch": 2.81, |
| "learning_rate": 2.171844162570824e-07, |
| "loss": 0.9681, |
| "step": 2192 |
| }, |
| { |
| "epoch": 2.81, |
| "learning_rate": 2.143276127626437e-07, |
| "loss": 0.8642, |
| "step": 2193 |
| }, |
| { |
| "epoch": 2.81, |
| "learning_rate": 2.1148951926418193e-07, |
| "loss": 1.0066, |
| "step": 2194 |
| }, |
| { |
| "epoch": 2.81, |
| "learning_rate": 2.0867014118806983e-07, |
| "loss": 0.9553, |
| "step": 2195 |
| }, |
| { |
| "epoch": 2.81, |
| "learning_rate": 2.0586948392489868e-07, |
| "loss": 0.8729, |
| "step": 2196 |
| }, |
| { |
| "epoch": 2.81, |
| "learning_rate": 2.0308755282946624e-07, |
| "loss": 1.0024, |
| "step": 2197 |
| }, |
| { |
| "epoch": 2.81, |
| "learning_rate": 2.0032435322076437e-07, |
| "loss": 0.879, |
| "step": 2198 |
| }, |
| { |
| "epoch": 2.81, |
| "learning_rate": 1.9757989038197146e-07, |
| "loss": 0.9339, |
| "step": 2199 |
| }, |
| { |
| "epoch": 2.82, |
| "learning_rate": 1.9485416956044222e-07, |
| "loss": 0.9289, |
| "step": 2200 |
| }, |
| { |
| "epoch": 2.82, |
| "learning_rate": 1.921471959676957e-07, |
| "loss": 0.921, |
| "step": 2201 |
| }, |
| { |
| "epoch": 2.82, |
| "learning_rate": 1.8945897477940844e-07, |
| "loss": 0.9297, |
| "step": 2202 |
| }, |
| { |
| "epoch": 2.82, |
| "learning_rate": 1.8678951113540012e-07, |
| "loss": 0.9944, |
| "step": 2203 |
| }, |
| { |
| "epoch": 2.82, |
| "learning_rate": 1.841388101396291e-07, |
| "loss": 1.0247, |
| "step": 2204 |
| }, |
| { |
| "epoch": 2.82, |
| "learning_rate": 1.815068768601791e-07, |
| "loss": 0.7998, |
| "step": 2205 |
| }, |
| { |
| "epoch": 2.82, |
| "learning_rate": 1.7889371632924924e-07, |
| "loss": 0.9197, |
| "step": 2206 |
| }, |
| { |
| "epoch": 2.82, |
| "learning_rate": 1.7629933354314733e-07, |
| "loss": 0.9439, |
| "step": 2207 |
| }, |
| { |
| "epoch": 2.83, |
| "learning_rate": 1.7372373346227767e-07, |
| "loss": 0.9658, |
| "step": 2208 |
| }, |
| { |
| "epoch": 2.83, |
| "learning_rate": 1.711669210111322e-07, |
| "loss": 0.8774, |
| "step": 2209 |
| }, |
| { |
| "epoch": 2.83, |
| "learning_rate": 1.6862890107828266e-07, |
| "loss": 0.891, |
| "step": 2210 |
| }, |
| { |
| "epoch": 2.83, |
| "learning_rate": 1.6610967851636629e-07, |
| "loss": 0.8791, |
| "step": 2211 |
| }, |
| { |
| "epoch": 2.83, |
| "learning_rate": 1.636092581420845e-07, |
| "loss": 0.977, |
| "step": 2212 |
| }, |
| { |
| "epoch": 2.83, |
| "learning_rate": 1.6112764473618647e-07, |
| "loss": 0.9197, |
| "step": 2213 |
| }, |
| { |
| "epoch": 2.83, |
| "learning_rate": 1.5866484304346564e-07, |
| "loss": 0.8602, |
| "step": 2214 |
| }, |
| { |
| "epoch": 2.84, |
| "learning_rate": 1.562208577727442e-07, |
| "loss": 0.8004, |
| "step": 2215 |
| }, |
| { |
| "epoch": 2.84, |
| "learning_rate": 1.5379569359686985e-07, |
| "loss": 0.9049, |
| "step": 2216 |
| }, |
| { |
| "epoch": 2.84, |
| "learning_rate": 1.513893551527057e-07, |
| "loss": 0.9723, |
| "step": 2217 |
| }, |
| { |
| "epoch": 2.84, |
| "learning_rate": 1.490018470411192e-07, |
| "loss": 0.8713, |
| "step": 2218 |
| }, |
| { |
| "epoch": 2.84, |
| "learning_rate": 1.4663317382697328e-07, |
| "loss": 0.804, |
| "step": 2219 |
| }, |
| { |
| "epoch": 2.84, |
| "learning_rate": 1.4428334003912192e-07, |
| "loss": 0.9583, |
| "step": 2220 |
| }, |
| { |
| "epoch": 2.84, |
| "learning_rate": 1.4195235017039566e-07, |
| "loss": 0.8844, |
| "step": 2221 |
| }, |
| { |
| "epoch": 2.84, |
| "learning_rate": 1.396402086775983e-07, |
| "loss": 0.9986, |
| "step": 2222 |
| }, |
| { |
| "epoch": 2.85, |
| "learning_rate": 1.3734691998149473e-07, |
| "loss": 0.8811, |
| "step": 2223 |
| }, |
| { |
| "epoch": 2.85, |
| "learning_rate": 1.350724884668031e-07, |
| "loss": 0.9256, |
| "step": 2224 |
| }, |
| { |
| "epoch": 2.85, |
| "learning_rate": 1.3281691848218813e-07, |
| "loss": 0.8379, |
| "step": 2225 |
| }, |
| { |
| "epoch": 2.85, |
| "learning_rate": 1.3058021434025126e-07, |
| "loss": 0.9923, |
| "step": 2226 |
| }, |
| { |
| "epoch": 2.85, |
| "learning_rate": 1.2836238031752269e-07, |
| "loss": 0.8646, |
| "step": 2227 |
| }, |
| { |
| "epoch": 2.85, |
| "learning_rate": 1.2616342065445485e-07, |
| "loss": 0.8732, |
| "step": 2228 |
| }, |
| { |
| "epoch": 2.85, |
| "learning_rate": 1.2398333955540908e-07, |
| "loss": 1.0197, |
| "step": 2229 |
| }, |
| { |
| "epoch": 2.85, |
| "learning_rate": 1.2182214118865666e-07, |
| "loss": 0.9308, |
| "step": 2230 |
| }, |
| { |
| "epoch": 2.86, |
| "learning_rate": 1.1967982968635994e-07, |
| "loss": 0.9724, |
| "step": 2231 |
| }, |
| { |
| "epoch": 2.86, |
| "learning_rate": 1.1755640914457355e-07, |
| "loss": 0.8944, |
| "step": 2232 |
| }, |
| { |
| "epoch": 2.86, |
| "learning_rate": 1.1545188362323323e-07, |
| "loss": 1.0192, |
| "step": 2233 |
| }, |
| { |
| "epoch": 2.86, |
| "learning_rate": 1.1336625714614469e-07, |
| "loss": 0.9587, |
| "step": 2234 |
| }, |
| { |
| "epoch": 2.86, |
| "learning_rate": 1.1129953370098256e-07, |
| "loss": 0.8751, |
| "step": 2235 |
| }, |
| { |
| "epoch": 2.86, |
| "learning_rate": 1.0925171723927819e-07, |
| "loss": 0.9521, |
| "step": 2236 |
| }, |
| { |
| "epoch": 2.86, |
| "learning_rate": 1.0722281167641069e-07, |
| "loss": 0.9035, |
| "step": 2237 |
| }, |
| { |
| "epoch": 2.86, |
| "learning_rate": 1.052128208916059e-07, |
| "loss": 0.867, |
| "step": 2238 |
| }, |
| { |
| "epoch": 2.87, |
| "learning_rate": 1.032217487279219e-07, |
| "loss": 0.8952, |
| "step": 2239 |
| }, |
| { |
| "epoch": 2.87, |
| "learning_rate": 1.0124959899224795e-07, |
| "loss": 0.9427, |
| "step": 2240 |
| }, |
| { |
| "epoch": 2.87, |
| "learning_rate": 9.929637545529003e-08, |
| "loss": 0.9899, |
| "step": 2241 |
| }, |
| { |
| "epoch": 2.87, |
| "learning_rate": 9.73620818515697e-08, |
| "loss": 0.8969, |
| "step": 2242 |
| }, |
| { |
| "epoch": 2.87, |
| "learning_rate": 9.54467218794164e-08, |
| "loss": 0.8914, |
| "step": 2243 |
| }, |
| { |
| "epoch": 2.87, |
| "learning_rate": 9.355029920095738e-08, |
| "loss": 0.9875, |
| "step": 2244 |
| }, |
| { |
| "epoch": 2.87, |
| "learning_rate": 9.167281744210999e-08, |
| "loss": 0.8659, |
| "step": 2245 |
| }, |
| { |
| "epoch": 2.87, |
| "learning_rate": 8.981428019258165e-08, |
| "loss": 0.913, |
| "step": 2246 |
| }, |
| { |
| "epoch": 2.88, |
| "learning_rate": 8.797469100585432e-08, |
| "loss": 1.0101, |
| "step": 2247 |
| }, |
| { |
| "epoch": 2.88, |
| "learning_rate": 8.615405339918337e-08, |
| "loss": 1.0033, |
| "step": 2248 |
| }, |
| { |
| "epoch": 2.88, |
| "learning_rate": 8.435237085358872e-08, |
| "loss": 0.8884, |
| "step": 2249 |
| }, |
| { |
| "epoch": 2.88, |
| "learning_rate": 8.256964681384927e-08, |
| "loss": 0.9305, |
| "step": 2250 |
| }, |
| { |
| "epoch": 2.88, |
| "learning_rate": 8.080588468849515e-08, |
| "loss": 0.8898, |
| "step": 2251 |
| }, |
| { |
| "epoch": 2.88, |
| "learning_rate": 7.906108784980216e-08, |
| "loss": 0.9422, |
| "step": 2252 |
| }, |
| { |
| "epoch": 2.88, |
| "learning_rate": 7.733525963378286e-08, |
| "loss": 0.7957, |
| "step": 2253 |
| }, |
| { |
| "epoch": 2.89, |
| "learning_rate": 7.56284033401855e-08, |
| "loss": 0.9738, |
| "step": 2254 |
| }, |
| { |
| "epoch": 2.89, |
| "learning_rate": 7.394052223248182e-08, |
| "loss": 0.8419, |
| "step": 2255 |
| }, |
| { |
| "epoch": 2.89, |
| "learning_rate": 7.227161953786588e-08, |
| "loss": 0.9483, |
| "step": 2256 |
| }, |
| { |
| "epoch": 2.89, |
| "learning_rate": 7.062169844724186e-08, |
| "loss": 0.889, |
| "step": 2257 |
| }, |
| { |
| "epoch": 2.89, |
| "learning_rate": 6.899076211522415e-08, |
| "loss": 0.9358, |
| "step": 2258 |
| }, |
| { |
| "epoch": 2.89, |
| "learning_rate": 6.737881366012833e-08, |
| "loss": 0.9565, |
| "step": 2259 |
| }, |
| { |
| "epoch": 2.89, |
| "learning_rate": 6.578585616396571e-08, |
| "loss": 0.9177, |
| "step": 2260 |
| }, |
| { |
| "epoch": 2.89, |
| "learning_rate": 6.421189267243555e-08, |
| "loss": 0.8695, |
| "step": 2261 |
| }, |
| { |
| "epoch": 2.9, |
| "learning_rate": 6.265692619492503e-08, |
| "loss": 0.8254, |
| "step": 2262 |
| }, |
| { |
| "epoch": 2.9, |
| "learning_rate": 6.112095970449261e-08, |
| "loss": 0.8282, |
| "step": 2263 |
| }, |
| { |
| "epoch": 2.9, |
| "learning_rate": 5.960399613787693e-08, |
| "loss": 0.8546, |
| "step": 2264 |
| }, |
| { |
| "epoch": 2.9, |
| "learning_rate": 5.810603839547901e-08, |
| "loss": 0.9774, |
| "step": 2265 |
| }, |
| { |
| "epoch": 2.9, |
| "learning_rate": 5.6627089341361186e-08, |
| "loss": 0.9648, |
| "step": 2266 |
| }, |
| { |
| "epoch": 2.9, |
| "learning_rate": 5.516715180324261e-08, |
| "loss": 0.934, |
| "step": 2267 |
| }, |
| { |
| "epoch": 2.9, |
| "learning_rate": 5.3726228572492656e-08, |
| "loss": 0.9614, |
| "step": 2268 |
| }, |
| { |
| "epoch": 2.9, |
| "learning_rate": 5.230432240412531e-08, |
| "loss": 0.9708, |
| "step": 2269 |
| }, |
| { |
| "epoch": 2.91, |
| "learning_rate": 5.090143601679587e-08, |
| "loss": 0.9307, |
| "step": 2270 |
| }, |
| { |
| "epoch": 2.91, |
| "learning_rate": 4.9517572092790954e-08, |
| "loss": 0.9711, |
| "step": 2271 |
| }, |
| { |
| "epoch": 2.91, |
| "learning_rate": 4.815273327803183e-08, |
| "loss": 0.9268, |
| "step": 2272 |
| }, |
| { |
| "epoch": 2.91, |
| "learning_rate": 4.680692218205773e-08, |
| "loss": 0.9245, |
| "step": 2273 |
| }, |
| { |
| "epoch": 2.91, |
| "learning_rate": 4.548014137803258e-08, |
| "loss": 0.9894, |
| "step": 2274 |
| }, |
| { |
| "epoch": 2.91, |
| "learning_rate": 4.4172393402732717e-08, |
| "loss": 0.9744, |
| "step": 2275 |
| }, |
| { |
| "epoch": 2.91, |
| "learning_rate": 4.28836807565447e-08, |
| "loss": 0.9679, |
| "step": 2276 |
| }, |
| { |
| "epoch": 2.91, |
| "learning_rate": 4.161400590345755e-08, |
| "loss": 0.9537, |
| "step": 2277 |
| }, |
| { |
| "epoch": 2.92, |
| "learning_rate": 4.0363371271064934e-08, |
| "loss": 0.9047, |
| "step": 2278 |
| }, |
| { |
| "epoch": 2.92, |
| "learning_rate": 3.913177925055189e-08, |
| "loss": 0.9199, |
| "step": 2279 |
| }, |
| { |
| "epoch": 2.92, |
| "learning_rate": 3.791923219669591e-08, |
| "loss": 0.8895, |
| "step": 2280 |
| }, |
| { |
| "epoch": 2.92, |
| "learning_rate": 3.672573242786137e-08, |
| "loss": 0.9589, |
| "step": 2281 |
| }, |
| { |
| "epoch": 2.92, |
| "learning_rate": 3.555128222599513e-08, |
| "loss": 0.8466, |
| "step": 2282 |
| }, |
| { |
| "epoch": 2.92, |
| "learning_rate": 3.4395883836618735e-08, |
| "loss": 1.0267, |
| "step": 2283 |
| }, |
| { |
| "epoch": 2.92, |
| "learning_rate": 3.325953946883065e-08, |
| "loss": 0.8811, |
| "step": 2284 |
| }, |
| { |
| "epoch": 2.92, |
| "learning_rate": 3.214225129529735e-08, |
| "loss": 0.9944, |
| "step": 2285 |
| }, |
| { |
| "epoch": 2.93, |
| "learning_rate": 3.104402145224894e-08, |
| "loss": 1.0145, |
| "step": 2286 |
| }, |
| { |
| "epoch": 2.93, |
| "learning_rate": 2.9964852039476854e-08, |
| "loss": 0.9561, |
| "step": 2287 |
| }, |
| { |
| "epoch": 2.93, |
| "learning_rate": 2.8904745120331702e-08, |
| "loss": 0.8804, |
| "step": 2288 |
| }, |
| { |
| "epoch": 2.93, |
| "learning_rate": 2.7863702721714348e-08, |
| "loss": 0.9863, |
| "step": 2289 |
| }, |
| { |
| "epoch": 2.93, |
| "learning_rate": 2.684172683407482e-08, |
| "loss": 0.9417, |
| "step": 2290 |
| }, |
| { |
| "epoch": 2.93, |
| "learning_rate": 2.5838819411408977e-08, |
| "loss": 0.9043, |
| "step": 2291 |
| }, |
| { |
| "epoch": 2.93, |
| "learning_rate": 2.4854982371256275e-08, |
| "loss": 0.8557, |
| "step": 2292 |
| }, |
| { |
| "epoch": 2.94, |
| "learning_rate": 2.3890217594689792e-08, |
| "loss": 0.8708, |
| "step": 2293 |
| }, |
| { |
| "epoch": 2.94, |
| "learning_rate": 2.2944526926321765e-08, |
| "loss": 0.8445, |
| "step": 2294 |
| }, |
| { |
| "epoch": 2.94, |
| "learning_rate": 2.201791217428917e-08, |
| "loss": 0.8733, |
| "step": 2295 |
| }, |
| { |
| "epoch": 2.94, |
| "learning_rate": 2.1110375110262592e-08, |
| "loss": 0.903, |
| "step": 2296 |
| }, |
| { |
| "epoch": 2.94, |
| "learning_rate": 2.0221917469430696e-08, |
| "loss": 0.881, |
| "step": 2297 |
| }, |
| { |
| "epoch": 2.94, |
| "learning_rate": 1.935254095050798e-08, |
| "loss": 1.0253, |
| "step": 2298 |
| }, |
| { |
| "epoch": 2.94, |
| "learning_rate": 1.850224721572258e-08, |
| "loss": 0.9571, |
| "step": 2299 |
| }, |
| { |
| "epoch": 2.94, |
| "learning_rate": 1.7671037890817366e-08, |
| "loss": 0.9932, |
| "step": 2300 |
| }, |
| { |
| "epoch": 2.95, |
| "learning_rate": 1.6858914565047736e-08, |
| "loss": 0.9456, |
| "step": 2301 |
| }, |
| { |
| "epoch": 2.95, |
| "learning_rate": 1.6065878791176048e-08, |
| "loss": 0.9261, |
| "step": 2302 |
| }, |
| { |
| "epoch": 2.95, |
| "learning_rate": 1.5291932085468308e-08, |
| "loss": 0.9156, |
| "step": 2303 |
| }, |
| { |
| "epoch": 2.95, |
| "learning_rate": 1.4537075927696375e-08, |
| "loss": 0.9729, |
| "step": 2304 |
| }, |
| { |
| "epoch": 2.95, |
| "learning_rate": 1.3801311761126868e-08, |
| "loss": 0.9212, |
| "step": 2305 |
| }, |
| { |
| "epoch": 2.95, |
| "learning_rate": 1.3084640992526708e-08, |
| "loss": 0.962, |
| "step": 2306 |
| }, |
| { |
| "epoch": 2.95, |
| "learning_rate": 1.2387064992154253e-08, |
| "loss": 1.0292, |
| "step": 2307 |
| }, |
| { |
| "epoch": 2.95, |
| "learning_rate": 1.1708585093759272e-08, |
| "loss": 0.9347, |
| "step": 2308 |
| }, |
| { |
| "epoch": 2.96, |
| "learning_rate": 1.1049202594581865e-08, |
| "loss": 0.919, |
| "step": 2309 |
| }, |
| { |
| "epoch": 2.96, |
| "learning_rate": 1.0408918755347996e-08, |
| "loss": 0.93, |
| "step": 2310 |
| }, |
| { |
| "epoch": 2.96, |
| "learning_rate": 9.78773480026396e-09, |
| "loss": 1.0261, |
| "step": 2311 |
| }, |
| { |
| "epoch": 2.96, |
| "learning_rate": 9.185651917023031e-09, |
| "loss": 0.9381, |
| "step": 2312 |
| }, |
| { |
| "epoch": 2.96, |
| "learning_rate": 8.60267125679215e-09, |
| "loss": 0.9209, |
| "step": 2313 |
| }, |
| { |
| "epoch": 2.96, |
| "learning_rate": 8.038793934220802e-09, |
| "loss": 0.9639, |
| "step": 2314 |
| }, |
| { |
| "epoch": 2.96, |
| "learning_rate": 7.494021027427689e-09, |
| "loss": 0.9684, |
| "step": 2315 |
| }, |
| { |
| "epoch": 2.96, |
| "learning_rate": 6.968353578007403e-09, |
| "loss": 0.955, |
| "step": 2316 |
| }, |
| { |
| "epoch": 2.97, |
| "learning_rate": 6.4617925910270825e-09, |
| "loss": 0.8382, |
| "step": 2317 |
| }, |
| { |
| "epoch": 2.97, |
| "learning_rate": 5.97433903501754e-09, |
| "loss": 0.9086, |
| "step": 2318 |
| }, |
| { |
| "epoch": 2.97, |
| "learning_rate": 5.505993841982138e-09, |
| "loss": 0.9283, |
| "step": 2319 |
| }, |
| { |
| "epoch": 2.97, |
| "learning_rate": 5.05675790738458e-09, |
| "loss": 0.9291, |
| "step": 2320 |
| }, |
| { |
| "epoch": 2.97, |
| "learning_rate": 4.62663209015557e-09, |
| "loss": 0.9277, |
| "step": 2321 |
| }, |
| { |
| "epoch": 2.97, |
| "learning_rate": 4.215617212686151e-09, |
| "loss": 0.8915, |
| "step": 2322 |
| }, |
| { |
| "epoch": 2.97, |
| "learning_rate": 3.8237140608277105e-09, |
| "loss": 0.965, |
| "step": 2323 |
| }, |
| { |
| "epoch": 2.97, |
| "learning_rate": 3.450923383891969e-09, |
| "loss": 0.9895, |
| "step": 2324 |
| }, |
| { |
| "epoch": 2.98, |
| "learning_rate": 3.0972458946454396e-09, |
| "loss": 0.8938, |
| "step": 2325 |
| }, |
| { |
| "epoch": 2.98, |
| "learning_rate": 2.762682269311645e-09, |
| "loss": 0.8063, |
| "step": 2326 |
| }, |
| { |
| "epoch": 2.98, |
| "learning_rate": 2.447233147570005e-09, |
| "loss": 0.9711, |
| "step": 2327 |
| }, |
| { |
| "epoch": 2.98, |
| "learning_rate": 2.1508991325525087e-09, |
| "loss": 1.048, |
| "step": 2328 |
| }, |
| { |
| "epoch": 2.98, |
| "learning_rate": 1.8736807908426027e-09, |
| "loss": 0.9388, |
| "step": 2329 |
| }, |
| { |
| "epoch": 2.98, |
| "learning_rate": 1.6155786524763018e-09, |
| "loss": 0.983, |
| "step": 2330 |
| }, |
| { |
| "epoch": 2.98, |
| "learning_rate": 1.376593210938859e-09, |
| "loss": 0.9322, |
| "step": 2331 |
| }, |
| { |
| "epoch": 2.98, |
| "learning_rate": 1.156724923165875e-09, |
| "loss": 0.972, |
| "step": 2332 |
| }, |
| { |
| "epoch": 2.99, |
| "learning_rate": 9.559742095410774e-10, |
| "loss": 0.9069, |
| "step": 2333 |
| }, |
| { |
| "epoch": 2.99, |
| "learning_rate": 7.74341453895211e-10, |
| "loss": 1.0329, |
| "step": 2334 |
| }, |
| { |
| "epoch": 2.99, |
| "learning_rate": 6.118270035071483e-10, |
| "loss": 1.0599, |
| "step": 2335 |
| }, |
| { |
| "epoch": 2.99, |
| "learning_rate": 4.684311690983378e-10, |
| "loss": 0.9053, |
| "step": 2336 |
| }, |
| { |
| "epoch": 2.99, |
| "learning_rate": 3.4415422484168626e-10, |
| "loss": 0.9449, |
| "step": 2337 |
| }, |
| { |
| "epoch": 2.99, |
| "learning_rate": 2.3899640834934567e-10, |
| "loss": 0.8478, |
| "step": 2338 |
| }, |
| { |
| "epoch": 2.99, |
| "learning_rate": 1.5295792068270586e-10, |
| "loss": 0.8927, |
| "step": 2339 |
| }, |
| { |
| "epoch": 3.0, |
| "learning_rate": 8.603892634462263e-11, |
| "loss": 0.9631, |
| "step": 2340 |
| }, |
| { |
| "epoch": 3.0, |
| "learning_rate": 3.823955328385864e-11, |
| "loss": 0.9813, |
| "step": 2341 |
| }, |
| { |
| "epoch": 3.0, |
| "learning_rate": 9.559892890642631e-12, |
| "loss": 0.8734, |
| "step": 2342 |
| }, |
| { |
| "epoch": 3.0, |
| "learning_rate": 0.0, |
| "loss": 1.0041, |
| "step": 2343 |
| }, |
| { |
| "epoch": 3.0, |
| "step": 2343, |
| "total_flos": 1.4406392337976525e+18, |
| "train_loss": 1.2288865842882304, |
| "train_runtime": 12734.9896, |
| "train_samples_per_second": 47.114, |
| "train_steps_per_second": 0.184 |
| } |
| ], |
| "max_steps": 2343, |
| "num_train_epochs": 3, |
| "total_flos": 1.4406392337976525e+18, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|