{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 0.7772272418148256, "eval_steps": 500, "global_step": 2000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0032, "grad_norm": 13.684800148010254, "learning_rate": 0.0, "loss": 2.3276, "step": 1 }, { "epoch": 0.0064, "grad_norm": 13.660787582397461, "learning_rate": 4e-05, "loss": 2.2792, "step": 2 }, { "epoch": 0.0096, "grad_norm": 13.35280704498291, "learning_rate": 8e-05, "loss": 2.4151, "step": 3 }, { "epoch": 0.0128, "grad_norm": 6.15027379989624, "learning_rate": 0.00012, "loss": 1.7812, "step": 4 }, { "epoch": 0.016, "grad_norm": 1.3168226480484009, "learning_rate": 0.00016, "loss": 1.4536, "step": 5 }, { "epoch": 0.0192, "grad_norm": 0.9872580170631409, "learning_rate": 0.0002, "loss": 1.4171, "step": 6 }, { "epoch": 0.0224, "grad_norm": 0.7496100664138794, "learning_rate": 0.00019935064935064936, "loss": 1.4168, "step": 7 }, { "epoch": 0.0256, "grad_norm": 0.7376005053520203, "learning_rate": 0.00019870129870129872, "loss": 1.3659, "step": 8 }, { "epoch": 0.0288, "grad_norm": 0.5281137824058533, "learning_rate": 0.00019805194805194807, "loss": 1.2566, "step": 9 }, { "epoch": 0.032, "grad_norm": 0.5485746264457703, "learning_rate": 0.00019740259740259742, "loss": 1.3761, "step": 10 }, { "epoch": 0.0352, "grad_norm": 0.5506592392921448, "learning_rate": 0.00019675324675324675, "loss": 1.3327, "step": 11 }, { "epoch": 0.0384, "grad_norm": 0.49382686614990234, "learning_rate": 0.00019610389610389613, "loss": 1.3727, "step": 12 }, { "epoch": 0.0416, "grad_norm": 0.36203011870384216, "learning_rate": 0.00019545454545454548, "loss": 1.1515, "step": 13 }, { "epoch": 0.0448, "grad_norm": 0.3528599739074707, "learning_rate": 0.0001948051948051948, "loss": 1.2636, "step": 14 }, { "epoch": 0.048, "grad_norm": 0.31244418025016785, "learning_rate": 0.00019415584415584416, "loss": 1.1873, "step": 15 }, { "epoch": 0.0512, "grad_norm": 0.3379523754119873, "learning_rate": 0.00019350649350649354, "loss": 1.2657, "step": 16 }, { "epoch": 0.0544, "grad_norm": 0.3025083839893341, "learning_rate": 0.00019285714285714286, "loss": 1.2846, "step": 17 }, { "epoch": 0.0576, "grad_norm": 0.2560190260410309, "learning_rate": 0.00019220779220779222, "loss": 1.1587, "step": 18 }, { "epoch": 0.0608, "grad_norm": 0.2554129958152771, "learning_rate": 0.00019155844155844157, "loss": 1.2812, "step": 19 }, { "epoch": 0.064, "grad_norm": 0.22662702202796936, "learning_rate": 0.00019090909090909092, "loss": 1.1664, "step": 20 }, { "epoch": 0.0672, "grad_norm": 0.2515714168548584, "learning_rate": 0.00019025974025974027, "loss": 1.2177, "step": 21 }, { "epoch": 0.0704, "grad_norm": 0.24396637082099915, "learning_rate": 0.00018961038961038963, "loss": 1.2053, "step": 22 }, { "epoch": 0.0736, "grad_norm": 0.24488303065299988, "learning_rate": 0.00018896103896103895, "loss": 1.2074, "step": 23 }, { "epoch": 0.0768, "grad_norm": 0.2168620079755783, "learning_rate": 0.00018831168831168833, "loss": 1.1284, "step": 24 }, { "epoch": 0.08, "grad_norm": 0.24021224677562714, "learning_rate": 0.00018766233766233769, "loss": 1.2169, "step": 25 }, { "epoch": 0.0832, "grad_norm": 0.20057056844234467, "learning_rate": 0.000187012987012987, "loss": 1.1031, "step": 26 }, { "epoch": 0.0864, "grad_norm": 0.19900795817375183, "learning_rate": 0.00018636363636363636, "loss": 1.1004, "step": 27 }, { "epoch": 0.0896, "grad_norm": 0.2019268423318863, "learning_rate": 0.00018571428571428572, "loss": 1.1476, "step": 28 }, { "epoch": 0.0928, "grad_norm": 0.1996479034423828, "learning_rate": 0.00018506493506493507, "loss": 1.1455, "step": 29 }, { "epoch": 0.096, "grad_norm": 0.25262022018432617, "learning_rate": 0.00018441558441558442, "loss": 1.1025, "step": 30 }, { "epoch": 0.0992, "grad_norm": 0.225438192486763, "learning_rate": 0.00018376623376623378, "loss": 1.1954, "step": 31 }, { "epoch": 0.1024, "grad_norm": 0.17834505438804626, "learning_rate": 0.00018311688311688313, "loss": 1.0934, "step": 32 }, { "epoch": 0.1056, "grad_norm": 0.20071206986904144, "learning_rate": 0.00018246753246753248, "loss": 1.0488, "step": 33 }, { "epoch": 0.1088, "grad_norm": 0.1920139640569687, "learning_rate": 0.00018181818181818183, "loss": 1.123, "step": 34 }, { "epoch": 0.112, "grad_norm": 0.18714852631092072, "learning_rate": 0.0001811688311688312, "loss": 1.0798, "step": 35 }, { "epoch": 0.1152, "grad_norm": 0.18315713107585907, "learning_rate": 0.00018051948051948054, "loss": 1.1107, "step": 36 }, { "epoch": 0.1184, "grad_norm": 0.19156870245933533, "learning_rate": 0.00017987012987012987, "loss": 1.1125, "step": 37 }, { "epoch": 0.1216, "grad_norm": 0.21527768671512604, "learning_rate": 0.00017922077922077922, "loss": 1.1346, "step": 38 }, { "epoch": 0.1248, "grad_norm": 0.1871163249015808, "learning_rate": 0.0001785714285714286, "loss": 1.0742, "step": 39 }, { "epoch": 0.128, "grad_norm": 0.17750784754753113, "learning_rate": 0.00017792207792207792, "loss": 1.1323, "step": 40 }, { "epoch": 0.1312, "grad_norm": 0.177419051527977, "learning_rate": 0.00017727272727272728, "loss": 1.1405, "step": 41 }, { "epoch": 0.1344, "grad_norm": 0.16714292764663696, "learning_rate": 0.00017662337662337663, "loss": 1.1084, "step": 42 }, { "epoch": 0.1376, "grad_norm": 0.1610356718301773, "learning_rate": 0.00017597402597402598, "loss": 1.1125, "step": 43 }, { "epoch": 0.1408, "grad_norm": 0.2548656761646271, "learning_rate": 0.00017532467532467534, "loss": 1.1114, "step": 44 }, { "epoch": 0.144, "grad_norm": 0.1731044203042984, "learning_rate": 0.0001746753246753247, "loss": 1.1197, "step": 45 }, { "epoch": 0.1472, "grad_norm": 0.1739533394575119, "learning_rate": 0.00017402597402597401, "loss": 1.1777, "step": 46 }, { "epoch": 0.1504, "grad_norm": 0.2178352177143097, "learning_rate": 0.0001733766233766234, "loss": 1.1111, "step": 47 }, { "epoch": 0.1536, "grad_norm": 0.17247150838375092, "learning_rate": 0.00017272727272727275, "loss": 1.1253, "step": 48 }, { "epoch": 0.1568, "grad_norm": 0.18075324594974518, "learning_rate": 0.00017207792207792207, "loss": 1.1358, "step": 49 }, { "epoch": 0.16, "grad_norm": 0.15898071229457855, "learning_rate": 0.00017142857142857143, "loss": 1.0606, "step": 50 }, { "epoch": 0.1632, "grad_norm": 0.16518613696098328, "learning_rate": 0.0001707792207792208, "loss": 1.0944, "step": 51 }, { "epoch": 0.1664, "grad_norm": 0.16035063564777374, "learning_rate": 0.00017012987012987013, "loss": 1.0554, "step": 52 }, { "epoch": 0.1696, "grad_norm": 0.1686483472585678, "learning_rate": 0.00016948051948051948, "loss": 1.0384, "step": 53 }, { "epoch": 0.1728, "grad_norm": 0.16575631499290466, "learning_rate": 0.00016883116883116884, "loss": 1.0243, "step": 54 }, { "epoch": 0.176, "grad_norm": 0.16840039193630219, "learning_rate": 0.0001681818181818182, "loss": 1.117, "step": 55 }, { "epoch": 0.1792, "grad_norm": 0.17616064846515656, "learning_rate": 0.00016753246753246754, "loss": 1.0743, "step": 56 }, { "epoch": 0.1824, "grad_norm": 0.168218195438385, "learning_rate": 0.0001668831168831169, "loss": 1.0627, "step": 57 }, { "epoch": 0.1856, "grad_norm": 0.17026656866073608, "learning_rate": 0.00016623376623376625, "loss": 1.0059, "step": 58 }, { "epoch": 0.1888, "grad_norm": 0.16454458236694336, "learning_rate": 0.0001655844155844156, "loss": 0.9943, "step": 59 }, { "epoch": 0.192, "grad_norm": 0.17185136675834656, "learning_rate": 0.00016493506493506495, "loss": 1.1545, "step": 60 }, { "epoch": 0.1952, "grad_norm": 0.17822986841201782, "learning_rate": 0.00016428571428571428, "loss": 1.073, "step": 61 }, { "epoch": 0.1984, "grad_norm": 0.1676608771085739, "learning_rate": 0.00016363636363636366, "loss": 1.0886, "step": 62 }, { "epoch": 0.2016, "grad_norm": 0.1727771908044815, "learning_rate": 0.000162987012987013, "loss": 1.0432, "step": 63 }, { "epoch": 0.2048, "grad_norm": 0.17827573418617249, "learning_rate": 0.00016233766233766234, "loss": 1.083, "step": 64 }, { "epoch": 0.208, "grad_norm": 0.19807517528533936, "learning_rate": 0.0001616883116883117, "loss": 1.1208, "step": 65 }, { "epoch": 0.2112, "grad_norm": 0.17693684995174408, "learning_rate": 0.00016103896103896104, "loss": 1.089, "step": 66 }, { "epoch": 0.2144, "grad_norm": 0.15489234030246735, "learning_rate": 0.0001603896103896104, "loss": 0.9707, "step": 67 }, { "epoch": 0.2176, "grad_norm": 0.16443990170955658, "learning_rate": 0.00015974025974025975, "loss": 1.0643, "step": 68 }, { "epoch": 0.2208, "grad_norm": 0.2051103413105011, "learning_rate": 0.0001590909090909091, "loss": 1.1246, "step": 69 }, { "epoch": 0.224, "grad_norm": 0.18824075162410736, "learning_rate": 0.00015844155844155845, "loss": 1.0855, "step": 70 }, { "epoch": 0.2272, "grad_norm": 0.18659448623657227, "learning_rate": 0.0001577922077922078, "loss": 1.1412, "step": 71 }, { "epoch": 0.2304, "grad_norm": 0.1854114979505539, "learning_rate": 0.00015714285714285716, "loss": 1.0249, "step": 72 }, { "epoch": 0.2336, "grad_norm": 0.1876193732023239, "learning_rate": 0.00015649350649350649, "loss": 1.1029, "step": 73 }, { "epoch": 0.2368, "grad_norm": 0.1888684630393982, "learning_rate": 0.00015584415584415587, "loss": 1.0789, "step": 74 }, { "epoch": 0.24, "grad_norm": 0.20240606367588043, "learning_rate": 0.0001551948051948052, "loss": 1.0495, "step": 75 }, { "epoch": 0.2432, "grad_norm": 0.232120081782341, "learning_rate": 0.00015454545454545454, "loss": 1.0735, "step": 76 }, { "epoch": 0.2464, "grad_norm": 0.16897843778133392, "learning_rate": 0.0001538961038961039, "loss": 1.0164, "step": 77 }, { "epoch": 0.2496, "grad_norm": 0.18796634674072266, "learning_rate": 0.00015324675324675325, "loss": 1.0676, "step": 78 }, { "epoch": 0.2528, "grad_norm": 0.19574032723903656, "learning_rate": 0.0001525974025974026, "loss": 1.0456, "step": 79 }, { "epoch": 0.256, "grad_norm": 0.18007811903953552, "learning_rate": 0.00015194805194805196, "loss": 1.0894, "step": 80 }, { "epoch": 0.2592, "grad_norm": 0.18932929635047913, "learning_rate": 0.0001512987012987013, "loss": 1.0729, "step": 81 }, { "epoch": 0.2624, "grad_norm": 0.20614288747310638, "learning_rate": 0.00015064935064935066, "loss": 1.0854, "step": 82 }, { "epoch": 0.2656, "grad_norm": 0.19291089475154877, "learning_rate": 0.00015000000000000001, "loss": 1.1217, "step": 83 }, { "epoch": 0.2688, "grad_norm": 0.18916529417037964, "learning_rate": 0.00014935064935064934, "loss": 1.0963, "step": 84 }, { "epoch": 0.272, "grad_norm": 0.20306220650672913, "learning_rate": 0.00014870129870129872, "loss": 1.0898, "step": 85 }, { "epoch": 0.2752, "grad_norm": 0.17870067059993744, "learning_rate": 0.00014805194805194807, "loss": 1.0213, "step": 86 }, { "epoch": 0.2784, "grad_norm": 0.18411923944950104, "learning_rate": 0.0001474025974025974, "loss": 1.0844, "step": 87 }, { "epoch": 0.2816, "grad_norm": 0.18788227438926697, "learning_rate": 0.00014675324675324675, "loss": 1.0338, "step": 88 }, { "epoch": 0.2848, "grad_norm": 0.23874884843826294, "learning_rate": 0.00014610389610389613, "loss": 1.1118, "step": 89 }, { "epoch": 0.288, "grad_norm": 0.19380499422550201, "learning_rate": 0.00014545454545454546, "loss": 1.0464, "step": 90 }, { "epoch": 0.2912, "grad_norm": 0.18968750536441803, "learning_rate": 0.0001448051948051948, "loss": 1.0569, "step": 91 }, { "epoch": 0.2944, "grad_norm": 0.19545753300189972, "learning_rate": 0.00014415584415584416, "loss": 1.1225, "step": 92 }, { "epoch": 0.2976, "grad_norm": 0.19170494377613068, "learning_rate": 0.00014350649350649352, "loss": 1.0602, "step": 93 }, { "epoch": 0.3008, "grad_norm": 0.17953918874263763, "learning_rate": 0.00014285714285714287, "loss": 1.032, "step": 94 }, { "epoch": 0.304, "grad_norm": 0.1822536289691925, "learning_rate": 0.00014220779220779222, "loss": 1.0559, "step": 95 }, { "epoch": 0.3072, "grad_norm": 0.18591298162937164, "learning_rate": 0.00014155844155844155, "loss": 1.031, "step": 96 }, { "epoch": 0.3104, "grad_norm": 0.2129002958536148, "learning_rate": 0.00014090909090909093, "loss": 1.1391, "step": 97 }, { "epoch": 0.3136, "grad_norm": 0.18386681377887726, "learning_rate": 0.00014025974025974028, "loss": 0.9919, "step": 98 }, { "epoch": 0.3168, "grad_norm": 0.18314239382743835, "learning_rate": 0.0001396103896103896, "loss": 1.0445, "step": 99 }, { "epoch": 0.32, "grad_norm": 0.1999066174030304, "learning_rate": 0.00013896103896103896, "loss": 1.0538, "step": 100 }, { "epoch": 0.3232, "grad_norm": 0.18741188943386078, "learning_rate": 0.00013831168831168834, "loss": 1.0722, "step": 101 }, { "epoch": 0.3264, "grad_norm": 0.19351010024547577, "learning_rate": 0.00013766233766233766, "loss": 1.0491, "step": 102 }, { "epoch": 0.3296, "grad_norm": 0.18859203159809113, "learning_rate": 0.00013701298701298702, "loss": 1.0593, "step": 103 }, { "epoch": 0.3328, "grad_norm": 0.1962767392396927, "learning_rate": 0.00013636363636363637, "loss": 1.1344, "step": 104 }, { "epoch": 0.336, "grad_norm": 0.20819440484046936, "learning_rate": 0.00013571428571428572, "loss": 1.1137, "step": 105 }, { "epoch": 0.3392, "grad_norm": 0.19590184092521667, "learning_rate": 0.00013506493506493507, "loss": 1.0624, "step": 106 }, { "epoch": 0.3424, "grad_norm": 0.18631424009799957, "learning_rate": 0.00013441558441558443, "loss": 1.0587, "step": 107 }, { "epoch": 0.3456, "grad_norm": 0.19572143256664276, "learning_rate": 0.00013376623376623375, "loss": 1.0494, "step": 108 }, { "epoch": 0.3488, "grad_norm": 0.1910988837480545, "learning_rate": 0.00013311688311688313, "loss": 1.0481, "step": 109 }, { "epoch": 0.352, "grad_norm": 0.19455869495868683, "learning_rate": 0.00013246753246753249, "loss": 1.029, "step": 110 }, { "epoch": 0.3552, "grad_norm": 0.18669827282428741, "learning_rate": 0.0001318181818181818, "loss": 1.0513, "step": 111 }, { "epoch": 0.3584, "grad_norm": 0.17523664236068726, "learning_rate": 0.0001311688311688312, "loss": 1.0126, "step": 112 }, { "epoch": 0.3616, "grad_norm": 0.17929129302501678, "learning_rate": 0.00013051948051948052, "loss": 1.0717, "step": 113 }, { "epoch": 0.3648, "grad_norm": 0.19380168616771698, "learning_rate": 0.00012987012987012987, "loss": 1.0324, "step": 114 }, { "epoch": 0.368, "grad_norm": 0.18090228736400604, "learning_rate": 0.00012922077922077922, "loss": 1.0515, "step": 115 }, { "epoch": 0.3712, "grad_norm": 0.2067340910434723, "learning_rate": 0.00012857142857142858, "loss": 1.0939, "step": 116 }, { "epoch": 0.3744, "grad_norm": 0.1880485862493515, "learning_rate": 0.00012792207792207793, "loss": 1.0986, "step": 117 }, { "epoch": 0.3776, "grad_norm": 0.182168647646904, "learning_rate": 0.00012727272727272728, "loss": 1.0109, "step": 118 }, { "epoch": 0.3808, "grad_norm": 0.20187129080295563, "learning_rate": 0.00012662337662337663, "loss": 1.0668, "step": 119 }, { "epoch": 0.384, "grad_norm": 0.2082669734954834, "learning_rate": 0.000125974025974026, "loss": 1.054, "step": 120 }, { "epoch": 0.3872, "grad_norm": 0.18294434249401093, "learning_rate": 0.00012532467532467534, "loss": 1.0397, "step": 121 }, { "epoch": 0.3904, "grad_norm": 0.20515067875385284, "learning_rate": 0.00012467532467532467, "loss": 1.1092, "step": 122 }, { "epoch": 0.3936, "grad_norm": 0.1758790761232376, "learning_rate": 0.00012402597402597402, "loss": 0.9755, "step": 123 }, { "epoch": 0.3968, "grad_norm": 0.2170792669057846, "learning_rate": 0.0001233766233766234, "loss": 1.0434, "step": 124 }, { "epoch": 0.4, "grad_norm": 0.202157124876976, "learning_rate": 0.00012272727272727272, "loss": 1.1129, "step": 125 }, { "epoch": 0.4032, "grad_norm": 0.18556398153305054, "learning_rate": 0.00012207792207792208, "loss": 1.0665, "step": 126 }, { "epoch": 0.4064, "grad_norm": 0.20196087658405304, "learning_rate": 0.00012142857142857143, "loss": 1.1, "step": 127 }, { "epoch": 0.4096, "grad_norm": 0.1921566128730774, "learning_rate": 0.0001207792207792208, "loss": 1.0918, "step": 128 }, { "epoch": 0.4128, "grad_norm": 0.18866224586963654, "learning_rate": 0.00012012987012987014, "loss": 1.0014, "step": 129 }, { "epoch": 0.416, "grad_norm": 0.207601398229599, "learning_rate": 0.00011948051948051949, "loss": 1.0726, "step": 130 }, { "epoch": 0.4192, "grad_norm": 0.21592366695404053, "learning_rate": 0.00011883116883116883, "loss": 1.1379, "step": 131 }, { "epoch": 0.4224, "grad_norm": 0.2016124576330185, "learning_rate": 0.0001181818181818182, "loss": 1.1428, "step": 132 }, { "epoch": 0.4256, "grad_norm": 0.20478437840938568, "learning_rate": 0.00011753246753246753, "loss": 1.121, "step": 133 }, { "epoch": 0.4288, "grad_norm": 0.22730594873428345, "learning_rate": 0.00011688311688311689, "loss": 1.0319, "step": 134 }, { "epoch": 0.432, "grad_norm": 0.22592711448669434, "learning_rate": 0.00011623376623376625, "loss": 1.1264, "step": 135 }, { "epoch": 0.4352, "grad_norm": 0.20035041868686676, "learning_rate": 0.00011558441558441559, "loss": 1.0686, "step": 136 }, { "epoch": 0.4384, "grad_norm": 0.20648567378520966, "learning_rate": 0.00011493506493506494, "loss": 1.0817, "step": 137 }, { "epoch": 0.4416, "grad_norm": 0.21222743391990662, "learning_rate": 0.00011428571428571428, "loss": 1.0678, "step": 138 }, { "epoch": 0.4448, "grad_norm": 0.2075391560792923, "learning_rate": 0.00011363636363636365, "loss": 1.0897, "step": 139 }, { "epoch": 0.448, "grad_norm": 0.1964101791381836, "learning_rate": 0.000112987012987013, "loss": 1.0906, "step": 140 }, { "epoch": 0.4512, "grad_norm": 0.22406511008739471, "learning_rate": 0.00011233766233766234, "loss": 1.0594, "step": 141 }, { "epoch": 0.4544, "grad_norm": 0.23787978291511536, "learning_rate": 0.00011168831168831168, "loss": 1.1053, "step": 142 }, { "epoch": 0.4576, "grad_norm": 0.21196185052394867, "learning_rate": 0.00011103896103896105, "loss": 1.0923, "step": 143 }, { "epoch": 0.4608, "grad_norm": 0.21042804419994354, "learning_rate": 0.0001103896103896104, "loss": 1.0381, "step": 144 }, { "epoch": 0.464, "grad_norm": 0.2267436534166336, "learning_rate": 0.00010974025974025974, "loss": 1.0818, "step": 145 }, { "epoch": 0.4672, "grad_norm": 0.23742735385894775, "learning_rate": 0.00010909090909090909, "loss": 1.0872, "step": 146 }, { "epoch": 0.4704, "grad_norm": 0.17787213623523712, "learning_rate": 0.00010844155844155846, "loss": 1.03, "step": 147 }, { "epoch": 0.4736, "grad_norm": 0.22422832250595093, "learning_rate": 0.0001077922077922078, "loss": 1.0738, "step": 148 }, { "epoch": 0.4768, "grad_norm": 0.22946301102638245, "learning_rate": 0.00010714285714285715, "loss": 1.0274, "step": 149 }, { "epoch": 0.48, "grad_norm": 0.2137996405363083, "learning_rate": 0.00010649350649350649, "loss": 1.0539, "step": 150 }, { "epoch": 0.4832, "grad_norm": 0.1748756766319275, "learning_rate": 0.00010584415584415586, "loss": 1.0355, "step": 151 }, { "epoch": 0.4864, "grad_norm": 0.22275175154209137, "learning_rate": 0.0001051948051948052, "loss": 1.1696, "step": 152 }, { "epoch": 0.4896, "grad_norm": 0.20996077358722687, "learning_rate": 0.00010454545454545455, "loss": 1.0303, "step": 153 }, { "epoch": 0.4928, "grad_norm": 0.1945938766002655, "learning_rate": 0.00010389610389610389, "loss": 0.9747, "step": 154 }, { "epoch": 0.496, "grad_norm": 0.1970377266407013, "learning_rate": 0.00010324675324675325, "loss": 1.0358, "step": 155 }, { "epoch": 0.4992, "grad_norm": 0.18814732134342194, "learning_rate": 0.00010259740259740261, "loss": 0.9612, "step": 156 }, { "epoch": 0.5024, "grad_norm": 0.2153233289718628, "learning_rate": 0.00010194805194805195, "loss": 1.0749, "step": 157 }, { "epoch": 0.5056, "grad_norm": 0.21788008511066437, "learning_rate": 0.0001012987012987013, "loss": 1.0883, "step": 158 }, { "epoch": 0.5088, "grad_norm": 0.214650496840477, "learning_rate": 0.00010064935064935067, "loss": 1.0539, "step": 159 }, { "epoch": 0.512, "grad_norm": 0.19312834739685059, "learning_rate": 0.0001, "loss": 1.0657, "step": 160 }, { "epoch": 0.5152, "grad_norm": 0.19916598498821259, "learning_rate": 9.935064935064936e-05, "loss": 1.0478, "step": 161 }, { "epoch": 0.5184, "grad_norm": 0.2057606726884842, "learning_rate": 9.870129870129871e-05, "loss": 1.0094, "step": 162 }, { "epoch": 0.5216, "grad_norm": 0.22159607708454132, "learning_rate": 9.805194805194806e-05, "loss": 1.0952, "step": 163 }, { "epoch": 0.5248, "grad_norm": 0.18274275958538055, "learning_rate": 9.74025974025974e-05, "loss": 1.0065, "step": 164 }, { "epoch": 0.528, "grad_norm": 0.19835162162780762, "learning_rate": 9.675324675324677e-05, "loss": 1.0742, "step": 165 }, { "epoch": 0.5312, "grad_norm": 0.2114904820919037, "learning_rate": 9.610389610389611e-05, "loss": 1.1109, "step": 166 }, { "epoch": 0.5344, "grad_norm": 0.21488523483276367, "learning_rate": 9.545454545454546e-05, "loss": 1.0465, "step": 167 }, { "epoch": 0.5376, "grad_norm": 0.19870303571224213, "learning_rate": 9.480519480519481e-05, "loss": 1.0318, "step": 168 }, { "epoch": 0.5408, "grad_norm": 0.20413029193878174, "learning_rate": 9.415584415584417e-05, "loss": 1.0817, "step": 169 }, { "epoch": 0.544, "grad_norm": 0.1847231239080429, "learning_rate": 9.35064935064935e-05, "loss": 1.0144, "step": 170 }, { "epoch": 0.5472, "grad_norm": 0.2715964913368225, "learning_rate": 9.285714285714286e-05, "loss": 0.9832, "step": 171 }, { "epoch": 0.5504, "grad_norm": 0.2225002497434616, "learning_rate": 9.220779220779221e-05, "loss": 1.1051, "step": 172 }, { "epoch": 0.5536, "grad_norm": 0.22931510210037231, "learning_rate": 9.155844155844156e-05, "loss": 1.1042, "step": 173 }, { "epoch": 0.5568, "grad_norm": 0.21848627924919128, "learning_rate": 9.090909090909092e-05, "loss": 1.1151, "step": 174 }, { "epoch": 0.56, "grad_norm": 0.19852259755134583, "learning_rate": 9.025974025974027e-05, "loss": 1.0889, "step": 175 }, { "epoch": 0.5632, "grad_norm": 0.2080363780260086, "learning_rate": 8.961038961038961e-05, "loss": 1.0777, "step": 176 }, { "epoch": 0.5664, "grad_norm": 0.22391024231910706, "learning_rate": 8.896103896103896e-05, "loss": 1.1092, "step": 177 }, { "epoch": 0.5696, "grad_norm": 0.21793846786022186, "learning_rate": 8.831168831168831e-05, "loss": 1.044, "step": 178 }, { "epoch": 0.5728, "grad_norm": 0.2009749859571457, "learning_rate": 8.766233766233767e-05, "loss": 1.0198, "step": 179 }, { "epoch": 0.576, "grad_norm": 0.19432318210601807, "learning_rate": 8.701298701298701e-05, "loss": 1.075, "step": 180 }, { "epoch": 0.5792, "grad_norm": 0.18634547293186188, "learning_rate": 8.636363636363637e-05, "loss": 0.9964, "step": 181 }, { "epoch": 0.5824, "grad_norm": 0.1947103589773178, "learning_rate": 8.571428571428571e-05, "loss": 1.0025, "step": 182 }, { "epoch": 0.5856, "grad_norm": 0.23098671436309814, "learning_rate": 8.506493506493507e-05, "loss": 1.0562, "step": 183 }, { "epoch": 0.5888, "grad_norm": 0.19686414301395416, "learning_rate": 8.441558441558442e-05, "loss": 1.0285, "step": 184 }, { "epoch": 0.592, "grad_norm": 0.19852428138256073, "learning_rate": 8.376623376623377e-05, "loss": 1.0054, "step": 185 }, { "epoch": 0.5952, "grad_norm": 0.21483510732650757, "learning_rate": 8.311688311688312e-05, "loss": 1.108, "step": 186 }, { "epoch": 0.5984, "grad_norm": 0.23313644528388977, "learning_rate": 8.246753246753248e-05, "loss": 1.1383, "step": 187 }, { "epoch": 0.6016, "grad_norm": 0.21453145146369934, "learning_rate": 8.181818181818183e-05, "loss": 1.0911, "step": 188 }, { "epoch": 0.6048, "grad_norm": 0.20268195867538452, "learning_rate": 8.116883116883117e-05, "loss": 1.0145, "step": 189 }, { "epoch": 0.608, "grad_norm": 0.20576398074626923, "learning_rate": 8.051948051948052e-05, "loss": 1.0829, "step": 190 }, { "epoch": 0.6112, "grad_norm": 0.21732626855373383, "learning_rate": 7.987012987012987e-05, "loss": 1.0152, "step": 191 }, { "epoch": 0.6144, "grad_norm": 0.22046895325183868, "learning_rate": 7.922077922077923e-05, "loss": 1.1311, "step": 192 }, { "epoch": 0.6176, "grad_norm": 0.19727715849876404, "learning_rate": 7.857142857142858e-05, "loss": 1.0364, "step": 193 }, { "epoch": 0.6208, "grad_norm": 0.20861488580703735, "learning_rate": 7.792207792207793e-05, "loss": 1.0435, "step": 194 }, { "epoch": 0.624, "grad_norm": 0.18545083701610565, "learning_rate": 7.727272727272727e-05, "loss": 1.0299, "step": 195 }, { "epoch": 0.6272, "grad_norm": 0.19965052604675293, "learning_rate": 7.662337662337662e-05, "loss": 1.0511, "step": 196 }, { "epoch": 0.6304, "grad_norm": 0.23673909902572632, "learning_rate": 7.597402597402598e-05, "loss": 1.081, "step": 197 }, { "epoch": 0.6336, "grad_norm": 0.17583179473876953, "learning_rate": 7.532467532467533e-05, "loss": 0.9808, "step": 198 }, { "epoch": 0.6368, "grad_norm": 0.2129366099834442, "learning_rate": 7.467532467532467e-05, "loss": 1.0522, "step": 199 }, { "epoch": 0.64, "grad_norm": 0.21679140627384186, "learning_rate": 7.402597402597404e-05, "loss": 1.0567, "step": 200 }, { "epoch": 0.6432, "grad_norm": 0.2032000720500946, "learning_rate": 7.337662337662338e-05, "loss": 1.0466, "step": 201 }, { "epoch": 0.6464, "grad_norm": 0.1887970268726349, "learning_rate": 7.272727272727273e-05, "loss": 1.0329, "step": 202 }, { "epoch": 0.6496, "grad_norm": 0.21060192584991455, "learning_rate": 7.207792207792208e-05, "loss": 1.1021, "step": 203 }, { "epoch": 0.6528, "grad_norm": 0.21191425621509552, "learning_rate": 7.142857142857143e-05, "loss": 0.99, "step": 204 }, { "epoch": 0.656, "grad_norm": 0.1995989829301834, "learning_rate": 7.077922077922077e-05, "loss": 1.0526, "step": 205 }, { "epoch": 0.6592, "grad_norm": 0.1849513053894043, "learning_rate": 7.012987012987014e-05, "loss": 0.9998, "step": 206 }, { "epoch": 0.6624, "grad_norm": 0.1948779672384262, "learning_rate": 6.948051948051948e-05, "loss": 1.075, "step": 207 }, { "epoch": 0.6656, "grad_norm": 0.20374052226543427, "learning_rate": 6.883116883116883e-05, "loss": 1.0933, "step": 208 }, { "epoch": 0.6688, "grad_norm": 0.2102465033531189, "learning_rate": 6.818181818181818e-05, "loss": 1.1123, "step": 209 }, { "epoch": 0.672, "grad_norm": 0.21376173198223114, "learning_rate": 6.753246753246754e-05, "loss": 1.1233, "step": 210 }, { "epoch": 0.6752, "grad_norm": 0.20934203267097473, "learning_rate": 6.688311688311688e-05, "loss": 1.1374, "step": 211 }, { "epoch": 0.6784, "grad_norm": 0.18604128062725067, "learning_rate": 6.623376623376624e-05, "loss": 1.0213, "step": 212 }, { "epoch": 0.6816, "grad_norm": 0.19644233584403992, "learning_rate": 6.55844155844156e-05, "loss": 1.0046, "step": 213 }, { "epoch": 0.6848, "grad_norm": 0.18479463458061218, "learning_rate": 6.493506493506494e-05, "loss": 0.9792, "step": 214 }, { "epoch": 0.688, "grad_norm": 0.1945149153470993, "learning_rate": 6.428571428571429e-05, "loss": 1.0584, "step": 215 }, { "epoch": 0.6912, "grad_norm": 0.2070147544145584, "learning_rate": 6.363636363636364e-05, "loss": 1.071, "step": 216 }, { "epoch": 0.6944, "grad_norm": 0.19645985960960388, "learning_rate": 6.2987012987013e-05, "loss": 1.0721, "step": 217 }, { "epoch": 0.6976, "grad_norm": 0.1960117667913437, "learning_rate": 6.233766233766233e-05, "loss": 1.071, "step": 218 }, { "epoch": 0.7008, "grad_norm": 0.20168261229991913, "learning_rate": 6.16883116883117e-05, "loss": 1.0808, "step": 219 }, { "epoch": 0.704, "grad_norm": 0.21254412829875946, "learning_rate": 6.103896103896104e-05, "loss": 1.0287, "step": 220 }, { "epoch": 0.7072, "grad_norm": 0.21271063387393951, "learning_rate": 6.03896103896104e-05, "loss": 1.0605, "step": 221 }, { "epoch": 0.7104, "grad_norm": 0.2081408053636551, "learning_rate": 5.9740259740259744e-05, "loss": 1.091, "step": 222 }, { "epoch": 0.7136, "grad_norm": 0.21113798022270203, "learning_rate": 5.90909090909091e-05, "loss": 1.1323, "step": 223 }, { "epoch": 0.7168, "grad_norm": 0.20670844614505768, "learning_rate": 5.844155844155844e-05, "loss": 1.0955, "step": 224 }, { "epoch": 0.72, "grad_norm": 0.2010120451450348, "learning_rate": 5.7792207792207796e-05, "loss": 1.1068, "step": 225 }, { "epoch": 0.7232, "grad_norm": 0.20379121601581573, "learning_rate": 5.714285714285714e-05, "loss": 1.0419, "step": 226 }, { "epoch": 0.7264, "grad_norm": 0.22799807786941528, "learning_rate": 5.64935064935065e-05, "loss": 1.0904, "step": 227 }, { "epoch": 0.7296, "grad_norm": 0.2005995213985443, "learning_rate": 5.584415584415584e-05, "loss": 1.078, "step": 228 }, { "epoch": 0.7328, "grad_norm": 0.20329605042934418, "learning_rate": 5.51948051948052e-05, "loss": 1.0245, "step": 229 }, { "epoch": 0.736, "grad_norm": 0.19283504784107208, "learning_rate": 5.4545454545454546e-05, "loss": 1.0367, "step": 230 }, { "epoch": 0.7392, "grad_norm": 0.20624355971813202, "learning_rate": 5.38961038961039e-05, "loss": 1.1046, "step": 231 }, { "epoch": 0.7424, "grad_norm": 0.21362991631031036, "learning_rate": 5.3246753246753245e-05, "loss": 1.1104, "step": 232 }, { "epoch": 0.7456, "grad_norm": 0.20447863638401031, "learning_rate": 5.25974025974026e-05, "loss": 1.0514, "step": 233 }, { "epoch": 0.7488, "grad_norm": 0.1974381059408188, "learning_rate": 5.1948051948051944e-05, "loss": 1.0048, "step": 234 }, { "epoch": 0.752, "grad_norm": 0.21237170696258545, "learning_rate": 5.1298701298701304e-05, "loss": 1.1299, "step": 235 }, { "epoch": 0.7552, "grad_norm": 0.21224971115589142, "learning_rate": 5.064935064935065e-05, "loss": 1.05, "step": 236 }, { "epoch": 0.7584, "grad_norm": 0.19865018129348755, "learning_rate": 5e-05, "loss": 1.0665, "step": 237 }, { "epoch": 0.7616, "grad_norm": 0.19199275970458984, "learning_rate": 4.9350649350649355e-05, "loss": 0.9531, "step": 238 }, { "epoch": 0.7648, "grad_norm": 0.19573214650154114, "learning_rate": 4.87012987012987e-05, "loss": 1.0318, "step": 239 }, { "epoch": 0.768, "grad_norm": 0.21338805556297302, "learning_rate": 4.8051948051948054e-05, "loss": 1.0343, "step": 240 }, { "epoch": 0.7712, "grad_norm": 0.2254691869020462, "learning_rate": 4.740259740259741e-05, "loss": 1.0472, "step": 241 }, { "epoch": 0.7744, "grad_norm": 0.18101665377616882, "learning_rate": 4.675324675324675e-05, "loss": 1.017, "step": 242 }, { "epoch": 0.7776, "grad_norm": 0.22090592980384827, "learning_rate": 4.6103896103896106e-05, "loss": 1.0389, "step": 243 }, { "epoch": 0.7808, "grad_norm": 0.20865507423877716, "learning_rate": 4.545454545454546e-05, "loss": 1.0369, "step": 244 }, { "epoch": 0.784, "grad_norm": 0.21619610488414764, "learning_rate": 4.4805194805194805e-05, "loss": 1.109, "step": 245 }, { "epoch": 0.7872, "grad_norm": 0.21694771945476532, "learning_rate": 4.415584415584416e-05, "loss": 1.0525, "step": 246 }, { "epoch": 0.7904, "grad_norm": 0.2182662934064865, "learning_rate": 4.3506493506493503e-05, "loss": 1.0331, "step": 247 }, { "epoch": 0.7936, "grad_norm": 0.2026486098766327, "learning_rate": 4.2857142857142856e-05, "loss": 1.027, "step": 248 }, { "epoch": 0.7968, "grad_norm": 0.19606547057628632, "learning_rate": 4.220779220779221e-05, "loss": 1.0242, "step": 249 }, { "epoch": 0.8, "grad_norm": 0.22107470035552979, "learning_rate": 4.155844155844156e-05, "loss": 1.0924, "step": 250 }, { "epoch": 0.8032, "grad_norm": 0.19960008561611176, "learning_rate": 4.0909090909090915e-05, "loss": 1.0384, "step": 251 }, { "epoch": 0.8064, "grad_norm": 0.1945488154888153, "learning_rate": 4.025974025974026e-05, "loss": 1.0673, "step": 252 }, { "epoch": 0.8096, "grad_norm": 0.22067414224147797, "learning_rate": 3.9610389610389614e-05, "loss": 1.0426, "step": 253 }, { "epoch": 0.8128, "grad_norm": 0.19010980427265167, "learning_rate": 3.8961038961038966e-05, "loss": 1.0617, "step": 254 }, { "epoch": 0.816, "grad_norm": 0.18781176209449768, "learning_rate": 3.831168831168831e-05, "loss": 1.0243, "step": 255 }, { "epoch": 0.8192, "grad_norm": 0.20388829708099365, "learning_rate": 3.7662337662337665e-05, "loss": 1.0476, "step": 256 }, { "epoch": 0.8224, "grad_norm": 0.19911155104637146, "learning_rate": 3.701298701298702e-05, "loss": 1.0324, "step": 257 }, { "epoch": 0.8256, "grad_norm": 0.19884039461612701, "learning_rate": 3.6363636363636364e-05, "loss": 1.0242, "step": 258 }, { "epoch": 0.8288, "grad_norm": 0.19036105275154114, "learning_rate": 3.571428571428572e-05, "loss": 1.0323, "step": 259 }, { "epoch": 0.832, "grad_norm": 0.20039844512939453, "learning_rate": 3.506493506493507e-05, "loss": 1.0749, "step": 260 }, { "epoch": 0.8352, "grad_norm": 0.1899934560060501, "learning_rate": 3.4415584415584416e-05, "loss": 1.0115, "step": 261 }, { "epoch": 0.8384, "grad_norm": 0.20019090175628662, "learning_rate": 3.376623376623377e-05, "loss": 1.0782, "step": 262 }, { "epoch": 0.8416, "grad_norm": 0.2020583152770996, "learning_rate": 3.311688311688312e-05, "loss": 1.0687, "step": 263 }, { "epoch": 0.8448, "grad_norm": 0.21407337486743927, "learning_rate": 3.246753246753247e-05, "loss": 1.1015, "step": 264 }, { "epoch": 0.848, "grad_norm": 0.1871640682220459, "learning_rate": 3.181818181818182e-05, "loss": 0.9637, "step": 265 }, { "epoch": 0.8512, "grad_norm": 0.21622811257839203, "learning_rate": 3.1168831168831166e-05, "loss": 1.1222, "step": 266 }, { "epoch": 0.8544, "grad_norm": 0.22504661977291107, "learning_rate": 3.051948051948052e-05, "loss": 1.132, "step": 267 }, { "epoch": 0.8576, "grad_norm": 0.19177629053592682, "learning_rate": 2.9870129870129872e-05, "loss": 1.0281, "step": 268 }, { "epoch": 0.8608, "grad_norm": 0.1970544159412384, "learning_rate": 2.922077922077922e-05, "loss": 1.0393, "step": 269 }, { "epoch": 0.864, "grad_norm": 0.21554522216320038, "learning_rate": 2.857142857142857e-05, "loss": 1.074, "step": 270 }, { "epoch": 0.8672, "grad_norm": 0.21131229400634766, "learning_rate": 2.792207792207792e-05, "loss": 1.054, "step": 271 }, { "epoch": 0.8704, "grad_norm": 0.19816523790359497, "learning_rate": 2.7272727272727273e-05, "loss": 1.0456, "step": 272 }, { "epoch": 0.8736, "grad_norm": 0.21075209975242615, "learning_rate": 2.6623376623376623e-05, "loss": 1.0758, "step": 273 }, { "epoch": 0.8768, "grad_norm": 0.2296527624130249, "learning_rate": 2.5974025974025972e-05, "loss": 1.0917, "step": 274 }, { "epoch": 0.88, "grad_norm": 0.19722610712051392, "learning_rate": 2.5324675324675325e-05, "loss": 1.0704, "step": 275 }, { "epoch": 0.8832, "grad_norm": 0.18721099197864532, "learning_rate": 2.4675324675324678e-05, "loss": 0.9919, "step": 276 }, { "epoch": 0.8864, "grad_norm": 0.20244193077087402, "learning_rate": 2.4025974025974027e-05, "loss": 1.0368, "step": 277 }, { "epoch": 0.8896, "grad_norm": 0.19518914818763733, "learning_rate": 2.3376623376623376e-05, "loss": 1.0436, "step": 278 }, { "epoch": 0.8928, "grad_norm": 0.19650357961654663, "learning_rate": 2.272727272727273e-05, "loss": 1.0306, "step": 279 }, { "epoch": 0.896, "grad_norm": 0.20320096611976624, "learning_rate": 2.207792207792208e-05, "loss": 1.0941, "step": 280 }, { "epoch": 0.8992, "grad_norm": 0.18296951055526733, "learning_rate": 2.1428571428571428e-05, "loss": 0.9802, "step": 281 }, { "epoch": 0.9024, "grad_norm": 0.21357610821723938, "learning_rate": 2.077922077922078e-05, "loss": 1.0449, "step": 282 }, { "epoch": 0.9056, "grad_norm": 0.193921759724617, "learning_rate": 2.012987012987013e-05, "loss": 1.0116, "step": 283 }, { "epoch": 0.9088, "grad_norm": 0.1953902244567871, "learning_rate": 1.9480519480519483e-05, "loss": 1.0105, "step": 284 }, { "epoch": 0.912, "grad_norm": 0.19440975785255432, "learning_rate": 1.8831168831168833e-05, "loss": 0.9952, "step": 285 }, { "epoch": 0.9152, "grad_norm": 0.21054105460643768, "learning_rate": 1.8181818181818182e-05, "loss": 1.0701, "step": 286 }, { "epoch": 0.9184, "grad_norm": 0.18844804167747498, "learning_rate": 1.7532467532467535e-05, "loss": 1.0146, "step": 287 }, { "epoch": 0.9216, "grad_norm": 0.2067311704158783, "learning_rate": 1.6883116883116884e-05, "loss": 1.0781, "step": 288 }, { "epoch": 0.9248, "grad_norm": 0.1941213756799698, "learning_rate": 1.6233766233766234e-05, "loss": 0.9814, "step": 289 }, { "epoch": 0.928, "grad_norm": 0.22726193070411682, "learning_rate": 1.5584415584415583e-05, "loss": 1.1431, "step": 290 }, { "epoch": 0.9312, "grad_norm": 0.18025581538677216, "learning_rate": 1.4935064935064936e-05, "loss": 0.9649, "step": 291 }, { "epoch": 0.9344, "grad_norm": 0.21535000205039978, "learning_rate": 1.4285714285714285e-05, "loss": 1.0441, "step": 292 }, { "epoch": 0.9376, "grad_norm": 0.20014546811580658, "learning_rate": 1.3636363636363637e-05, "loss": 1.0166, "step": 293 }, { "epoch": 0.9408, "grad_norm": 0.22738787531852722, "learning_rate": 1.2987012987012986e-05, "loss": 1.0564, "step": 294 }, { "epoch": 0.944, "grad_norm": 0.2020861804485321, "learning_rate": 1.2337662337662339e-05, "loss": 1.1241, "step": 295 }, { "epoch": 0.9472, "grad_norm": 0.19888809323310852, "learning_rate": 1.1688311688311688e-05, "loss": 1.1114, "step": 296 }, { "epoch": 0.9504, "grad_norm": 0.20912377536296844, "learning_rate": 1.103896103896104e-05, "loss": 1.0971, "step": 297 }, { "epoch": 0.9536, "grad_norm": 0.21206621825695038, "learning_rate": 1.038961038961039e-05, "loss": 1.0601, "step": 298 }, { "epoch": 0.9568, "grad_norm": 0.18667680025100708, "learning_rate": 9.740259740259742e-06, "loss": 1.0291, "step": 299 }, { "epoch": 0.96, "grad_norm": 0.21125559508800507, "learning_rate": 9.090909090909091e-06, "loss": 1.0483, "step": 300 }, { "epoch": 0.9632, "grad_norm": 0.21776145696640015, "learning_rate": 8.441558441558442e-06, "loss": 0.9912, "step": 301 }, { "epoch": 0.9664, "grad_norm": 0.20144303143024445, "learning_rate": 7.792207792207792e-06, "loss": 1.0357, "step": 302 }, { "epoch": 0.9696, "grad_norm": 0.1984029859304428, "learning_rate": 7.142857142857143e-06, "loss": 1.0648, "step": 303 }, { "epoch": 0.9728, "grad_norm": 0.17972829937934875, "learning_rate": 6.493506493506493e-06, "loss": 1.0033, "step": 304 }, { "epoch": 0.976, "grad_norm": 0.1818286031484604, "learning_rate": 5.844155844155844e-06, "loss": 0.997, "step": 305 }, { "epoch": 0.9792, "grad_norm": 0.19670912623405457, "learning_rate": 5.194805194805195e-06, "loss": 1.0256, "step": 306 }, { "epoch": 0.9824, "grad_norm": 0.20527283847332, "learning_rate": 4.5454545454545455e-06, "loss": 1.0348, "step": 307 }, { "epoch": 0.9856, "grad_norm": 0.19025909900665283, "learning_rate": 3.896103896103896e-06, "loss": 1.0682, "step": 308 }, { "epoch": 0.9888, "grad_norm": 0.19544818997383118, "learning_rate": 3.2467532467532465e-06, "loss": 0.9872, "step": 309 }, { "epoch": 0.992, "grad_norm": 0.22112183272838593, "learning_rate": 2.5974025974025976e-06, "loss": 1.0661, "step": 310 }, { "epoch": 0.9952, "grad_norm": 0.23328153789043427, "learning_rate": 1.948051948051948e-06, "loss": 1.0691, "step": 311 }, { "epoch": 0.9984, "grad_norm": 0.20181375741958618, "learning_rate": 1.2987012987012988e-06, "loss": 0.9416, "step": 312 }, { "epoch": 1.0, "grad_norm": 0.29312625527381897, "learning_rate": 6.493506493506494e-07, "loss": 1.1216, "step": 313 }, { "epoch": 0.12202467696492762, "grad_norm": 0.2231415957212448, "learning_rate": 0.0, "loss": 1.0468, "step": 314 }, { "epoch": 0.12241329058583503, "grad_norm": 0.22263288497924805, "learning_rate": 0.00017594394706111328, "loss": 1.0399, "step": 315 }, { "epoch": 0.12280190420674245, "grad_norm": 0.22909891605377197, "learning_rate": 0.00017586609575710393, "loss": 1.1069, "step": 316 }, { "epoch": 0.12319051782764986, "grad_norm": 0.23951445519924164, "learning_rate": 0.0001757882444530946, "loss": 1.1036, "step": 317 }, { "epoch": 0.12357913144855727, "grad_norm": 0.2409268021583557, "learning_rate": 0.00017571039314908526, "loss": 1.1114, "step": 318 }, { "epoch": 0.12396774506946469, "grad_norm": 0.23753899335861206, "learning_rate": 0.00017563254184507592, "loss": 1.1297, "step": 319 }, { "epoch": 0.12435635869037209, "grad_norm": 0.2823902666568756, "learning_rate": 0.00017555469054106657, "loss": 1.1293, "step": 320 }, { "epoch": 0.12474497231127951, "grad_norm": 0.24093545973300934, "learning_rate": 0.00017547683923705722, "loss": 1.0678, "step": 321 }, { "epoch": 0.12513358593218693, "grad_norm": 0.22565563023090363, "learning_rate": 0.0001753989879330479, "loss": 1.1408, "step": 322 }, { "epoch": 0.12552219955309435, "grad_norm": 0.22569572925567627, "learning_rate": 0.00017532113662903855, "loss": 1.0543, "step": 323 }, { "epoch": 0.12591081317400174, "grad_norm": 0.24962866306304932, "learning_rate": 0.0001752432853250292, "loss": 1.0818, "step": 324 }, { "epoch": 0.12629942679490916, "grad_norm": 0.22184576094150543, "learning_rate": 0.00017516543402101986, "loss": 1.0835, "step": 325 }, { "epoch": 0.12668804041581658, "grad_norm": 0.2572194039821625, "learning_rate": 0.0001750875827170105, "loss": 1.0767, "step": 326 }, { "epoch": 0.127076654036724, "grad_norm": 0.24131342768669128, "learning_rate": 0.00017500973141300116, "loss": 1.0981, "step": 327 }, { "epoch": 0.1274652676576314, "grad_norm": 0.2386389970779419, "learning_rate": 0.00017493188010899184, "loss": 1.0828, "step": 328 }, { "epoch": 0.1278538812785388, "grad_norm": 0.2654125690460205, "learning_rate": 0.0001748540288049825, "loss": 1.1266, "step": 329 }, { "epoch": 0.12824249489944622, "grad_norm": 0.2925739884376526, "learning_rate": 0.00017477617750097314, "loss": 1.0983, "step": 330 }, { "epoch": 0.12863110852035364, "grad_norm": 0.26589342951774597, "learning_rate": 0.0001746983261969638, "loss": 1.1029, "step": 331 }, { "epoch": 0.12901972214126106, "grad_norm": 0.24565957486629486, "learning_rate": 0.00017462047489295445, "loss": 1.0975, "step": 332 }, { "epoch": 0.12940833576216845, "grad_norm": 0.2459682673215866, "learning_rate": 0.00017454262358894513, "loss": 1.0566, "step": 333 }, { "epoch": 0.12979694938307587, "grad_norm": 0.23349183797836304, "learning_rate": 0.00017446477228493578, "loss": 1.0833, "step": 334 }, { "epoch": 0.1301855630039833, "grad_norm": 0.26166337728500366, "learning_rate": 0.00017438692098092643, "loss": 1.1598, "step": 335 }, { "epoch": 0.1305741766248907, "grad_norm": 0.24188168346881866, "learning_rate": 0.00017430906967691708, "loss": 1.0728, "step": 336 }, { "epoch": 0.13096279024579813, "grad_norm": 0.22922398149967194, "learning_rate": 0.00017423121837290773, "loss": 1.0311, "step": 337 }, { "epoch": 0.13135140386670552, "grad_norm": 0.2652754485607147, "learning_rate": 0.00017415336706889841, "loss": 1.1096, "step": 338 }, { "epoch": 0.13174001748761294, "grad_norm": 0.2355881780385971, "learning_rate": 0.00017407551576488907, "loss": 1.0964, "step": 339 }, { "epoch": 0.13212863110852036, "grad_norm": 0.244523823261261, "learning_rate": 0.00017399766446087972, "loss": 1.142, "step": 340 }, { "epoch": 0.13251724472942777, "grad_norm": 0.24705976247787476, "learning_rate": 0.00017391981315687037, "loss": 1.0943, "step": 341 }, { "epoch": 0.13290585835033517, "grad_norm": 0.22817552089691162, "learning_rate": 0.00017384196185286102, "loss": 1.0621, "step": 342 }, { "epoch": 0.13329447197124258, "grad_norm": 0.22605225443840027, "learning_rate": 0.0001737641105488517, "loss": 1.0714, "step": 343 }, { "epoch": 0.13368308559215, "grad_norm": 0.2584545314311981, "learning_rate": 0.00017368625924484235, "loss": 1.1367, "step": 344 }, { "epoch": 0.13407169921305742, "grad_norm": 0.2248220443725586, "learning_rate": 0.000173608407940833, "loss": 1.0872, "step": 345 }, { "epoch": 0.13446031283396484, "grad_norm": 0.2141868770122528, "learning_rate": 0.00017353055663682368, "loss": 1.0572, "step": 346 }, { "epoch": 0.13484892645487223, "grad_norm": 0.2615523934364319, "learning_rate": 0.00017345270533281434, "loss": 1.1048, "step": 347 }, { "epoch": 0.13523754007577965, "grad_norm": 0.22990448772907257, "learning_rate": 0.000173374854028805, "loss": 1.0528, "step": 348 }, { "epoch": 0.13562615369668707, "grad_norm": 0.2132262885570526, "learning_rate": 0.00017329700272479564, "loss": 1.0476, "step": 349 }, { "epoch": 0.1360147673175945, "grad_norm": 0.2578272819519043, "learning_rate": 0.00017321915142078632, "loss": 1.0852, "step": 350 }, { "epoch": 0.1364033809385019, "grad_norm": 0.22881457209587097, "learning_rate": 0.00017314130011677697, "loss": 1.1017, "step": 351 }, { "epoch": 0.1367919945594093, "grad_norm": 0.21067696809768677, "learning_rate": 0.00017306344881276762, "loss": 1.0444, "step": 352 }, { "epoch": 0.13718060818031672, "grad_norm": 0.2304215282201767, "learning_rate": 0.0001729855975087583, "loss": 1.0737, "step": 353 }, { "epoch": 0.13756922180122413, "grad_norm": 0.2031925916671753, "learning_rate": 0.00017290774620474895, "loss": 1.0036, "step": 354 }, { "epoch": 0.13795783542213155, "grad_norm": 0.27281051874160767, "learning_rate": 0.0001728298949007396, "loss": 1.148, "step": 355 }, { "epoch": 0.13834644904303897, "grad_norm": 0.204191654920578, "learning_rate": 0.00017275204359673026, "loss": 0.9607, "step": 356 }, { "epoch": 0.13873506266394636, "grad_norm": 0.221976637840271, "learning_rate": 0.0001726741922927209, "loss": 1.1068, "step": 357 }, { "epoch": 0.13912367628485378, "grad_norm": 0.20831729471683502, "learning_rate": 0.0001725963409887116, "loss": 1.034, "step": 358 }, { "epoch": 0.1395122899057612, "grad_norm": 0.21639779210090637, "learning_rate": 0.00017251848968470224, "loss": 1.0613, "step": 359 }, { "epoch": 0.13990090352666862, "grad_norm": 0.1959424465894699, "learning_rate": 0.0001724406383806929, "loss": 1.0506, "step": 360 }, { "epoch": 0.140289517147576, "grad_norm": 0.2044398933649063, "learning_rate": 0.00017236278707668355, "loss": 1.0316, "step": 361 }, { "epoch": 0.14067813076848343, "grad_norm": 0.21483004093170166, "learning_rate": 0.0001722849357726742, "loss": 1.0361, "step": 362 }, { "epoch": 0.14106674438939085, "grad_norm": 0.237701416015625, "learning_rate": 0.00017220708446866485, "loss": 1.1264, "step": 363 }, { "epoch": 0.14145535801029827, "grad_norm": 0.20750795304775238, "learning_rate": 0.00017212923316465553, "loss": 1.0523, "step": 364 }, { "epoch": 0.14184397163120568, "grad_norm": 0.2252965271472931, "learning_rate": 0.00017205138186064618, "loss": 1.0764, "step": 365 }, { "epoch": 0.14223258525211308, "grad_norm": 0.2033565789461136, "learning_rate": 0.00017197353055663683, "loss": 1.064, "step": 366 }, { "epoch": 0.1426211988730205, "grad_norm": 0.21123190224170685, "learning_rate": 0.00017189567925262749, "loss": 1.0515, "step": 367 }, { "epoch": 0.1430098124939279, "grad_norm": 0.20646221935749054, "learning_rate": 0.00017181782794861814, "loss": 1.0617, "step": 368 }, { "epoch": 0.14339842611483533, "grad_norm": 0.2079589068889618, "learning_rate": 0.00017173997664460882, "loss": 1.0569, "step": 369 }, { "epoch": 0.14378703973574275, "grad_norm": 0.216246098279953, "learning_rate": 0.00017166212534059947, "loss": 1.0986, "step": 370 }, { "epoch": 0.14417565335665014, "grad_norm": 0.20711806416511536, "learning_rate": 0.00017158427403659012, "loss": 1.1342, "step": 371 }, { "epoch": 0.14456426697755756, "grad_norm": 0.235435351729393, "learning_rate": 0.00017150642273258077, "loss": 1.1082, "step": 372 }, { "epoch": 0.14495288059846498, "grad_norm": 0.2273191511631012, "learning_rate": 0.00017142857142857143, "loss": 1.1064, "step": 373 }, { "epoch": 0.1453414942193724, "grad_norm": 0.2075672745704651, "learning_rate": 0.0001713507201245621, "loss": 1.0536, "step": 374 }, { "epoch": 0.14573010784027982, "grad_norm": 0.20764274895191193, "learning_rate": 0.00017127286882055276, "loss": 1.0673, "step": 375 }, { "epoch": 0.1461187214611872, "grad_norm": 0.2441243678331375, "learning_rate": 0.0001711950175165434, "loss": 1.1271, "step": 376 }, { "epoch": 0.14650733508209463, "grad_norm": 0.2383374124765396, "learning_rate": 0.00017111716621253406, "loss": 1.083, "step": 377 }, { "epoch": 0.14689594870300204, "grad_norm": 0.2172410786151886, "learning_rate": 0.0001710393149085247, "loss": 1.0605, "step": 378 }, { "epoch": 0.14728456232390946, "grad_norm": 0.22591541707515717, "learning_rate": 0.0001709614636045154, "loss": 1.0931, "step": 379 }, { "epoch": 0.14767317594481685, "grad_norm": 0.23099495470523834, "learning_rate": 0.00017088361230050604, "loss": 1.1021, "step": 380 }, { "epoch": 0.14806178956572427, "grad_norm": 0.21461094915866852, "learning_rate": 0.0001708057609964967, "loss": 1.0959, "step": 381 }, { "epoch": 0.1484504031866317, "grad_norm": 0.21557241678237915, "learning_rate": 0.00017072790969248735, "loss": 1.0155, "step": 382 }, { "epoch": 0.1488390168075391, "grad_norm": 0.234396293759346, "learning_rate": 0.000170650058388478, "loss": 1.1289, "step": 383 }, { "epoch": 0.14922763042844653, "grad_norm": 0.22895503044128418, "learning_rate": 0.00017057220708446868, "loss": 0.9919, "step": 384 }, { "epoch": 0.14961624404935392, "grad_norm": 0.2054683268070221, "learning_rate": 0.00017049435578045933, "loss": 1.0607, "step": 385 }, { "epoch": 0.15000485767026134, "grad_norm": 0.25569215416908264, "learning_rate": 0.00017041650447644998, "loss": 1.0517, "step": 386 }, { "epoch": 0.15039347129116876, "grad_norm": 0.2222641259431839, "learning_rate": 0.00017033865317244064, "loss": 1.0404, "step": 387 }, { "epoch": 0.15078208491207618, "grad_norm": 0.20501169562339783, "learning_rate": 0.0001702608018684313, "loss": 0.9897, "step": 388 }, { "epoch": 0.1511706985329836, "grad_norm": 0.22080403566360474, "learning_rate": 0.00017018295056442197, "loss": 1.1013, "step": 389 }, { "epoch": 0.15155931215389098, "grad_norm": 0.21218529343605042, "learning_rate": 0.00017010509926041262, "loss": 1.0541, "step": 390 }, { "epoch": 0.1519479257747984, "grad_norm": 0.23064807057380676, "learning_rate": 0.00017002724795640327, "loss": 1.037, "step": 391 }, { "epoch": 0.15233653939570582, "grad_norm": 0.21164493262767792, "learning_rate": 0.00016994939665239392, "loss": 1.0769, "step": 392 }, { "epoch": 0.15272515301661324, "grad_norm": 0.22565549612045288, "learning_rate": 0.00016987154534838457, "loss": 1.0638, "step": 393 }, { "epoch": 0.15311376663752063, "grad_norm": 0.22492647171020508, "learning_rate": 0.00016979369404437525, "loss": 1.063, "step": 394 }, { "epoch": 0.15350238025842805, "grad_norm": 0.22335395216941833, "learning_rate": 0.0001697158427403659, "loss": 1.1032, "step": 395 }, { "epoch": 0.15389099387933547, "grad_norm": 0.2164154201745987, "learning_rate": 0.00016963799143635656, "loss": 1.1275, "step": 396 }, { "epoch": 0.1542796075002429, "grad_norm": 0.22547736763954163, "learning_rate": 0.0001695601401323472, "loss": 1.1324, "step": 397 }, { "epoch": 0.1546682211211503, "grad_norm": 0.2028045952320099, "learning_rate": 0.0001694822888283379, "loss": 1.0057, "step": 398 }, { "epoch": 0.1550568347420577, "grad_norm": 0.20770573616027832, "learning_rate": 0.00016940443752432854, "loss": 1.0311, "step": 399 }, { "epoch": 0.15544544836296512, "grad_norm": 0.2231476902961731, "learning_rate": 0.0001693265862203192, "loss": 1.0535, "step": 400 }, { "epoch": 0.15583406198387253, "grad_norm": 0.21618099510669708, "learning_rate": 0.00016924873491630987, "loss": 1.0616, "step": 401 }, { "epoch": 0.15622267560477995, "grad_norm": 0.24024419486522675, "learning_rate": 0.00016917088361230052, "loss": 1.1324, "step": 402 }, { "epoch": 0.15661128922568737, "grad_norm": 0.2002171128988266, "learning_rate": 0.00016909303230829118, "loss": 1.015, "step": 403 }, { "epoch": 0.15699990284659476, "grad_norm": 0.21771477162837982, "learning_rate": 0.00016901518100428183, "loss": 1.0817, "step": 404 }, { "epoch": 0.15738851646750218, "grad_norm": 0.22052259743213654, "learning_rate": 0.0001689373297002725, "loss": 1.0836, "step": 405 }, { "epoch": 0.1577771300884096, "grad_norm": 0.1964062750339508, "learning_rate": 0.00016885947839626316, "loss": 1.0505, "step": 406 }, { "epoch": 0.15816574370931702, "grad_norm": 0.22714298963546753, "learning_rate": 0.0001687816270922538, "loss": 1.0702, "step": 407 }, { "epoch": 0.15855435733022444, "grad_norm": 0.20647728443145752, "learning_rate": 0.00016870377578824446, "loss": 1.0349, "step": 408 }, { "epoch": 0.15894297095113183, "grad_norm": 0.2355160117149353, "learning_rate": 0.00016862592448423512, "loss": 1.0305, "step": 409 }, { "epoch": 0.15933158457203925, "grad_norm": 0.22890770435333252, "learning_rate": 0.0001685480731802258, "loss": 1.0854, "step": 410 }, { "epoch": 0.15972019819294667, "grad_norm": 0.21947838366031647, "learning_rate": 0.00016847022187621645, "loss": 1.0948, "step": 411 }, { "epoch": 0.16010881181385409, "grad_norm": 0.22334899008274078, "learning_rate": 0.0001683923705722071, "loss": 1.006, "step": 412 }, { "epoch": 0.16049742543476148, "grad_norm": 0.22324936091899872, "learning_rate": 0.00016831451926819775, "loss": 1.0402, "step": 413 }, { "epoch": 0.1608860390556689, "grad_norm": 0.21462097764015198, "learning_rate": 0.0001682366679641884, "loss": 1.077, "step": 414 }, { "epoch": 0.1612746526765763, "grad_norm": 0.24567006528377533, "learning_rate": 0.00016815881666017908, "loss": 1.15, "step": 415 }, { "epoch": 0.16166326629748373, "grad_norm": 0.26437243819236755, "learning_rate": 0.00016808096535616973, "loss": 1.1251, "step": 416 }, { "epoch": 0.16205187991839115, "grad_norm": 0.2217959761619568, "learning_rate": 0.00016800311405216039, "loss": 1.1103, "step": 417 }, { "epoch": 0.16244049353929854, "grad_norm": 0.24402475357055664, "learning_rate": 0.00016792526274815104, "loss": 1.0672, "step": 418 }, { "epoch": 0.16282910716020596, "grad_norm": 0.21609526872634888, "learning_rate": 0.0001678474114441417, "loss": 1.0291, "step": 419 }, { "epoch": 0.16321772078111338, "grad_norm": 0.20054642856121063, "learning_rate": 0.00016776956014013237, "loss": 1.0704, "step": 420 }, { "epoch": 0.1636063344020208, "grad_norm": 0.22864869236946106, "learning_rate": 0.00016769170883612302, "loss": 1.0612, "step": 421 }, { "epoch": 0.16399494802292822, "grad_norm": 0.22651974856853485, "learning_rate": 0.00016761385753211367, "loss": 1.0749, "step": 422 }, { "epoch": 0.1643835616438356, "grad_norm": 0.21587328612804413, "learning_rate": 0.00016753600622810433, "loss": 1.0398, "step": 423 }, { "epoch": 0.16477217526474303, "grad_norm": 0.1953774094581604, "learning_rate": 0.00016745815492409498, "loss": 1.0275, "step": 424 }, { "epoch": 0.16516078888565044, "grad_norm": 0.21803410351276398, "learning_rate": 0.00016738030362008566, "loss": 1.1219, "step": 425 }, { "epoch": 0.16554940250655786, "grad_norm": 0.2034682035446167, "learning_rate": 0.0001673024523160763, "loss": 1.0342, "step": 426 }, { "epoch": 0.16593801612746525, "grad_norm": 0.20135951042175293, "learning_rate": 0.00016722460101206696, "loss": 0.9802, "step": 427 }, { "epoch": 0.16632662974837267, "grad_norm": 0.23310376703739166, "learning_rate": 0.0001671467497080576, "loss": 1.0789, "step": 428 }, { "epoch": 0.1667152433692801, "grad_norm": 0.21475404500961304, "learning_rate": 0.00016706889840404827, "loss": 1.0416, "step": 429 }, { "epoch": 0.1671038569901875, "grad_norm": 0.21661072969436646, "learning_rate": 0.00016699104710003894, "loss": 1.0568, "step": 430 }, { "epoch": 0.16749247061109493, "grad_norm": 0.20310629904270172, "learning_rate": 0.0001669131957960296, "loss": 0.9968, "step": 431 }, { "epoch": 0.16788108423200232, "grad_norm": 0.2596947252750397, "learning_rate": 0.00016683534449202025, "loss": 1.0478, "step": 432 }, { "epoch": 0.16826969785290974, "grad_norm": 0.22226987779140472, "learning_rate": 0.0001667574931880109, "loss": 1.0898, "step": 433 }, { "epoch": 0.16865831147381716, "grad_norm": 0.22499911487102509, "learning_rate": 0.00016667964188400155, "loss": 1.07, "step": 434 }, { "epoch": 0.16904692509472458, "grad_norm": 0.2717292308807373, "learning_rate": 0.0001666017905799922, "loss": 1.0562, "step": 435 }, { "epoch": 0.169435538715632, "grad_norm": 0.22052323818206787, "learning_rate": 0.00016652393927598288, "loss": 1.0732, "step": 436 }, { "epoch": 0.16982415233653939, "grad_norm": 0.21741728484630585, "learning_rate": 0.00016644608797197354, "loss": 1.0409, "step": 437 }, { "epoch": 0.1702127659574468, "grad_norm": 0.20701193809509277, "learning_rate": 0.0001663682366679642, "loss": 1.0731, "step": 438 }, { "epoch": 0.17060137957835422, "grad_norm": 0.22071130573749542, "learning_rate": 0.00016629038536395484, "loss": 1.0992, "step": 439 }, { "epoch": 0.17098999319926164, "grad_norm": 0.20261412858963013, "learning_rate": 0.0001662125340599455, "loss": 1.0051, "step": 440 }, { "epoch": 0.17137860682016906, "grad_norm": 0.2082947939634323, "learning_rate": 0.00016613468275593617, "loss": 1.0477, "step": 441 }, { "epoch": 0.17176722044107645, "grad_norm": 0.22534717619419098, "learning_rate": 0.00016605683145192682, "loss": 1.041, "step": 442 }, { "epoch": 0.17215583406198387, "grad_norm": 0.21547731757164001, "learning_rate": 0.00016597898014791748, "loss": 1.0528, "step": 443 }, { "epoch": 0.1725444476828913, "grad_norm": 0.24141089618206024, "learning_rate": 0.00016590112884390813, "loss": 1.0928, "step": 444 }, { "epoch": 0.1729330613037987, "grad_norm": 0.21910884976387024, "learning_rate": 0.00016582327753989878, "loss": 1.063, "step": 445 }, { "epoch": 0.1733216749247061, "grad_norm": 0.21782316267490387, "learning_rate": 0.00016574542623588946, "loss": 1.0976, "step": 446 }, { "epoch": 0.17371028854561352, "grad_norm": 0.21771778166294098, "learning_rate": 0.0001656675749318801, "loss": 1.0677, "step": 447 }, { "epoch": 0.17409890216652094, "grad_norm": 0.22117659449577332, "learning_rate": 0.00016558972362787076, "loss": 1.0669, "step": 448 }, { "epoch": 0.17448751578742835, "grad_norm": 0.21918092668056488, "learning_rate": 0.00016551187232386141, "loss": 1.0955, "step": 449 }, { "epoch": 0.17487612940833577, "grad_norm": 0.22027818858623505, "learning_rate": 0.0001654340210198521, "loss": 1.0201, "step": 450 }, { "epoch": 0.17526474302924316, "grad_norm": 0.2042885720729828, "learning_rate": 0.00016535616971584275, "loss": 1.0881, "step": 451 }, { "epoch": 0.17565335665015058, "grad_norm": 0.21788261830806732, "learning_rate": 0.0001652783184118334, "loss": 1.0918, "step": 452 }, { "epoch": 0.176041970271058, "grad_norm": 0.23332571983337402, "learning_rate": 0.00016520046710782408, "loss": 1.091, "step": 453 }, { "epoch": 0.17643058389196542, "grad_norm": 0.20204192399978638, "learning_rate": 0.00016512261580381473, "loss": 1.0366, "step": 454 }, { "epoch": 0.17681919751287284, "grad_norm": 0.21761906147003174, "learning_rate": 0.00016504476449980538, "loss": 1.0131, "step": 455 }, { "epoch": 0.17720781113378023, "grad_norm": 0.2152051478624344, "learning_rate": 0.00016496691319579606, "loss": 1.0868, "step": 456 }, { "epoch": 0.17759642475468765, "grad_norm": 0.22776494920253754, "learning_rate": 0.0001648890618917867, "loss": 1.0807, "step": 457 }, { "epoch": 0.17798503837559507, "grad_norm": 0.2171342968940735, "learning_rate": 0.00016481121058777736, "loss": 1.0537, "step": 458 }, { "epoch": 0.17837365199650249, "grad_norm": 0.2046273946762085, "learning_rate": 0.00016473335928376802, "loss": 1.0097, "step": 459 }, { "epoch": 0.17876226561740988, "grad_norm": 0.2047681361436844, "learning_rate": 0.00016465550797975867, "loss": 1.0204, "step": 460 }, { "epoch": 0.1791508792383173, "grad_norm": 0.1876862645149231, "learning_rate": 0.00016457765667574935, "loss": 0.9383, "step": 461 }, { "epoch": 0.17953949285922471, "grad_norm": 0.218430757522583, "learning_rate": 0.00016449980537174, "loss": 1.0721, "step": 462 }, { "epoch": 0.17992810648013213, "grad_norm": 0.2245480865240097, "learning_rate": 0.00016442195406773065, "loss": 1.0859, "step": 463 }, { "epoch": 0.18031672010103955, "grad_norm": 0.22577151656150818, "learning_rate": 0.0001643441027637213, "loss": 1.0825, "step": 464 }, { "epoch": 0.18070533372194694, "grad_norm": 0.20132745802402496, "learning_rate": 0.00016426625145971196, "loss": 1.0615, "step": 465 }, { "epoch": 0.18109394734285436, "grad_norm": 0.2277505248785019, "learning_rate": 0.00016418840015570263, "loss": 1.0426, "step": 466 }, { "epoch": 0.18148256096376178, "grad_norm": 0.22540105879306793, "learning_rate": 0.0001641105488516933, "loss": 1.0481, "step": 467 }, { "epoch": 0.1818711745846692, "grad_norm": 0.20358088612556458, "learning_rate": 0.00016403269754768394, "loss": 1.0286, "step": 468 }, { "epoch": 0.18225978820557662, "grad_norm": 0.22534145414829254, "learning_rate": 0.0001639548462436746, "loss": 1.1183, "step": 469 }, { "epoch": 0.182648401826484, "grad_norm": 0.2188873142004013, "learning_rate": 0.00016387699493966524, "loss": 1.0439, "step": 470 }, { "epoch": 0.18303701544739143, "grad_norm": 0.2128048539161682, "learning_rate": 0.00016379914363565592, "loss": 1.027, "step": 471 }, { "epoch": 0.18342562906829885, "grad_norm": 0.2518141567707062, "learning_rate": 0.00016372129233164657, "loss": 1.0468, "step": 472 }, { "epoch": 0.18381424268920626, "grad_norm": 0.2189142256975174, "learning_rate": 0.00016364344102763723, "loss": 1.0581, "step": 473 }, { "epoch": 0.18420285631011368, "grad_norm": 0.31266725063323975, "learning_rate": 0.00016356558972362788, "loss": 1.0554, "step": 474 }, { "epoch": 0.18459146993102107, "grad_norm": 0.21343916654586792, "learning_rate": 0.00016348773841961853, "loss": 1.0795, "step": 475 }, { "epoch": 0.1849800835519285, "grad_norm": 0.22907280921936035, "learning_rate": 0.00016340988711560918, "loss": 1.0304, "step": 476 }, { "epoch": 0.1853686971728359, "grad_norm": 0.2105257511138916, "learning_rate": 0.00016333203581159986, "loss": 1.0231, "step": 477 }, { "epoch": 0.18575731079374333, "grad_norm": 0.19537831842899323, "learning_rate": 0.00016325418450759051, "loss": 1.0103, "step": 478 }, { "epoch": 0.18614592441465072, "grad_norm": 0.20522372424602509, "learning_rate": 0.00016317633320358117, "loss": 1.0196, "step": 479 }, { "epoch": 0.18653453803555814, "grad_norm": 0.21646477282047272, "learning_rate": 0.00016309848189957182, "loss": 1.0579, "step": 480 }, { "epoch": 0.18692315165646556, "grad_norm": 0.21077193319797516, "learning_rate": 0.00016302063059556247, "loss": 1.0638, "step": 481 }, { "epoch": 0.18731176527737298, "grad_norm": 0.20357473194599152, "learning_rate": 0.00016294277929155315, "loss": 1.0635, "step": 482 }, { "epoch": 0.1877003788982804, "grad_norm": 0.2188001275062561, "learning_rate": 0.0001628649279875438, "loss": 1.0267, "step": 483 }, { "epoch": 0.1880889925191878, "grad_norm": 0.2128928154706955, "learning_rate": 0.00016278707668353445, "loss": 0.9706, "step": 484 }, { "epoch": 0.1884776061400952, "grad_norm": 0.22081372141838074, "learning_rate": 0.0001627092253795251, "loss": 1.08, "step": 485 }, { "epoch": 0.18886621976100262, "grad_norm": 0.2250615805387497, "learning_rate": 0.00016263137407551576, "loss": 1.1451, "step": 486 }, { "epoch": 0.18925483338191004, "grad_norm": 0.1984967589378357, "learning_rate": 0.00016255352277150644, "loss": 1.0744, "step": 487 }, { "epoch": 0.18964344700281746, "grad_norm": 0.20778900384902954, "learning_rate": 0.0001624756714674971, "loss": 1.0623, "step": 488 }, { "epoch": 0.19003206062372485, "grad_norm": 0.2026563137769699, "learning_rate": 0.00016239782016348774, "loss": 1.0714, "step": 489 }, { "epoch": 0.19042067424463227, "grad_norm": 0.21598374843597412, "learning_rate": 0.0001623199688594784, "loss": 1.0869, "step": 490 }, { "epoch": 0.1908092878655397, "grad_norm": 0.18944978713989258, "learning_rate": 0.00016224211755546904, "loss": 1.055, "step": 491 }, { "epoch": 0.1911979014864471, "grad_norm": 0.20698946714401245, "learning_rate": 0.00016216426625145972, "loss": 1.0392, "step": 492 }, { "epoch": 0.1915865151073545, "grad_norm": 0.22395353019237518, "learning_rate": 0.00016208641494745038, "loss": 1.0681, "step": 493 }, { "epoch": 0.19197512872826192, "grad_norm": 0.22372962534427643, "learning_rate": 0.00016200856364344103, "loss": 1.0767, "step": 494 }, { "epoch": 0.19236374234916934, "grad_norm": 0.2066701054573059, "learning_rate": 0.00016193071233943168, "loss": 1.0061, "step": 495 }, { "epoch": 0.19275235597007676, "grad_norm": 0.19716408848762512, "learning_rate": 0.00016185286103542233, "loss": 1.039, "step": 496 }, { "epoch": 0.19314096959098417, "grad_norm": 0.22159601747989655, "learning_rate": 0.000161775009731413, "loss": 1.0832, "step": 497 }, { "epoch": 0.19352958321189156, "grad_norm": 0.21509626507759094, "learning_rate": 0.00016169715842740366, "loss": 1.0264, "step": 498 }, { "epoch": 0.19391819683279898, "grad_norm": 0.21598199009895325, "learning_rate": 0.00016161930712339431, "loss": 1.049, "step": 499 }, { "epoch": 0.1943068104537064, "grad_norm": 0.20279590785503387, "learning_rate": 0.00016154145581938497, "loss": 1.0505, "step": 500 }, { "epoch": 0.19469542407461382, "grad_norm": 0.21796855330467224, "learning_rate": 0.00016146360451537565, "loss": 1.0885, "step": 501 }, { "epoch": 0.19508403769552124, "grad_norm": 0.22128933668136597, "learning_rate": 0.0001613857532113663, "loss": 1.0903, "step": 502 }, { "epoch": 0.19547265131642863, "grad_norm": 0.2032536417245865, "learning_rate": 0.00016130790190735695, "loss": 1.0285, "step": 503 }, { "epoch": 0.19586126493733605, "grad_norm": 0.23738974332809448, "learning_rate": 0.0001612300506033476, "loss": 1.1188, "step": 504 }, { "epoch": 0.19624987855824347, "grad_norm": 0.19614790380001068, "learning_rate": 0.00016115219929933828, "loss": 1.04, "step": 505 }, { "epoch": 0.1966384921791509, "grad_norm": 0.2198178917169571, "learning_rate": 0.00016107434799532893, "loss": 1.0696, "step": 506 }, { "epoch": 0.1970271058000583, "grad_norm": 0.18814648687839508, "learning_rate": 0.00016099649669131959, "loss": 1.0203, "step": 507 }, { "epoch": 0.1974157194209657, "grad_norm": 0.20699037611484528, "learning_rate": 0.00016091864538731026, "loss": 1.1074, "step": 508 }, { "epoch": 0.19780433304187311, "grad_norm": 0.21490445733070374, "learning_rate": 0.00016084079408330092, "loss": 1.0682, "step": 509 }, { "epoch": 0.19819294666278053, "grad_norm": 0.2363848090171814, "learning_rate": 0.00016076294277929157, "loss": 1.0408, "step": 510 }, { "epoch": 0.19858156028368795, "grad_norm": 0.20186659693717957, "learning_rate": 0.00016068509147528222, "loss": 1.026, "step": 511 }, { "epoch": 0.19897017390459534, "grad_norm": 0.21564024686813354, "learning_rate": 0.00016060724017127287, "loss": 1.0418, "step": 512 }, { "epoch": 0.19935878752550276, "grad_norm": 0.19151560962200165, "learning_rate": 0.00016052938886726355, "loss": 1.0037, "step": 513 }, { "epoch": 0.19974740114641018, "grad_norm": 0.21038194000720978, "learning_rate": 0.0001604515375632542, "loss": 1.0545, "step": 514 }, { "epoch": 0.2001360147673176, "grad_norm": 0.20496582984924316, "learning_rate": 0.00016037368625924486, "loss": 1.0543, "step": 515 }, { "epoch": 0.20052462838822502, "grad_norm": 0.20689113438129425, "learning_rate": 0.0001602958349552355, "loss": 1.0905, "step": 516 }, { "epoch": 0.2009132420091324, "grad_norm": 0.2284041792154312, "learning_rate": 0.00016021798365122616, "loss": 1.0717, "step": 517 }, { "epoch": 0.20130185563003983, "grad_norm": 0.23457761108875275, "learning_rate": 0.00016014013234721684, "loss": 1.106, "step": 518 }, { "epoch": 0.20169046925094725, "grad_norm": 0.2088528722524643, "learning_rate": 0.0001600622810432075, "loss": 1.0428, "step": 519 }, { "epoch": 0.20207908287185467, "grad_norm": 0.2170068770647049, "learning_rate": 0.00015998442973919814, "loss": 0.9875, "step": 520 }, { "epoch": 0.20246769649276208, "grad_norm": 0.2270561158657074, "learning_rate": 0.0001599065784351888, "loss": 1.0676, "step": 521 }, { "epoch": 0.20285631011366947, "grad_norm": 0.2151324599981308, "learning_rate": 0.00015982872713117945, "loss": 1.0675, "step": 522 }, { "epoch": 0.2032449237345769, "grad_norm": 0.23113249242305756, "learning_rate": 0.00015975087582717013, "loss": 1.0608, "step": 523 }, { "epoch": 0.2036335373554843, "grad_norm": 0.2587106227874756, "learning_rate": 0.00015967302452316078, "loss": 1.0867, "step": 524 }, { "epoch": 0.20402215097639173, "grad_norm": 0.21842992305755615, "learning_rate": 0.00015959517321915143, "loss": 1.0726, "step": 525 }, { "epoch": 0.20441076459729912, "grad_norm": 0.20867805182933807, "learning_rate": 0.00015951732191514208, "loss": 1.0578, "step": 526 }, { "epoch": 0.20479937821820654, "grad_norm": 0.2396962195634842, "learning_rate": 0.00015943947061113273, "loss": 1.0292, "step": 527 }, { "epoch": 0.20518799183911396, "grad_norm": 0.221155047416687, "learning_rate": 0.00015936161930712341, "loss": 1.0019, "step": 528 }, { "epoch": 0.20557660546002138, "grad_norm": 0.20032119750976562, "learning_rate": 0.00015928376800311407, "loss": 1.0435, "step": 529 }, { "epoch": 0.2059652190809288, "grad_norm": 0.24095888435840607, "learning_rate": 0.00015920591669910472, "loss": 1.0355, "step": 530 }, { "epoch": 0.2063538327018362, "grad_norm": 0.2286604344844818, "learning_rate": 0.00015912806539509537, "loss": 0.9989, "step": 531 }, { "epoch": 0.2067424463227436, "grad_norm": 0.21537137031555176, "learning_rate": 0.00015905021409108602, "loss": 1.0642, "step": 532 }, { "epoch": 0.20713105994365102, "grad_norm": 0.22447925806045532, "learning_rate": 0.0001589723627870767, "loss": 1.1244, "step": 533 }, { "epoch": 0.20751967356455844, "grad_norm": 0.21077273786067963, "learning_rate": 0.00015889451148306735, "loss": 1.0167, "step": 534 }, { "epoch": 0.20790828718546586, "grad_norm": 0.22340558469295502, "learning_rate": 0.000158816660179058, "loss": 1.0991, "step": 535 }, { "epoch": 0.20829690080637325, "grad_norm": 0.223599374294281, "learning_rate": 0.00015873880887504866, "loss": 1.086, "step": 536 }, { "epoch": 0.20868551442728067, "grad_norm": 0.2615208923816681, "learning_rate": 0.0001586609575710393, "loss": 1.0584, "step": 537 }, { "epoch": 0.2090741280481881, "grad_norm": 0.2085907757282257, "learning_rate": 0.00015858310626703, "loss": 1.0994, "step": 538 }, { "epoch": 0.2094627416690955, "grad_norm": 0.2170211672782898, "learning_rate": 0.00015850525496302064, "loss": 1.1105, "step": 539 }, { "epoch": 0.20985135529000293, "grad_norm": 0.21978625655174255, "learning_rate": 0.0001584274036590113, "loss": 1.002, "step": 540 }, { "epoch": 0.21023996891091032, "grad_norm": 0.23684021830558777, "learning_rate": 0.00015834955235500194, "loss": 1.1216, "step": 541 }, { "epoch": 0.21062858253181774, "grad_norm": 0.220269113779068, "learning_rate": 0.0001582717010509926, "loss": 1.0773, "step": 542 }, { "epoch": 0.21101719615272516, "grad_norm": 0.22447973489761353, "learning_rate": 0.00015819384974698328, "loss": 1.0941, "step": 543 }, { "epoch": 0.21140580977363257, "grad_norm": 0.22435730695724487, "learning_rate": 0.00015811599844297393, "loss": 1.0138, "step": 544 }, { "epoch": 0.21179442339453997, "grad_norm": 0.2230793684720993, "learning_rate": 0.00015803814713896458, "loss": 1.0343, "step": 545 }, { "epoch": 0.21218303701544738, "grad_norm": 0.23491905629634857, "learning_rate": 0.00015796029583495523, "loss": 1.11, "step": 546 }, { "epoch": 0.2125716506363548, "grad_norm": 0.213560551404953, "learning_rate": 0.00015788244453094588, "loss": 1.0615, "step": 547 }, { "epoch": 0.21296026425726222, "grad_norm": 0.21392837166786194, "learning_rate": 0.00015780459322693654, "loss": 1.0872, "step": 548 }, { "epoch": 0.21334887787816964, "grad_norm": 0.20007692277431488, "learning_rate": 0.00015772674192292722, "loss": 1.0394, "step": 549 }, { "epoch": 0.21373749149907703, "grad_norm": 0.1969841718673706, "learning_rate": 0.00015764889061891787, "loss": 1.0381, "step": 550 }, { "epoch": 0.21412610511998445, "grad_norm": 0.21874025464057922, "learning_rate": 0.00015757103931490852, "loss": 1.0822, "step": 551 }, { "epoch": 0.21451471874089187, "grad_norm": 0.21824273467063904, "learning_rate": 0.00015749318801089917, "loss": 1.0802, "step": 552 }, { "epoch": 0.2149033323617993, "grad_norm": 0.20942047238349915, "learning_rate": 0.00015741533670688985, "loss": 1.0634, "step": 553 }, { "epoch": 0.2152919459827067, "grad_norm": 0.1940152943134308, "learning_rate": 0.0001573374854028805, "loss": 1.0264, "step": 554 }, { "epoch": 0.2156805596036141, "grad_norm": 0.19859059154987335, "learning_rate": 0.00015725963409887115, "loss": 0.9701, "step": 555 }, { "epoch": 0.21606917322452152, "grad_norm": 0.22239404916763306, "learning_rate": 0.0001571817827948618, "loss": 1.1282, "step": 556 }, { "epoch": 0.21645778684542893, "grad_norm": 0.23820599913597107, "learning_rate": 0.00015710393149085249, "loss": 1.1123, "step": 557 }, { "epoch": 0.21684640046633635, "grad_norm": 0.21279917657375336, "learning_rate": 0.00015702608018684314, "loss": 1.0542, "step": 558 }, { "epoch": 0.21723501408724374, "grad_norm": 0.2065514773130417, "learning_rate": 0.0001569482288828338, "loss": 1.0685, "step": 559 }, { "epoch": 0.21762362770815116, "grad_norm": 0.20130831003189087, "learning_rate": 0.00015687037757882447, "loss": 0.9869, "step": 560 }, { "epoch": 0.21801224132905858, "grad_norm": 0.2187541127204895, "learning_rate": 0.00015679252627481512, "loss": 1.1095, "step": 561 }, { "epoch": 0.218400854949966, "grad_norm": 0.21028277277946472, "learning_rate": 0.00015671467497080577, "loss": 1.0804, "step": 562 }, { "epoch": 0.21878946857087342, "grad_norm": 0.8187636733055115, "learning_rate": 0.00015663682366679643, "loss": 1.0782, "step": 563 }, { "epoch": 0.2191780821917808, "grad_norm": 0.20059974491596222, "learning_rate": 0.0001565589723627871, "loss": 1.0279, "step": 564 }, { "epoch": 0.21956669581268823, "grad_norm": 0.20440839231014252, "learning_rate": 0.00015648112105877776, "loss": 0.9863, "step": 565 }, { "epoch": 0.21995530943359565, "grad_norm": 0.21423624455928802, "learning_rate": 0.0001564032697547684, "loss": 1.0685, "step": 566 }, { "epoch": 0.22034392305450307, "grad_norm": 0.22430062294006348, "learning_rate": 0.00015632541845075906, "loss": 1.0761, "step": 567 }, { "epoch": 0.22073253667541048, "grad_norm": 0.22782258689403534, "learning_rate": 0.0001562475671467497, "loss": 1.1024, "step": 568 }, { "epoch": 0.22112115029631788, "grad_norm": 0.21150320768356323, "learning_rate": 0.0001561697158427404, "loss": 1.0621, "step": 569 }, { "epoch": 0.2215097639172253, "grad_norm": 0.20342351496219635, "learning_rate": 0.00015609186453873104, "loss": 1.0667, "step": 570 }, { "epoch": 0.2218983775381327, "grad_norm": 0.22866711020469666, "learning_rate": 0.0001560140132347217, "loss": 1.0631, "step": 571 }, { "epoch": 0.22228699115904013, "grad_norm": 0.2200063169002533, "learning_rate": 0.00015593616193071235, "loss": 1.0448, "step": 572 }, { "epoch": 0.22267560477994755, "grad_norm": 0.19440248608589172, "learning_rate": 0.000155858310626703, "loss": 1.037, "step": 573 }, { "epoch": 0.22306421840085494, "grad_norm": 0.205752432346344, "learning_rate": 0.00015578045932269368, "loss": 1.0465, "step": 574 }, { "epoch": 0.22345283202176236, "grad_norm": 0.22247998416423798, "learning_rate": 0.00015570260801868433, "loss": 0.997, "step": 575 }, { "epoch": 0.22384144564266978, "grad_norm": 0.22199274599552155, "learning_rate": 0.00015562475671467498, "loss": 1.0178, "step": 576 }, { "epoch": 0.2242300592635772, "grad_norm": 0.2114989310503006, "learning_rate": 0.00015554690541066564, "loss": 1.0457, "step": 577 }, { "epoch": 0.2246186728844846, "grad_norm": 0.24248506128787994, "learning_rate": 0.0001554690541066563, "loss": 1.002, "step": 578 }, { "epoch": 0.225007286505392, "grad_norm": 0.2565505802631378, "learning_rate": 0.00015539120280264697, "loss": 1.0541, "step": 579 }, { "epoch": 0.22539590012629943, "grad_norm": 0.22799409925937653, "learning_rate": 0.00015531335149863762, "loss": 1.0788, "step": 580 }, { "epoch": 0.22578451374720684, "grad_norm": 0.2196080982685089, "learning_rate": 0.00015523550019462827, "loss": 1.0877, "step": 581 }, { "epoch": 0.22617312736811426, "grad_norm": 0.21992824971675873, "learning_rate": 0.00015515764889061892, "loss": 1.0213, "step": 582 }, { "epoch": 0.22656174098902165, "grad_norm": 0.22793298959732056, "learning_rate": 0.00015507979758660957, "loss": 1.0633, "step": 583 }, { "epoch": 0.22695035460992907, "grad_norm": 0.21707972884178162, "learning_rate": 0.00015500194628260023, "loss": 1.081, "step": 584 }, { "epoch": 0.2273389682308365, "grad_norm": 0.220685675740242, "learning_rate": 0.0001549240949785909, "loss": 1.0658, "step": 585 }, { "epoch": 0.2277275818517439, "grad_norm": 0.22576668858528137, "learning_rate": 0.00015484624367458156, "loss": 1.0795, "step": 586 }, { "epoch": 0.22811619547265133, "grad_norm": 0.21778982877731323, "learning_rate": 0.0001547683923705722, "loss": 1.033, "step": 587 }, { "epoch": 0.22850480909355872, "grad_norm": 0.22748610377311707, "learning_rate": 0.00015469054106656286, "loss": 1.0948, "step": 588 }, { "epoch": 0.22889342271446614, "grad_norm": 0.21561284363269806, "learning_rate": 0.00015461268976255351, "loss": 1.0022, "step": 589 }, { "epoch": 0.22928203633537356, "grad_norm": 0.2419756054878235, "learning_rate": 0.0001545348384585442, "loss": 1.0786, "step": 590 }, { "epoch": 0.22967064995628098, "grad_norm": 0.20479315519332886, "learning_rate": 0.00015445698715453485, "loss": 1.027, "step": 591 }, { "epoch": 0.2300592635771884, "grad_norm": 0.21365883946418762, "learning_rate": 0.0001543791358505255, "loss": 1.0773, "step": 592 }, { "epoch": 0.23044787719809579, "grad_norm": 0.23133166134357452, "learning_rate": 0.00015430128454651615, "loss": 1.0877, "step": 593 }, { "epoch": 0.2308364908190032, "grad_norm": 0.2110515981912613, "learning_rate": 0.0001542234332425068, "loss": 1.0509, "step": 594 }, { "epoch": 0.23122510443991062, "grad_norm": 0.20658442378044128, "learning_rate": 0.00015414558193849748, "loss": 1.0623, "step": 595 }, { "epoch": 0.23161371806081804, "grad_norm": 0.21831996738910675, "learning_rate": 0.00015406773063448813, "loss": 1.021, "step": 596 }, { "epoch": 0.23200233168172543, "grad_norm": 0.23015642166137695, "learning_rate": 0.00015398987933047878, "loss": 1.0358, "step": 597 }, { "epoch": 0.23239094530263285, "grad_norm": 0.23071645200252533, "learning_rate": 0.00015391202802646944, "loss": 1.1255, "step": 598 }, { "epoch": 0.23277955892354027, "grad_norm": 0.19513486325740814, "learning_rate": 0.0001538341767224601, "loss": 1.0189, "step": 599 }, { "epoch": 0.2331681725444477, "grad_norm": 0.20821452140808105, "learning_rate": 0.00015375632541845077, "loss": 1.0843, "step": 600 }, { "epoch": 0.2335567861653551, "grad_norm": 0.20563223958015442, "learning_rate": 0.00015367847411444142, "loss": 1.0012, "step": 601 }, { "epoch": 0.2339453997862625, "grad_norm": 0.22674202919006348, "learning_rate": 0.00015360062281043207, "loss": 1.0371, "step": 602 }, { "epoch": 0.23433401340716992, "grad_norm": 0.20744135975837708, "learning_rate": 0.00015352277150642272, "loss": 1.0466, "step": 603 }, { "epoch": 0.23472262702807734, "grad_norm": 0.22103577852249146, "learning_rate": 0.00015344492020241338, "loss": 1.0942, "step": 604 }, { "epoch": 0.23511124064898475, "grad_norm": 0.20643098652362823, "learning_rate": 0.00015336706889840406, "loss": 1.0682, "step": 605 }, { "epoch": 0.23549985426989217, "grad_norm": 0.23436777293682098, "learning_rate": 0.0001532892175943947, "loss": 1.0613, "step": 606 }, { "epoch": 0.23588846789079956, "grad_norm": 0.21898899972438812, "learning_rate": 0.00015321136629038536, "loss": 1.0571, "step": 607 }, { "epoch": 0.23627708151170698, "grad_norm": 0.20569247007369995, "learning_rate": 0.00015313351498637604, "loss": 1.061, "step": 608 }, { "epoch": 0.2366656951326144, "grad_norm": 0.2099207490682602, "learning_rate": 0.0001530556636823667, "loss": 1.0776, "step": 609 }, { "epoch": 0.23705430875352182, "grad_norm": 0.20078738033771515, "learning_rate": 0.00015297781237835734, "loss": 1.0341, "step": 610 }, { "epoch": 0.2374429223744292, "grad_norm": 0.20327065885066986, "learning_rate": 0.000152899961074348, "loss": 1.0168, "step": 611 }, { "epoch": 0.23783153599533663, "grad_norm": 0.21741214394569397, "learning_rate": 0.00015282210977033867, "loss": 1.0726, "step": 612 }, { "epoch": 0.23822014961624405, "grad_norm": 0.2065727263689041, "learning_rate": 0.00015274425846632933, "loss": 1.0474, "step": 613 }, { "epoch": 0.23860876323715147, "grad_norm": 0.21241194009780884, "learning_rate": 0.00015266640716231998, "loss": 1.0666, "step": 614 }, { "epoch": 0.23899737685805889, "grad_norm": 0.2194201797246933, "learning_rate": 0.00015258855585831066, "loss": 1.1411, "step": 615 }, { "epoch": 0.23938599047896628, "grad_norm": 0.21537193655967712, "learning_rate": 0.0001525107045543013, "loss": 1.081, "step": 616 }, { "epoch": 0.2397746040998737, "grad_norm": 0.21125951409339905, "learning_rate": 0.00015243285325029196, "loss": 1.0679, "step": 617 }, { "epoch": 0.2401632177207811, "grad_norm": 0.21342721581459045, "learning_rate": 0.0001523550019462826, "loss": 1.0564, "step": 618 }, { "epoch": 0.24055183134168853, "grad_norm": 0.2223503291606903, "learning_rate": 0.00015227715064227327, "loss": 1.1163, "step": 619 }, { "epoch": 0.24094044496259595, "grad_norm": 0.21626527607440948, "learning_rate": 0.00015219929933826394, "loss": 1.0793, "step": 620 }, { "epoch": 0.24132905858350334, "grad_norm": 0.21899500489234924, "learning_rate": 0.0001521214480342546, "loss": 1.0864, "step": 621 }, { "epoch": 0.24171767220441076, "grad_norm": 0.2499915212392807, "learning_rate": 0.00015204359673024525, "loss": 1.1381, "step": 622 }, { "epoch": 0.24210628582531818, "grad_norm": 0.2108345925807953, "learning_rate": 0.0001519657454262359, "loss": 1.0534, "step": 623 }, { "epoch": 0.2424948994462256, "grad_norm": 0.2224910855293274, "learning_rate": 0.00015188789412222655, "loss": 1.0235, "step": 624 }, { "epoch": 0.24288351306713302, "grad_norm": 0.22163094580173492, "learning_rate": 0.0001518100428182172, "loss": 1.0143, "step": 625 }, { "epoch": 0.2432721266880404, "grad_norm": 0.20709283649921417, "learning_rate": 0.00015173219151420788, "loss": 1.0506, "step": 626 }, { "epoch": 0.24366074030894783, "grad_norm": 0.2112802267074585, "learning_rate": 0.00015165434021019854, "loss": 1.0692, "step": 627 }, { "epoch": 0.24404935392985525, "grad_norm": 0.23622830212116241, "learning_rate": 0.0001515764889061892, "loss": 1.0769, "step": 628 }, { "epoch": 0.24443796755076266, "grad_norm": 0.23328271508216858, "learning_rate": 0.00015149863760217984, "loss": 1.1158, "step": 629 }, { "epoch": 0.24482658117167005, "grad_norm": 0.2071760892868042, "learning_rate": 0.0001514207862981705, "loss": 1.0133, "step": 630 }, { "epoch": 0.24521519479257747, "grad_norm": 0.21428920328617096, "learning_rate": 0.00015134293499416117, "loss": 1.0342, "step": 631 }, { "epoch": 0.2456038084134849, "grad_norm": 0.22225375473499298, "learning_rate": 0.00015126508369015182, "loss": 1.1054, "step": 632 }, { "epoch": 0.2459924220343923, "grad_norm": 0.2096671611070633, "learning_rate": 0.00015118723238614248, "loss": 1.0229, "step": 633 }, { "epoch": 0.24638103565529973, "grad_norm": 0.21473252773284912, "learning_rate": 0.00015110938108213313, "loss": 1.0915, "step": 634 }, { "epoch": 0.24676964927620712, "grad_norm": 0.2071562111377716, "learning_rate": 0.00015103152977812378, "loss": 1.047, "step": 635 }, { "epoch": 0.24715826289711454, "grad_norm": 0.19868609309196472, "learning_rate": 0.00015095367847411446, "loss": 1.0073, "step": 636 }, { "epoch": 0.24754687651802196, "grad_norm": 0.20937366783618927, "learning_rate": 0.0001508758271701051, "loss": 1.0155, "step": 637 }, { "epoch": 0.24793549013892938, "grad_norm": 0.19225911796092987, "learning_rate": 0.00015079797586609576, "loss": 1.0163, "step": 638 }, { "epoch": 0.2483241037598368, "grad_norm": 0.20427283644676208, "learning_rate": 0.00015072012456208641, "loss": 1.062, "step": 639 }, { "epoch": 0.24871271738074419, "grad_norm": 0.21640253067016602, "learning_rate": 0.00015064227325807707, "loss": 1.025, "step": 640 }, { "epoch": 0.2491013310016516, "grad_norm": 0.20416739583015442, "learning_rate": 0.00015056442195406775, "loss": 1.0635, "step": 641 }, { "epoch": 0.24948994462255902, "grad_norm": 0.1990521252155304, "learning_rate": 0.0001504865706500584, "loss": 1.0757, "step": 642 }, { "epoch": 0.24987855824346644, "grad_norm": 0.21636444330215454, "learning_rate": 0.00015040871934604905, "loss": 1.0441, "step": 643 }, { "epoch": 0.25026717186437386, "grad_norm": 0.21253719925880432, "learning_rate": 0.0001503308680420397, "loss": 1.0574, "step": 644 }, { "epoch": 0.2506557854852813, "grad_norm": 0.2134159356355667, "learning_rate": 0.00015025301673803035, "loss": 1.0396, "step": 645 }, { "epoch": 0.2510443991061887, "grad_norm": 0.2018527239561081, "learning_rate": 0.00015017516543402103, "loss": 1.0606, "step": 646 }, { "epoch": 0.25143301272709606, "grad_norm": 0.20320741832256317, "learning_rate": 0.00015009731413001169, "loss": 1.0093, "step": 647 }, { "epoch": 0.2518216263480035, "grad_norm": 0.21007056534290314, "learning_rate": 0.00015001946282600234, "loss": 1.0284, "step": 648 }, { "epoch": 0.2522102399689109, "grad_norm": 0.22453372180461884, "learning_rate": 0.000149941611521993, "loss": 1.0271, "step": 649 }, { "epoch": 0.2525988535898183, "grad_norm": 0.19889335334300995, "learning_rate": 0.00014986376021798364, "loss": 1.0238, "step": 650 }, { "epoch": 0.25298746721072574, "grad_norm": 0.19339965283870697, "learning_rate": 0.00014978590891397432, "loss": 1.024, "step": 651 }, { "epoch": 0.25337608083163315, "grad_norm": 0.22362011671066284, "learning_rate": 0.00014970805760996497, "loss": 1.0722, "step": 652 }, { "epoch": 0.2537646944525406, "grad_norm": 0.2110588103532791, "learning_rate": 0.00014963020630595562, "loss": 1.0541, "step": 653 }, { "epoch": 0.254153308073448, "grad_norm": 0.203025683760643, "learning_rate": 0.00014955235500194628, "loss": 1.0335, "step": 654 }, { "epoch": 0.2545419216943554, "grad_norm": 0.20884902775287628, "learning_rate": 0.00014947450369793693, "loss": 1.0507, "step": 655 }, { "epoch": 0.2549305353152628, "grad_norm": 0.21234256029129028, "learning_rate": 0.0001493966523939276, "loss": 1.0372, "step": 656 }, { "epoch": 0.2553191489361702, "grad_norm": 0.1984352171421051, "learning_rate": 0.00014931880108991826, "loss": 0.9979, "step": 657 }, { "epoch": 0.2557077625570776, "grad_norm": 0.18848282098770142, "learning_rate": 0.0001492409497859089, "loss": 0.9973, "step": 658 }, { "epoch": 0.25609637617798503, "grad_norm": 0.2201709896326065, "learning_rate": 0.00014916309848189956, "loss": 1.0386, "step": 659 }, { "epoch": 0.25648498979889245, "grad_norm": 0.23094095289707184, "learning_rate": 0.00014908524717789024, "loss": 1.1205, "step": 660 }, { "epoch": 0.25687360341979987, "grad_norm": 0.21087734401226044, "learning_rate": 0.0001490073958738809, "loss": 1.0231, "step": 661 }, { "epoch": 0.2572622170407073, "grad_norm": 0.24970979988574982, "learning_rate": 0.00014892954456987155, "loss": 1.0421, "step": 662 }, { "epoch": 0.2576508306616147, "grad_norm": 0.22024711966514587, "learning_rate": 0.00014885169326586223, "loss": 1.1033, "step": 663 }, { "epoch": 0.2580394442825221, "grad_norm": 0.2195248156785965, "learning_rate": 0.00014877384196185288, "loss": 1.089, "step": 664 }, { "epoch": 0.25842805790342954, "grad_norm": 0.20236417651176453, "learning_rate": 0.00014869599065784353, "loss": 1.0196, "step": 665 }, { "epoch": 0.2588166715243369, "grad_norm": 0.21973329782485962, "learning_rate": 0.00014861813935383418, "loss": 1.0844, "step": 666 }, { "epoch": 0.2592052851452443, "grad_norm": 0.2069879174232483, "learning_rate": 0.00014854028804982486, "loss": 1.0312, "step": 667 }, { "epoch": 0.25959389876615174, "grad_norm": 0.2037455290555954, "learning_rate": 0.00014846243674581551, "loss": 1.0018, "step": 668 }, { "epoch": 0.25998251238705916, "grad_norm": 0.24176378548145294, "learning_rate": 0.00014838458544180617, "loss": 1.0749, "step": 669 }, { "epoch": 0.2603711260079666, "grad_norm": 0.2007879763841629, "learning_rate": 0.00014830673413779682, "loss": 1.0443, "step": 670 }, { "epoch": 0.260759739628874, "grad_norm": 0.23503245413303375, "learning_rate": 0.00014822888283378747, "loss": 1.0674, "step": 671 }, { "epoch": 0.2611483532497814, "grad_norm": 0.2166167050600052, "learning_rate": 0.00014815103152977815, "loss": 1.079, "step": 672 }, { "epoch": 0.26153696687068884, "grad_norm": 0.2293982058763504, "learning_rate": 0.0001480731802257688, "loss": 1.0517, "step": 673 }, { "epoch": 0.26192558049159625, "grad_norm": 0.21040330827236176, "learning_rate": 0.00014799532892175945, "loss": 1.0475, "step": 674 }, { "epoch": 0.2623141941125036, "grad_norm": 0.20750463008880615, "learning_rate": 0.0001479174776177501, "loss": 1.025, "step": 675 }, { "epoch": 0.26270280773341104, "grad_norm": 0.2748873233795166, "learning_rate": 0.00014783962631374076, "loss": 1.0212, "step": 676 }, { "epoch": 0.26309142135431846, "grad_norm": 0.19212333858013153, "learning_rate": 0.00014776177500973144, "loss": 1.0049, "step": 677 }, { "epoch": 0.2634800349752259, "grad_norm": 0.207731693983078, "learning_rate": 0.0001476839237057221, "loss": 1.0062, "step": 678 }, { "epoch": 0.2638686485961333, "grad_norm": 0.2177981585264206, "learning_rate": 0.00014760607240171274, "loss": 1.0489, "step": 679 }, { "epoch": 0.2642572622170407, "grad_norm": 0.23239290714263916, "learning_rate": 0.0001475282210977034, "loss": 1.0856, "step": 680 }, { "epoch": 0.26464587583794813, "grad_norm": 0.2033151388168335, "learning_rate": 0.00014745036979369404, "loss": 1.0389, "step": 681 }, { "epoch": 0.26503448945885555, "grad_norm": 0.20917408168315887, "learning_rate": 0.00014737251848968472, "loss": 1.1208, "step": 682 }, { "epoch": 0.26542310307976297, "grad_norm": 0.22075454890727997, "learning_rate": 0.00014729466718567538, "loss": 1.0435, "step": 683 }, { "epoch": 0.26581171670067033, "grad_norm": 0.23094993829727173, "learning_rate": 0.00014721681588166603, "loss": 1.0649, "step": 684 }, { "epoch": 0.26620033032157775, "grad_norm": 0.21209536492824554, "learning_rate": 0.00014713896457765668, "loss": 1.0578, "step": 685 }, { "epoch": 0.26658894394248517, "grad_norm": 0.21412219107151031, "learning_rate": 0.00014706111327364733, "loss": 1.1137, "step": 686 }, { "epoch": 0.2669775575633926, "grad_norm": 0.21175475418567657, "learning_rate": 0.000146983261969638, "loss": 1.023, "step": 687 }, { "epoch": 0.2673661711843, "grad_norm": 0.21968993544578552, "learning_rate": 0.00014690541066562866, "loss": 1.1183, "step": 688 }, { "epoch": 0.2677547848052074, "grad_norm": 0.20414218306541443, "learning_rate": 0.00014682755936161932, "loss": 1.078, "step": 689 }, { "epoch": 0.26814339842611484, "grad_norm": 0.18986597657203674, "learning_rate": 0.00014674970805760997, "loss": 1.0029, "step": 690 }, { "epoch": 0.26853201204702226, "grad_norm": 0.21215832233428955, "learning_rate": 0.00014667185675360062, "loss": 1.0759, "step": 691 }, { "epoch": 0.2689206256679297, "grad_norm": 0.2113744169473648, "learning_rate": 0.0001465940054495913, "loss": 1.1027, "step": 692 }, { "epoch": 0.2693092392888371, "grad_norm": 0.22010880708694458, "learning_rate": 0.00014651615414558195, "loss": 1.0984, "step": 693 }, { "epoch": 0.26969785290974446, "grad_norm": 0.203857421875, "learning_rate": 0.0001464383028415726, "loss": 1.0407, "step": 694 }, { "epoch": 0.2700864665306519, "grad_norm": 0.21120867133140564, "learning_rate": 0.00014636045153756325, "loss": 1.0521, "step": 695 }, { "epoch": 0.2704750801515593, "grad_norm": 0.20039112865924835, "learning_rate": 0.0001462826002335539, "loss": 1.0897, "step": 696 }, { "epoch": 0.2708636937724667, "grad_norm": 0.22893202304840088, "learning_rate": 0.00014620474892954456, "loss": 1.0903, "step": 697 }, { "epoch": 0.27125230739337414, "grad_norm": 0.19886267185211182, "learning_rate": 0.00014612689762553524, "loss": 1.0889, "step": 698 }, { "epoch": 0.27164092101428156, "grad_norm": 0.18892349302768707, "learning_rate": 0.0001460490463215259, "loss": 0.981, "step": 699 }, { "epoch": 0.272029534635189, "grad_norm": 0.20602507889270782, "learning_rate": 0.00014597119501751654, "loss": 1.0223, "step": 700 }, { "epoch": 0.2724181482560964, "grad_norm": 0.21480505168437958, "learning_rate": 0.0001458933437135072, "loss": 1.0355, "step": 701 }, { "epoch": 0.2728067618770038, "grad_norm": 0.21011753380298615, "learning_rate": 0.00014581549240949785, "loss": 1.0613, "step": 702 }, { "epoch": 0.2731953754979112, "grad_norm": 0.19350819289684296, "learning_rate": 0.00014573764110548853, "loss": 1.0144, "step": 703 }, { "epoch": 0.2735839891188186, "grad_norm": 0.207548126578331, "learning_rate": 0.00014565978980147918, "loss": 1.0465, "step": 704 }, { "epoch": 0.273972602739726, "grad_norm": 0.22220565378665924, "learning_rate": 0.00014558193849746983, "loss": 1.1073, "step": 705 }, { "epoch": 0.27436121636063343, "grad_norm": 0.193622425198555, "learning_rate": 0.00014550408719346048, "loss": 1.0357, "step": 706 }, { "epoch": 0.27474982998154085, "grad_norm": 0.2067158818244934, "learning_rate": 0.00014542623588945113, "loss": 1.0502, "step": 707 }, { "epoch": 0.27513844360244827, "grad_norm": 0.2218742072582245, "learning_rate": 0.0001453483845854418, "loss": 0.9934, "step": 708 }, { "epoch": 0.2755270572233557, "grad_norm": 0.22316142916679382, "learning_rate": 0.00014527053328143246, "loss": 1.0707, "step": 709 }, { "epoch": 0.2759156708442631, "grad_norm": 0.21004025638103485, "learning_rate": 0.00014519268197742312, "loss": 1.0543, "step": 710 }, { "epoch": 0.2763042844651705, "grad_norm": 0.22070440649986267, "learning_rate": 0.00014511483067341377, "loss": 1.0467, "step": 711 }, { "epoch": 0.27669289808607794, "grad_norm": 0.21463747322559357, "learning_rate": 0.00014503697936940445, "loss": 1.0793, "step": 712 }, { "epoch": 0.2770815117069853, "grad_norm": 0.23452533781528473, "learning_rate": 0.0001449591280653951, "loss": 1.043, "step": 713 }, { "epoch": 0.2774701253278927, "grad_norm": 0.2405795156955719, "learning_rate": 0.00014488127676138575, "loss": 1.0752, "step": 714 }, { "epoch": 0.27785873894880014, "grad_norm": 0.21546585857868195, "learning_rate": 0.00014480342545737643, "loss": 1.0834, "step": 715 }, { "epoch": 0.27824735256970756, "grad_norm": 0.22675828635692596, "learning_rate": 0.00014472557415336708, "loss": 1.055, "step": 716 }, { "epoch": 0.278635966190615, "grad_norm": 0.2117871195077896, "learning_rate": 0.00014464772284935774, "loss": 1.03, "step": 717 }, { "epoch": 0.2790245798115224, "grad_norm": 0.2193155735731125, "learning_rate": 0.00014456987154534841, "loss": 1.0073, "step": 718 }, { "epoch": 0.2794131934324298, "grad_norm": 0.21447965502738953, "learning_rate": 0.00014449202024133907, "loss": 1.0174, "step": 719 }, { "epoch": 0.27980180705333724, "grad_norm": 0.22867532074451447, "learning_rate": 0.00014441416893732972, "loss": 1.0948, "step": 720 }, { "epoch": 0.28019042067424466, "grad_norm": 0.21570557355880737, "learning_rate": 0.00014433631763332037, "loss": 1.0105, "step": 721 }, { "epoch": 0.280579034295152, "grad_norm": 0.20787014067173004, "learning_rate": 0.00014425846632931102, "loss": 1.0384, "step": 722 }, { "epoch": 0.28096764791605944, "grad_norm": 0.19924762845039368, "learning_rate": 0.0001441806150253017, "loss": 1.0653, "step": 723 }, { "epoch": 0.28135626153696686, "grad_norm": 0.1996215283870697, "learning_rate": 0.00014410276372129235, "loss": 1.0439, "step": 724 }, { "epoch": 0.2817448751578743, "grad_norm": 0.2054813802242279, "learning_rate": 0.000144024912417283, "loss": 0.9895, "step": 725 }, { "epoch": 0.2821334887787817, "grad_norm": 0.2268310785293579, "learning_rate": 0.00014394706111327366, "loss": 1.0993, "step": 726 }, { "epoch": 0.2825221023996891, "grad_norm": 0.19867680966854095, "learning_rate": 0.0001438692098092643, "loss": 0.985, "step": 727 }, { "epoch": 0.28291071602059653, "grad_norm": 0.21099598705768585, "learning_rate": 0.000143791358505255, "loss": 1.0333, "step": 728 }, { "epoch": 0.28329932964150395, "grad_norm": 0.22479215264320374, "learning_rate": 0.00014371350720124564, "loss": 1.0449, "step": 729 }, { "epoch": 0.28368794326241137, "grad_norm": 0.22717688977718353, "learning_rate": 0.0001436356558972363, "loss": 1.0482, "step": 730 }, { "epoch": 0.2840765568833188, "grad_norm": 0.20389345288276672, "learning_rate": 0.00014355780459322695, "loss": 0.956, "step": 731 }, { "epoch": 0.28446517050422615, "grad_norm": 0.21583619713783264, "learning_rate": 0.0001434799532892176, "loss": 1.0154, "step": 732 }, { "epoch": 0.28485378412513357, "grad_norm": 0.2219148874282837, "learning_rate": 0.00014340210198520825, "loss": 1.0553, "step": 733 }, { "epoch": 0.285242397746041, "grad_norm": 0.19920189678668976, "learning_rate": 0.00014332425068119893, "loss": 0.9881, "step": 734 }, { "epoch": 0.2856310113669484, "grad_norm": 0.2295670360326767, "learning_rate": 0.00014324639937718958, "loss": 1.0529, "step": 735 }, { "epoch": 0.2860196249878558, "grad_norm": 0.21271567046642303, "learning_rate": 0.00014316854807318023, "loss": 1.037, "step": 736 }, { "epoch": 0.28640823860876324, "grad_norm": 0.21304361522197723, "learning_rate": 0.00014309069676917088, "loss": 1.048, "step": 737 }, { "epoch": 0.28679685222967066, "grad_norm": 0.19902732968330383, "learning_rate": 0.00014301284546516154, "loss": 1.0306, "step": 738 }, { "epoch": 0.2871854658505781, "grad_norm": 0.1995929330587387, "learning_rate": 0.00014293499416115222, "loss": 1.0394, "step": 739 }, { "epoch": 0.2875740794714855, "grad_norm": 0.20426060259342194, "learning_rate": 0.00014285714285714287, "loss": 1.0052, "step": 740 }, { "epoch": 0.28796269309239286, "grad_norm": 0.20284566283226013, "learning_rate": 0.00014277929155313352, "loss": 1.0115, "step": 741 }, { "epoch": 0.2883513067133003, "grad_norm": 0.2041557878255844, "learning_rate": 0.00014270144024912417, "loss": 1.0473, "step": 742 }, { "epoch": 0.2887399203342077, "grad_norm": 0.2152249962091446, "learning_rate": 0.00014262358894511482, "loss": 1.0802, "step": 743 }, { "epoch": 0.2891285339551151, "grad_norm": 0.20569871366024017, "learning_rate": 0.0001425457376411055, "loss": 1.0203, "step": 744 }, { "epoch": 0.28951714757602254, "grad_norm": 0.21128378808498383, "learning_rate": 0.00014246788633709616, "loss": 1.108, "step": 745 }, { "epoch": 0.28990576119692996, "grad_norm": 0.19587135314941406, "learning_rate": 0.0001423900350330868, "loss": 1.0427, "step": 746 }, { "epoch": 0.2902943748178374, "grad_norm": 0.22052550315856934, "learning_rate": 0.00014231218372907746, "loss": 1.055, "step": 747 }, { "epoch": 0.2906829884387448, "grad_norm": 0.21291717886924744, "learning_rate": 0.0001422343324250681, "loss": 1.0591, "step": 748 }, { "epoch": 0.2910716020596522, "grad_norm": 0.20634084939956665, "learning_rate": 0.0001421564811210588, "loss": 1.0527, "step": 749 }, { "epoch": 0.29146021568055963, "grad_norm": 0.2075488269329071, "learning_rate": 0.00014207862981704944, "loss": 1.0786, "step": 750 }, { "epoch": 0.291848829301467, "grad_norm": 0.19780080020427704, "learning_rate": 0.0001420007785130401, "loss": 1.059, "step": 751 }, { "epoch": 0.2922374429223744, "grad_norm": 0.21212074160575867, "learning_rate": 0.00014192292720903075, "loss": 1.0346, "step": 752 }, { "epoch": 0.29262605654328183, "grad_norm": 0.2218451350927353, "learning_rate": 0.0001418450759050214, "loss": 1.0908, "step": 753 }, { "epoch": 0.29301467016418925, "grad_norm": 0.20107759535312653, "learning_rate": 0.00014176722460101208, "loss": 1.0202, "step": 754 }, { "epoch": 0.29340328378509667, "grad_norm": 0.20933273434638977, "learning_rate": 0.00014168937329700273, "loss": 1.0719, "step": 755 }, { "epoch": 0.2937918974060041, "grad_norm": 0.22369107604026794, "learning_rate": 0.00014161152199299338, "loss": 1.0433, "step": 756 }, { "epoch": 0.2941805110269115, "grad_norm": 0.2113707810640335, "learning_rate": 0.00014153367068898403, "loss": 1.0637, "step": 757 }, { "epoch": 0.2945691246478189, "grad_norm": 0.21105700731277466, "learning_rate": 0.00014145581938497469, "loss": 1.0468, "step": 758 }, { "epoch": 0.29495773826872634, "grad_norm": 0.20189693570137024, "learning_rate": 0.00014137796808096537, "loss": 1.0281, "step": 759 }, { "epoch": 0.2953463518896337, "grad_norm": 0.1954152137041092, "learning_rate": 0.00014130011677695602, "loss": 1.0519, "step": 760 }, { "epoch": 0.2957349655105411, "grad_norm": 0.24295592308044434, "learning_rate": 0.00014122226547294667, "loss": 1.1303, "step": 761 }, { "epoch": 0.29612357913144854, "grad_norm": 0.20158620178699493, "learning_rate": 0.00014114441416893732, "loss": 1.0367, "step": 762 }, { "epoch": 0.29651219275235596, "grad_norm": 0.20734666287899017, "learning_rate": 0.00014106656286492797, "loss": 1.0392, "step": 763 }, { "epoch": 0.2969008063732634, "grad_norm": 0.2177533656358719, "learning_rate": 0.00014098871156091865, "loss": 1.0619, "step": 764 }, { "epoch": 0.2972894199941708, "grad_norm": 0.1961720883846283, "learning_rate": 0.0001409108602569093, "loss": 0.9872, "step": 765 }, { "epoch": 0.2976780336150782, "grad_norm": 0.21530941128730774, "learning_rate": 0.00014083300895289996, "loss": 1.1246, "step": 766 }, { "epoch": 0.29806664723598564, "grad_norm": 0.2039783000946045, "learning_rate": 0.00014075515764889064, "loss": 1.0789, "step": 767 }, { "epoch": 0.29845526085689306, "grad_norm": 0.20641569793224335, "learning_rate": 0.0001406773063448813, "loss": 1.05, "step": 768 }, { "epoch": 0.2988438744778004, "grad_norm": 0.2071225494146347, "learning_rate": 0.00014059945504087194, "loss": 1.047, "step": 769 }, { "epoch": 0.29923248809870784, "grad_norm": 0.20367531478405, "learning_rate": 0.00014052160373686262, "loss": 1.0734, "step": 770 }, { "epoch": 0.29962110171961526, "grad_norm": 0.21718619763851166, "learning_rate": 0.00014044375243285327, "loss": 1.0613, "step": 771 }, { "epoch": 0.3000097153405227, "grad_norm": 0.21649087965488434, "learning_rate": 0.00014036590112884392, "loss": 1.0671, "step": 772 }, { "epoch": 0.3003983289614301, "grad_norm": 0.22223225235939026, "learning_rate": 0.00014028804982483458, "loss": 1.0977, "step": 773 }, { "epoch": 0.3007869425823375, "grad_norm": 0.23101870715618134, "learning_rate": 0.00014021019852082523, "loss": 1.1236, "step": 774 }, { "epoch": 0.30117555620324493, "grad_norm": 0.22855506837368011, "learning_rate": 0.0001401323472168159, "loss": 1.0517, "step": 775 }, { "epoch": 0.30156416982415235, "grad_norm": 0.20862117409706116, "learning_rate": 0.00014005449591280656, "loss": 1.0493, "step": 776 }, { "epoch": 0.30195278344505977, "grad_norm": 0.21692048013210297, "learning_rate": 0.0001399766446087972, "loss": 1.0681, "step": 777 }, { "epoch": 0.3023413970659672, "grad_norm": 0.21541331708431244, "learning_rate": 0.00013989879330478786, "loss": 1.0775, "step": 778 }, { "epoch": 0.30273001068687455, "grad_norm": 0.21221749484539032, "learning_rate": 0.00013982094200077851, "loss": 1.0421, "step": 779 }, { "epoch": 0.30311862430778197, "grad_norm": 0.22497743368148804, "learning_rate": 0.0001397430906967692, "loss": 1.1115, "step": 780 }, { "epoch": 0.3035072379286894, "grad_norm": 0.1974119246006012, "learning_rate": 0.00013966523939275985, "loss": 1.0264, "step": 781 }, { "epoch": 0.3038958515495968, "grad_norm": 0.20349323749542236, "learning_rate": 0.0001395873880887505, "loss": 1.0512, "step": 782 }, { "epoch": 0.3042844651705042, "grad_norm": 0.21116937696933746, "learning_rate": 0.00013950953678474115, "loss": 1.0135, "step": 783 }, { "epoch": 0.30467307879141164, "grad_norm": 0.2133677899837494, "learning_rate": 0.0001394316854807318, "loss": 1.0694, "step": 784 }, { "epoch": 0.30506169241231906, "grad_norm": 0.20406191051006317, "learning_rate": 0.00013935383417672248, "loss": 1.0179, "step": 785 }, { "epoch": 0.3054503060332265, "grad_norm": 0.21428678929805756, "learning_rate": 0.00013927598287271313, "loss": 1.0577, "step": 786 }, { "epoch": 0.3058389196541339, "grad_norm": 0.20878921449184418, "learning_rate": 0.00013919813156870379, "loss": 1.0311, "step": 787 }, { "epoch": 0.30622753327504126, "grad_norm": 0.19033175706863403, "learning_rate": 0.00013912028026469444, "loss": 0.976, "step": 788 }, { "epoch": 0.3066161468959487, "grad_norm": 0.22138020396232605, "learning_rate": 0.0001390424289606851, "loss": 1.0438, "step": 789 }, { "epoch": 0.3070047605168561, "grad_norm": 0.20765596628189087, "learning_rate": 0.00013896457765667577, "loss": 1.0865, "step": 790 }, { "epoch": 0.3073933741377635, "grad_norm": 0.209733247756958, "learning_rate": 0.00013888672635266642, "loss": 1.0648, "step": 791 }, { "epoch": 0.30778198775867094, "grad_norm": 0.1896686851978302, "learning_rate": 0.00013880887504865707, "loss": 1.0133, "step": 792 }, { "epoch": 0.30817060137957836, "grad_norm": 0.21651998162269592, "learning_rate": 0.00013873102374464772, "loss": 1.0729, "step": 793 }, { "epoch": 0.3085592150004858, "grad_norm": 0.21751996874809265, "learning_rate": 0.00013865317244063838, "loss": 1.0444, "step": 794 }, { "epoch": 0.3089478286213932, "grad_norm": 0.20593520998954773, "learning_rate": 0.00013857532113662906, "loss": 1.0304, "step": 795 }, { "epoch": 0.3093364422423006, "grad_norm": 0.19937261939048767, "learning_rate": 0.0001384974698326197, "loss": 1.0017, "step": 796 }, { "epoch": 0.30972505586320803, "grad_norm": 0.18901696801185608, "learning_rate": 0.00013841961852861036, "loss": 1.0362, "step": 797 }, { "epoch": 0.3101136694841154, "grad_norm": 0.2079760730266571, "learning_rate": 0.000138341767224601, "loss": 1.0784, "step": 798 }, { "epoch": 0.3105022831050228, "grad_norm": 0.24873265624046326, "learning_rate": 0.00013826391592059166, "loss": 1.1026, "step": 799 }, { "epoch": 0.31089089672593023, "grad_norm": 0.20185396075248718, "learning_rate": 0.00013818606461658234, "loss": 1.0235, "step": 800 }, { "epoch": 0.31127951034683765, "grad_norm": 0.211393803358078, "learning_rate": 0.000138108213312573, "loss": 1.0999, "step": 801 }, { "epoch": 0.31166812396774507, "grad_norm": 0.19948823750019073, "learning_rate": 0.00013803036200856365, "loss": 1.0242, "step": 802 }, { "epoch": 0.3120567375886525, "grad_norm": 0.21470944583415985, "learning_rate": 0.0001379525107045543, "loss": 1.0736, "step": 803 }, { "epoch": 0.3124453512095599, "grad_norm": 0.2195902317762375, "learning_rate": 0.00013787465940054495, "loss": 1.0368, "step": 804 }, { "epoch": 0.3128339648304673, "grad_norm": 0.22142355144023895, "learning_rate": 0.00013779680809653563, "loss": 1.1022, "step": 805 }, { "epoch": 0.31322257845137474, "grad_norm": 0.20487886667251587, "learning_rate": 0.00013771895679252628, "loss": 1.0478, "step": 806 }, { "epoch": 0.3136111920722821, "grad_norm": 0.217549130320549, "learning_rate": 0.00013764110548851693, "loss": 1.0526, "step": 807 }, { "epoch": 0.3139998056931895, "grad_norm": 0.20199982821941376, "learning_rate": 0.0001375632541845076, "loss": 0.9992, "step": 808 }, { "epoch": 0.31438841931409695, "grad_norm": 0.19496634602546692, "learning_rate": 0.00013748540288049824, "loss": 1.0179, "step": 809 }, { "epoch": 0.31477703293500436, "grad_norm": 0.21999460458755493, "learning_rate": 0.0001374075515764889, "loss": 1.0547, "step": 810 }, { "epoch": 0.3151656465559118, "grad_norm": 0.21421074867248535, "learning_rate": 0.00013732970027247957, "loss": 1.0283, "step": 811 }, { "epoch": 0.3155542601768192, "grad_norm": 0.1913364827632904, "learning_rate": 0.00013725184896847022, "loss": 0.9826, "step": 812 }, { "epoch": 0.3159428737977266, "grad_norm": 0.20509806275367737, "learning_rate": 0.00013717399766446087, "loss": 1.0303, "step": 813 }, { "epoch": 0.31633148741863404, "grad_norm": 0.20309868454933167, "learning_rate": 0.00013709614636045153, "loss": 1.0479, "step": 814 }, { "epoch": 0.31672010103954146, "grad_norm": 0.2274443656206131, "learning_rate": 0.0001370182950564422, "loss": 1.1311, "step": 815 }, { "epoch": 0.3171087146604489, "grad_norm": 0.22785170376300812, "learning_rate": 0.00013694044375243286, "loss": 1.1009, "step": 816 }, { "epoch": 0.31749732828135624, "grad_norm": 0.2105439007282257, "learning_rate": 0.0001368625924484235, "loss": 1.0251, "step": 817 }, { "epoch": 0.31788594190226366, "grad_norm": 0.20583970844745636, "learning_rate": 0.00013678474114441416, "loss": 1.0833, "step": 818 }, { "epoch": 0.3182745555231711, "grad_norm": 0.21091191470623016, "learning_rate": 0.00013670688984040484, "loss": 1.071, "step": 819 }, { "epoch": 0.3186631691440785, "grad_norm": 0.20645928382873535, "learning_rate": 0.0001366290385363955, "loss": 1.0605, "step": 820 }, { "epoch": 0.3190517827649859, "grad_norm": 0.1990513950586319, "learning_rate": 0.00013655118723238614, "loss": 1.0461, "step": 821 }, { "epoch": 0.31944039638589333, "grad_norm": 0.2192249745130539, "learning_rate": 0.00013647333592837682, "loss": 1.0975, "step": 822 }, { "epoch": 0.31982901000680075, "grad_norm": 0.2157617211341858, "learning_rate": 0.00013639548462436748, "loss": 1.091, "step": 823 }, { "epoch": 0.32021762362770817, "grad_norm": 0.21964526176452637, "learning_rate": 0.00013631763332035813, "loss": 1.0286, "step": 824 }, { "epoch": 0.3206062372486156, "grad_norm": 0.2079797089099884, "learning_rate": 0.00013623978201634878, "loss": 1.0257, "step": 825 }, { "epoch": 0.32099485086952295, "grad_norm": 0.21220168471336365, "learning_rate": 0.00013616193071233946, "loss": 1.0046, "step": 826 }, { "epoch": 0.32138346449043037, "grad_norm": 0.2885231673717499, "learning_rate": 0.0001360840794083301, "loss": 1.1442, "step": 827 }, { "epoch": 0.3217720781113378, "grad_norm": 0.2096511274576187, "learning_rate": 0.00013600622810432076, "loss": 1.0209, "step": 828 }, { "epoch": 0.3221606917322452, "grad_norm": 0.2179451286792755, "learning_rate": 0.00013592837680031142, "loss": 1.0548, "step": 829 }, { "epoch": 0.3225493053531526, "grad_norm": 0.2096329927444458, "learning_rate": 0.00013585052549630207, "loss": 1.0279, "step": 830 }, { "epoch": 0.32293791897406005, "grad_norm": 0.22531811892986298, "learning_rate": 0.00013577267419229275, "loss": 1.0463, "step": 831 }, { "epoch": 0.32332653259496746, "grad_norm": 0.22516901791095734, "learning_rate": 0.0001356948228882834, "loss": 1.1127, "step": 832 }, { "epoch": 0.3237151462158749, "grad_norm": 0.22487780451774597, "learning_rate": 0.00013561697158427405, "loss": 1.0707, "step": 833 }, { "epoch": 0.3241037598367823, "grad_norm": 0.20976543426513672, "learning_rate": 0.0001355391202802647, "loss": 1.0217, "step": 834 }, { "epoch": 0.32449237345768966, "grad_norm": 0.19849295914173126, "learning_rate": 0.00013546126897625535, "loss": 1.021, "step": 835 }, { "epoch": 0.3248809870785971, "grad_norm": 0.21772268414497375, "learning_rate": 0.00013538341767224603, "loss": 1.0605, "step": 836 }, { "epoch": 0.3252696006995045, "grad_norm": 0.19670265913009644, "learning_rate": 0.00013530556636823669, "loss": 1.0165, "step": 837 }, { "epoch": 0.3256582143204119, "grad_norm": 0.19339734315872192, "learning_rate": 0.00013522771506422734, "loss": 1.0203, "step": 838 }, { "epoch": 0.32604682794131934, "grad_norm": 0.21289557218551636, "learning_rate": 0.000135149863760218, "loss": 1.0252, "step": 839 }, { "epoch": 0.32643544156222676, "grad_norm": 0.1964789777994156, "learning_rate": 0.00013507201245620864, "loss": 1.0392, "step": 840 }, { "epoch": 0.3268240551831342, "grad_norm": 0.20783716440200806, "learning_rate": 0.00013499416115219932, "loss": 1.0569, "step": 841 }, { "epoch": 0.3272126688040416, "grad_norm": 0.22782161831855774, "learning_rate": 0.00013491630984818997, "loss": 1.0555, "step": 842 }, { "epoch": 0.327601282424949, "grad_norm": 0.22771142423152924, "learning_rate": 0.00013483845854418063, "loss": 1.085, "step": 843 }, { "epoch": 0.32798989604585643, "grad_norm": 0.19773711264133453, "learning_rate": 0.00013476060724017128, "loss": 1.008, "step": 844 }, { "epoch": 0.3283785096667638, "grad_norm": 0.22399166226387024, "learning_rate": 0.00013468275593616193, "loss": 1.0511, "step": 845 }, { "epoch": 0.3287671232876712, "grad_norm": 0.20488236844539642, "learning_rate": 0.00013460490463215258, "loss": 1.0883, "step": 846 }, { "epoch": 0.32915573690857863, "grad_norm": 0.21387654542922974, "learning_rate": 0.00013452705332814326, "loss": 1.0808, "step": 847 }, { "epoch": 0.32954435052948605, "grad_norm": 0.1972568780183792, "learning_rate": 0.0001344492020241339, "loss": 1.0555, "step": 848 }, { "epoch": 0.32993296415039347, "grad_norm": 0.20835663378238678, "learning_rate": 0.00013437135072012456, "loss": 1.0473, "step": 849 }, { "epoch": 0.3303215777713009, "grad_norm": 0.19707520306110382, "learning_rate": 0.00013429349941611522, "loss": 0.9585, "step": 850 }, { "epoch": 0.3307101913922083, "grad_norm": 0.19163411855697632, "learning_rate": 0.00013421564811210587, "loss": 1.0025, "step": 851 }, { "epoch": 0.3310988050131157, "grad_norm": 0.19730083644390106, "learning_rate": 0.00013413779680809655, "loss": 1.0696, "step": 852 }, { "epoch": 0.33148741863402315, "grad_norm": 0.19537493586540222, "learning_rate": 0.0001340599455040872, "loss": 1.0466, "step": 853 }, { "epoch": 0.3318760322549305, "grad_norm": 0.2255164235830307, "learning_rate": 0.00013398209420007785, "loss": 1.0659, "step": 854 }, { "epoch": 0.3322646458758379, "grad_norm": 0.19774770736694336, "learning_rate": 0.0001339042428960685, "loss": 1.0326, "step": 855 }, { "epoch": 0.33265325949674535, "grad_norm": 0.2004510909318924, "learning_rate": 0.00013382639159205916, "loss": 1.0327, "step": 856 }, { "epoch": 0.33304187311765276, "grad_norm": 0.19187591969966888, "learning_rate": 0.00013374854028804984, "loss": 1.0069, "step": 857 }, { "epoch": 0.3334304867385602, "grad_norm": 0.18775832653045654, "learning_rate": 0.0001336706889840405, "loss": 1.0083, "step": 858 }, { "epoch": 0.3338191003594676, "grad_norm": 0.2005717158317566, "learning_rate": 0.00013359283768003114, "loss": 1.0398, "step": 859 }, { "epoch": 0.334207713980375, "grad_norm": 0.19705893099308014, "learning_rate": 0.0001335149863760218, "loss": 1.0031, "step": 860 }, { "epoch": 0.33459632760128244, "grad_norm": 0.19589562714099884, "learning_rate": 0.00013343713507201244, "loss": 0.9831, "step": 861 }, { "epoch": 0.33498494122218986, "grad_norm": 0.19302591681480408, "learning_rate": 0.00013335928376800312, "loss": 1.0009, "step": 862 }, { "epoch": 0.3353735548430973, "grad_norm": 0.20499618351459503, "learning_rate": 0.00013328143246399377, "loss": 1.0205, "step": 863 }, { "epoch": 0.33576216846400464, "grad_norm": 0.20514456927776337, "learning_rate": 0.00013320358115998443, "loss": 1.0837, "step": 864 }, { "epoch": 0.33615078208491206, "grad_norm": 0.19285848736763, "learning_rate": 0.00013312572985597508, "loss": 1.0167, "step": 865 }, { "epoch": 0.3365393957058195, "grad_norm": 0.20891553163528442, "learning_rate": 0.00013304787855196573, "loss": 1.0127, "step": 866 }, { "epoch": 0.3369280093267269, "grad_norm": 0.20511706173419952, "learning_rate": 0.0001329700272479564, "loss": 0.964, "step": 867 }, { "epoch": 0.3373166229476343, "grad_norm": 0.1855512261390686, "learning_rate": 0.00013289217594394706, "loss": 0.9721, "step": 868 }, { "epoch": 0.33770523656854173, "grad_norm": 0.20010098814964294, "learning_rate": 0.00013281432463993771, "loss": 1.0411, "step": 869 }, { "epoch": 0.33809385018944915, "grad_norm": 0.1991325318813324, "learning_rate": 0.0001327364733359284, "loss": 0.9658, "step": 870 }, { "epoch": 0.33848246381035657, "grad_norm": 0.19895736873149872, "learning_rate": 0.00013265862203191905, "loss": 1.0744, "step": 871 }, { "epoch": 0.338871077431264, "grad_norm": 0.2091255635023117, "learning_rate": 0.0001325807707279097, "loss": 1.0375, "step": 872 }, { "epoch": 0.33925969105217135, "grad_norm": 0.21355532109737396, "learning_rate": 0.00013250291942390035, "loss": 1.09, "step": 873 }, { "epoch": 0.33964830467307877, "grad_norm": 0.21844851970672607, "learning_rate": 0.00013242506811989103, "loss": 1.0769, "step": 874 }, { "epoch": 0.3400369182939862, "grad_norm": 0.1877543330192566, "learning_rate": 0.00013234721681588168, "loss": 1.0199, "step": 875 }, { "epoch": 0.3404255319148936, "grad_norm": 0.2020038366317749, "learning_rate": 0.00013226936551187233, "loss": 1.0218, "step": 876 }, { "epoch": 0.340814145535801, "grad_norm": 0.20682141184806824, "learning_rate": 0.000132191514207863, "loss": 1.0891, "step": 877 }, { "epoch": 0.34120275915670845, "grad_norm": 0.21942824125289917, "learning_rate": 0.00013211366290385366, "loss": 0.9877, "step": 878 }, { "epoch": 0.34159137277761586, "grad_norm": 0.21150313317775726, "learning_rate": 0.00013203581159984432, "loss": 1.0815, "step": 879 }, { "epoch": 0.3419799863985233, "grad_norm": 0.2073293924331665, "learning_rate": 0.00013195796029583497, "loss": 1.0579, "step": 880 }, { "epoch": 0.3423686000194307, "grad_norm": 0.221574068069458, "learning_rate": 0.00013188010899182562, "loss": 1.0279, "step": 881 }, { "epoch": 0.3427572136403381, "grad_norm": 0.22334492206573486, "learning_rate": 0.00013180225768781627, "loss": 1.0837, "step": 882 }, { "epoch": 0.3431458272612455, "grad_norm": 0.18817654252052307, "learning_rate": 0.00013172440638380695, "loss": 1.0262, "step": 883 }, { "epoch": 0.3435344408821529, "grad_norm": 0.20126822590827942, "learning_rate": 0.0001316465550797976, "loss": 1.0679, "step": 884 }, { "epoch": 0.3439230545030603, "grad_norm": 0.2128864973783493, "learning_rate": 0.00013156870377578825, "loss": 1.0316, "step": 885 }, { "epoch": 0.34431166812396774, "grad_norm": 0.20054499804973602, "learning_rate": 0.0001314908524717789, "loss": 1.0024, "step": 886 }, { "epoch": 0.34470028174487516, "grad_norm": 0.21358034014701843, "learning_rate": 0.00013141300116776956, "loss": 1.0475, "step": 887 }, { "epoch": 0.3450888953657826, "grad_norm": 0.21377703547477722, "learning_rate": 0.00013133514986376024, "loss": 1.0957, "step": 888 }, { "epoch": 0.34547750898669, "grad_norm": 0.20166514813899994, "learning_rate": 0.0001312572985597509, "loss": 1.0189, "step": 889 }, { "epoch": 0.3458661226075974, "grad_norm": 0.20424878597259521, "learning_rate": 0.00013117944725574154, "loss": 1.0896, "step": 890 }, { "epoch": 0.34625473622850483, "grad_norm": 0.19028648734092712, "learning_rate": 0.0001311015959517322, "loss": 0.9881, "step": 891 }, { "epoch": 0.3466433498494122, "grad_norm": 0.20828665792942047, "learning_rate": 0.00013102374464772285, "loss": 0.9932, "step": 892 }, { "epoch": 0.3470319634703196, "grad_norm": 0.20756572484970093, "learning_rate": 0.00013094589334371353, "loss": 1.0406, "step": 893 }, { "epoch": 0.34742057709122703, "grad_norm": 0.20768921077251434, "learning_rate": 0.00013086804203970418, "loss": 0.9652, "step": 894 }, { "epoch": 0.34780919071213445, "grad_norm": 0.20660027861595154, "learning_rate": 0.00013079019073569483, "loss": 1.0728, "step": 895 }, { "epoch": 0.34819780433304187, "grad_norm": 0.20186837017536163, "learning_rate": 0.00013071233943168548, "loss": 1.0407, "step": 896 }, { "epoch": 0.3485864179539493, "grad_norm": 0.20880667865276337, "learning_rate": 0.00013063448812767613, "loss": 1.0275, "step": 897 }, { "epoch": 0.3489750315748567, "grad_norm": 0.22212949395179749, "learning_rate": 0.0001305566368236668, "loss": 1.0293, "step": 898 }, { "epoch": 0.3493636451957641, "grad_norm": 0.20552745461463928, "learning_rate": 0.00013047878551965746, "loss": 1.0434, "step": 899 }, { "epoch": 0.34975225881667155, "grad_norm": 0.21239839494228363, "learning_rate": 0.00013040093421564812, "loss": 1.052, "step": 900 }, { "epoch": 0.3501408724375789, "grad_norm": 0.22420544922351837, "learning_rate": 0.00013032308291163877, "loss": 1.0236, "step": 901 }, { "epoch": 0.35052948605848633, "grad_norm": 0.23435090482234955, "learning_rate": 0.00013024523160762942, "loss": 1.0876, "step": 902 }, { "epoch": 0.35091809967939375, "grad_norm": 0.22763386368751526, "learning_rate": 0.0001301673803036201, "loss": 1.0636, "step": 903 }, { "epoch": 0.35130671330030117, "grad_norm": 0.20948883891105652, "learning_rate": 0.00013008952899961075, "loss": 1.0083, "step": 904 }, { "epoch": 0.3516953269212086, "grad_norm": 0.20408779382705688, "learning_rate": 0.0001300116776956014, "loss": 1.039, "step": 905 }, { "epoch": 0.352083940542116, "grad_norm": 0.2126050591468811, "learning_rate": 0.00012993382639159206, "loss": 1.0365, "step": 906 }, { "epoch": 0.3524725541630234, "grad_norm": 0.20314334332942963, "learning_rate": 0.0001298559750875827, "loss": 1.0474, "step": 907 }, { "epoch": 0.35286116778393084, "grad_norm": 0.23720984160900116, "learning_rate": 0.0001297781237835734, "loss": 1.0529, "step": 908 }, { "epoch": 0.35324978140483826, "grad_norm": 0.22642800211906433, "learning_rate": 0.00012970027247956404, "loss": 1.0586, "step": 909 }, { "epoch": 0.3536383950257457, "grad_norm": 0.20469972491264343, "learning_rate": 0.0001296224211755547, "loss": 1.0267, "step": 910 }, { "epoch": 0.35402700864665304, "grad_norm": 0.197368785738945, "learning_rate": 0.00012954456987154534, "loss": 1.0348, "step": 911 }, { "epoch": 0.35441562226756046, "grad_norm": 0.21924498677253723, "learning_rate": 0.000129466718567536, "loss": 1.0861, "step": 912 }, { "epoch": 0.3548042358884679, "grad_norm": 0.22006285190582275, "learning_rate": 0.00012938886726352667, "loss": 1.0545, "step": 913 }, { "epoch": 0.3551928495093753, "grad_norm": 0.22419220209121704, "learning_rate": 0.00012931101595951733, "loss": 1.0716, "step": 914 }, { "epoch": 0.3555814631302827, "grad_norm": 0.215990349650383, "learning_rate": 0.00012923316465550798, "loss": 1.0619, "step": 915 }, { "epoch": 0.35597007675119013, "grad_norm": 0.20783264935016632, "learning_rate": 0.00012915531335149863, "loss": 1.0412, "step": 916 }, { "epoch": 0.35635869037209755, "grad_norm": 0.24584618210792542, "learning_rate": 0.00012907746204748928, "loss": 1.1165, "step": 917 }, { "epoch": 0.35674730399300497, "grad_norm": 0.23146122694015503, "learning_rate": 0.00012899961074347996, "loss": 1.1111, "step": 918 }, { "epoch": 0.3571359176139124, "grad_norm": 0.19983729720115662, "learning_rate": 0.00012892175943947061, "loss": 1.0674, "step": 919 }, { "epoch": 0.35752453123481975, "grad_norm": 0.2161000818014145, "learning_rate": 0.00012884390813546127, "loss": 1.076, "step": 920 }, { "epoch": 0.35791314485572717, "grad_norm": 0.21042793989181519, "learning_rate": 0.00012876605683145192, "loss": 1.0535, "step": 921 }, { "epoch": 0.3583017584766346, "grad_norm": 0.20135439932346344, "learning_rate": 0.0001286882055274426, "loss": 1.0059, "step": 922 }, { "epoch": 0.358690372097542, "grad_norm": 0.19394971430301666, "learning_rate": 0.00012861035422343325, "loss": 1.0381, "step": 923 }, { "epoch": 0.35907898571844943, "grad_norm": 0.21171030402183533, "learning_rate": 0.0001285325029194239, "loss": 1.0513, "step": 924 }, { "epoch": 0.35946759933935685, "grad_norm": 0.19476690888404846, "learning_rate": 0.00012845465161541458, "loss": 1.0003, "step": 925 }, { "epoch": 0.35985621296026427, "grad_norm": 0.20468670129776, "learning_rate": 0.00012837680031140523, "loss": 1.0608, "step": 926 }, { "epoch": 0.3602448265811717, "grad_norm": 0.21159446239471436, "learning_rate": 0.00012829894900739588, "loss": 1.0734, "step": 927 }, { "epoch": 0.3606334402020791, "grad_norm": 0.21179519593715668, "learning_rate": 0.00012822109770338654, "loss": 1.0957, "step": 928 }, { "epoch": 0.3610220538229865, "grad_norm": 0.20997527241706848, "learning_rate": 0.00012814324639937722, "loss": 1.0644, "step": 929 }, { "epoch": 0.3614106674438939, "grad_norm": 0.21178296208381653, "learning_rate": 0.00012806539509536787, "loss": 1.0208, "step": 930 }, { "epoch": 0.3617992810648013, "grad_norm": 0.20890356600284576, "learning_rate": 0.00012798754379135852, "loss": 1.0888, "step": 931 }, { "epoch": 0.3621878946857087, "grad_norm": 0.20177409052848816, "learning_rate": 0.00012790969248734917, "loss": 0.9741, "step": 932 }, { "epoch": 0.36257650830661614, "grad_norm": 0.23504556715488434, "learning_rate": 0.00012783184118333982, "loss": 1.1048, "step": 933 }, { "epoch": 0.36296512192752356, "grad_norm": 0.22829356789588928, "learning_rate": 0.0001277539898793305, "loss": 1.0798, "step": 934 }, { "epoch": 0.363353735548431, "grad_norm": 0.2068483531475067, "learning_rate": 0.00012767613857532116, "loss": 1.0452, "step": 935 }, { "epoch": 0.3637423491693384, "grad_norm": 0.2093171775341034, "learning_rate": 0.0001275982872713118, "loss": 1.0742, "step": 936 }, { "epoch": 0.3641309627902458, "grad_norm": 0.21478736400604248, "learning_rate": 0.00012752043596730246, "loss": 1.0572, "step": 937 }, { "epoch": 0.36451957641115323, "grad_norm": 0.1906953752040863, "learning_rate": 0.0001274425846632931, "loss": 1.0107, "step": 938 }, { "epoch": 0.3649081900320606, "grad_norm": 0.20580604672431946, "learning_rate": 0.0001273647333592838, "loss": 1.0677, "step": 939 }, { "epoch": 0.365296803652968, "grad_norm": 0.22586850821971893, "learning_rate": 0.00012728688205527444, "loss": 1.0389, "step": 940 }, { "epoch": 0.36568541727387543, "grad_norm": 0.199899360537529, "learning_rate": 0.0001272090307512651, "loss": 1.0462, "step": 941 }, { "epoch": 0.36607403089478285, "grad_norm": 0.19881689548492432, "learning_rate": 0.00012713117944725575, "loss": 1.0565, "step": 942 }, { "epoch": 0.3664626445156903, "grad_norm": 0.21748925745487213, "learning_rate": 0.0001270533281432464, "loss": 1.0659, "step": 943 }, { "epoch": 0.3668512581365977, "grad_norm": 0.19363689422607422, "learning_rate": 0.00012697547683923708, "loss": 1.0307, "step": 944 }, { "epoch": 0.3672398717575051, "grad_norm": 0.21701784431934357, "learning_rate": 0.00012689762553522773, "loss": 1.0684, "step": 945 }, { "epoch": 0.36762848537841253, "grad_norm": 0.21406958997249603, "learning_rate": 0.00012681977423121838, "loss": 1.0703, "step": 946 }, { "epoch": 0.36801709899931995, "grad_norm": 0.23539729416370392, "learning_rate": 0.00012674192292720903, "loss": 1.1537, "step": 947 }, { "epoch": 0.36840571262022737, "grad_norm": 0.2177354395389557, "learning_rate": 0.00012666407162319969, "loss": 1.0131, "step": 948 }, { "epoch": 0.36879432624113473, "grad_norm": 0.255346417427063, "learning_rate": 0.00012658622031919037, "loss": 0.9807, "step": 949 }, { "epoch": 0.36918293986204215, "grad_norm": 0.2139921486377716, "learning_rate": 0.00012650836901518102, "loss": 1.0392, "step": 950 }, { "epoch": 0.36957155348294957, "grad_norm": 0.22490833699703217, "learning_rate": 0.00012643051771117167, "loss": 1.0512, "step": 951 }, { "epoch": 0.369960167103857, "grad_norm": 0.20698820054531097, "learning_rate": 0.00012635266640716232, "loss": 1.0391, "step": 952 }, { "epoch": 0.3703487807247644, "grad_norm": 0.2276201844215393, "learning_rate": 0.00012627481510315297, "loss": 1.0513, "step": 953 }, { "epoch": 0.3707373943456718, "grad_norm": 0.2493600994348526, "learning_rate": 0.00012619696379914365, "loss": 1.0136, "step": 954 }, { "epoch": 0.37112600796657924, "grad_norm": 0.2155001014471054, "learning_rate": 0.0001261191124951343, "loss": 1.0523, "step": 955 }, { "epoch": 0.37151462158748666, "grad_norm": 0.21571211516857147, "learning_rate": 0.00012604126119112496, "loss": 1.0288, "step": 956 }, { "epoch": 0.3719032352083941, "grad_norm": 0.23238877952098846, "learning_rate": 0.0001259634098871156, "loss": 1.0638, "step": 957 }, { "epoch": 0.37229184882930144, "grad_norm": 0.2002813220024109, "learning_rate": 0.00012588555858310626, "loss": 0.9665, "step": 958 }, { "epoch": 0.37268046245020886, "grad_norm": 0.21712858974933624, "learning_rate": 0.0001258077072790969, "loss": 1.0469, "step": 959 }, { "epoch": 0.3730690760711163, "grad_norm": 0.2178192287683487, "learning_rate": 0.0001257298559750876, "loss": 1.0267, "step": 960 }, { "epoch": 0.3734576896920237, "grad_norm": 0.25488024950027466, "learning_rate": 0.00012565200467107824, "loss": 1.0153, "step": 961 }, { "epoch": 0.3738463033129311, "grad_norm": 0.20070038735866547, "learning_rate": 0.0001255741533670689, "loss": 1.0279, "step": 962 }, { "epoch": 0.37423491693383854, "grad_norm": 0.21885356307029724, "learning_rate": 0.00012549630206305955, "loss": 1.0395, "step": 963 }, { "epoch": 0.37462353055474595, "grad_norm": 0.2407921701669693, "learning_rate": 0.0001254184507590502, "loss": 1.0767, "step": 964 }, { "epoch": 0.3750121441756534, "grad_norm": 0.20645053684711456, "learning_rate": 0.00012534059945504088, "loss": 1.0318, "step": 965 }, { "epoch": 0.3754007577965608, "grad_norm": 0.21275092661380768, "learning_rate": 0.00012526274815103153, "loss": 1.0546, "step": 966 }, { "epoch": 0.3757893714174682, "grad_norm": 0.21574917435646057, "learning_rate": 0.00012518489684702218, "loss": 1.032, "step": 967 }, { "epoch": 0.3761779850383756, "grad_norm": 0.21589480340480804, "learning_rate": 0.00012510704554301284, "loss": 1.0834, "step": 968 }, { "epoch": 0.376566598659283, "grad_norm": 0.19576796889305115, "learning_rate": 0.0001250291942390035, "loss": 1.0178, "step": 969 }, { "epoch": 0.3769552122801904, "grad_norm": 0.20941287279129028, "learning_rate": 0.00012495134293499417, "loss": 1.0712, "step": 970 }, { "epoch": 0.37734382590109783, "grad_norm": 0.22585494816303253, "learning_rate": 0.00012487349163098482, "loss": 1.0401, "step": 971 }, { "epoch": 0.37773243952200525, "grad_norm": 0.21093420684337616, "learning_rate": 0.00012479564032697547, "loss": 1.0569, "step": 972 }, { "epoch": 0.37812105314291267, "grad_norm": 0.22375014424324036, "learning_rate": 0.00012471778902296612, "loss": 1.0687, "step": 973 }, { "epoch": 0.3785096667638201, "grad_norm": 0.19787487387657166, "learning_rate": 0.0001246399377189568, "loss": 1.0266, "step": 974 }, { "epoch": 0.3788982803847275, "grad_norm": 0.20633013546466827, "learning_rate": 0.00012456208641494745, "loss": 0.9996, "step": 975 }, { "epoch": 0.3792868940056349, "grad_norm": 0.21559873223304749, "learning_rate": 0.0001244842351109381, "loss": 1.0851, "step": 976 }, { "epoch": 0.3796755076265423, "grad_norm": 0.2166333943605423, "learning_rate": 0.00012440638380692879, "loss": 1.0859, "step": 977 }, { "epoch": 0.3800641212474497, "grad_norm": 0.18558773398399353, "learning_rate": 0.00012432853250291944, "loss": 0.9534, "step": 978 }, { "epoch": 0.3804527348683571, "grad_norm": 0.2086942344903946, "learning_rate": 0.0001242506811989101, "loss": 1.0786, "step": 979 }, { "epoch": 0.38084134848926454, "grad_norm": 0.2207823544740677, "learning_rate": 0.00012417282989490074, "loss": 1.0626, "step": 980 }, { "epoch": 0.38122996211017196, "grad_norm": 0.21255749464035034, "learning_rate": 0.00012409497859089142, "loss": 1.063, "step": 981 }, { "epoch": 0.3816185757310794, "grad_norm": 0.20682042837142944, "learning_rate": 0.00012401712728688207, "loss": 1.034, "step": 982 }, { "epoch": 0.3820071893519868, "grad_norm": 0.2084134966135025, "learning_rate": 0.00012393927598287272, "loss": 1.0481, "step": 983 }, { "epoch": 0.3823958029728942, "grad_norm": 0.1922312080860138, "learning_rate": 0.00012386142467886338, "loss": 1.0461, "step": 984 }, { "epoch": 0.38278441659380164, "grad_norm": 0.20893707871437073, "learning_rate": 0.00012378357337485406, "loss": 1.0797, "step": 985 }, { "epoch": 0.383173030214709, "grad_norm": 0.19717541337013245, "learning_rate": 0.0001237057220708447, "loss": 1.0028, "step": 986 }, { "epoch": 0.3835616438356164, "grad_norm": 0.20688053965568542, "learning_rate": 0.00012362787076683536, "loss": 0.989, "step": 987 }, { "epoch": 0.38395025745652384, "grad_norm": 0.20580583810806274, "learning_rate": 0.000123550019462826, "loss": 1.06, "step": 988 }, { "epoch": 0.38433887107743125, "grad_norm": 0.2151709794998169, "learning_rate": 0.00012347216815881666, "loss": 1.0685, "step": 989 }, { "epoch": 0.3847274846983387, "grad_norm": 0.19573980569839478, "learning_rate": 0.00012339431685480734, "loss": 1.0072, "step": 990 }, { "epoch": 0.3851160983192461, "grad_norm": 0.1949119120836258, "learning_rate": 0.000123316465550798, "loss": 0.9995, "step": 991 }, { "epoch": 0.3855047119401535, "grad_norm": 0.2062375247478485, "learning_rate": 0.00012323861424678865, "loss": 1.0694, "step": 992 }, { "epoch": 0.38589332556106093, "grad_norm": 0.2007209211587906, "learning_rate": 0.0001231607629427793, "loss": 1.0397, "step": 993 }, { "epoch": 0.38628193918196835, "grad_norm": 0.2231544405221939, "learning_rate": 0.00012308291163876995, "loss": 1.0755, "step": 994 }, { "epoch": 0.38667055280287577, "grad_norm": 0.2103337049484253, "learning_rate": 0.0001230050603347606, "loss": 1.0505, "step": 995 }, { "epoch": 0.38705916642378313, "grad_norm": 0.20178386569023132, "learning_rate": 0.00012292720903075128, "loss": 1.0696, "step": 996 }, { "epoch": 0.38744778004469055, "grad_norm": 0.21268007159233093, "learning_rate": 0.00012284935772674193, "loss": 1.0262, "step": 997 }, { "epoch": 0.38783639366559797, "grad_norm": 0.21439722180366516, "learning_rate": 0.0001227715064227326, "loss": 1.0718, "step": 998 }, { "epoch": 0.3882250072865054, "grad_norm": 0.19691336154937744, "learning_rate": 0.00012269365511872324, "loss": 0.9663, "step": 999 }, { "epoch": 0.3886136209074128, "grad_norm": 0.2165926694869995, "learning_rate": 0.0001226158038147139, "loss": 1.0432, "step": 1000 }, { "epoch": 0.3890022345283202, "grad_norm": 0.20730604231357574, "learning_rate": 0.00012253795251070457, "loss": 1.0386, "step": 1001 }, { "epoch": 0.38939084814922764, "grad_norm": 0.2138068974018097, "learning_rate": 0.00012246010120669522, "loss": 1.0683, "step": 1002 }, { "epoch": 0.38977946177013506, "grad_norm": 0.2118951678276062, "learning_rate": 0.00012238224990268587, "loss": 1.0393, "step": 1003 }, { "epoch": 0.3901680753910425, "grad_norm": 0.20879961550235748, "learning_rate": 0.00012230439859867653, "loss": 1.0349, "step": 1004 }, { "epoch": 0.39055668901194984, "grad_norm": 0.19588464498519897, "learning_rate": 0.00012222654729466718, "loss": 1.0226, "step": 1005 }, { "epoch": 0.39094530263285726, "grad_norm": 0.2059485912322998, "learning_rate": 0.00012214869599065786, "loss": 1.052, "step": 1006 }, { "epoch": 0.3913339162537647, "grad_norm": 0.2299761176109314, "learning_rate": 0.0001220708446866485, "loss": 1.1055, "step": 1007 }, { "epoch": 0.3917225298746721, "grad_norm": 0.20196737349033356, "learning_rate": 0.00012199299338263916, "loss": 1.0497, "step": 1008 }, { "epoch": 0.3921111434955795, "grad_norm": 0.20615293085575104, "learning_rate": 0.00012191514207862981, "loss": 1.047, "step": 1009 }, { "epoch": 0.39249975711648694, "grad_norm": 0.20265278220176697, "learning_rate": 0.00012183729077462047, "loss": 1.0035, "step": 1010 }, { "epoch": 0.39288837073739435, "grad_norm": 0.20197926461696625, "learning_rate": 0.00012175943947061114, "loss": 0.9847, "step": 1011 }, { "epoch": 0.3932769843583018, "grad_norm": 0.19974152743816376, "learning_rate": 0.0001216815881666018, "loss": 1.0669, "step": 1012 }, { "epoch": 0.3936655979792092, "grad_norm": 0.21684005856513977, "learning_rate": 0.00012160373686259245, "loss": 1.0562, "step": 1013 }, { "epoch": 0.3940542116001166, "grad_norm": 0.2030404955148697, "learning_rate": 0.00012152588555858311, "loss": 1.0159, "step": 1014 }, { "epoch": 0.394442825221024, "grad_norm": 0.2123572677373886, "learning_rate": 0.00012144803425457377, "loss": 1.0757, "step": 1015 }, { "epoch": 0.3948314388419314, "grad_norm": 0.20320011675357819, "learning_rate": 0.00012137018295056443, "loss": 1.038, "step": 1016 }, { "epoch": 0.3952200524628388, "grad_norm": 0.20120739936828613, "learning_rate": 0.00012129233164655508, "loss": 1.1015, "step": 1017 }, { "epoch": 0.39560866608374623, "grad_norm": 0.19862449169158936, "learning_rate": 0.00012121448034254575, "loss": 1.0328, "step": 1018 }, { "epoch": 0.39599727970465365, "grad_norm": 0.19761312007904053, "learning_rate": 0.0001211366290385364, "loss": 0.997, "step": 1019 }, { "epoch": 0.39638589332556107, "grad_norm": 0.1943569928407669, "learning_rate": 0.00012105877773452705, "loss": 1.0099, "step": 1020 }, { "epoch": 0.3967745069464685, "grad_norm": 0.2109062373638153, "learning_rate": 0.00012098092643051773, "loss": 1.1039, "step": 1021 }, { "epoch": 0.3971631205673759, "grad_norm": 0.20966266095638275, "learning_rate": 0.00012090307512650839, "loss": 1.1208, "step": 1022 }, { "epoch": 0.3975517341882833, "grad_norm": 0.19208088517189026, "learning_rate": 0.00012082522382249904, "loss": 1.0147, "step": 1023 }, { "epoch": 0.3979403478091907, "grad_norm": 0.21821236610412598, "learning_rate": 0.00012074737251848969, "loss": 1.0615, "step": 1024 }, { "epoch": 0.3983289614300981, "grad_norm": 0.20031368732452393, "learning_rate": 0.00012066952121448034, "loss": 1.0303, "step": 1025 }, { "epoch": 0.3987175750510055, "grad_norm": 0.22910597920417786, "learning_rate": 0.00012059166991047102, "loss": 1.0182, "step": 1026 }, { "epoch": 0.39910618867191294, "grad_norm": 0.20816978812217712, "learning_rate": 0.00012051381860646167, "loss": 1.0142, "step": 1027 }, { "epoch": 0.39949480229282036, "grad_norm": 0.20989780128002167, "learning_rate": 0.00012043596730245232, "loss": 1.0676, "step": 1028 }, { "epoch": 0.3998834159137278, "grad_norm": 0.21894055604934692, "learning_rate": 0.00012035811599844298, "loss": 1.0222, "step": 1029 }, { "epoch": 0.4002720295346352, "grad_norm": 0.2170870155096054, "learning_rate": 0.00012028026469443363, "loss": 1.0319, "step": 1030 }, { "epoch": 0.4006606431555426, "grad_norm": 0.20869679749011993, "learning_rate": 0.00012020241339042428, "loss": 1.055, "step": 1031 }, { "epoch": 0.40104925677645004, "grad_norm": 0.18850640952587128, "learning_rate": 0.00012012456208641496, "loss": 0.9993, "step": 1032 }, { "epoch": 0.40143787039735745, "grad_norm": 0.21462580561637878, "learning_rate": 0.00012004671078240561, "loss": 1.0115, "step": 1033 }, { "epoch": 0.4018264840182648, "grad_norm": 0.2008499950170517, "learning_rate": 0.00011996885947839626, "loss": 1.0229, "step": 1034 }, { "epoch": 0.40221509763917224, "grad_norm": 0.20063354074954987, "learning_rate": 0.00011989100817438692, "loss": 1.0295, "step": 1035 }, { "epoch": 0.40260371126007966, "grad_norm": 0.20655786991119385, "learning_rate": 0.00011981315687037757, "loss": 1.0044, "step": 1036 }, { "epoch": 0.4029923248809871, "grad_norm": 0.1985999196767807, "learning_rate": 0.00011973530556636825, "loss": 1.0063, "step": 1037 }, { "epoch": 0.4033809385018945, "grad_norm": 0.2039060890674591, "learning_rate": 0.0001196574542623589, "loss": 1.044, "step": 1038 }, { "epoch": 0.4037695521228019, "grad_norm": 0.21838189661502838, "learning_rate": 0.00011957960295834955, "loss": 1.1101, "step": 1039 }, { "epoch": 0.40415816574370933, "grad_norm": 0.21508415043354034, "learning_rate": 0.00011950175165434022, "loss": 1.0764, "step": 1040 }, { "epoch": 0.40454677936461675, "grad_norm": 0.2089119255542755, "learning_rate": 0.00011942390035033087, "loss": 0.9986, "step": 1041 }, { "epoch": 0.40493539298552417, "grad_norm": 0.19859452545642853, "learning_rate": 0.00011934604904632153, "loss": 1.0122, "step": 1042 }, { "epoch": 0.40532400660643153, "grad_norm": 0.2018653154373169, "learning_rate": 0.00011926819774231219, "loss": 1.0187, "step": 1043 }, { "epoch": 0.40571262022733895, "grad_norm": 0.19892063736915588, "learning_rate": 0.00011919034643830285, "loss": 1.0029, "step": 1044 }, { "epoch": 0.40610123384824637, "grad_norm": 0.20355650782585144, "learning_rate": 0.0001191124951342935, "loss": 1.0484, "step": 1045 }, { "epoch": 0.4064898474691538, "grad_norm": 0.2033994495868683, "learning_rate": 0.00011903464383028416, "loss": 1.087, "step": 1046 }, { "epoch": 0.4068784610900612, "grad_norm": 0.2047330141067505, "learning_rate": 0.00011895679252627484, "loss": 1.0774, "step": 1047 }, { "epoch": 0.4072670747109686, "grad_norm": 0.21420112252235413, "learning_rate": 0.00011887894122226549, "loss": 1.0252, "step": 1048 }, { "epoch": 0.40765568833187604, "grad_norm": 0.2030097395181656, "learning_rate": 0.00011880108991825614, "loss": 1.0501, "step": 1049 }, { "epoch": 0.40804430195278346, "grad_norm": 0.2128026783466339, "learning_rate": 0.00011872323861424679, "loss": 1.1031, "step": 1050 }, { "epoch": 0.4084329155736909, "grad_norm": 0.20724938809871674, "learning_rate": 0.00011864538731023744, "loss": 1.0327, "step": 1051 }, { "epoch": 0.40882152919459824, "grad_norm": 0.20344072580337524, "learning_rate": 0.00011856753600622812, "loss": 1.0719, "step": 1052 }, { "epoch": 0.40921014281550566, "grad_norm": 0.2145012468099594, "learning_rate": 0.00011848968470221877, "loss": 1.0582, "step": 1053 }, { "epoch": 0.4095987564364131, "grad_norm": 0.220048725605011, "learning_rate": 0.00011841183339820943, "loss": 1.0825, "step": 1054 }, { "epoch": 0.4099873700573205, "grad_norm": 0.19074465334415436, "learning_rate": 0.00011833398209420008, "loss": 0.9657, "step": 1055 }, { "epoch": 0.4103759836782279, "grad_norm": 0.1958267241716385, "learning_rate": 0.00011825613079019073, "loss": 0.9864, "step": 1056 }, { "epoch": 0.41076459729913534, "grad_norm": 0.21768233180046082, "learning_rate": 0.00011817827948618141, "loss": 0.9997, "step": 1057 }, { "epoch": 0.41115321092004276, "grad_norm": 0.20218704640865326, "learning_rate": 0.00011810042818217206, "loss": 1.072, "step": 1058 }, { "epoch": 0.4115418245409502, "grad_norm": 0.2035023719072342, "learning_rate": 0.00011802257687816271, "loss": 1.0415, "step": 1059 }, { "epoch": 0.4119304381618576, "grad_norm": 0.22603970766067505, "learning_rate": 0.00011794472557415337, "loss": 1.0751, "step": 1060 }, { "epoch": 0.412319051782765, "grad_norm": 0.2125842273235321, "learning_rate": 0.00011786687427014402, "loss": 1.0727, "step": 1061 }, { "epoch": 0.4127076654036724, "grad_norm": 0.2005981206893921, "learning_rate": 0.0001177890229661347, "loss": 1.0191, "step": 1062 }, { "epoch": 0.4130962790245798, "grad_norm": 0.22252701222896576, "learning_rate": 0.00011771117166212535, "loss": 1.0591, "step": 1063 }, { "epoch": 0.4134848926454872, "grad_norm": 0.22205251455307007, "learning_rate": 0.000117633320358116, "loss": 1.1198, "step": 1064 }, { "epoch": 0.41387350626639463, "grad_norm": 0.20037783682346344, "learning_rate": 0.00011755546905410665, "loss": 1.0548, "step": 1065 }, { "epoch": 0.41426211988730205, "grad_norm": 0.21737834811210632, "learning_rate": 0.00011747761775009732, "loss": 1.0922, "step": 1066 }, { "epoch": 0.41465073350820947, "grad_norm": 0.19312533736228943, "learning_rate": 0.00011739976644608798, "loss": 0.9836, "step": 1067 }, { "epoch": 0.4150393471291169, "grad_norm": 0.22055000066757202, "learning_rate": 0.00011732191514207864, "loss": 1.0383, "step": 1068 }, { "epoch": 0.4154279607500243, "grad_norm": 0.22623857855796814, "learning_rate": 0.0001172440638380693, "loss": 1.0704, "step": 1069 }, { "epoch": 0.4158165743709317, "grad_norm": 0.21481367945671082, "learning_rate": 0.00011716621253405995, "loss": 1.052, "step": 1070 }, { "epoch": 0.4162051879918391, "grad_norm": 0.21022087335586548, "learning_rate": 0.0001170883612300506, "loss": 1.1021, "step": 1071 }, { "epoch": 0.4165938016127465, "grad_norm": 0.2154620885848999, "learning_rate": 0.00011701050992604126, "loss": 1.0128, "step": 1072 }, { "epoch": 0.4169824152336539, "grad_norm": 0.20545578002929688, "learning_rate": 0.00011693265862203194, "loss": 1.0058, "step": 1073 }, { "epoch": 0.41737102885456134, "grad_norm": 0.21726195514202118, "learning_rate": 0.00011685480731802259, "loss": 1.0753, "step": 1074 }, { "epoch": 0.41775964247546876, "grad_norm": 0.2067115604877472, "learning_rate": 0.00011677695601401324, "loss": 1.0594, "step": 1075 }, { "epoch": 0.4181482560963762, "grad_norm": 0.23024648427963257, "learning_rate": 0.0001166991047100039, "loss": 1.1039, "step": 1076 }, { "epoch": 0.4185368697172836, "grad_norm": 0.20692144334316254, "learning_rate": 0.00011662125340599455, "loss": 1.0598, "step": 1077 }, { "epoch": 0.418925483338191, "grad_norm": 0.19839999079704285, "learning_rate": 0.00011654340210198522, "loss": 1.054, "step": 1078 }, { "epoch": 0.41931409695909844, "grad_norm": 0.19227825105190277, "learning_rate": 0.00011646555079797588, "loss": 0.9453, "step": 1079 }, { "epoch": 0.41970271058000586, "grad_norm": 0.2112567275762558, "learning_rate": 0.00011638769949396653, "loss": 1.023, "step": 1080 }, { "epoch": 0.4200913242009132, "grad_norm": 0.185299351811409, "learning_rate": 0.00011630984818995718, "loss": 0.9752, "step": 1081 }, { "epoch": 0.42047993782182064, "grad_norm": 0.20148858428001404, "learning_rate": 0.00011623199688594783, "loss": 1.0659, "step": 1082 }, { "epoch": 0.42086855144272806, "grad_norm": 0.1935974359512329, "learning_rate": 0.00011615414558193851, "loss": 1.0116, "step": 1083 }, { "epoch": 0.4212571650636355, "grad_norm": 0.20433953404426575, "learning_rate": 0.00011607629427792916, "loss": 1.0671, "step": 1084 }, { "epoch": 0.4216457786845429, "grad_norm": 0.20729799568653107, "learning_rate": 0.00011599844297391982, "loss": 1.0341, "step": 1085 }, { "epoch": 0.4220343923054503, "grad_norm": 0.2126002460718155, "learning_rate": 0.00011592059166991047, "loss": 1.0188, "step": 1086 }, { "epoch": 0.42242300592635773, "grad_norm": 0.19453707337379456, "learning_rate": 0.00011584274036590112, "loss": 1.0331, "step": 1087 }, { "epoch": 0.42281161954726515, "grad_norm": 0.20909856259822845, "learning_rate": 0.0001157648890618918, "loss": 0.9984, "step": 1088 }, { "epoch": 0.42320023316817257, "grad_norm": 0.19596272706985474, "learning_rate": 0.00011568703775788245, "loss": 1.0121, "step": 1089 }, { "epoch": 0.42358884678907993, "grad_norm": 0.22045716643333435, "learning_rate": 0.0001156091864538731, "loss": 1.0591, "step": 1090 }, { "epoch": 0.42397746040998735, "grad_norm": 0.22624897956848145, "learning_rate": 0.00011553133514986376, "loss": 1.0565, "step": 1091 }, { "epoch": 0.42436607403089477, "grad_norm": 0.20263417065143585, "learning_rate": 0.00011545348384585442, "loss": 1.024, "step": 1092 }, { "epoch": 0.4247546876518022, "grad_norm": 0.20179417729377747, "learning_rate": 0.00011537563254184509, "loss": 0.9806, "step": 1093 }, { "epoch": 0.4251433012727096, "grad_norm": 0.30221593379974365, "learning_rate": 0.00011529778123783574, "loss": 1.0683, "step": 1094 }, { "epoch": 0.425531914893617, "grad_norm": 0.21195146441459656, "learning_rate": 0.0001152199299338264, "loss": 1.1283, "step": 1095 }, { "epoch": 0.42592052851452444, "grad_norm": 0.21860192716121674, "learning_rate": 0.00011514207862981706, "loss": 1.0046, "step": 1096 }, { "epoch": 0.42630914213543186, "grad_norm": 0.2234150469303131, "learning_rate": 0.00011506422732580771, "loss": 1.0461, "step": 1097 }, { "epoch": 0.4266977557563393, "grad_norm": 0.21535125374794006, "learning_rate": 0.00011498637602179837, "loss": 1.0593, "step": 1098 }, { "epoch": 0.4270863693772467, "grad_norm": 0.19313789904117584, "learning_rate": 0.00011490852471778904, "loss": 1.0357, "step": 1099 }, { "epoch": 0.42747498299815406, "grad_norm": 0.19886989891529083, "learning_rate": 0.00011483067341377969, "loss": 0.9946, "step": 1100 }, { "epoch": 0.4278635966190615, "grad_norm": 0.21028490364551544, "learning_rate": 0.00011475282210977034, "loss": 1.0765, "step": 1101 }, { "epoch": 0.4282522102399689, "grad_norm": 0.2066621333360672, "learning_rate": 0.000114674970805761, "loss": 1.0405, "step": 1102 }, { "epoch": 0.4286408238608763, "grad_norm": 0.18400220572948456, "learning_rate": 0.00011459711950175168, "loss": 0.9404, "step": 1103 }, { "epoch": 0.42902943748178374, "grad_norm": 0.2058599591255188, "learning_rate": 0.00011451926819774233, "loss": 1.0505, "step": 1104 }, { "epoch": 0.42941805110269116, "grad_norm": 0.19696786999702454, "learning_rate": 0.00011444141689373298, "loss": 1.032, "step": 1105 }, { "epoch": 0.4298066647235986, "grad_norm": 0.2082854062318802, "learning_rate": 0.00011436356558972363, "loss": 1.0914, "step": 1106 }, { "epoch": 0.430195278344506, "grad_norm": 0.20155015587806702, "learning_rate": 0.00011428571428571428, "loss": 1.0541, "step": 1107 }, { "epoch": 0.4305838919654134, "grad_norm": 0.23419982194900513, "learning_rate": 0.00011420786298170494, "loss": 1.0684, "step": 1108 }, { "epoch": 0.4309725055863208, "grad_norm": 0.23493975400924683, "learning_rate": 0.00011413001167769561, "loss": 1.0509, "step": 1109 }, { "epoch": 0.4313611192072282, "grad_norm": 0.2089843600988388, "learning_rate": 0.00011405216037368627, "loss": 1.0479, "step": 1110 }, { "epoch": 0.4317497328281356, "grad_norm": 0.21076850593090057, "learning_rate": 0.00011397430906967692, "loss": 1.064, "step": 1111 }, { "epoch": 0.43213834644904303, "grad_norm": 0.20307987928390503, "learning_rate": 0.00011389645776566757, "loss": 1.0416, "step": 1112 }, { "epoch": 0.43252696006995045, "grad_norm": 0.20955562591552734, "learning_rate": 0.00011381860646165822, "loss": 1.0158, "step": 1113 }, { "epoch": 0.43291557369085787, "grad_norm": 0.2074531465768814, "learning_rate": 0.0001137407551576489, "loss": 1.0486, "step": 1114 }, { "epoch": 0.4333041873117653, "grad_norm": 0.20907235145568848, "learning_rate": 0.00011366290385363955, "loss": 1.0352, "step": 1115 }, { "epoch": 0.4336928009326727, "grad_norm": 0.21726477146148682, "learning_rate": 0.0001135850525496302, "loss": 1.0068, "step": 1116 }, { "epoch": 0.4340814145535801, "grad_norm": 0.20231984555721283, "learning_rate": 0.00011350720124562086, "loss": 0.9757, "step": 1117 }, { "epoch": 0.4344700281744875, "grad_norm": 0.23485834896564484, "learning_rate": 0.00011342934994161152, "loss": 1.0681, "step": 1118 }, { "epoch": 0.4348586417953949, "grad_norm": 0.21286556124687195, "learning_rate": 0.00011335149863760219, "loss": 1.0399, "step": 1119 }, { "epoch": 0.4352472554163023, "grad_norm": 0.2097872495651245, "learning_rate": 0.00011327364733359284, "loss": 1.0435, "step": 1120 }, { "epoch": 0.43563586903720974, "grad_norm": 0.2224377542734146, "learning_rate": 0.00011319579602958351, "loss": 1.1664, "step": 1121 }, { "epoch": 0.43602448265811716, "grad_norm": 0.19213411211967468, "learning_rate": 0.00011311794472557416, "loss": 1.0424, "step": 1122 }, { "epoch": 0.4364130962790246, "grad_norm": 0.20974959433078766, "learning_rate": 0.00011304009342156481, "loss": 1.0943, "step": 1123 }, { "epoch": 0.436801709899932, "grad_norm": 0.19943708181381226, "learning_rate": 0.00011296224211755549, "loss": 1.0652, "step": 1124 }, { "epoch": 0.4371903235208394, "grad_norm": 0.1832750141620636, "learning_rate": 0.00011288439081354614, "loss": 0.9883, "step": 1125 }, { "epoch": 0.43757893714174684, "grad_norm": 0.2205052226781845, "learning_rate": 0.0001128065395095368, "loss": 1.0733, "step": 1126 }, { "epoch": 0.43796755076265426, "grad_norm": 0.2082854062318802, "learning_rate": 0.00011272868820552745, "loss": 1.0141, "step": 1127 }, { "epoch": 0.4383561643835616, "grad_norm": 0.22755026817321777, "learning_rate": 0.0001126508369015181, "loss": 1.0942, "step": 1128 }, { "epoch": 0.43874477800446904, "grad_norm": 0.2098863571882248, "learning_rate": 0.00011257298559750878, "loss": 0.9987, "step": 1129 }, { "epoch": 0.43913339162537646, "grad_norm": 0.20559263229370117, "learning_rate": 0.00011249513429349943, "loss": 1.0345, "step": 1130 }, { "epoch": 0.4395220052462839, "grad_norm": 0.21955084800720215, "learning_rate": 0.00011241728298949008, "loss": 1.1068, "step": 1131 }, { "epoch": 0.4399106188671913, "grad_norm": 0.21353478729724884, "learning_rate": 0.00011233943168548073, "loss": 1.0094, "step": 1132 }, { "epoch": 0.4402992324880987, "grad_norm": 0.19822491705417633, "learning_rate": 0.00011226158038147139, "loss": 0.9758, "step": 1133 }, { "epoch": 0.44068784610900613, "grad_norm": 0.20079441368579865, "learning_rate": 0.00011218372907746206, "loss": 1.0202, "step": 1134 }, { "epoch": 0.44107645972991355, "grad_norm": 0.2261926829814911, "learning_rate": 0.00011210587777345272, "loss": 0.9877, "step": 1135 }, { "epoch": 0.44146507335082097, "grad_norm": 0.2264915257692337, "learning_rate": 0.00011202802646944337, "loss": 0.9887, "step": 1136 }, { "epoch": 0.44185368697172833, "grad_norm": 0.21853779256343842, "learning_rate": 0.00011195017516543402, "loss": 1.0535, "step": 1137 }, { "epoch": 0.44224230059263575, "grad_norm": 0.21332694590091705, "learning_rate": 0.00011187232386142467, "loss": 1.0824, "step": 1138 }, { "epoch": 0.44263091421354317, "grad_norm": 0.21350236237049103, "learning_rate": 0.00011179447255741535, "loss": 1.0758, "step": 1139 }, { "epoch": 0.4430195278344506, "grad_norm": 0.21305765211582184, "learning_rate": 0.000111716621253406, "loss": 1.035, "step": 1140 }, { "epoch": 0.443408141455358, "grad_norm": 0.20486389100551605, "learning_rate": 0.00011163876994939666, "loss": 1.0413, "step": 1141 }, { "epoch": 0.4437967550762654, "grad_norm": 0.19255472719669342, "learning_rate": 0.00011156091864538731, "loss": 0.9583, "step": 1142 }, { "epoch": 0.44418536869717284, "grad_norm": 0.19824008643627167, "learning_rate": 0.00011148306734137796, "loss": 1.0331, "step": 1143 }, { "epoch": 0.44457398231808026, "grad_norm": 0.20308080315589905, "learning_rate": 0.00011140521603736863, "loss": 1.0399, "step": 1144 }, { "epoch": 0.4449625959389877, "grad_norm": 0.2193964123725891, "learning_rate": 0.00011132736473335929, "loss": 1.063, "step": 1145 }, { "epoch": 0.4453512095598951, "grad_norm": 0.2151576578617096, "learning_rate": 0.00011124951342934994, "loss": 1.0795, "step": 1146 }, { "epoch": 0.44573982318080246, "grad_norm": 0.23056697845458984, "learning_rate": 0.00011117166212534061, "loss": 1.0351, "step": 1147 }, { "epoch": 0.4461284368017099, "grad_norm": 0.1973094493150711, "learning_rate": 0.00011109381082133126, "loss": 0.9866, "step": 1148 }, { "epoch": 0.4465170504226173, "grad_norm": 0.2119562178850174, "learning_rate": 0.00011101595951732191, "loss": 1.0591, "step": 1149 }, { "epoch": 0.4469056640435247, "grad_norm": 0.20407763123512268, "learning_rate": 0.00011093810821331259, "loss": 0.988, "step": 1150 }, { "epoch": 0.44729427766443214, "grad_norm": 0.19474107027053833, "learning_rate": 0.00011086025690930324, "loss": 0.9729, "step": 1151 }, { "epoch": 0.44768289128533956, "grad_norm": 0.2179928421974182, "learning_rate": 0.0001107824056052939, "loss": 1.0558, "step": 1152 }, { "epoch": 0.448071504906247, "grad_norm": 0.44306451082229614, "learning_rate": 0.00011070455430128455, "loss": 1.0901, "step": 1153 }, { "epoch": 0.4484601185271544, "grad_norm": 0.22060540318489075, "learning_rate": 0.0001106267029972752, "loss": 1.0009, "step": 1154 }, { "epoch": 0.4488487321480618, "grad_norm": 0.20534972846508026, "learning_rate": 0.00011054885169326588, "loss": 0.9741, "step": 1155 }, { "epoch": 0.4492373457689692, "grad_norm": 0.19488993287086487, "learning_rate": 0.00011047100038925653, "loss": 1.0, "step": 1156 }, { "epoch": 0.4496259593898766, "grad_norm": 0.20462395250797272, "learning_rate": 0.00011039314908524718, "loss": 1.0309, "step": 1157 }, { "epoch": 0.450014573010784, "grad_norm": 0.2170749306678772, "learning_rate": 0.00011031529778123784, "loss": 1.0726, "step": 1158 }, { "epoch": 0.45040318663169143, "grad_norm": 0.2066730111837387, "learning_rate": 0.00011023744647722849, "loss": 1.0227, "step": 1159 }, { "epoch": 0.45079180025259885, "grad_norm": 0.20625676214694977, "learning_rate": 0.00011015959517321917, "loss": 1.0287, "step": 1160 }, { "epoch": 0.45118041387350627, "grad_norm": 0.19483047723770142, "learning_rate": 0.00011008174386920982, "loss": 0.9639, "step": 1161 }, { "epoch": 0.4515690274944137, "grad_norm": 0.24705417454242706, "learning_rate": 0.00011000389256520047, "loss": 0.9903, "step": 1162 }, { "epoch": 0.4519576411153211, "grad_norm": 0.2109205424785614, "learning_rate": 0.00010992604126119112, "loss": 1.054, "step": 1163 }, { "epoch": 0.4523462547362285, "grad_norm": 0.20904991030693054, "learning_rate": 0.00010984818995718178, "loss": 1.0416, "step": 1164 }, { "epoch": 0.45273486835713594, "grad_norm": 0.19841328263282776, "learning_rate": 0.00010977033865317245, "loss": 0.9986, "step": 1165 }, { "epoch": 0.4531234819780433, "grad_norm": 0.20545506477355957, "learning_rate": 0.0001096924873491631, "loss": 1.0337, "step": 1166 }, { "epoch": 0.4535120955989507, "grad_norm": 0.208644837141037, "learning_rate": 0.00010961463604515376, "loss": 1.0304, "step": 1167 }, { "epoch": 0.45390070921985815, "grad_norm": 0.2111911028623581, "learning_rate": 0.00010953678474114441, "loss": 1.0398, "step": 1168 }, { "epoch": 0.45428932284076556, "grad_norm": 0.2600184381008148, "learning_rate": 0.00010945893343713506, "loss": 1.0509, "step": 1169 }, { "epoch": 0.454677936461673, "grad_norm": 0.2059030532836914, "learning_rate": 0.00010938108213312574, "loss": 0.9347, "step": 1170 }, { "epoch": 0.4550665500825804, "grad_norm": 0.19232551753520966, "learning_rate": 0.0001093032308291164, "loss": 1.0162, "step": 1171 }, { "epoch": 0.4554551637034878, "grad_norm": 0.19147330522537231, "learning_rate": 0.00010922537952510705, "loss": 0.9872, "step": 1172 }, { "epoch": 0.45584377732439524, "grad_norm": 0.2599676251411438, "learning_rate": 0.00010914752822109771, "loss": 1.0402, "step": 1173 }, { "epoch": 0.45623239094530266, "grad_norm": 0.2159397304058075, "learning_rate": 0.00010906967691708836, "loss": 1.0411, "step": 1174 }, { "epoch": 0.45662100456621, "grad_norm": 0.23864266276359558, "learning_rate": 0.00010899182561307903, "loss": 1.054, "step": 1175 }, { "epoch": 0.45700961818711744, "grad_norm": 0.2027217596769333, "learning_rate": 0.0001089139743090697, "loss": 0.9713, "step": 1176 }, { "epoch": 0.45739823180802486, "grad_norm": 0.1837588995695114, "learning_rate": 0.00010883612300506035, "loss": 0.9698, "step": 1177 }, { "epoch": 0.4577868454289323, "grad_norm": 0.20038527250289917, "learning_rate": 0.000108758271701051, "loss": 1.0456, "step": 1178 }, { "epoch": 0.4581754590498397, "grad_norm": 0.21525044739246368, "learning_rate": 0.00010868042039704165, "loss": 1.021, "step": 1179 }, { "epoch": 0.4585640726707471, "grad_norm": 0.18813730776309967, "learning_rate": 0.0001086025690930323, "loss": 0.9673, "step": 1180 }, { "epoch": 0.45895268629165453, "grad_norm": 0.2056179642677307, "learning_rate": 0.00010852471778902298, "loss": 1.0119, "step": 1181 }, { "epoch": 0.45934129991256195, "grad_norm": 0.21599683165550232, "learning_rate": 0.00010844686648501363, "loss": 1.0537, "step": 1182 }, { "epoch": 0.45972991353346937, "grad_norm": 0.19750265777111053, "learning_rate": 0.00010836901518100429, "loss": 1.0203, "step": 1183 }, { "epoch": 0.4601185271543768, "grad_norm": 0.22186161577701569, "learning_rate": 0.00010829116387699494, "loss": 1.0583, "step": 1184 }, { "epoch": 0.46050714077528415, "grad_norm": 0.2109905481338501, "learning_rate": 0.00010821331257298559, "loss": 1.0022, "step": 1185 }, { "epoch": 0.46089575439619157, "grad_norm": 0.2032858431339264, "learning_rate": 0.00010813546126897627, "loss": 0.9774, "step": 1186 }, { "epoch": 0.461284368017099, "grad_norm": 0.20381197333335876, "learning_rate": 0.00010805760996496692, "loss": 0.9768, "step": 1187 }, { "epoch": 0.4616729816380064, "grad_norm": 0.20488987863063812, "learning_rate": 0.00010797975866095757, "loss": 1.0448, "step": 1188 }, { "epoch": 0.4620615952589138, "grad_norm": 0.20257477462291718, "learning_rate": 0.00010790190735694823, "loss": 1.0157, "step": 1189 }, { "epoch": 0.46245020887982125, "grad_norm": 0.20761239528656006, "learning_rate": 0.00010782405605293888, "loss": 1.0328, "step": 1190 }, { "epoch": 0.46283882250072866, "grad_norm": 0.22062581777572632, "learning_rate": 0.00010774620474892956, "loss": 1.0362, "step": 1191 }, { "epoch": 0.4632274361216361, "grad_norm": 0.19970272481441498, "learning_rate": 0.00010766835344492021, "loss": 1.0783, "step": 1192 }, { "epoch": 0.4636160497425435, "grad_norm": 0.2221893072128296, "learning_rate": 0.00010759050214091086, "loss": 1.0136, "step": 1193 }, { "epoch": 0.46400466336345086, "grad_norm": 0.2124665081501007, "learning_rate": 0.00010751265083690151, "loss": 1.0528, "step": 1194 }, { "epoch": 0.4643932769843583, "grad_norm": 0.2001204937696457, "learning_rate": 0.00010743479953289218, "loss": 1.0495, "step": 1195 }, { "epoch": 0.4647818906052657, "grad_norm": 0.20979635417461395, "learning_rate": 0.00010735694822888284, "loss": 1.0664, "step": 1196 }, { "epoch": 0.4651705042261731, "grad_norm": 0.190982848405838, "learning_rate": 0.0001072790969248735, "loss": 1.0256, "step": 1197 }, { "epoch": 0.46555911784708054, "grad_norm": 0.19910745322704315, "learning_rate": 0.00010720124562086415, "loss": 1.0263, "step": 1198 }, { "epoch": 0.46594773146798796, "grad_norm": 0.21624085307121277, "learning_rate": 0.00010712339431685481, "loss": 1.0768, "step": 1199 }, { "epoch": 0.4663363450888954, "grad_norm": 0.20857703685760498, "learning_rate": 0.00010704554301284547, "loss": 1.0892, "step": 1200 }, { "epoch": 0.4667249587098028, "grad_norm": 0.21897061169147491, "learning_rate": 0.00010696769170883613, "loss": 1.0873, "step": 1201 }, { "epoch": 0.4671135723307102, "grad_norm": 0.1943386346101761, "learning_rate": 0.0001068898404048268, "loss": 1.0116, "step": 1202 }, { "epoch": 0.4675021859516176, "grad_norm": 0.22607874870300293, "learning_rate": 0.00010681198910081745, "loss": 1.0328, "step": 1203 }, { "epoch": 0.467890799572525, "grad_norm": 0.1898999959230423, "learning_rate": 0.0001067341377968081, "loss": 0.9791, "step": 1204 }, { "epoch": 0.4682794131934324, "grad_norm": 0.2193334400653839, "learning_rate": 0.00010665628649279875, "loss": 1.0742, "step": 1205 }, { "epoch": 0.46866802681433983, "grad_norm": 0.2096349149942398, "learning_rate": 0.00010657843518878943, "loss": 1.0683, "step": 1206 }, { "epoch": 0.46905664043524725, "grad_norm": 0.2040576934814453, "learning_rate": 0.00010650058388478008, "loss": 1.0516, "step": 1207 }, { "epoch": 0.46944525405615467, "grad_norm": 0.20619645714759827, "learning_rate": 0.00010642273258077074, "loss": 1.0429, "step": 1208 }, { "epoch": 0.4698338676770621, "grad_norm": 0.19753660261631012, "learning_rate": 0.00010634488127676139, "loss": 1.0268, "step": 1209 }, { "epoch": 0.4702224812979695, "grad_norm": 0.2201426476240158, "learning_rate": 0.00010626702997275204, "loss": 1.0879, "step": 1210 }, { "epoch": 0.4706110949188769, "grad_norm": 0.21307805180549622, "learning_rate": 0.00010618917866874272, "loss": 1.0186, "step": 1211 }, { "epoch": 0.47099970853978435, "grad_norm": 0.21142373979091644, "learning_rate": 0.00010611132736473337, "loss": 1.0417, "step": 1212 }, { "epoch": 0.4713883221606917, "grad_norm": 0.20523706078529358, "learning_rate": 0.00010603347606072402, "loss": 1.0372, "step": 1213 }, { "epoch": 0.4717769357815991, "grad_norm": 0.19843094050884247, "learning_rate": 0.00010595562475671468, "loss": 1.0062, "step": 1214 }, { "epoch": 0.47216554940250655, "grad_norm": 0.2146739959716797, "learning_rate": 0.00010587777345270533, "loss": 1.0528, "step": 1215 }, { "epoch": 0.47255416302341396, "grad_norm": 0.2136303037405014, "learning_rate": 0.00010579992214869601, "loss": 1.0521, "step": 1216 }, { "epoch": 0.4729427766443214, "grad_norm": 0.21379397809505463, "learning_rate": 0.00010572207084468666, "loss": 1.0362, "step": 1217 }, { "epoch": 0.4733313902652288, "grad_norm": 0.20459088683128357, "learning_rate": 0.00010564421954067731, "loss": 1.0455, "step": 1218 }, { "epoch": 0.4737200038861362, "grad_norm": 0.20667988061904907, "learning_rate": 0.00010556636823666796, "loss": 1.0284, "step": 1219 }, { "epoch": 0.47410861750704364, "grad_norm": 0.21820449829101562, "learning_rate": 0.00010548851693265862, "loss": 1.0584, "step": 1220 }, { "epoch": 0.47449723112795106, "grad_norm": 0.19705156981945038, "learning_rate": 0.00010541066562864928, "loss": 1.004, "step": 1221 }, { "epoch": 0.4748858447488584, "grad_norm": 0.19806528091430664, "learning_rate": 0.00010533281432463995, "loss": 1.0519, "step": 1222 }, { "epoch": 0.47527445836976584, "grad_norm": 0.2006833702325821, "learning_rate": 0.0001052549630206306, "loss": 1.0119, "step": 1223 }, { "epoch": 0.47566307199067326, "grad_norm": 0.21757058799266815, "learning_rate": 0.00010517711171662125, "loss": 1.0961, "step": 1224 }, { "epoch": 0.4760516856115807, "grad_norm": 0.2015775889158249, "learning_rate": 0.00010509926041261192, "loss": 1.0419, "step": 1225 }, { "epoch": 0.4764402992324881, "grad_norm": 0.19691923260688782, "learning_rate": 0.00010502140910860257, "loss": 1.0555, "step": 1226 }, { "epoch": 0.4768289128533955, "grad_norm": 0.19924800097942352, "learning_rate": 0.00010494355780459323, "loss": 1.0106, "step": 1227 }, { "epoch": 0.47721752647430293, "grad_norm": 0.21416346728801727, "learning_rate": 0.0001048657065005839, "loss": 1.0741, "step": 1228 }, { "epoch": 0.47760614009521035, "grad_norm": 0.21823547780513763, "learning_rate": 0.00010478785519657455, "loss": 1.023, "step": 1229 }, { "epoch": 0.47799475371611777, "grad_norm": 0.2083735466003418, "learning_rate": 0.0001047100038925652, "loss": 1.0424, "step": 1230 }, { "epoch": 0.4783833673370252, "grad_norm": 0.2219141572713852, "learning_rate": 0.00010463215258855586, "loss": 1.0839, "step": 1231 }, { "epoch": 0.47877198095793255, "grad_norm": 0.21334600448608398, "learning_rate": 0.00010455430128454653, "loss": 0.9888, "step": 1232 }, { "epoch": 0.47916059457883997, "grad_norm": 0.2140086442232132, "learning_rate": 0.00010447644998053719, "loss": 1.0119, "step": 1233 }, { "epoch": 0.4795492081997474, "grad_norm": 0.25360551476478577, "learning_rate": 0.00010439859867652784, "loss": 1.0026, "step": 1234 }, { "epoch": 0.4799378218206548, "grad_norm": 0.20200380682945251, "learning_rate": 0.00010432074737251849, "loss": 1.0, "step": 1235 }, { "epoch": 0.4803264354415622, "grad_norm": 0.22641289234161377, "learning_rate": 0.00010424289606850914, "loss": 1.1022, "step": 1236 }, { "epoch": 0.48071504906246965, "grad_norm": 0.20538561046123505, "learning_rate": 0.00010416504476449982, "loss": 0.9847, "step": 1237 }, { "epoch": 0.48110366268337706, "grad_norm": 0.206883504986763, "learning_rate": 0.00010408719346049047, "loss": 1.0152, "step": 1238 }, { "epoch": 0.4814922763042845, "grad_norm": 0.21584320068359375, "learning_rate": 0.00010400934215648113, "loss": 1.0361, "step": 1239 }, { "epoch": 0.4818808899251919, "grad_norm": 0.20963703095912933, "learning_rate": 0.00010393149085247178, "loss": 1.0814, "step": 1240 }, { "epoch": 0.48226950354609927, "grad_norm": 0.1965872198343277, "learning_rate": 0.00010385363954846243, "loss": 1.0365, "step": 1241 }, { "epoch": 0.4826581171670067, "grad_norm": 0.2030191719532013, "learning_rate": 0.00010377578824445311, "loss": 1.0374, "step": 1242 }, { "epoch": 0.4830467307879141, "grad_norm": 0.21448804438114166, "learning_rate": 0.00010369793694044376, "loss": 0.9686, "step": 1243 }, { "epoch": 0.4834353444088215, "grad_norm": 0.2181752622127533, "learning_rate": 0.00010362008563643441, "loss": 1.0812, "step": 1244 }, { "epoch": 0.48382395802972894, "grad_norm": 0.19887101650238037, "learning_rate": 0.00010354223433242507, "loss": 1.036, "step": 1245 }, { "epoch": 0.48421257165063636, "grad_norm": 0.19007287919521332, "learning_rate": 0.00010346438302841572, "loss": 1.0292, "step": 1246 }, { "epoch": 0.4846011852715438, "grad_norm": 0.21390347182750702, "learning_rate": 0.0001033865317244064, "loss": 1.0284, "step": 1247 }, { "epoch": 0.4849897988924512, "grad_norm": 0.23822663724422455, "learning_rate": 0.00010330868042039705, "loss": 1.1044, "step": 1248 }, { "epoch": 0.4853784125133586, "grad_norm": 0.20779070258140564, "learning_rate": 0.0001032308291163877, "loss": 1.0475, "step": 1249 }, { "epoch": 0.48576702613426603, "grad_norm": 0.19232134521007538, "learning_rate": 0.00010315297781237835, "loss": 0.9945, "step": 1250 }, { "epoch": 0.4861556397551734, "grad_norm": 0.22378556430339813, "learning_rate": 0.00010307512650836902, "loss": 1.0462, "step": 1251 }, { "epoch": 0.4865442533760808, "grad_norm": 0.22156798839569092, "learning_rate": 0.00010299727520435968, "loss": 1.051, "step": 1252 }, { "epoch": 0.48693286699698823, "grad_norm": 0.19885733723640442, "learning_rate": 0.00010291942390035034, "loss": 1.0593, "step": 1253 }, { "epoch": 0.48732148061789565, "grad_norm": 0.2172418236732483, "learning_rate": 0.000102841572596341, "loss": 1.0513, "step": 1254 }, { "epoch": 0.48771009423880307, "grad_norm": 0.22136956453323364, "learning_rate": 0.00010276372129233165, "loss": 1.0438, "step": 1255 }, { "epoch": 0.4880987078597105, "grad_norm": 0.21337302029132843, "learning_rate": 0.0001026858699883223, "loss": 1.0551, "step": 1256 }, { "epoch": 0.4884873214806179, "grad_norm": 0.21376267075538635, "learning_rate": 0.00010260801868431296, "loss": 1.054, "step": 1257 }, { "epoch": 0.4888759351015253, "grad_norm": 0.19498860836029053, "learning_rate": 0.00010253016738030364, "loss": 1.0045, "step": 1258 }, { "epoch": 0.48926454872243275, "grad_norm": 0.22354961931705475, "learning_rate": 0.00010245231607629429, "loss": 1.096, "step": 1259 }, { "epoch": 0.4896531623433401, "grad_norm": 0.2078939527273178, "learning_rate": 0.00010237446477228494, "loss": 1.0102, "step": 1260 }, { "epoch": 0.49004177596424753, "grad_norm": 0.20992495119571686, "learning_rate": 0.00010229661346827559, "loss": 0.9814, "step": 1261 }, { "epoch": 0.49043038958515495, "grad_norm": 0.2178875207901001, "learning_rate": 0.00010221876216426625, "loss": 1.0489, "step": 1262 }, { "epoch": 0.49081900320606237, "grad_norm": 0.22152946889400482, "learning_rate": 0.00010214091086025692, "loss": 1.0808, "step": 1263 }, { "epoch": 0.4912076168269698, "grad_norm": 0.21179009974002838, "learning_rate": 0.00010206305955624758, "loss": 1.0323, "step": 1264 }, { "epoch": 0.4915962304478772, "grad_norm": 0.2126997411251068, "learning_rate": 0.00010198520825223823, "loss": 1.0093, "step": 1265 }, { "epoch": 0.4919848440687846, "grad_norm": 0.20912809669971466, "learning_rate": 0.00010190735694822888, "loss": 1.0343, "step": 1266 }, { "epoch": 0.49237345768969204, "grad_norm": 0.2231636494398117, "learning_rate": 0.00010182950564421953, "loss": 1.0587, "step": 1267 }, { "epoch": 0.49276207131059946, "grad_norm": 0.1954583376646042, "learning_rate": 0.00010175165434021021, "loss": 0.9566, "step": 1268 }, { "epoch": 0.4931506849315068, "grad_norm": 0.20520909130573273, "learning_rate": 0.00010167380303620086, "loss": 1.024, "step": 1269 }, { "epoch": 0.49353929855241424, "grad_norm": 0.21736180782318115, "learning_rate": 0.00010159595173219152, "loss": 1.0434, "step": 1270 }, { "epoch": 0.49392791217332166, "grad_norm": 0.2360561490058899, "learning_rate": 0.00010151810042818217, "loss": 1.114, "step": 1271 }, { "epoch": 0.4943165257942291, "grad_norm": 0.20595967769622803, "learning_rate": 0.00010144024912417282, "loss": 0.9909, "step": 1272 }, { "epoch": 0.4947051394151365, "grad_norm": 0.2161860466003418, "learning_rate": 0.0001013623978201635, "loss": 1.0536, "step": 1273 }, { "epoch": 0.4950937530360439, "grad_norm": 0.19852355122566223, "learning_rate": 0.00010128454651615415, "loss": 1.0001, "step": 1274 }, { "epoch": 0.49548236665695133, "grad_norm": 0.21081402897834778, "learning_rate": 0.0001012066952121448, "loss": 1.0151, "step": 1275 }, { "epoch": 0.49587098027785875, "grad_norm": 0.2053362876176834, "learning_rate": 0.00010112884390813547, "loss": 1.018, "step": 1276 }, { "epoch": 0.49625959389876617, "grad_norm": 0.21205593645572662, "learning_rate": 0.00010105099260412612, "loss": 0.9912, "step": 1277 }, { "epoch": 0.4966482075196736, "grad_norm": 0.2005016952753067, "learning_rate": 0.00010097314130011679, "loss": 1.0069, "step": 1278 }, { "epoch": 0.49703682114058095, "grad_norm": 0.21688181161880493, "learning_rate": 0.00010089528999610744, "loss": 1.0364, "step": 1279 }, { "epoch": 0.49742543476148837, "grad_norm": 0.20582237839698792, "learning_rate": 0.0001008174386920981, "loss": 1.0138, "step": 1280 }, { "epoch": 0.4978140483823958, "grad_norm": 0.20824448764324188, "learning_rate": 0.00010073958738808876, "loss": 0.9941, "step": 1281 }, { "epoch": 0.4982026620033032, "grad_norm": 0.20749075710773468, "learning_rate": 0.00010066173608407941, "loss": 1.0478, "step": 1282 }, { "epoch": 0.49859127562421063, "grad_norm": 0.20012183487415314, "learning_rate": 0.00010058388478007009, "loss": 0.995, "step": 1283 }, { "epoch": 0.49897988924511805, "grad_norm": 0.20275959372520447, "learning_rate": 0.00010050603347606074, "loss": 1.097, "step": 1284 }, { "epoch": 0.49936850286602547, "grad_norm": 0.19588243961334229, "learning_rate": 0.00010042818217205139, "loss": 1.0, "step": 1285 }, { "epoch": 0.4997571164869329, "grad_norm": 0.20693185925483704, "learning_rate": 0.00010035033086804204, "loss": 1.0527, "step": 1286 }, { "epoch": 0.5001457301078402, "grad_norm": 0.20330573618412018, "learning_rate": 0.0001002724795640327, "loss": 1.0137, "step": 1287 }, { "epoch": 0.5005343437287477, "grad_norm": 0.19123876094818115, "learning_rate": 0.00010019462826002337, "loss": 0.9688, "step": 1288 }, { "epoch": 0.5009229573496551, "grad_norm": 0.2184276431798935, "learning_rate": 0.00010011677695601403, "loss": 1.0367, "step": 1289 }, { "epoch": 0.5013115709705626, "grad_norm": 0.21642108261585236, "learning_rate": 0.00010003892565200468, "loss": 1.102, "step": 1290 }, { "epoch": 0.5017001845914699, "grad_norm": 0.20351074635982513, "learning_rate": 9.996107434799533e-05, "loss": 1.0327, "step": 1291 }, { "epoch": 0.5020887982123774, "grad_norm": 0.22771553695201874, "learning_rate": 9.9883223043986e-05, "loss": 1.104, "step": 1292 }, { "epoch": 0.5024774118332848, "grad_norm": 0.2271403968334198, "learning_rate": 9.980537173997665e-05, "loss": 1.1313, "step": 1293 }, { "epoch": 0.5028660254541921, "grad_norm": 0.2157830148935318, "learning_rate": 9.97275204359673e-05, "loss": 1.0203, "step": 1294 }, { "epoch": 0.5032546390750996, "grad_norm": 0.19555307924747467, "learning_rate": 9.964966913195797e-05, "loss": 1.0194, "step": 1295 }, { "epoch": 0.503643252696007, "grad_norm": 0.1898549199104309, "learning_rate": 9.957181782794862e-05, "loss": 1.0034, "step": 1296 }, { "epoch": 0.5040318663169144, "grad_norm": 0.23555906116962433, "learning_rate": 9.949396652393928e-05, "loss": 1.0298, "step": 1297 }, { "epoch": 0.5044204799378218, "grad_norm": 0.20434850454330444, "learning_rate": 9.941611521992994e-05, "loss": 0.9999, "step": 1298 }, { "epoch": 0.5048090935587293, "grad_norm": 0.21015289425849915, "learning_rate": 9.933826391592059e-05, "loss": 1.006, "step": 1299 }, { "epoch": 0.5051977071796366, "grad_norm": 0.21147851645946503, "learning_rate": 9.926041261191125e-05, "loss": 1.0854, "step": 1300 }, { "epoch": 0.5055863208005441, "grad_norm": 0.19666944444179535, "learning_rate": 9.91825613079019e-05, "loss": 1.0057, "step": 1301 }, { "epoch": 0.5059749344214515, "grad_norm": 0.21233728528022766, "learning_rate": 9.910471000389257e-05, "loss": 1.0675, "step": 1302 }, { "epoch": 0.5063635480423588, "grad_norm": 0.21905581653118134, "learning_rate": 9.902685869988322e-05, "loss": 1.0054, "step": 1303 }, { "epoch": 0.5067521616632663, "grad_norm": 0.23434993624687195, "learning_rate": 9.894900739587389e-05, "loss": 0.9915, "step": 1304 }, { "epoch": 0.5071407752841737, "grad_norm": 0.21684227883815765, "learning_rate": 9.887115609186454e-05, "loss": 1.1131, "step": 1305 }, { "epoch": 0.5075293889050811, "grad_norm": 0.21699552237987518, "learning_rate": 9.87933047878552e-05, "loss": 1.0782, "step": 1306 }, { "epoch": 0.5079180025259885, "grad_norm": 0.2218221127986908, "learning_rate": 9.871545348384586e-05, "loss": 1.0388, "step": 1307 }, { "epoch": 0.508306616146896, "grad_norm": 0.20104359090328217, "learning_rate": 9.863760217983652e-05, "loss": 1.0336, "step": 1308 }, { "epoch": 0.5086952297678033, "grad_norm": 0.21907050907611847, "learning_rate": 9.855975087582718e-05, "loss": 1.0587, "step": 1309 }, { "epoch": 0.5090838433887108, "grad_norm": 0.2140391767024994, "learning_rate": 9.848189957181784e-05, "loss": 1.0351, "step": 1310 }, { "epoch": 0.5094724570096182, "grad_norm": 0.33287563920021057, "learning_rate": 9.84040482678085e-05, "loss": 0.9908, "step": 1311 }, { "epoch": 0.5098610706305255, "grad_norm": 0.2706705927848816, "learning_rate": 9.832619696379915e-05, "loss": 1.0078, "step": 1312 }, { "epoch": 0.510249684251433, "grad_norm": 0.20216278731822968, "learning_rate": 9.824834565978981e-05, "loss": 1.0253, "step": 1313 }, { "epoch": 0.5106382978723404, "grad_norm": 0.20736576616764069, "learning_rate": 9.817049435578046e-05, "loss": 1.0217, "step": 1314 }, { "epoch": 0.5110269114932479, "grad_norm": 0.2275344580411911, "learning_rate": 9.809264305177113e-05, "loss": 1.0139, "step": 1315 }, { "epoch": 0.5114155251141552, "grad_norm": 0.22243620455265045, "learning_rate": 9.801479174776178e-05, "loss": 1.0427, "step": 1316 }, { "epoch": 0.5118041387350627, "grad_norm": 0.198841854929924, "learning_rate": 9.793694044375243e-05, "loss": 1.0231, "step": 1317 }, { "epoch": 0.5121927523559701, "grad_norm": 0.2031068503856659, "learning_rate": 9.78590891397431e-05, "loss": 1.0184, "step": 1318 }, { "epoch": 0.5125813659768775, "grad_norm": 0.21712587773799896, "learning_rate": 9.778123783573375e-05, "loss": 1.0205, "step": 1319 }, { "epoch": 0.5129699795977849, "grad_norm": 0.19366060197353363, "learning_rate": 9.77033865317244e-05, "loss": 0.9623, "step": 1320 }, { "epoch": 0.5133585932186923, "grad_norm": 0.19845952093601227, "learning_rate": 9.762553522771507e-05, "loss": 1.0209, "step": 1321 }, { "epoch": 0.5137472068395997, "grad_norm": 0.19700276851654053, "learning_rate": 9.754768392370572e-05, "loss": 0.9506, "step": 1322 }, { "epoch": 0.5141358204605071, "grad_norm": 0.19797460734844208, "learning_rate": 9.746983261969639e-05, "loss": 1.0928, "step": 1323 }, { "epoch": 0.5145244340814146, "grad_norm": 0.20470699667930603, "learning_rate": 9.739198131568704e-05, "loss": 1.0835, "step": 1324 }, { "epoch": 0.5149130477023219, "grad_norm": 0.19121742248535156, "learning_rate": 9.731413001167769e-05, "loss": 0.9877, "step": 1325 }, { "epoch": 0.5153016613232294, "grad_norm": 0.20026616752147675, "learning_rate": 9.723627870766836e-05, "loss": 1.0094, "step": 1326 }, { "epoch": 0.5156902749441368, "grad_norm": 0.2214539796113968, "learning_rate": 9.715842740365901e-05, "loss": 0.9867, "step": 1327 }, { "epoch": 0.5160788885650442, "grad_norm": 0.22674603760242462, "learning_rate": 9.708057609964967e-05, "loss": 1.0738, "step": 1328 }, { "epoch": 0.5164675021859516, "grad_norm": 0.21274834871292114, "learning_rate": 9.700272479564033e-05, "loss": 1.0458, "step": 1329 }, { "epoch": 0.5168561158068591, "grad_norm": 0.20305052399635315, "learning_rate": 9.692487349163099e-05, "loss": 1.0041, "step": 1330 }, { "epoch": 0.5172447294277664, "grad_norm": 0.1840772181749344, "learning_rate": 9.684702218762166e-05, "loss": 0.9498, "step": 1331 }, { "epoch": 0.5176333430486738, "grad_norm": 0.2055782824754715, "learning_rate": 9.676917088361231e-05, "loss": 1.0223, "step": 1332 }, { "epoch": 0.5180219566695813, "grad_norm": 0.21826402842998505, "learning_rate": 9.669131957960297e-05, "loss": 1.1068, "step": 1333 }, { "epoch": 0.5184105702904886, "grad_norm": 0.22516922652721405, "learning_rate": 9.661346827559363e-05, "loss": 1.0957, "step": 1334 }, { "epoch": 0.5187991839113961, "grad_norm": 0.21044284105300903, "learning_rate": 9.653561697158428e-05, "loss": 1.0384, "step": 1335 }, { "epoch": 0.5191877975323035, "grad_norm": 0.20275571942329407, "learning_rate": 9.645776566757494e-05, "loss": 0.9978, "step": 1336 }, { "epoch": 0.519576411153211, "grad_norm": 0.2077122926712036, "learning_rate": 9.63799143635656e-05, "loss": 1.0418, "step": 1337 }, { "epoch": 0.5199650247741183, "grad_norm": 0.19158867001533508, "learning_rate": 9.630206305955625e-05, "loss": 1.0527, "step": 1338 }, { "epoch": 0.5203536383950258, "grad_norm": 0.1932496577501297, "learning_rate": 9.622421175554691e-05, "loss": 1.0039, "step": 1339 }, { "epoch": 0.5207422520159332, "grad_norm": 0.21937766671180725, "learning_rate": 9.614636045153757e-05, "loss": 1.0373, "step": 1340 }, { "epoch": 0.5211308656368405, "grad_norm": 0.2268432229757309, "learning_rate": 9.606850914752823e-05, "loss": 1.0815, "step": 1341 }, { "epoch": 0.521519479257748, "grad_norm": 0.2147454470396042, "learning_rate": 9.599065784351888e-05, "loss": 1.0331, "step": 1342 }, { "epoch": 0.5219080928786554, "grad_norm": 0.19899709522724152, "learning_rate": 9.591280653950954e-05, "loss": 1.032, "step": 1343 }, { "epoch": 0.5222967064995628, "grad_norm": 0.19646069407463074, "learning_rate": 9.58349552355002e-05, "loss": 0.9788, "step": 1344 }, { "epoch": 0.5226853201204702, "grad_norm": 0.2146075963973999, "learning_rate": 9.575710393149085e-05, "loss": 1.0201, "step": 1345 }, { "epoch": 0.5230739337413777, "grad_norm": 0.1968650370836258, "learning_rate": 9.567925262748152e-05, "loss": 0.9894, "step": 1346 }, { "epoch": 0.523462547362285, "grad_norm": 0.21111296117305756, "learning_rate": 9.560140132347217e-05, "loss": 1.0961, "step": 1347 }, { "epoch": 0.5238511609831925, "grad_norm": 0.20917272567749023, "learning_rate": 9.552355001946282e-05, "loss": 1.0435, "step": 1348 }, { "epoch": 0.5242397746040999, "grad_norm": 0.2029752880334854, "learning_rate": 9.544569871545349e-05, "loss": 1.0328, "step": 1349 }, { "epoch": 0.5246283882250072, "grad_norm": 0.20726613700389862, "learning_rate": 9.536784741144414e-05, "loss": 1.0465, "step": 1350 }, { "epoch": 0.5250170018459147, "grad_norm": 0.19778740406036377, "learning_rate": 9.52899961074348e-05, "loss": 1.0058, "step": 1351 }, { "epoch": 0.5254056154668221, "grad_norm": 0.19958540797233582, "learning_rate": 9.521214480342546e-05, "loss": 1.0164, "step": 1352 }, { "epoch": 0.5257942290877295, "grad_norm": 0.2151395082473755, "learning_rate": 9.513429349941611e-05, "loss": 1.0703, "step": 1353 }, { "epoch": 0.5261828427086369, "grad_norm": 0.2366979569196701, "learning_rate": 9.505644219540678e-05, "loss": 0.9832, "step": 1354 }, { "epoch": 0.5265714563295444, "grad_norm": 0.22064165771007538, "learning_rate": 9.497859089139743e-05, "loss": 1.0181, "step": 1355 }, { "epoch": 0.5269600699504517, "grad_norm": 0.20221936702728271, "learning_rate": 9.49007395873881e-05, "loss": 1.0424, "step": 1356 }, { "epoch": 0.5273486835713592, "grad_norm": 0.19608759880065918, "learning_rate": 9.482288828337876e-05, "loss": 1.0074, "step": 1357 }, { "epoch": 0.5277372971922666, "grad_norm": 0.20686689019203186, "learning_rate": 9.474503697936941e-05, "loss": 1.0213, "step": 1358 }, { "epoch": 0.528125910813174, "grad_norm": 0.223610520362854, "learning_rate": 9.466718567536008e-05, "loss": 1.05, "step": 1359 }, { "epoch": 0.5285145244340814, "grad_norm": 0.2135966569185257, "learning_rate": 9.458933437135073e-05, "loss": 1.034, "step": 1360 }, { "epoch": 0.5289031380549888, "grad_norm": 0.1933239996433258, "learning_rate": 9.451148306734138e-05, "loss": 0.9883, "step": 1361 }, { "epoch": 0.5292917516758963, "grad_norm": 0.20794694125652313, "learning_rate": 9.443363176333205e-05, "loss": 1.0103, "step": 1362 }, { "epoch": 0.5296803652968036, "grad_norm": 0.20128493010997772, "learning_rate": 9.43557804593227e-05, "loss": 1.015, "step": 1363 }, { "epoch": 0.5300689789177111, "grad_norm": 0.2128933072090149, "learning_rate": 9.427792915531336e-05, "loss": 1.0038, "step": 1364 }, { "epoch": 0.5304575925386185, "grad_norm": 0.2046983689069748, "learning_rate": 9.420007785130402e-05, "loss": 0.9948, "step": 1365 }, { "epoch": 0.5308462061595259, "grad_norm": 0.20909680426120758, "learning_rate": 9.412222654729467e-05, "loss": 1.0308, "step": 1366 }, { "epoch": 0.5312348197804333, "grad_norm": 0.2182164192199707, "learning_rate": 9.404437524328533e-05, "loss": 1.0018, "step": 1367 }, { "epoch": 0.5316234334013407, "grad_norm": 0.2107028216123581, "learning_rate": 9.396652393927599e-05, "loss": 1.0419, "step": 1368 }, { "epoch": 0.5320120470222481, "grad_norm": 0.24631445109844208, "learning_rate": 9.388867263526665e-05, "loss": 1.0171, "step": 1369 }, { "epoch": 0.5324006606431555, "grad_norm": 0.20331013202667236, "learning_rate": 9.38108213312573e-05, "loss": 1.0592, "step": 1370 }, { "epoch": 0.532789274264063, "grad_norm": 0.19266058504581451, "learning_rate": 9.373297002724796e-05, "loss": 0.9912, "step": 1371 }, { "epoch": 0.5331778878849703, "grad_norm": 0.22874227166175842, "learning_rate": 9.365511872323862e-05, "loss": 1.0533, "step": 1372 }, { "epoch": 0.5335665015058778, "grad_norm": 0.2088235765695572, "learning_rate": 9.357726741922927e-05, "loss": 1.0464, "step": 1373 }, { "epoch": 0.5339551151267852, "grad_norm": 0.2112397700548172, "learning_rate": 9.349941611521994e-05, "loss": 1.0503, "step": 1374 }, { "epoch": 0.5343437287476926, "grad_norm": 0.20712170004844666, "learning_rate": 9.342156481121059e-05, "loss": 1.0237, "step": 1375 }, { "epoch": 0.5347323423686, "grad_norm": 0.20077116787433624, "learning_rate": 9.334371350720124e-05, "loss": 1.0467, "step": 1376 }, { "epoch": 0.5351209559895075, "grad_norm": 0.20394501090049744, "learning_rate": 9.326586220319191e-05, "loss": 1.0054, "step": 1377 }, { "epoch": 0.5355095696104148, "grad_norm": 0.19459395110607147, "learning_rate": 9.318801089918256e-05, "loss": 0.9792, "step": 1378 }, { "epoch": 0.5358981832313222, "grad_norm": 0.2116049826145172, "learning_rate": 9.311015959517321e-05, "loss": 1.0345, "step": 1379 }, { "epoch": 0.5362867968522297, "grad_norm": 0.21672269701957703, "learning_rate": 9.303230829116388e-05, "loss": 1.0709, "step": 1380 }, { "epoch": 0.536675410473137, "grad_norm": 0.20358407497406006, "learning_rate": 9.295445698715453e-05, "loss": 1.0534, "step": 1381 }, { "epoch": 0.5370640240940445, "grad_norm": 0.19512853026390076, "learning_rate": 9.28766056831452e-05, "loss": 0.9397, "step": 1382 }, { "epoch": 0.5374526377149519, "grad_norm": 0.2140122503042221, "learning_rate": 9.279875437913586e-05, "loss": 1.0164, "step": 1383 }, { "epoch": 0.5378412513358594, "grad_norm": 0.20486049354076385, "learning_rate": 9.272090307512651e-05, "loss": 0.9892, "step": 1384 }, { "epoch": 0.5382298649567667, "grad_norm": 0.20023222267627716, "learning_rate": 9.264305177111718e-05, "loss": 1.0019, "step": 1385 }, { "epoch": 0.5386184785776742, "grad_norm": 0.20024439692497253, "learning_rate": 9.256520046710783e-05, "loss": 0.9717, "step": 1386 }, { "epoch": 0.5390070921985816, "grad_norm": 0.21021386981010437, "learning_rate": 9.24873491630985e-05, "loss": 1.028, "step": 1387 }, { "epoch": 0.5393957058194889, "grad_norm": 0.18508704006671906, "learning_rate": 9.240949785908915e-05, "loss": 1.0008, "step": 1388 }, { "epoch": 0.5397843194403964, "grad_norm": 0.19351208209991455, "learning_rate": 9.23316465550798e-05, "loss": 0.9898, "step": 1389 }, { "epoch": 0.5401729330613038, "grad_norm": 0.20341919362545013, "learning_rate": 9.225379525107047e-05, "loss": 1.0203, "step": 1390 }, { "epoch": 0.5405615466822112, "grad_norm": 0.1942797303199768, "learning_rate": 9.217594394706112e-05, "loss": 1.003, "step": 1391 }, { "epoch": 0.5409501603031186, "grad_norm": 0.2056138813495636, "learning_rate": 9.209809264305178e-05, "loss": 1.0149, "step": 1392 }, { "epoch": 0.5413387739240261, "grad_norm": 0.21572062373161316, "learning_rate": 9.202024133904244e-05, "loss": 0.9808, "step": 1393 }, { "epoch": 0.5417273875449334, "grad_norm": 0.19841499626636505, "learning_rate": 9.194239003503309e-05, "loss": 1.0467, "step": 1394 }, { "epoch": 0.5421160011658409, "grad_norm": 0.20452147722244263, "learning_rate": 9.186453873102375e-05, "loss": 1.0378, "step": 1395 }, { "epoch": 0.5425046147867483, "grad_norm": 0.2090451419353485, "learning_rate": 9.17866874270144e-05, "loss": 1.0823, "step": 1396 }, { "epoch": 0.5428932284076556, "grad_norm": 0.215814009308815, "learning_rate": 9.170883612300506e-05, "loss": 1.0994, "step": 1397 }, { "epoch": 0.5432818420285631, "grad_norm": 0.19924724102020264, "learning_rate": 9.163098481899572e-05, "loss": 1.0099, "step": 1398 }, { "epoch": 0.5436704556494705, "grad_norm": 0.20074865221977234, "learning_rate": 9.155313351498638e-05, "loss": 1.0163, "step": 1399 }, { "epoch": 0.544059069270378, "grad_norm": 0.21737203001976013, "learning_rate": 9.147528221097704e-05, "loss": 1.0527, "step": 1400 }, { "epoch": 0.5444476828912853, "grad_norm": 0.2036885768175125, "learning_rate": 9.139743090696769e-05, "loss": 1.0208, "step": 1401 }, { "epoch": 0.5448362965121928, "grad_norm": 0.20861585438251495, "learning_rate": 9.131957960295835e-05, "loss": 1.0175, "step": 1402 }, { "epoch": 0.5452249101331001, "grad_norm": 0.23425570130348206, "learning_rate": 9.124172829894901e-05, "loss": 1.053, "step": 1403 }, { "epoch": 0.5456135237540076, "grad_norm": 0.20389291644096375, "learning_rate": 9.116387699493966e-05, "loss": 1.0479, "step": 1404 }, { "epoch": 0.546002137374915, "grad_norm": 0.20166678726673126, "learning_rate": 9.108602569093033e-05, "loss": 1.0064, "step": 1405 }, { "epoch": 0.5463907509958223, "grad_norm": 0.21419203281402588, "learning_rate": 9.100817438692098e-05, "loss": 1.0122, "step": 1406 }, { "epoch": 0.5467793646167298, "grad_norm": 0.20541758835315704, "learning_rate": 9.093032308291165e-05, "loss": 1.0355, "step": 1407 }, { "epoch": 0.5471679782376372, "grad_norm": 0.21865367889404297, "learning_rate": 9.08524717789023e-05, "loss": 1.0201, "step": 1408 }, { "epoch": 0.5475565918585447, "grad_norm": 0.21181468665599823, "learning_rate": 9.077462047489296e-05, "loss": 1.0501, "step": 1409 }, { "epoch": 0.547945205479452, "grad_norm": 0.21016767621040344, "learning_rate": 9.069676917088362e-05, "loss": 1.0452, "step": 1410 }, { "epoch": 0.5483338191003595, "grad_norm": 0.21119755506515503, "learning_rate": 9.061891786687428e-05, "loss": 1.0935, "step": 1411 }, { "epoch": 0.5487224327212669, "grad_norm": 0.20688095688819885, "learning_rate": 9.054106656286493e-05, "loss": 1.0526, "step": 1412 }, { "epoch": 0.5491110463421743, "grad_norm": 0.21857528388500214, "learning_rate": 9.04632152588556e-05, "loss": 1.0067, "step": 1413 }, { "epoch": 0.5494996599630817, "grad_norm": 0.2196548581123352, "learning_rate": 9.038536395484625e-05, "loss": 1.0263, "step": 1414 }, { "epoch": 0.5498882735839892, "grad_norm": 0.21952040493488312, "learning_rate": 9.03075126508369e-05, "loss": 1.0009, "step": 1415 }, { "epoch": 0.5502768872048965, "grad_norm": 0.20059294998645782, "learning_rate": 9.022966134682757e-05, "loss": 1.0481, "step": 1416 }, { "epoch": 0.5506655008258039, "grad_norm": 0.1960824728012085, "learning_rate": 9.015181004281822e-05, "loss": 1.0003, "step": 1417 }, { "epoch": 0.5510541144467114, "grad_norm": 0.19051724672317505, "learning_rate": 9.007395873880889e-05, "loss": 0.9556, "step": 1418 }, { "epoch": 0.5514427280676187, "grad_norm": 0.21008028090000153, "learning_rate": 8.999610743479954e-05, "loss": 1.0457, "step": 1419 }, { "epoch": 0.5518313416885262, "grad_norm": 0.21465444564819336, "learning_rate": 8.991825613079019e-05, "loss": 1.0196, "step": 1420 }, { "epoch": 0.5522199553094336, "grad_norm": 0.2062770277261734, "learning_rate": 8.984040482678086e-05, "loss": 1.0501, "step": 1421 }, { "epoch": 0.552608568930341, "grad_norm": 0.21400012075901031, "learning_rate": 8.976255352277151e-05, "loss": 1.0711, "step": 1422 }, { "epoch": 0.5529971825512484, "grad_norm": 0.19617624580860138, "learning_rate": 8.968470221876217e-05, "loss": 0.9858, "step": 1423 }, { "epoch": 0.5533857961721559, "grad_norm": 0.20835624635219574, "learning_rate": 8.960685091475283e-05, "loss": 1.0122, "step": 1424 }, { "epoch": 0.5537744097930632, "grad_norm": 0.21708111464977264, "learning_rate": 8.952899961074348e-05, "loss": 1.0108, "step": 1425 }, { "epoch": 0.5541630234139706, "grad_norm": 0.20877864956855774, "learning_rate": 8.945114830673414e-05, "loss": 1.0389, "step": 1426 }, { "epoch": 0.5545516370348781, "grad_norm": 0.1924441158771515, "learning_rate": 8.93732970027248e-05, "loss": 1.0088, "step": 1427 }, { "epoch": 0.5549402506557854, "grad_norm": 0.20288826525211334, "learning_rate": 8.929544569871546e-05, "loss": 1.0296, "step": 1428 }, { "epoch": 0.5553288642766929, "grad_norm": 0.2008143663406372, "learning_rate": 8.921759439470611e-05, "loss": 1.0521, "step": 1429 }, { "epoch": 0.5557174778976003, "grad_norm": 0.24407047033309937, "learning_rate": 8.913974309069677e-05, "loss": 1.1038, "step": 1430 }, { "epoch": 0.5561060915185078, "grad_norm": 0.2172536998987198, "learning_rate": 8.906189178668743e-05, "loss": 1.0811, "step": 1431 }, { "epoch": 0.5564947051394151, "grad_norm": 0.21712054312229156, "learning_rate": 8.898404048267808e-05, "loss": 1.0642, "step": 1432 }, { "epoch": 0.5568833187603226, "grad_norm": 0.22482797503471375, "learning_rate": 8.890618917866875e-05, "loss": 1.0742, "step": 1433 }, { "epoch": 0.55727193238123, "grad_norm": 0.1974876970052719, "learning_rate": 8.88283378746594e-05, "loss": 0.9954, "step": 1434 }, { "epoch": 0.5576605460021373, "grad_norm": 0.19162166118621826, "learning_rate": 8.875048657065007e-05, "loss": 1.0074, "step": 1435 }, { "epoch": 0.5580491596230448, "grad_norm": 0.20439045131206512, "learning_rate": 8.867263526664072e-05, "loss": 1.026, "step": 1436 }, { "epoch": 0.5584377732439522, "grad_norm": 0.1947651207447052, "learning_rate": 8.859478396263138e-05, "loss": 0.9848, "step": 1437 }, { "epoch": 0.5588263868648596, "grad_norm": 0.21434316039085388, "learning_rate": 8.851693265862204e-05, "loss": 1.0843, "step": 1438 }, { "epoch": 0.559215000485767, "grad_norm": 1.3314417600631714, "learning_rate": 8.84390813546127e-05, "loss": 1.0356, "step": 1439 }, { "epoch": 0.5596036141066745, "grad_norm": 0.20131289958953857, "learning_rate": 8.836123005060335e-05, "loss": 1.0214, "step": 1440 }, { "epoch": 0.5599922277275818, "grad_norm": 0.21596461534500122, "learning_rate": 8.828337874659402e-05, "loss": 1.0962, "step": 1441 }, { "epoch": 0.5603808413484893, "grad_norm": 0.20477193593978882, "learning_rate": 8.820552744258467e-05, "loss": 1.0643, "step": 1442 }, { "epoch": 0.5607694549693967, "grad_norm": 0.1978107988834381, "learning_rate": 8.812767613857532e-05, "loss": 1.0054, "step": 1443 }, { "epoch": 0.561158068590304, "grad_norm": 0.219422847032547, "learning_rate": 8.804982483456599e-05, "loss": 1.0009, "step": 1444 }, { "epoch": 0.5615466822112115, "grad_norm": 0.21489015221595764, "learning_rate": 8.797197353055664e-05, "loss": 1.052, "step": 1445 }, { "epoch": 0.5619352958321189, "grad_norm": 0.2235930860042572, "learning_rate": 8.78941222265473e-05, "loss": 1.037, "step": 1446 }, { "epoch": 0.5623239094530263, "grad_norm": 0.19922038912773132, "learning_rate": 8.781627092253796e-05, "loss": 1.0006, "step": 1447 }, { "epoch": 0.5627125230739337, "grad_norm": 0.24740247428417206, "learning_rate": 8.773841961852861e-05, "loss": 1.0753, "step": 1448 }, { "epoch": 0.5631011366948412, "grad_norm": 0.2148803174495697, "learning_rate": 8.766056831451928e-05, "loss": 1.0712, "step": 1449 }, { "epoch": 0.5634897503157485, "grad_norm": 0.19838745892047882, "learning_rate": 8.758271701050993e-05, "loss": 1.027, "step": 1450 }, { "epoch": 0.563878363936656, "grad_norm": 0.20328201353549957, "learning_rate": 8.750486570650058e-05, "loss": 1.0117, "step": 1451 }, { "epoch": 0.5642669775575634, "grad_norm": 0.21230114996433258, "learning_rate": 8.742701440249125e-05, "loss": 1.0658, "step": 1452 }, { "epoch": 0.5646555911784708, "grad_norm": 0.2030259519815445, "learning_rate": 8.73491630984819e-05, "loss": 1.0002, "step": 1453 }, { "epoch": 0.5650442047993782, "grad_norm": 0.21404659748077393, "learning_rate": 8.727131179447256e-05, "loss": 1.0572, "step": 1454 }, { "epoch": 0.5654328184202856, "grad_norm": 0.2148464322090149, "learning_rate": 8.719346049046322e-05, "loss": 1.0164, "step": 1455 }, { "epoch": 0.5658214320411931, "grad_norm": 0.22083118557929993, "learning_rate": 8.711560918645387e-05, "loss": 0.9704, "step": 1456 }, { "epoch": 0.5662100456621004, "grad_norm": 0.19305935502052307, "learning_rate": 8.703775788244453e-05, "loss": 1.0034, "step": 1457 }, { "epoch": 0.5665986592830079, "grad_norm": 0.2100098729133606, "learning_rate": 8.695990657843518e-05, "loss": 1.0907, "step": 1458 }, { "epoch": 0.5669872729039153, "grad_norm": 0.18947799503803253, "learning_rate": 8.688205527442585e-05, "loss": 0.9664, "step": 1459 }, { "epoch": 0.5673758865248227, "grad_norm": 0.22341710329055786, "learning_rate": 8.68042039704165e-05, "loss": 1.0551, "step": 1460 }, { "epoch": 0.5677645001457301, "grad_norm": 0.219679057598114, "learning_rate": 8.672635266640717e-05, "loss": 1.0398, "step": 1461 }, { "epoch": 0.5681531137666376, "grad_norm": 0.22389841079711914, "learning_rate": 8.664850136239782e-05, "loss": 1.0472, "step": 1462 }, { "epoch": 0.5685417273875449, "grad_norm": 0.21402975916862488, "learning_rate": 8.657065005838849e-05, "loss": 1.0224, "step": 1463 }, { "epoch": 0.5689303410084523, "grad_norm": 0.20917154848575592, "learning_rate": 8.649279875437915e-05, "loss": 1.0526, "step": 1464 }, { "epoch": 0.5693189546293598, "grad_norm": 0.2252056896686554, "learning_rate": 8.64149474503698e-05, "loss": 1.1064, "step": 1465 }, { "epoch": 0.5697075682502671, "grad_norm": 0.21834802627563477, "learning_rate": 8.633709614636046e-05, "loss": 1.0318, "step": 1466 }, { "epoch": 0.5700961818711746, "grad_norm": 0.21882353723049164, "learning_rate": 8.625924484235112e-05, "loss": 1.0285, "step": 1467 }, { "epoch": 0.570484795492082, "grad_norm": 0.2028426229953766, "learning_rate": 8.618139353834177e-05, "loss": 1.0356, "step": 1468 }, { "epoch": 0.5708734091129894, "grad_norm": 0.22297166287899017, "learning_rate": 8.610354223433243e-05, "loss": 1.0804, "step": 1469 }, { "epoch": 0.5712620227338968, "grad_norm": 0.21775268018245697, "learning_rate": 8.602569093032309e-05, "loss": 0.9978, "step": 1470 }, { "epoch": 0.5716506363548043, "grad_norm": 0.20362353324890137, "learning_rate": 8.594783962631374e-05, "loss": 0.9982, "step": 1471 }, { "epoch": 0.5720392499757117, "grad_norm": 0.21854591369628906, "learning_rate": 8.586998832230441e-05, "loss": 1.0465, "step": 1472 }, { "epoch": 0.572427863596619, "grad_norm": 0.20501428842544556, "learning_rate": 8.579213701829506e-05, "loss": 1.0468, "step": 1473 }, { "epoch": 0.5728164772175265, "grad_norm": 0.21606214344501495, "learning_rate": 8.571428571428571e-05, "loss": 1.0477, "step": 1474 }, { "epoch": 0.5732050908384339, "grad_norm": 0.2100660502910614, "learning_rate": 8.563643441027638e-05, "loss": 1.0071, "step": 1475 }, { "epoch": 0.5735937044593413, "grad_norm": 0.21008896827697754, "learning_rate": 8.555858310626703e-05, "loss": 0.9914, "step": 1476 }, { "epoch": 0.5739823180802487, "grad_norm": 0.22192159295082092, "learning_rate": 8.54807318022577e-05, "loss": 1.0385, "step": 1477 }, { "epoch": 0.5743709317011562, "grad_norm": 0.20123356580734253, "learning_rate": 8.540288049824835e-05, "loss": 1.0062, "step": 1478 }, { "epoch": 0.5747595453220635, "grad_norm": 0.201947420835495, "learning_rate": 8.5325029194239e-05, "loss": 1.0218, "step": 1479 }, { "epoch": 0.575148158942971, "grad_norm": 0.22804415225982666, "learning_rate": 8.524717789022967e-05, "loss": 1.0445, "step": 1480 }, { "epoch": 0.5755367725638784, "grad_norm": 0.20527036488056183, "learning_rate": 8.516932658622032e-05, "loss": 0.9972, "step": 1481 }, { "epoch": 0.5759253861847857, "grad_norm": 0.20298773050308228, "learning_rate": 8.509147528221098e-05, "loss": 1.0272, "step": 1482 }, { "epoch": 0.5763139998056932, "grad_norm": 0.22500957548618317, "learning_rate": 8.501362397820164e-05, "loss": 1.0982, "step": 1483 }, { "epoch": 0.5767026134266006, "grad_norm": 0.1950521320104599, "learning_rate": 8.493577267419229e-05, "loss": 0.9848, "step": 1484 }, { "epoch": 0.577091227047508, "grad_norm": 0.21087585389614105, "learning_rate": 8.485792137018295e-05, "loss": 1.0125, "step": 1485 }, { "epoch": 0.5774798406684154, "grad_norm": 0.20122238993644714, "learning_rate": 8.47800700661736e-05, "loss": 1.0533, "step": 1486 }, { "epoch": 0.5778684542893229, "grad_norm": 0.20149008929729462, "learning_rate": 8.470221876216427e-05, "loss": 1.0719, "step": 1487 }, { "epoch": 0.5782570679102302, "grad_norm": 0.21307213604450226, "learning_rate": 8.462436745815494e-05, "loss": 1.0522, "step": 1488 }, { "epoch": 0.5786456815311377, "grad_norm": 0.21828554570674896, "learning_rate": 8.454651615414559e-05, "loss": 1.0184, "step": 1489 }, { "epoch": 0.5790342951520451, "grad_norm": 0.22002705931663513, "learning_rate": 8.446866485013625e-05, "loss": 1.0101, "step": 1490 }, { "epoch": 0.5794229087729524, "grad_norm": 0.19479142129421234, "learning_rate": 8.43908135461269e-05, "loss": 0.9889, "step": 1491 }, { "epoch": 0.5798115223938599, "grad_norm": 0.21346086263656616, "learning_rate": 8.431296224211756e-05, "loss": 1.0373, "step": 1492 }, { "epoch": 0.5802001360147673, "grad_norm": 0.20177558064460754, "learning_rate": 8.423511093810822e-05, "loss": 1.0215, "step": 1493 }, { "epoch": 0.5805887496356748, "grad_norm": 0.2117915153503418, "learning_rate": 8.415725963409888e-05, "loss": 1.0321, "step": 1494 }, { "epoch": 0.5809773632565821, "grad_norm": 0.21304374933242798, "learning_rate": 8.407940833008954e-05, "loss": 1.0123, "step": 1495 }, { "epoch": 0.5813659768774896, "grad_norm": 0.21173715591430664, "learning_rate": 8.400155702608019e-05, "loss": 1.0696, "step": 1496 }, { "epoch": 0.581754590498397, "grad_norm": 0.20407019555568695, "learning_rate": 8.392370572207085e-05, "loss": 1.0086, "step": 1497 }, { "epoch": 0.5821432041193044, "grad_norm": 0.209481880068779, "learning_rate": 8.384585441806151e-05, "loss": 0.9975, "step": 1498 }, { "epoch": 0.5825318177402118, "grad_norm": 0.22184531390666962, "learning_rate": 8.376800311405216e-05, "loss": 1.0956, "step": 1499 }, { "epoch": 0.5829204313611193, "grad_norm": 0.21344684064388275, "learning_rate": 8.369015181004283e-05, "loss": 1.0685, "step": 1500 }, { "epoch": 0.5833090449820266, "grad_norm": 0.19837221503257751, "learning_rate": 8.361230050603348e-05, "loss": 1.0149, "step": 1501 }, { "epoch": 0.583697658602934, "grad_norm": 0.2133672833442688, "learning_rate": 8.353444920202413e-05, "loss": 1.0453, "step": 1502 }, { "epoch": 0.5840862722238415, "grad_norm": 0.21944090723991394, "learning_rate": 8.34565978980148e-05, "loss": 1.04, "step": 1503 }, { "epoch": 0.5844748858447488, "grad_norm": 0.1983667016029358, "learning_rate": 8.337874659400545e-05, "loss": 0.9919, "step": 1504 }, { "epoch": 0.5848634994656563, "grad_norm": 0.2025303989648819, "learning_rate": 8.33008952899961e-05, "loss": 1.0021, "step": 1505 }, { "epoch": 0.5852521130865637, "grad_norm": 0.2015170007944107, "learning_rate": 8.322304398598677e-05, "loss": 0.9945, "step": 1506 }, { "epoch": 0.5856407267074711, "grad_norm": 0.20768272876739502, "learning_rate": 8.314519268197742e-05, "loss": 1.0465, "step": 1507 }, { "epoch": 0.5860293403283785, "grad_norm": 0.20513412356376648, "learning_rate": 8.306734137796809e-05, "loss": 1.0124, "step": 1508 }, { "epoch": 0.586417953949286, "grad_norm": 0.20268471539020538, "learning_rate": 8.298949007395874e-05, "loss": 1.0586, "step": 1509 }, { "epoch": 0.5868065675701933, "grad_norm": 0.20915938913822174, "learning_rate": 8.291163876994939e-05, "loss": 1.0047, "step": 1510 }, { "epoch": 0.5871951811911007, "grad_norm": 0.2161451131105423, "learning_rate": 8.283378746594006e-05, "loss": 1.0184, "step": 1511 }, { "epoch": 0.5875837948120082, "grad_norm": 0.1915571093559265, "learning_rate": 8.275593616193071e-05, "loss": 1.0187, "step": 1512 }, { "epoch": 0.5879724084329155, "grad_norm": 0.20907992124557495, "learning_rate": 8.267808485792137e-05, "loss": 1.0212, "step": 1513 }, { "epoch": 0.588361022053823, "grad_norm": 0.20140786468982697, "learning_rate": 8.260023355391204e-05, "loss": 1.014, "step": 1514 }, { "epoch": 0.5887496356747304, "grad_norm": 0.208252415060997, "learning_rate": 8.252238224990269e-05, "loss": 1.0806, "step": 1515 }, { "epoch": 0.5891382492956379, "grad_norm": 0.20596125721931458, "learning_rate": 8.244453094589336e-05, "loss": 0.9823, "step": 1516 }, { "epoch": 0.5895268629165452, "grad_norm": 0.18832452595233917, "learning_rate": 8.236667964188401e-05, "loss": 0.9925, "step": 1517 }, { "epoch": 0.5899154765374527, "grad_norm": 0.2078334391117096, "learning_rate": 8.228882833787467e-05, "loss": 1.0587, "step": 1518 }, { "epoch": 0.59030409015836, "grad_norm": 0.20121365785598755, "learning_rate": 8.221097703386533e-05, "loss": 1.0607, "step": 1519 }, { "epoch": 0.5906927037792674, "grad_norm": 0.19666099548339844, "learning_rate": 8.213312572985598e-05, "loss": 1.0124, "step": 1520 }, { "epoch": 0.5910813174001749, "grad_norm": 0.20176006853580475, "learning_rate": 8.205527442584664e-05, "loss": 1.0297, "step": 1521 }, { "epoch": 0.5914699310210823, "grad_norm": 0.2038574516773224, "learning_rate": 8.19774231218373e-05, "loss": 1.0311, "step": 1522 }, { "epoch": 0.5918585446419897, "grad_norm": 0.19517424702644348, "learning_rate": 8.189957181782796e-05, "loss": 0.9945, "step": 1523 }, { "epoch": 0.5922471582628971, "grad_norm": 0.19599094986915588, "learning_rate": 8.182172051381861e-05, "loss": 1.0255, "step": 1524 }, { "epoch": 0.5926357718838046, "grad_norm": 0.21409402787685394, "learning_rate": 8.174386920980927e-05, "loss": 1.0868, "step": 1525 }, { "epoch": 0.5930243855047119, "grad_norm": 0.19567830860614777, "learning_rate": 8.166601790579993e-05, "loss": 0.9654, "step": 1526 }, { "epoch": 0.5934129991256194, "grad_norm": 0.2275007963180542, "learning_rate": 8.158816660179058e-05, "loss": 1.0867, "step": 1527 }, { "epoch": 0.5938016127465268, "grad_norm": 0.19826427102088928, "learning_rate": 8.151031529778123e-05, "loss": 1.0301, "step": 1528 }, { "epoch": 0.5941902263674341, "grad_norm": 0.2051352709531784, "learning_rate": 8.14324639937719e-05, "loss": 1.023, "step": 1529 }, { "epoch": 0.5945788399883416, "grad_norm": 0.19492043554782867, "learning_rate": 8.135461268976255e-05, "loss": 0.9608, "step": 1530 }, { "epoch": 0.594967453609249, "grad_norm": 0.21521608531475067, "learning_rate": 8.127676138575322e-05, "loss": 1.0612, "step": 1531 }, { "epoch": 0.5953560672301564, "grad_norm": 0.22739367187023163, "learning_rate": 8.119891008174387e-05, "loss": 1.0603, "step": 1532 }, { "epoch": 0.5957446808510638, "grad_norm": 0.20334595441818237, "learning_rate": 8.112105877773452e-05, "loss": 1.0191, "step": 1533 }, { "epoch": 0.5961332944719713, "grad_norm": 0.20985397696495056, "learning_rate": 8.104320747372519e-05, "loss": 1.0721, "step": 1534 }, { "epoch": 0.5965219080928786, "grad_norm": 0.20472954213619232, "learning_rate": 8.096535616971584e-05, "loss": 1.0556, "step": 1535 }, { "epoch": 0.5969105217137861, "grad_norm": 0.2112964689731598, "learning_rate": 8.08875048657065e-05, "loss": 1.0016, "step": 1536 }, { "epoch": 0.5972991353346935, "grad_norm": 0.21330617368221283, "learning_rate": 8.080965356169716e-05, "loss": 1.0783, "step": 1537 }, { "epoch": 0.5976877489556008, "grad_norm": 0.20907814800739288, "learning_rate": 8.073180225768782e-05, "loss": 1.071, "step": 1538 }, { "epoch": 0.5980763625765083, "grad_norm": 0.2038964033126831, "learning_rate": 8.065395095367848e-05, "loss": 1.0039, "step": 1539 }, { "epoch": 0.5984649761974157, "grad_norm": 0.2175542712211609, "learning_rate": 8.057609964966914e-05, "loss": 1.0015, "step": 1540 }, { "epoch": 0.5988535898183232, "grad_norm": 0.21474529802799225, "learning_rate": 8.049824834565979e-05, "loss": 1.0273, "step": 1541 }, { "epoch": 0.5992422034392305, "grad_norm": 0.21428482234477997, "learning_rate": 8.042039704165046e-05, "loss": 1.0767, "step": 1542 }, { "epoch": 0.599630817060138, "grad_norm": 0.20287524163722992, "learning_rate": 8.034254573764111e-05, "loss": 1.064, "step": 1543 }, { "epoch": 0.6000194306810454, "grad_norm": 0.20689848065376282, "learning_rate": 8.026469443363178e-05, "loss": 1.0084, "step": 1544 }, { "epoch": 0.6004080443019528, "grad_norm": 0.22451332211494446, "learning_rate": 8.018684312962243e-05, "loss": 1.1039, "step": 1545 }, { "epoch": 0.6007966579228602, "grad_norm": 0.21381956338882446, "learning_rate": 8.010899182561308e-05, "loss": 1.0551, "step": 1546 }, { "epoch": 0.6011852715437677, "grad_norm": 0.20108483731746674, "learning_rate": 8.003114052160375e-05, "loss": 1.0326, "step": 1547 }, { "epoch": 0.601573885164675, "grad_norm": 0.19739678502082825, "learning_rate": 7.99532892175944e-05, "loss": 1.0319, "step": 1548 }, { "epoch": 0.6019624987855824, "grad_norm": 0.21635359525680542, "learning_rate": 7.987543791358506e-05, "loss": 1.0465, "step": 1549 }, { "epoch": 0.6023511124064899, "grad_norm": 0.1949319988489151, "learning_rate": 7.979758660957572e-05, "loss": 1.0026, "step": 1550 }, { "epoch": 0.6027397260273972, "grad_norm": 0.1989699900150299, "learning_rate": 7.971973530556637e-05, "loss": 1.021, "step": 1551 }, { "epoch": 0.6031283396483047, "grad_norm": 0.24031391739845276, "learning_rate": 7.964188400155703e-05, "loss": 1.0293, "step": 1552 }, { "epoch": 0.6035169532692121, "grad_norm": 0.21247251331806183, "learning_rate": 7.956403269754769e-05, "loss": 1.023, "step": 1553 }, { "epoch": 0.6039055668901195, "grad_norm": 0.21565628051757812, "learning_rate": 7.948618139353835e-05, "loss": 1.1027, "step": 1554 }, { "epoch": 0.6042941805110269, "grad_norm": 0.21207931637763977, "learning_rate": 7.9408330089529e-05, "loss": 1.0634, "step": 1555 }, { "epoch": 0.6046827941319344, "grad_norm": 0.21354155242443085, "learning_rate": 7.933047878551965e-05, "loss": 1.0433, "step": 1556 }, { "epoch": 0.6050714077528417, "grad_norm": 0.21708370745182037, "learning_rate": 7.925262748151032e-05, "loss": 1.0499, "step": 1557 }, { "epoch": 0.6054600213737491, "grad_norm": 0.2051447182893753, "learning_rate": 7.917477617750097e-05, "loss": 1.0042, "step": 1558 }, { "epoch": 0.6058486349946566, "grad_norm": 0.18768000602722168, "learning_rate": 7.909692487349164e-05, "loss": 1.009, "step": 1559 }, { "epoch": 0.6062372486155639, "grad_norm": 0.2142931967973709, "learning_rate": 7.901907356948229e-05, "loss": 1.0458, "step": 1560 }, { "epoch": 0.6066258622364714, "grad_norm": 0.21006444096565247, "learning_rate": 7.894122226547294e-05, "loss": 1.0286, "step": 1561 }, { "epoch": 0.6070144758573788, "grad_norm": 0.2187039703130722, "learning_rate": 7.886337096146361e-05, "loss": 1.0103, "step": 1562 }, { "epoch": 0.6074030894782863, "grad_norm": 0.19863669574260712, "learning_rate": 7.878551965745426e-05, "loss": 0.9925, "step": 1563 }, { "epoch": 0.6077917030991936, "grad_norm": 0.21771976351737976, "learning_rate": 7.870766835344493e-05, "loss": 0.9853, "step": 1564 }, { "epoch": 0.6081803167201011, "grad_norm": 0.21714983880519867, "learning_rate": 7.862981704943558e-05, "loss": 1.0123, "step": 1565 }, { "epoch": 0.6085689303410085, "grad_norm": 0.2251398265361786, "learning_rate": 7.855196574542624e-05, "loss": 1.0265, "step": 1566 }, { "epoch": 0.6089575439619158, "grad_norm": 0.22089716792106628, "learning_rate": 7.84741144414169e-05, "loss": 1.0689, "step": 1567 }, { "epoch": 0.6093461575828233, "grad_norm": 0.2453841269016266, "learning_rate": 7.839626313740756e-05, "loss": 1.0185, "step": 1568 }, { "epoch": 0.6097347712037307, "grad_norm": 0.21866528689861298, "learning_rate": 7.831841183339821e-05, "loss": 1.0361, "step": 1569 }, { "epoch": 0.6101233848246381, "grad_norm": 0.22421486675739288, "learning_rate": 7.824056052938888e-05, "loss": 1.024, "step": 1570 }, { "epoch": 0.6105119984455455, "grad_norm": 0.21107137203216553, "learning_rate": 7.816270922537953e-05, "loss": 1.0335, "step": 1571 }, { "epoch": 0.610900612066453, "grad_norm": 0.20731772482395172, "learning_rate": 7.80848579213702e-05, "loss": 1.0563, "step": 1572 }, { "epoch": 0.6112892256873603, "grad_norm": 0.19535884261131287, "learning_rate": 7.800700661736085e-05, "loss": 0.9698, "step": 1573 }, { "epoch": 0.6116778393082678, "grad_norm": 0.20449021458625793, "learning_rate": 7.79291553133515e-05, "loss": 1.0125, "step": 1574 }, { "epoch": 0.6120664529291752, "grad_norm": 0.19576509296894073, "learning_rate": 7.785130400934217e-05, "loss": 0.9326, "step": 1575 }, { "epoch": 0.6124550665500825, "grad_norm": 0.18914124369621277, "learning_rate": 7.777345270533282e-05, "loss": 0.9939, "step": 1576 }, { "epoch": 0.61284368017099, "grad_norm": 0.21239091455936432, "learning_rate": 7.769560140132348e-05, "loss": 1.0271, "step": 1577 }, { "epoch": 0.6132322937918974, "grad_norm": 0.22204811871051788, "learning_rate": 7.761775009731414e-05, "loss": 1.0524, "step": 1578 }, { "epoch": 0.6136209074128048, "grad_norm": 0.20047850906848907, "learning_rate": 7.753989879330479e-05, "loss": 1.0076, "step": 1579 }, { "epoch": 0.6140095210337122, "grad_norm": 0.22619746625423431, "learning_rate": 7.746204748929545e-05, "loss": 1.0611, "step": 1580 }, { "epoch": 0.6143981346546197, "grad_norm": 0.2500879466533661, "learning_rate": 7.73841961852861e-05, "loss": 1.0364, "step": 1581 }, { "epoch": 0.614786748275527, "grad_norm": 0.23486928641796112, "learning_rate": 7.730634488127676e-05, "loss": 1.0472, "step": 1582 }, { "epoch": 0.6151753618964345, "grad_norm": 0.19849038124084473, "learning_rate": 7.722849357726742e-05, "loss": 0.9847, "step": 1583 }, { "epoch": 0.6155639755173419, "grad_norm": 0.21516263484954834, "learning_rate": 7.715064227325807e-05, "loss": 1.0351, "step": 1584 }, { "epoch": 0.6159525891382492, "grad_norm": 0.20137760043144226, "learning_rate": 7.707279096924874e-05, "loss": 0.9879, "step": 1585 }, { "epoch": 0.6163412027591567, "grad_norm": 0.2146228402853012, "learning_rate": 7.699493966523939e-05, "loss": 1.0792, "step": 1586 }, { "epoch": 0.6167298163800641, "grad_norm": 0.19929760694503784, "learning_rate": 7.691708836123004e-05, "loss": 1.0313, "step": 1587 }, { "epoch": 0.6171184300009716, "grad_norm": 0.201123908162117, "learning_rate": 7.683923705722071e-05, "loss": 1.0279, "step": 1588 }, { "epoch": 0.6175070436218789, "grad_norm": 0.2154105007648468, "learning_rate": 7.676138575321136e-05, "loss": 1.075, "step": 1589 }, { "epoch": 0.6178956572427864, "grad_norm": 0.2028442770242691, "learning_rate": 7.668353444920203e-05, "loss": 0.9771, "step": 1590 }, { "epoch": 0.6182842708636938, "grad_norm": 0.18003074824810028, "learning_rate": 7.660568314519268e-05, "loss": 0.9677, "step": 1591 }, { "epoch": 0.6186728844846012, "grad_norm": 0.23250891268253326, "learning_rate": 7.652783184118335e-05, "loss": 1.015, "step": 1592 }, { "epoch": 0.6190614981055086, "grad_norm": 0.2047244906425476, "learning_rate": 7.6449980537174e-05, "loss": 1.0044, "step": 1593 }, { "epoch": 0.6194501117264161, "grad_norm": 0.20011259615421295, "learning_rate": 7.637212923316466e-05, "loss": 1.0089, "step": 1594 }, { "epoch": 0.6198387253473234, "grad_norm": 0.2212608903646469, "learning_rate": 7.629427792915533e-05, "loss": 1.0457, "step": 1595 }, { "epoch": 0.6202273389682308, "grad_norm": 0.22725115716457367, "learning_rate": 7.621642662514598e-05, "loss": 1.1198, "step": 1596 }, { "epoch": 0.6206159525891383, "grad_norm": 0.2065306007862091, "learning_rate": 7.613857532113663e-05, "loss": 1.0572, "step": 1597 }, { "epoch": 0.6210045662100456, "grad_norm": 0.2132783830165863, "learning_rate": 7.60607240171273e-05, "loss": 1.0332, "step": 1598 }, { "epoch": 0.6213931798309531, "grad_norm": 0.20527103543281555, "learning_rate": 7.598287271311795e-05, "loss": 1.0156, "step": 1599 }, { "epoch": 0.6217817934518605, "grad_norm": 0.23608024418354034, "learning_rate": 7.59050214091086e-05, "loss": 1.0379, "step": 1600 }, { "epoch": 0.6221704070727679, "grad_norm": 0.22227297723293304, "learning_rate": 7.582717010509927e-05, "loss": 1.0507, "step": 1601 }, { "epoch": 0.6225590206936753, "grad_norm": 0.22359615564346313, "learning_rate": 7.574931880108992e-05, "loss": 1.0705, "step": 1602 }, { "epoch": 0.6229476343145828, "grad_norm": 0.20478755235671997, "learning_rate": 7.567146749708059e-05, "loss": 1.0309, "step": 1603 }, { "epoch": 0.6233362479354901, "grad_norm": 0.2223423272371292, "learning_rate": 7.559361619307124e-05, "loss": 1.0386, "step": 1604 }, { "epoch": 0.6237248615563975, "grad_norm": 0.21232105791568756, "learning_rate": 7.551576488906189e-05, "loss": 1.0353, "step": 1605 }, { "epoch": 0.624113475177305, "grad_norm": 0.22431129217147827, "learning_rate": 7.543791358505256e-05, "loss": 1.1017, "step": 1606 }, { "epoch": 0.6245020887982123, "grad_norm": 0.20826031267642975, "learning_rate": 7.536006228104321e-05, "loss": 1.0172, "step": 1607 }, { "epoch": 0.6248907024191198, "grad_norm": 0.2803161144256592, "learning_rate": 7.528221097703387e-05, "loss": 1.0554, "step": 1608 }, { "epoch": 0.6252793160400272, "grad_norm": 0.2185174971818924, "learning_rate": 7.520435967302453e-05, "loss": 0.9842, "step": 1609 }, { "epoch": 0.6256679296609347, "grad_norm": 0.2091478854417801, "learning_rate": 7.512650836901518e-05, "loss": 0.9783, "step": 1610 }, { "epoch": 0.626056543281842, "grad_norm": 0.22342967987060547, "learning_rate": 7.504865706500584e-05, "loss": 0.9891, "step": 1611 }, { "epoch": 0.6264451569027495, "grad_norm": 0.195283442735672, "learning_rate": 7.49708057609965e-05, "loss": 0.9654, "step": 1612 }, { "epoch": 0.6268337705236569, "grad_norm": 0.21048255264759064, "learning_rate": 7.489295445698716e-05, "loss": 1.0112, "step": 1613 }, { "epoch": 0.6272223841445642, "grad_norm": 0.21405541896820068, "learning_rate": 7.481510315297781e-05, "loss": 1.0498, "step": 1614 }, { "epoch": 0.6276109977654717, "grad_norm": 0.2144453227519989, "learning_rate": 7.473725184896846e-05, "loss": 1.0487, "step": 1615 }, { "epoch": 0.627999611386379, "grad_norm": 0.21963326632976532, "learning_rate": 7.465940054495913e-05, "loss": 1.0634, "step": 1616 }, { "epoch": 0.6283882250072865, "grad_norm": 0.20100601017475128, "learning_rate": 7.458154924094978e-05, "loss": 1.0407, "step": 1617 }, { "epoch": 0.6287768386281939, "grad_norm": 0.19469478726387024, "learning_rate": 7.450369793694045e-05, "loss": 0.9923, "step": 1618 }, { "epoch": 0.6291654522491014, "grad_norm": 0.2114047408103943, "learning_rate": 7.442584663293111e-05, "loss": 1.0263, "step": 1619 }, { "epoch": 0.6295540658700087, "grad_norm": 0.21080389618873596, "learning_rate": 7.434799532892177e-05, "loss": 1.0012, "step": 1620 }, { "epoch": 0.6299426794909162, "grad_norm": 0.20366831123828888, "learning_rate": 7.427014402491243e-05, "loss": 1.0254, "step": 1621 }, { "epoch": 0.6303312931118236, "grad_norm": 0.209821879863739, "learning_rate": 7.419229272090308e-05, "loss": 0.9416, "step": 1622 }, { "epoch": 0.6307199067327309, "grad_norm": 0.2228868007659912, "learning_rate": 7.411444141689374e-05, "loss": 1.0128, "step": 1623 }, { "epoch": 0.6311085203536384, "grad_norm": 0.19673995673656464, "learning_rate": 7.40365901128844e-05, "loss": 0.9709, "step": 1624 }, { "epoch": 0.6314971339745458, "grad_norm": 0.21590839326381683, "learning_rate": 7.395873880887505e-05, "loss": 1.0251, "step": 1625 }, { "epoch": 0.6318857475954532, "grad_norm": 0.20200593769550323, "learning_rate": 7.388088750486572e-05, "loss": 1.0307, "step": 1626 }, { "epoch": 0.6322743612163606, "grad_norm": 0.19623909890651703, "learning_rate": 7.380303620085637e-05, "loss": 1.0375, "step": 1627 }, { "epoch": 0.6326629748372681, "grad_norm": 0.19878128170967102, "learning_rate": 7.372518489684702e-05, "loss": 0.9844, "step": 1628 }, { "epoch": 0.6330515884581754, "grad_norm": 0.21292422711849213, "learning_rate": 7.364733359283769e-05, "loss": 1.0228, "step": 1629 }, { "epoch": 0.6334402020790829, "grad_norm": 0.1915559619665146, "learning_rate": 7.356948228882834e-05, "loss": 0.9818, "step": 1630 }, { "epoch": 0.6338288156999903, "grad_norm": 0.2264430969953537, "learning_rate": 7.3491630984819e-05, "loss": 1.146, "step": 1631 }, { "epoch": 0.6342174293208978, "grad_norm": 0.19332270324230194, "learning_rate": 7.341377968080966e-05, "loss": 1.0007, "step": 1632 }, { "epoch": 0.6346060429418051, "grad_norm": 0.217147096991539, "learning_rate": 7.333592837680031e-05, "loss": 1.0498, "step": 1633 }, { "epoch": 0.6349946565627125, "grad_norm": 0.22200679779052734, "learning_rate": 7.325807707279098e-05, "loss": 1.0358, "step": 1634 }, { "epoch": 0.63538327018362, "grad_norm": 0.19485117495059967, "learning_rate": 7.318022576878163e-05, "loss": 0.9717, "step": 1635 }, { "epoch": 0.6357718838045273, "grad_norm": 0.20595680177211761, "learning_rate": 7.310237446477228e-05, "loss": 1.0195, "step": 1636 }, { "epoch": 0.6361604974254348, "grad_norm": 0.21184709668159485, "learning_rate": 7.302452316076294e-05, "loss": 1.0354, "step": 1637 }, { "epoch": 0.6365491110463422, "grad_norm": 0.22607794404029846, "learning_rate": 7.29466718567536e-05, "loss": 1.0217, "step": 1638 }, { "epoch": 0.6369377246672496, "grad_norm": 0.20236065983772278, "learning_rate": 7.286882055274426e-05, "loss": 1.0441, "step": 1639 }, { "epoch": 0.637326338288157, "grad_norm": 0.19979622960090637, "learning_rate": 7.279096924873491e-05, "loss": 1.0105, "step": 1640 }, { "epoch": 0.6377149519090645, "grad_norm": 0.2655459940433502, "learning_rate": 7.271311794472557e-05, "loss": 1.0726, "step": 1641 }, { "epoch": 0.6381035655299718, "grad_norm": 0.25107496976852417, "learning_rate": 7.263526664071623e-05, "loss": 1.037, "step": 1642 }, { "epoch": 0.6384921791508792, "grad_norm": 0.19250229001045227, "learning_rate": 7.255741533670688e-05, "loss": 0.9741, "step": 1643 }, { "epoch": 0.6388807927717867, "grad_norm": 0.19324181973934174, "learning_rate": 7.247956403269755e-05, "loss": 1.0333, "step": 1644 }, { "epoch": 0.639269406392694, "grad_norm": 0.22267483174800873, "learning_rate": 7.240171272868822e-05, "loss": 1.0313, "step": 1645 }, { "epoch": 0.6396580200136015, "grad_norm": 0.2775348722934723, "learning_rate": 7.232386142467887e-05, "loss": 1.0686, "step": 1646 }, { "epoch": 0.6400466336345089, "grad_norm": 0.1886623501777649, "learning_rate": 7.224601012066953e-05, "loss": 1.0029, "step": 1647 }, { "epoch": 0.6404352472554163, "grad_norm": 0.20303374528884888, "learning_rate": 7.216815881666019e-05, "loss": 1.0346, "step": 1648 }, { "epoch": 0.6408238608763237, "grad_norm": 0.20815756916999817, "learning_rate": 7.209030751265085e-05, "loss": 1.0258, "step": 1649 }, { "epoch": 0.6412124744972312, "grad_norm": 0.22055703401565552, "learning_rate": 7.20124562086415e-05, "loss": 1.0215, "step": 1650 }, { "epoch": 0.6416010881181385, "grad_norm": 0.20248562097549438, "learning_rate": 7.193460490463215e-05, "loss": 0.9979, "step": 1651 }, { "epoch": 0.6419897017390459, "grad_norm": 0.2093247026205063, "learning_rate": 7.185675360062282e-05, "loss": 1.0605, "step": 1652 }, { "epoch": 0.6423783153599534, "grad_norm": 0.22276204824447632, "learning_rate": 7.177890229661347e-05, "loss": 1.0788, "step": 1653 }, { "epoch": 0.6427669289808607, "grad_norm": 0.19959624111652374, "learning_rate": 7.170105099260412e-05, "loss": 0.9954, "step": 1654 }, { "epoch": 0.6431555426017682, "grad_norm": 0.20173248648643494, "learning_rate": 7.162319968859479e-05, "loss": 1.003, "step": 1655 }, { "epoch": 0.6435441562226756, "grad_norm": 0.207533061504364, "learning_rate": 7.154534838458544e-05, "loss": 1.043, "step": 1656 }, { "epoch": 0.643932769843583, "grad_norm": 0.21928350627422333, "learning_rate": 7.146749708057611e-05, "loss": 1.0472, "step": 1657 }, { "epoch": 0.6443213834644904, "grad_norm": 0.2567078173160553, "learning_rate": 7.138964577656676e-05, "loss": 1.0946, "step": 1658 }, { "epoch": 0.6447099970853979, "grad_norm": 0.19454176723957062, "learning_rate": 7.131179447255741e-05, "loss": 0.9437, "step": 1659 }, { "epoch": 0.6450986107063053, "grad_norm": 0.19198423624038696, "learning_rate": 7.123394316854808e-05, "loss": 0.9976, "step": 1660 }, { "epoch": 0.6454872243272126, "grad_norm": 0.1929445117712021, "learning_rate": 7.115609186453873e-05, "loss": 1.0279, "step": 1661 }, { "epoch": 0.6458758379481201, "grad_norm": 0.2041027694940567, "learning_rate": 7.10782405605294e-05, "loss": 1.0458, "step": 1662 }, { "epoch": 0.6462644515690275, "grad_norm": 0.23750995099544525, "learning_rate": 7.100038925652005e-05, "loss": 1.0916, "step": 1663 }, { "epoch": 0.6466530651899349, "grad_norm": 0.1971994787454605, "learning_rate": 7.09225379525107e-05, "loss": 0.951, "step": 1664 }, { "epoch": 0.6470416788108423, "grad_norm": 0.20459246635437012, "learning_rate": 7.084468664850136e-05, "loss": 0.9653, "step": 1665 }, { "epoch": 0.6474302924317498, "grad_norm": 0.2137187272310257, "learning_rate": 7.076683534449202e-05, "loss": 1.0291, "step": 1666 }, { "epoch": 0.6478189060526571, "grad_norm": 0.21235258877277374, "learning_rate": 7.068898404048268e-05, "loss": 1.0104, "step": 1667 }, { "epoch": 0.6482075196735646, "grad_norm": 0.23120944201946259, "learning_rate": 7.061113273647333e-05, "loss": 1.0693, "step": 1668 }, { "epoch": 0.648596133294472, "grad_norm": 1.38257896900177, "learning_rate": 7.053328143246399e-05, "loss": 1.0339, "step": 1669 }, { "epoch": 0.6489847469153793, "grad_norm": 0.20898790657520294, "learning_rate": 7.045543012845465e-05, "loss": 1.004, "step": 1670 }, { "epoch": 0.6493733605362868, "grad_norm": 0.20251236855983734, "learning_rate": 7.037757882444532e-05, "loss": 0.9992, "step": 1671 }, { "epoch": 0.6497619741571942, "grad_norm": 0.2358030527830124, "learning_rate": 7.029972752043597e-05, "loss": 0.9854, "step": 1672 }, { "epoch": 0.6501505877781016, "grad_norm": 0.18945704400539398, "learning_rate": 7.022187621642664e-05, "loss": 0.9677, "step": 1673 }, { "epoch": 0.650539201399009, "grad_norm": 0.1965213567018509, "learning_rate": 7.014402491241729e-05, "loss": 1.0118, "step": 1674 }, { "epoch": 0.6509278150199165, "grad_norm": 0.2340148687362671, "learning_rate": 7.006617360840795e-05, "loss": 1.0312, "step": 1675 }, { "epoch": 0.6513164286408238, "grad_norm": 0.1992296278476715, "learning_rate": 6.99883223043986e-05, "loss": 1.0155, "step": 1676 }, { "epoch": 0.6517050422617313, "grad_norm": 0.20410223305225372, "learning_rate": 6.991047100038926e-05, "loss": 1.0646, "step": 1677 }, { "epoch": 0.6520936558826387, "grad_norm": 0.19254536926746368, "learning_rate": 6.983261969637992e-05, "loss": 0.9538, "step": 1678 }, { "epoch": 0.6524822695035462, "grad_norm": 0.19980847835540771, "learning_rate": 6.975476839237057e-05, "loss": 0.9912, "step": 1679 }, { "epoch": 0.6528708831244535, "grad_norm": 0.19503261148929596, "learning_rate": 6.967691708836124e-05, "loss": 0.9844, "step": 1680 }, { "epoch": 0.6532594967453609, "grad_norm": 0.22375883162021637, "learning_rate": 6.959906578435189e-05, "loss": 1.1266, "step": 1681 }, { "epoch": 0.6536481103662684, "grad_norm": 0.21456514298915863, "learning_rate": 6.952121448034254e-05, "loss": 1.0902, "step": 1682 }, { "epoch": 0.6540367239871757, "grad_norm": 0.20348122715950012, "learning_rate": 6.944336317633321e-05, "loss": 1.0228, "step": 1683 }, { "epoch": 0.6544253376080832, "grad_norm": 0.21647393703460693, "learning_rate": 6.936551187232386e-05, "loss": 1.0653, "step": 1684 }, { "epoch": 0.6548139512289906, "grad_norm": 0.20160923898220062, "learning_rate": 6.928766056831453e-05, "loss": 1.0249, "step": 1685 }, { "epoch": 0.655202564849898, "grad_norm": 0.20070499181747437, "learning_rate": 6.920980926430518e-05, "loss": 1.0585, "step": 1686 }, { "epoch": 0.6555911784708054, "grad_norm": 0.2656902074813843, "learning_rate": 6.913195796029583e-05, "loss": 1.0042, "step": 1687 }, { "epoch": 0.6559797920917129, "grad_norm": 0.1934545785188675, "learning_rate": 6.90541066562865e-05, "loss": 0.9831, "step": 1688 }, { "epoch": 0.6563684057126202, "grad_norm": 0.21719245612621307, "learning_rate": 6.897625535227715e-05, "loss": 0.9934, "step": 1689 }, { "epoch": 0.6567570193335276, "grad_norm": 0.20906969904899597, "learning_rate": 6.889840404826782e-05, "loss": 1.023, "step": 1690 }, { "epoch": 0.6571456329544351, "grad_norm": 0.225227490067482, "learning_rate": 6.882055274425847e-05, "loss": 1.0265, "step": 1691 }, { "epoch": 0.6575342465753424, "grad_norm": 0.22766710817813873, "learning_rate": 6.874270144024912e-05, "loss": 1.0306, "step": 1692 }, { "epoch": 0.6579228601962499, "grad_norm": 0.20964065194129944, "learning_rate": 6.866485013623978e-05, "loss": 0.9431, "step": 1693 }, { "epoch": 0.6583114738171573, "grad_norm": 0.19821231067180634, "learning_rate": 6.858699883223044e-05, "loss": 0.9959, "step": 1694 }, { "epoch": 0.6587000874380647, "grad_norm": 0.2071307748556137, "learning_rate": 6.85091475282211e-05, "loss": 1.0332, "step": 1695 }, { "epoch": 0.6590887010589721, "grad_norm": 0.27962490916252136, "learning_rate": 6.843129622421175e-05, "loss": 0.9755, "step": 1696 }, { "epoch": 0.6594773146798796, "grad_norm": 0.21582698822021484, "learning_rate": 6.835344492020242e-05, "loss": 1.0305, "step": 1697 }, { "epoch": 0.6598659283007869, "grad_norm": 0.1872921586036682, "learning_rate": 6.827559361619307e-05, "loss": 0.9693, "step": 1698 }, { "epoch": 0.6602545419216943, "grad_norm": 0.27033379673957825, "learning_rate": 6.819774231218374e-05, "loss": 1.0756, "step": 1699 }, { "epoch": 0.6606431555426018, "grad_norm": 0.2010008543729782, "learning_rate": 6.811989100817439e-05, "loss": 1.0077, "step": 1700 }, { "epoch": 0.6610317691635091, "grad_norm": 0.20637495815753937, "learning_rate": 6.804203970416506e-05, "loss": 1.0208, "step": 1701 }, { "epoch": 0.6614203827844166, "grad_norm": 0.21331818401813507, "learning_rate": 6.796418840015571e-05, "loss": 1.0242, "step": 1702 }, { "epoch": 0.661808996405324, "grad_norm": 0.2092941552400589, "learning_rate": 6.788633709614637e-05, "loss": 1.0949, "step": 1703 }, { "epoch": 0.6621976100262315, "grad_norm": 0.22332265973091125, "learning_rate": 6.780848579213703e-05, "loss": 1.1068, "step": 1704 }, { "epoch": 0.6625862236471388, "grad_norm": 0.20077067613601685, "learning_rate": 6.773063448812768e-05, "loss": 0.9801, "step": 1705 }, { "epoch": 0.6629748372680463, "grad_norm": 0.2057008296251297, "learning_rate": 6.765278318411834e-05, "loss": 1.0058, "step": 1706 }, { "epoch": 0.6633634508889537, "grad_norm": 0.20337353646755219, "learning_rate": 6.7574931880109e-05, "loss": 1.0141, "step": 1707 }, { "epoch": 0.663752064509861, "grad_norm": 0.22756130993366241, "learning_rate": 6.749708057609966e-05, "loss": 1.0287, "step": 1708 }, { "epoch": 0.6641406781307685, "grad_norm": 0.2052423506975174, "learning_rate": 6.741922927209031e-05, "loss": 1.0069, "step": 1709 }, { "epoch": 0.6645292917516759, "grad_norm": 0.1988023817539215, "learning_rate": 6.734137796808096e-05, "loss": 0.9761, "step": 1710 }, { "epoch": 0.6649179053725833, "grad_norm": 0.20491188764572144, "learning_rate": 6.726352666407163e-05, "loss": 0.9767, "step": 1711 }, { "epoch": 0.6653065189934907, "grad_norm": 0.18790274858474731, "learning_rate": 6.718567536006228e-05, "loss": 0.9944, "step": 1712 }, { "epoch": 0.6656951326143982, "grad_norm": 0.19979891180992126, "learning_rate": 6.710782405605293e-05, "loss": 1.0842, "step": 1713 }, { "epoch": 0.6660837462353055, "grad_norm": 0.22204813361167908, "learning_rate": 6.70299727520436e-05, "loss": 1.0561, "step": 1714 }, { "epoch": 0.666472359856213, "grad_norm": 0.20182965695858002, "learning_rate": 6.695212144803425e-05, "loss": 1.0015, "step": 1715 }, { "epoch": 0.6668609734771204, "grad_norm": 0.20719997584819794, "learning_rate": 6.687427014402492e-05, "loss": 1.0144, "step": 1716 }, { "epoch": 0.6672495870980278, "grad_norm": 0.1944626122713089, "learning_rate": 6.679641884001557e-05, "loss": 1.0083, "step": 1717 }, { "epoch": 0.6676382007189352, "grad_norm": 0.2072264701128006, "learning_rate": 6.671856753600622e-05, "loss": 1.0246, "step": 1718 }, { "epoch": 0.6680268143398426, "grad_norm": 0.2134973257780075, "learning_rate": 6.664071623199689e-05, "loss": 1.0926, "step": 1719 }, { "epoch": 0.66841542796075, "grad_norm": 0.2119186669588089, "learning_rate": 6.656286492798754e-05, "loss": 1.0129, "step": 1720 }, { "epoch": 0.6688040415816574, "grad_norm": 0.21205540001392365, "learning_rate": 6.64850136239782e-05, "loss": 1.0611, "step": 1721 }, { "epoch": 0.6691926552025649, "grad_norm": 0.21632088720798492, "learning_rate": 6.640716231996886e-05, "loss": 1.0821, "step": 1722 }, { "epoch": 0.6695812688234722, "grad_norm": 0.21734434366226196, "learning_rate": 6.632931101595952e-05, "loss": 1.0821, "step": 1723 }, { "epoch": 0.6699698824443797, "grad_norm": 0.2030603289604187, "learning_rate": 6.625145971195017e-05, "loss": 0.9976, "step": 1724 }, { "epoch": 0.6703584960652871, "grad_norm": 0.19921456277370453, "learning_rate": 6.617360840794084e-05, "loss": 0.9187, "step": 1725 }, { "epoch": 0.6707471096861946, "grad_norm": 0.20548826456069946, "learning_rate": 6.60957571039315e-05, "loss": 1.0486, "step": 1726 }, { "epoch": 0.6711357233071019, "grad_norm": 0.21784676611423492, "learning_rate": 6.601790579992216e-05, "loss": 1.1089, "step": 1727 }, { "epoch": 0.6715243369280093, "grad_norm": 0.2137753963470459, "learning_rate": 6.594005449591281e-05, "loss": 1.0075, "step": 1728 }, { "epoch": 0.6719129505489168, "grad_norm": 0.20200639963150024, "learning_rate": 6.586220319190348e-05, "loss": 0.9915, "step": 1729 }, { "epoch": 0.6723015641698241, "grad_norm": 0.20898796617984772, "learning_rate": 6.578435188789413e-05, "loss": 1.0292, "step": 1730 }, { "epoch": 0.6726901777907316, "grad_norm": 0.22515977919101715, "learning_rate": 6.570650058388478e-05, "loss": 1.0118, "step": 1731 }, { "epoch": 0.673078791411639, "grad_norm": 0.2132793813943863, "learning_rate": 6.562864927987545e-05, "loss": 1.1097, "step": 1732 }, { "epoch": 0.6734674050325464, "grad_norm": 0.20358797907829285, "learning_rate": 6.55507979758661e-05, "loss": 1.0241, "step": 1733 }, { "epoch": 0.6738560186534538, "grad_norm": 0.21155016124248505, "learning_rate": 6.547294667185676e-05, "loss": 1.0235, "step": 1734 }, { "epoch": 0.6742446322743613, "grad_norm": 0.198009192943573, "learning_rate": 6.539509536784741e-05, "loss": 0.9542, "step": 1735 }, { "epoch": 0.6746332458952686, "grad_norm": 0.20318005979061127, "learning_rate": 6.531724406383807e-05, "loss": 0.9993, "step": 1736 }, { "epoch": 0.675021859516176, "grad_norm": 0.21384860575199127, "learning_rate": 6.523939275982873e-05, "loss": 1.1188, "step": 1737 }, { "epoch": 0.6754104731370835, "grad_norm": 0.18736955523490906, "learning_rate": 6.516154145581938e-05, "loss": 0.9832, "step": 1738 }, { "epoch": 0.6757990867579908, "grad_norm": 0.2002391368150711, "learning_rate": 6.508369015181005e-05, "loss": 1.0288, "step": 1739 }, { "epoch": 0.6761877003788983, "grad_norm": 0.20011006295681, "learning_rate": 6.50058388478007e-05, "loss": 0.9588, "step": 1740 }, { "epoch": 0.6765763139998057, "grad_norm": 0.20782291889190674, "learning_rate": 6.492798754379135e-05, "loss": 1.0033, "step": 1741 }, { "epoch": 0.6769649276207131, "grad_norm": 0.2056814581155777, "learning_rate": 6.485013623978202e-05, "loss": 1.0648, "step": 1742 }, { "epoch": 0.6773535412416205, "grad_norm": 0.2207457572221756, "learning_rate": 6.477228493577267e-05, "loss": 1.0758, "step": 1743 }, { "epoch": 0.677742154862528, "grad_norm": 0.20437198877334595, "learning_rate": 6.469443363176334e-05, "loss": 1.0253, "step": 1744 }, { "epoch": 0.6781307684834353, "grad_norm": 0.198721781373024, "learning_rate": 6.461658232775399e-05, "loss": 1.0087, "step": 1745 }, { "epoch": 0.6785193821043427, "grad_norm": 0.22781015932559967, "learning_rate": 6.453873102374464e-05, "loss": 1.0692, "step": 1746 }, { "epoch": 0.6789079957252502, "grad_norm": 0.21826857328414917, "learning_rate": 6.446087971973531e-05, "loss": 1.0232, "step": 1747 }, { "epoch": 0.6792966093461575, "grad_norm": 0.2156928926706314, "learning_rate": 6.438302841572596e-05, "loss": 1.0686, "step": 1748 }, { "epoch": 0.679685222967065, "grad_norm": 0.2161693125963211, "learning_rate": 6.430517711171662e-05, "loss": 1.0298, "step": 1749 }, { "epoch": 0.6800738365879724, "grad_norm": 0.19139425456523895, "learning_rate": 6.422732580770729e-05, "loss": 0.9545, "step": 1750 }, { "epoch": 0.6804624502088799, "grad_norm": 0.22626161575317383, "learning_rate": 6.414947450369794e-05, "loss": 1.0669, "step": 1751 }, { "epoch": 0.6808510638297872, "grad_norm": 0.2135801464319229, "learning_rate": 6.407162319968861e-05, "loss": 1.0187, "step": 1752 }, { "epoch": 0.6812396774506947, "grad_norm": 0.20803681015968323, "learning_rate": 6.399377189567926e-05, "loss": 1.0856, "step": 1753 }, { "epoch": 0.681628291071602, "grad_norm": 0.21317154169082642, "learning_rate": 6.391592059166991e-05, "loss": 1.1018, "step": 1754 }, { "epoch": 0.6820169046925094, "grad_norm": 0.20877891778945923, "learning_rate": 6.383806928766058e-05, "loss": 1.0383, "step": 1755 }, { "epoch": 0.6824055183134169, "grad_norm": 0.20769146084785461, "learning_rate": 6.376021798365123e-05, "loss": 1.0852, "step": 1756 }, { "epoch": 0.6827941319343243, "grad_norm": 0.2252657413482666, "learning_rate": 6.36823666796419e-05, "loss": 1.0749, "step": 1757 }, { "epoch": 0.6831827455552317, "grad_norm": 0.24453257024288177, "learning_rate": 6.360451537563255e-05, "loss": 1.1042, "step": 1758 }, { "epoch": 0.6835713591761391, "grad_norm": 0.2082965075969696, "learning_rate": 6.35266640716232e-05, "loss": 1.0729, "step": 1759 }, { "epoch": 0.6839599727970466, "grad_norm": 0.20121856033802032, "learning_rate": 6.344881276761387e-05, "loss": 1.038, "step": 1760 }, { "epoch": 0.6843485864179539, "grad_norm": 0.20096386969089508, "learning_rate": 6.337096146360452e-05, "loss": 0.9655, "step": 1761 }, { "epoch": 0.6847372000388614, "grad_norm": 0.20015959441661835, "learning_rate": 6.329311015959518e-05, "loss": 1.0187, "step": 1762 }, { "epoch": 0.6851258136597688, "grad_norm": 0.21056395769119263, "learning_rate": 6.321525885558583e-05, "loss": 1.0567, "step": 1763 }, { "epoch": 0.6855144272806762, "grad_norm": 0.2211030125617981, "learning_rate": 6.313740755157649e-05, "loss": 1.0588, "step": 1764 }, { "epoch": 0.6859030409015836, "grad_norm": 0.20809797942638397, "learning_rate": 6.305955624756715e-05, "loss": 0.9488, "step": 1765 }, { "epoch": 0.686291654522491, "grad_norm": 0.2331530600786209, "learning_rate": 6.29817049435578e-05, "loss": 1.0789, "step": 1766 }, { "epoch": 0.6866802681433984, "grad_norm": 0.21708674728870392, "learning_rate": 6.290385363954846e-05, "loss": 1.0518, "step": 1767 }, { "epoch": 0.6870688817643058, "grad_norm": 0.2088184356689453, "learning_rate": 6.282600233553912e-05, "loss": 1.0178, "step": 1768 }, { "epoch": 0.6874574953852133, "grad_norm": 0.20285943150520325, "learning_rate": 6.274815103152977e-05, "loss": 1.018, "step": 1769 }, { "epoch": 0.6878461090061206, "grad_norm": 0.211436927318573, "learning_rate": 6.267029972752044e-05, "loss": 1.0572, "step": 1770 }, { "epoch": 0.6882347226270281, "grad_norm": 0.21108384430408478, "learning_rate": 6.259244842351109e-05, "loss": 1.0227, "step": 1771 }, { "epoch": 0.6886233362479355, "grad_norm": 0.2060437649488449, "learning_rate": 6.251459711950174e-05, "loss": 1.0251, "step": 1772 }, { "epoch": 0.689011949868843, "grad_norm": 0.20819245278835297, "learning_rate": 6.243674581549241e-05, "loss": 1.0643, "step": 1773 }, { "epoch": 0.6894005634897503, "grad_norm": 0.2172113060951233, "learning_rate": 6.235889451148306e-05, "loss": 1.0869, "step": 1774 }, { "epoch": 0.6897891771106577, "grad_norm": 0.2087356299161911, "learning_rate": 6.228104320747373e-05, "loss": 1.0622, "step": 1775 }, { "epoch": 0.6901777907315652, "grad_norm": 0.1958473175764084, "learning_rate": 6.220319190346439e-05, "loss": 0.9542, "step": 1776 }, { "epoch": 0.6905664043524725, "grad_norm": 0.23630915582180023, "learning_rate": 6.212534059945504e-05, "loss": 1.0535, "step": 1777 }, { "epoch": 0.69095501797338, "grad_norm": 0.2127649188041687, "learning_rate": 6.204748929544571e-05, "loss": 0.972, "step": 1778 }, { "epoch": 0.6913436315942874, "grad_norm": 0.19873055815696716, "learning_rate": 6.196963799143636e-05, "loss": 0.9969, "step": 1779 }, { "epoch": 0.6917322452151948, "grad_norm": 0.2013067901134491, "learning_rate": 6.189178668742703e-05, "loss": 1.0399, "step": 1780 }, { "epoch": 0.6921208588361022, "grad_norm": 0.21300987899303436, "learning_rate": 6.181393538341768e-05, "loss": 1.0377, "step": 1781 }, { "epoch": 0.6925094724570097, "grad_norm": 0.21665994822978973, "learning_rate": 6.173608407940833e-05, "loss": 1.008, "step": 1782 }, { "epoch": 0.692898086077917, "grad_norm": 0.21622590720653534, "learning_rate": 6.1658232775399e-05, "loss": 1.1128, "step": 1783 }, { "epoch": 0.6932866996988244, "grad_norm": 0.2000272423028946, "learning_rate": 6.158038147138965e-05, "loss": 1.0115, "step": 1784 }, { "epoch": 0.6936753133197319, "grad_norm": 0.20774856209754944, "learning_rate": 6.15025301673803e-05, "loss": 1.066, "step": 1785 }, { "epoch": 0.6940639269406392, "grad_norm": 0.18497461080551147, "learning_rate": 6.142467886337097e-05, "loss": 0.9608, "step": 1786 }, { "epoch": 0.6944525405615467, "grad_norm": 0.19819007813930511, "learning_rate": 6.134682755936162e-05, "loss": 1.0114, "step": 1787 }, { "epoch": 0.6948411541824541, "grad_norm": 0.22013314068317413, "learning_rate": 6.126897625535229e-05, "loss": 0.976, "step": 1788 }, { "epoch": 0.6952297678033615, "grad_norm": 0.2066160887479782, "learning_rate": 6.119112495134294e-05, "loss": 1.0585, "step": 1789 }, { "epoch": 0.6956183814242689, "grad_norm": 0.21364475786685944, "learning_rate": 6.111327364733359e-05, "loss": 1.0842, "step": 1790 }, { "epoch": 0.6960069950451764, "grad_norm": 0.19731444120407104, "learning_rate": 6.103542234332425e-05, "loss": 0.9936, "step": 1791 }, { "epoch": 0.6963956086660837, "grad_norm": 0.2162671983242035, "learning_rate": 6.095757103931491e-05, "loss": 1.0446, "step": 1792 }, { "epoch": 0.6967842222869911, "grad_norm": 0.21486608684062958, "learning_rate": 6.087971973530557e-05, "loss": 1.0441, "step": 1793 }, { "epoch": 0.6971728359078986, "grad_norm": 0.20850563049316406, "learning_rate": 6.0801868431296224e-05, "loss": 1.0431, "step": 1794 }, { "epoch": 0.6975614495288059, "grad_norm": 0.20492027699947357, "learning_rate": 6.072401712728688e-05, "loss": 0.9845, "step": 1795 }, { "epoch": 0.6979500631497134, "grad_norm": 0.1986648142337799, "learning_rate": 6.064616582327754e-05, "loss": 0.9855, "step": 1796 }, { "epoch": 0.6983386767706208, "grad_norm": 0.20606310665607452, "learning_rate": 6.05683145192682e-05, "loss": 1.0608, "step": 1797 }, { "epoch": 0.6987272903915283, "grad_norm": 0.20496073365211487, "learning_rate": 6.0490463215258867e-05, "loss": 1.0311, "step": 1798 }, { "epoch": 0.6991159040124356, "grad_norm": 0.2153409719467163, "learning_rate": 6.041261191124952e-05, "loss": 1.0394, "step": 1799 }, { "epoch": 0.6995045176333431, "grad_norm": 0.21410655975341797, "learning_rate": 6.033476060724017e-05, "loss": 1.0229, "step": 1800 }, { "epoch": 0.6998931312542505, "grad_norm": 0.20418782532215118, "learning_rate": 6.0256909303230836e-05, "loss": 1.0382, "step": 1801 }, { "epoch": 0.7002817448751578, "grad_norm": 0.19154146313667297, "learning_rate": 6.017905799922149e-05, "loss": 0.9891, "step": 1802 }, { "epoch": 0.7006703584960653, "grad_norm": 0.19138328731060028, "learning_rate": 6.010120669521214e-05, "loss": 0.9638, "step": 1803 }, { "epoch": 0.7010589721169727, "grad_norm": 0.19704872369766235, "learning_rate": 6.0023355391202806e-05, "loss": 0.9835, "step": 1804 }, { "epoch": 0.7014475857378801, "grad_norm": 0.2175600379705429, "learning_rate": 5.994550408719346e-05, "loss": 1.1192, "step": 1805 }, { "epoch": 0.7018361993587875, "grad_norm": 0.21614274382591248, "learning_rate": 5.9867652783184124e-05, "loss": 1.0877, "step": 1806 }, { "epoch": 0.702224812979695, "grad_norm": 0.20461414754390717, "learning_rate": 5.9789801479174776e-05, "loss": 0.9706, "step": 1807 }, { "epoch": 0.7026134266006023, "grad_norm": 0.1989748477935791, "learning_rate": 5.9711950175165434e-05, "loss": 1.0004, "step": 1808 }, { "epoch": 0.7030020402215098, "grad_norm": 0.21304792165756226, "learning_rate": 5.963409887115609e-05, "loss": 1.0177, "step": 1809 }, { "epoch": 0.7033906538424172, "grad_norm": 0.19023855030536652, "learning_rate": 5.955624756714675e-05, "loss": 0.9759, "step": 1810 }, { "epoch": 0.7037792674633246, "grad_norm": 0.21915188431739807, "learning_rate": 5.947839626313742e-05, "loss": 1.0621, "step": 1811 }, { "epoch": 0.704167881084232, "grad_norm": 0.21626822650432587, "learning_rate": 5.940054495912807e-05, "loss": 1.0144, "step": 1812 }, { "epoch": 0.7045564947051394, "grad_norm": 0.20742040872573853, "learning_rate": 5.932269365511872e-05, "loss": 0.9778, "step": 1813 }, { "epoch": 0.7049451083260468, "grad_norm": 0.2172158658504486, "learning_rate": 5.924484235110939e-05, "loss": 1.0416, "step": 1814 }, { "epoch": 0.7053337219469542, "grad_norm": 0.209465891122818, "learning_rate": 5.916699104710004e-05, "loss": 1.0378, "step": 1815 }, { "epoch": 0.7057223355678617, "grad_norm": 0.2097882628440857, "learning_rate": 5.9089139743090705e-05, "loss": 1.0166, "step": 1816 }, { "epoch": 0.706110949188769, "grad_norm": 0.2251904308795929, "learning_rate": 5.901128843908136e-05, "loss": 1.0783, "step": 1817 }, { "epoch": 0.7064995628096765, "grad_norm": 0.1952916979789734, "learning_rate": 5.893343713507201e-05, "loss": 0.993, "step": 1818 }, { "epoch": 0.7068881764305839, "grad_norm": 0.20997455716133118, "learning_rate": 5.8855585831062675e-05, "loss": 1.0448, "step": 1819 }, { "epoch": 0.7072767900514914, "grad_norm": 0.20070020854473114, "learning_rate": 5.877773452705333e-05, "loss": 0.9603, "step": 1820 }, { "epoch": 0.7076654036723987, "grad_norm": 0.25765034556388855, "learning_rate": 5.869988322304399e-05, "loss": 1.0361, "step": 1821 }, { "epoch": 0.7080540172933061, "grad_norm": 0.21948982775211334, "learning_rate": 5.862203191903465e-05, "loss": 1.0668, "step": 1822 }, { "epoch": 0.7084426309142136, "grad_norm": 0.1867108792066574, "learning_rate": 5.85441806150253e-05, "loss": 0.9372, "step": 1823 }, { "epoch": 0.7088312445351209, "grad_norm": 0.2037520408630371, "learning_rate": 5.846632931101597e-05, "loss": 0.9905, "step": 1824 }, { "epoch": 0.7092198581560284, "grad_norm": 0.21352072060108185, "learning_rate": 5.838847800700662e-05, "loss": 1.0514, "step": 1825 }, { "epoch": 0.7096084717769358, "grad_norm": 0.1949845850467682, "learning_rate": 5.831062670299727e-05, "loss": 0.9636, "step": 1826 }, { "epoch": 0.7099970853978432, "grad_norm": 0.2092294692993164, "learning_rate": 5.823277539898794e-05, "loss": 1.0361, "step": 1827 }, { "epoch": 0.7103856990187506, "grad_norm": 0.20054267346858978, "learning_rate": 5.815492409497859e-05, "loss": 1.0195, "step": 1828 }, { "epoch": 0.7107743126396581, "grad_norm": 0.2202107012271881, "learning_rate": 5.8077072790969256e-05, "loss": 1.0918, "step": 1829 }, { "epoch": 0.7111629262605654, "grad_norm": 0.2001042366027832, "learning_rate": 5.799922148695991e-05, "loss": 1.0142, "step": 1830 }, { "epoch": 0.7115515398814728, "grad_norm": 0.2102631777524948, "learning_rate": 5.792137018295056e-05, "loss": 1.0231, "step": 1831 }, { "epoch": 0.7119401535023803, "grad_norm": 0.21717461943626404, "learning_rate": 5.7843518878941226e-05, "loss": 1.0295, "step": 1832 }, { "epoch": 0.7123287671232876, "grad_norm": 0.2001933753490448, "learning_rate": 5.776566757493188e-05, "loss": 1.022, "step": 1833 }, { "epoch": 0.7127173807441951, "grad_norm": 0.2218201756477356, "learning_rate": 5.7687816270922544e-05, "loss": 1.0762, "step": 1834 }, { "epoch": 0.7131059943651025, "grad_norm": 0.20680001378059387, "learning_rate": 5.76099649669132e-05, "loss": 1.0017, "step": 1835 }, { "epoch": 0.7134946079860099, "grad_norm": 0.21511508524417877, "learning_rate": 5.7532113662903854e-05, "loss": 1.048, "step": 1836 }, { "epoch": 0.7138832216069173, "grad_norm": 0.19720061123371124, "learning_rate": 5.745426235889452e-05, "loss": 0.9983, "step": 1837 }, { "epoch": 0.7142718352278248, "grad_norm": 0.2005409449338913, "learning_rate": 5.737641105488517e-05, "loss": 0.9941, "step": 1838 }, { "epoch": 0.7146604488487321, "grad_norm": 0.2222924679517746, "learning_rate": 5.729855975087584e-05, "loss": 1.0476, "step": 1839 }, { "epoch": 0.7150490624696395, "grad_norm": 0.21131208539009094, "learning_rate": 5.722070844686649e-05, "loss": 1.0208, "step": 1840 }, { "epoch": 0.715437676090547, "grad_norm": 0.2307305932044983, "learning_rate": 5.714285714285714e-05, "loss": 0.9867, "step": 1841 }, { "epoch": 0.7158262897114543, "grad_norm": 0.1974973827600479, "learning_rate": 5.706500583884781e-05, "loss": 1.0285, "step": 1842 }, { "epoch": 0.7162149033323618, "grad_norm": 0.2006559520959854, "learning_rate": 5.698715453483846e-05, "loss": 1.024, "step": 1843 }, { "epoch": 0.7166035169532692, "grad_norm": 0.21160584688186646, "learning_rate": 5.690930323082911e-05, "loss": 1.0256, "step": 1844 }, { "epoch": 0.7169921305741767, "grad_norm": 0.28184664249420166, "learning_rate": 5.683145192681978e-05, "loss": 1.0443, "step": 1845 }, { "epoch": 0.717380744195084, "grad_norm": 0.2206653356552124, "learning_rate": 5.675360062281043e-05, "loss": 1.0458, "step": 1846 }, { "epoch": 0.7177693578159915, "grad_norm": 0.21346066892147064, "learning_rate": 5.6675749318801095e-05, "loss": 1.0106, "step": 1847 }, { "epoch": 0.7181579714368989, "grad_norm": 0.20931747555732727, "learning_rate": 5.6597898014791753e-05, "loss": 0.9831, "step": 1848 }, { "epoch": 0.7185465850578063, "grad_norm": 0.2026771456003189, "learning_rate": 5.6520046710782406e-05, "loss": 1.0162, "step": 1849 }, { "epoch": 0.7189351986787137, "grad_norm": 0.21388716995716095, "learning_rate": 5.644219540677307e-05, "loss": 1.0867, "step": 1850 }, { "epoch": 0.7193238122996211, "grad_norm": 0.2039308398962021, "learning_rate": 5.636434410276372e-05, "loss": 1.0325, "step": 1851 }, { "epoch": 0.7197124259205285, "grad_norm": 0.21741114556789398, "learning_rate": 5.628649279875439e-05, "loss": 1.0251, "step": 1852 }, { "epoch": 0.7201010395414359, "grad_norm": 0.21343208849430084, "learning_rate": 5.620864149474504e-05, "loss": 1.0766, "step": 1853 }, { "epoch": 0.7204896531623434, "grad_norm": 0.21712560951709747, "learning_rate": 5.613079019073569e-05, "loss": 1.0643, "step": 1854 }, { "epoch": 0.7208782667832507, "grad_norm": 0.2176978886127472, "learning_rate": 5.605293888672636e-05, "loss": 1.0375, "step": 1855 }, { "epoch": 0.7212668804041582, "grad_norm": 0.2065533846616745, "learning_rate": 5.597508758271701e-05, "loss": 1.0385, "step": 1856 }, { "epoch": 0.7216554940250656, "grad_norm": 0.2169170081615448, "learning_rate": 5.5897236278707676e-05, "loss": 1.0197, "step": 1857 }, { "epoch": 0.722044107645973, "grad_norm": 0.2047201544046402, "learning_rate": 5.581938497469833e-05, "loss": 0.9794, "step": 1858 }, { "epoch": 0.7224327212668804, "grad_norm": 0.20898981392383575, "learning_rate": 5.574153367068898e-05, "loss": 1.032, "step": 1859 }, { "epoch": 0.7228213348877878, "grad_norm": 0.2090533971786499, "learning_rate": 5.5663682366679646e-05, "loss": 1.0694, "step": 1860 }, { "epoch": 0.7232099485086952, "grad_norm": 0.21963149309158325, "learning_rate": 5.5585831062670305e-05, "loss": 1.0367, "step": 1861 }, { "epoch": 0.7235985621296026, "grad_norm": 0.1974373459815979, "learning_rate": 5.550797975866096e-05, "loss": 1.0402, "step": 1862 }, { "epoch": 0.7239871757505101, "grad_norm": 0.1924194097518921, "learning_rate": 5.543012845465162e-05, "loss": 0.9647, "step": 1863 }, { "epoch": 0.7243757893714174, "grad_norm": 0.21366077661514282, "learning_rate": 5.5352277150642274e-05, "loss": 1.0139, "step": 1864 }, { "epoch": 0.7247644029923249, "grad_norm": 0.21722929179668427, "learning_rate": 5.527442584663294e-05, "loss": 1.0366, "step": 1865 }, { "epoch": 0.7251530166132323, "grad_norm": 0.20646587014198303, "learning_rate": 5.519657454262359e-05, "loss": 1.0465, "step": 1866 }, { "epoch": 0.7255416302341398, "grad_norm": 0.19144394993782043, "learning_rate": 5.5118723238614244e-05, "loss": 0.9645, "step": 1867 }, { "epoch": 0.7259302438550471, "grad_norm": 0.19553838670253754, "learning_rate": 5.504087193460491e-05, "loss": 0.98, "step": 1868 }, { "epoch": 0.7263188574759545, "grad_norm": 0.21739792823791504, "learning_rate": 5.496302063059556e-05, "loss": 1.002, "step": 1869 }, { "epoch": 0.726707471096862, "grad_norm": 0.1910562962293625, "learning_rate": 5.488516932658623e-05, "loss": 0.985, "step": 1870 }, { "epoch": 0.7270960847177693, "grad_norm": 0.2133384346961975, "learning_rate": 5.480731802257688e-05, "loss": 1.0325, "step": 1871 }, { "epoch": 0.7274846983386768, "grad_norm": 0.21884119510650635, "learning_rate": 5.472946671856753e-05, "loss": 1.0412, "step": 1872 }, { "epoch": 0.7278733119595842, "grad_norm": 0.21069306135177612, "learning_rate": 5.46516154145582e-05, "loss": 1.0474, "step": 1873 }, { "epoch": 0.7282619255804916, "grad_norm": 0.19266243278980255, "learning_rate": 5.4573764110548856e-05, "loss": 0.9941, "step": 1874 }, { "epoch": 0.728650539201399, "grad_norm": 0.21255099773406982, "learning_rate": 5.4495912806539515e-05, "loss": 1.0211, "step": 1875 }, { "epoch": 0.7290391528223065, "grad_norm": 0.1924402117729187, "learning_rate": 5.4418061502530173e-05, "loss": 1.0117, "step": 1876 }, { "epoch": 0.7294277664432138, "grad_norm": 0.2019895315170288, "learning_rate": 5.4340210198520825e-05, "loss": 0.9921, "step": 1877 }, { "epoch": 0.7298163800641212, "grad_norm": 0.20398026704788208, "learning_rate": 5.426235889451149e-05, "loss": 1.0423, "step": 1878 }, { "epoch": 0.7302049936850287, "grad_norm": 0.20153217017650604, "learning_rate": 5.418450759050214e-05, "loss": 1.0333, "step": 1879 }, { "epoch": 0.730593607305936, "grad_norm": 0.21259640157222748, "learning_rate": 5.4106656286492795e-05, "loss": 1.0689, "step": 1880 }, { "epoch": 0.7309822209268435, "grad_norm": 0.2037276029586792, "learning_rate": 5.402880498248346e-05, "loss": 1.0203, "step": 1881 }, { "epoch": 0.7313708345477509, "grad_norm": 0.19976729154586792, "learning_rate": 5.395095367847411e-05, "loss": 1.0173, "step": 1882 }, { "epoch": 0.7317594481686583, "grad_norm": 0.20481806993484497, "learning_rate": 5.387310237446478e-05, "loss": 0.9864, "step": 1883 }, { "epoch": 0.7321480617895657, "grad_norm": 0.21900932490825653, "learning_rate": 5.379525107045543e-05, "loss": 1.0519, "step": 1884 }, { "epoch": 0.7325366754104732, "grad_norm": 0.200319305062294, "learning_rate": 5.371739976644609e-05, "loss": 1.0834, "step": 1885 }, { "epoch": 0.7329252890313805, "grad_norm": 0.19662296772003174, "learning_rate": 5.363954846243675e-05, "loss": 0.9794, "step": 1886 }, { "epoch": 0.7333139026522879, "grad_norm": 0.2113952785730362, "learning_rate": 5.356169715842741e-05, "loss": 1.0763, "step": 1887 }, { "epoch": 0.7337025162731954, "grad_norm": 0.21348755061626434, "learning_rate": 5.3483845854418066e-05, "loss": 1.0781, "step": 1888 }, { "epoch": 0.7340911298941027, "grad_norm": 0.20673702657222748, "learning_rate": 5.3405994550408725e-05, "loss": 1.0513, "step": 1889 }, { "epoch": 0.7344797435150102, "grad_norm": 0.210855171084404, "learning_rate": 5.332814324639938e-05, "loss": 0.9972, "step": 1890 }, { "epoch": 0.7348683571359176, "grad_norm": 0.2136204093694687, "learning_rate": 5.325029194239004e-05, "loss": 1.03, "step": 1891 }, { "epoch": 0.7352569707568251, "grad_norm": 0.20035260915756226, "learning_rate": 5.3172440638380694e-05, "loss": 0.9739, "step": 1892 }, { "epoch": 0.7356455843777324, "grad_norm": 0.1943352371454239, "learning_rate": 5.309458933437136e-05, "loss": 0.9411, "step": 1893 }, { "epoch": 0.7360341979986399, "grad_norm": 0.3994326889514923, "learning_rate": 5.301673803036201e-05, "loss": 1.0714, "step": 1894 }, { "epoch": 0.7364228116195473, "grad_norm": 0.21691356599330902, "learning_rate": 5.2938886726352664e-05, "loss": 1.0648, "step": 1895 }, { "epoch": 0.7368114252404547, "grad_norm": 0.19853095710277557, "learning_rate": 5.286103542234333e-05, "loss": 0.983, "step": 1896 }, { "epoch": 0.7372000388613621, "grad_norm": 0.21836897730827332, "learning_rate": 5.278318411833398e-05, "loss": 1.0396, "step": 1897 }, { "epoch": 0.7375886524822695, "grad_norm": 0.19596605002880096, "learning_rate": 5.270533281432464e-05, "loss": 0.9593, "step": 1898 }, { "epoch": 0.7379772661031769, "grad_norm": 0.2141752541065216, "learning_rate": 5.26274815103153e-05, "loss": 1.0373, "step": 1899 }, { "epoch": 0.7383658797240843, "grad_norm": 0.20552939176559448, "learning_rate": 5.254963020630596e-05, "loss": 1.0352, "step": 1900 }, { "epoch": 0.7387544933449918, "grad_norm": 0.2095794975757599, "learning_rate": 5.247177890229662e-05, "loss": 1.0632, "step": 1901 }, { "epoch": 0.7391431069658991, "grad_norm": 0.19894710183143616, "learning_rate": 5.2393927598287276e-05, "loss": 0.9886, "step": 1902 }, { "epoch": 0.7395317205868066, "grad_norm": 0.22996319830417633, "learning_rate": 5.231607629427793e-05, "loss": 1.0826, "step": 1903 }, { "epoch": 0.739920334207714, "grad_norm": 0.21416957676410675, "learning_rate": 5.2238224990268593e-05, "loss": 1.0161, "step": 1904 }, { "epoch": 0.7403089478286214, "grad_norm": 0.21819345653057098, "learning_rate": 5.2160373686259245e-05, "loss": 1.0458, "step": 1905 }, { "epoch": 0.7406975614495288, "grad_norm": 0.21327044069766998, "learning_rate": 5.208252238224991e-05, "loss": 1.0721, "step": 1906 }, { "epoch": 0.7410861750704362, "grad_norm": 0.21436645090579987, "learning_rate": 5.200467107824056e-05, "loss": 1.0743, "step": 1907 }, { "epoch": 0.7414747886913436, "grad_norm": 0.215640127658844, "learning_rate": 5.1926819774231215e-05, "loss": 1.0274, "step": 1908 }, { "epoch": 0.741863402312251, "grad_norm": 0.2043589949607849, "learning_rate": 5.184896847022188e-05, "loss": 1.0618, "step": 1909 }, { "epoch": 0.7422520159331585, "grad_norm": 0.2014230340719223, "learning_rate": 5.177111716621253e-05, "loss": 0.9892, "step": 1910 }, { "epoch": 0.7426406295540658, "grad_norm": 0.19954468309879303, "learning_rate": 5.16932658622032e-05, "loss": 0.9815, "step": 1911 }, { "epoch": 0.7430292431749733, "grad_norm": 0.23119708895683289, "learning_rate": 5.161541455819385e-05, "loss": 1.0783, "step": 1912 }, { "epoch": 0.7434178567958807, "grad_norm": 0.20650482177734375, "learning_rate": 5.153756325418451e-05, "loss": 1.0162, "step": 1913 }, { "epoch": 0.7438064704167882, "grad_norm": 0.20021970570087433, "learning_rate": 5.145971195017517e-05, "loss": 1.0062, "step": 1914 }, { "epoch": 0.7441950840376955, "grad_norm": 0.23300811648368835, "learning_rate": 5.138186064616583e-05, "loss": 1.0049, "step": 1915 }, { "epoch": 0.7445836976586029, "grad_norm": 0.23268327116966248, "learning_rate": 5.130400934215648e-05, "loss": 1.0138, "step": 1916 }, { "epoch": 0.7449723112795104, "grad_norm": 0.20413407683372498, "learning_rate": 5.1226158038147145e-05, "loss": 0.9903, "step": 1917 }, { "epoch": 0.7453609249004177, "grad_norm": 0.20714978873729706, "learning_rate": 5.1148306734137797e-05, "loss": 1.0374, "step": 1918 }, { "epoch": 0.7457495385213252, "grad_norm": 0.2000850886106491, "learning_rate": 5.107045543012846e-05, "loss": 0.9885, "step": 1919 }, { "epoch": 0.7461381521422326, "grad_norm": 0.2054719179868698, "learning_rate": 5.0992604126119114e-05, "loss": 1.0551, "step": 1920 }, { "epoch": 0.74652676576314, "grad_norm": 0.2351357489824295, "learning_rate": 5.0914752822109766e-05, "loss": 1.0693, "step": 1921 }, { "epoch": 0.7469153793840474, "grad_norm": 0.22370338439941406, "learning_rate": 5.083690151810043e-05, "loss": 0.9781, "step": 1922 }, { "epoch": 0.7473039930049549, "grad_norm": 0.18734332919120789, "learning_rate": 5.0759050214091084e-05, "loss": 0.9329, "step": 1923 }, { "epoch": 0.7476926066258622, "grad_norm": 0.22099906206130981, "learning_rate": 5.068119891008175e-05, "loss": 1.0498, "step": 1924 }, { "epoch": 0.7480812202467696, "grad_norm": 0.20144490897655487, "learning_rate": 5.06033476060724e-05, "loss": 0.9865, "step": 1925 }, { "epoch": 0.7484698338676771, "grad_norm": 0.21770039200782776, "learning_rate": 5.052549630206306e-05, "loss": 1.0867, "step": 1926 }, { "epoch": 0.7488584474885844, "grad_norm": 0.19649921357631683, "learning_rate": 5.044764499805372e-05, "loss": 0.9887, "step": 1927 }, { "epoch": 0.7492470611094919, "grad_norm": 0.1940620392560959, "learning_rate": 5.036979369404438e-05, "loss": 1.0073, "step": 1928 }, { "epoch": 0.7496356747303993, "grad_norm": 0.20987650752067566, "learning_rate": 5.0291942390035044e-05, "loss": 1.046, "step": 1929 }, { "epoch": 0.7500242883513067, "grad_norm": 0.2116398960351944, "learning_rate": 5.0214091086025696e-05, "loss": 1.0423, "step": 1930 }, { "epoch": 0.7504129019722141, "grad_norm": 0.18996965885162354, "learning_rate": 5.013623978201635e-05, "loss": 0.9822, "step": 1931 }, { "epoch": 0.7508015155931216, "grad_norm": 0.20942547917366028, "learning_rate": 5.005838847800701e-05, "loss": 1.0472, "step": 1932 }, { "epoch": 0.751190129214029, "grad_norm": 0.19006839394569397, "learning_rate": 4.9980537173997665e-05, "loss": 0.993, "step": 1933 }, { "epoch": 0.7515787428349364, "grad_norm": 0.21508941054344177, "learning_rate": 4.9902685869988324e-05, "loss": 1.0406, "step": 1934 }, { "epoch": 0.7519673564558438, "grad_norm": 0.1989334225654602, "learning_rate": 4.982483456597898e-05, "loss": 0.9997, "step": 1935 }, { "epoch": 0.7523559700767511, "grad_norm": 0.19993600249290466, "learning_rate": 4.974698326196964e-05, "loss": 1.0139, "step": 1936 }, { "epoch": 0.7527445836976586, "grad_norm": 0.20927831530570984, "learning_rate": 4.9669131957960294e-05, "loss": 0.995, "step": 1937 }, { "epoch": 0.753133197318566, "grad_norm": 0.20963850617408752, "learning_rate": 4.959128065395095e-05, "loss": 1.0678, "step": 1938 }, { "epoch": 0.7535218109394735, "grad_norm": 0.19523034989833832, "learning_rate": 4.951342934994161e-05, "loss": 0.9883, "step": 1939 }, { "epoch": 0.7539104245603808, "grad_norm": 0.21588142216205597, "learning_rate": 4.943557804593227e-05, "loss": 1.0398, "step": 1940 }, { "epoch": 0.7542990381812883, "grad_norm": 0.19894704222679138, "learning_rate": 4.935772674192293e-05, "loss": 1.0125, "step": 1941 }, { "epoch": 0.7546876518021957, "grad_norm": 0.2155168056488037, "learning_rate": 4.927987543791359e-05, "loss": 1.0447, "step": 1942 }, { "epoch": 0.7550762654231031, "grad_norm": 0.212605819106102, "learning_rate": 4.920202413390425e-05, "loss": 1.077, "step": 1943 }, { "epoch": 0.7554648790440105, "grad_norm": 0.2168148010969162, "learning_rate": 4.9124172829894906e-05, "loss": 1.0029, "step": 1944 }, { "epoch": 0.7558534926649179, "grad_norm": 0.2020149528980255, "learning_rate": 4.9046321525885565e-05, "loss": 1.0684, "step": 1945 }, { "epoch": 0.7562421062858253, "grad_norm": 0.21063408255577087, "learning_rate": 4.8968470221876217e-05, "loss": 1.0147, "step": 1946 }, { "epoch": 0.7566307199067327, "grad_norm": 0.19599388539791107, "learning_rate": 4.8890618917866875e-05, "loss": 0.9719, "step": 1947 }, { "epoch": 0.7570193335276402, "grad_norm": 0.2158602923154831, "learning_rate": 4.8812767613857534e-05, "loss": 1.0439, "step": 1948 }, { "epoch": 0.7574079471485475, "grad_norm": 0.21013815701007843, "learning_rate": 4.873491630984819e-05, "loss": 1.0319, "step": 1949 }, { "epoch": 0.757796560769455, "grad_norm": 0.2020798772573471, "learning_rate": 4.8657065005838845e-05, "loss": 1.0037, "step": 1950 }, { "epoch": 0.7581851743903624, "grad_norm": 0.21202047169208527, "learning_rate": 4.8579213701829504e-05, "loss": 0.9823, "step": 1951 }, { "epoch": 0.7585737880112698, "grad_norm": 0.20750083029270172, "learning_rate": 4.850136239782016e-05, "loss": 1.0073, "step": 1952 }, { "epoch": 0.7589624016321772, "grad_norm": 0.20938372611999512, "learning_rate": 4.842351109381083e-05, "loss": 1.0326, "step": 1953 }, { "epoch": 0.7593510152530846, "grad_norm": 0.21984544396400452, "learning_rate": 4.834565978980149e-05, "loss": 1.0363, "step": 1954 }, { "epoch": 0.759739628873992, "grad_norm": 0.20306189358234406, "learning_rate": 4.826780848579214e-05, "loss": 1.0374, "step": 1955 }, { "epoch": 0.7601282424948994, "grad_norm": 0.20631705224514008, "learning_rate": 4.81899571817828e-05, "loss": 1.0985, "step": 1956 }, { "epoch": 0.7605168561158069, "grad_norm": 0.22092190384864807, "learning_rate": 4.811210587777346e-05, "loss": 1.0216, "step": 1957 }, { "epoch": 0.7609054697367142, "grad_norm": 0.21419481933116913, "learning_rate": 4.8034254573764116e-05, "loss": 1.0327, "step": 1958 }, { "epoch": 0.7612940833576217, "grad_norm": 0.1954476237297058, "learning_rate": 4.795640326975477e-05, "loss": 1.0139, "step": 1959 }, { "epoch": 0.7616826969785291, "grad_norm": 0.21092113852500916, "learning_rate": 4.7878551965745427e-05, "loss": 1.0934, "step": 1960 }, { "epoch": 0.7620713105994366, "grad_norm": 0.1998988837003708, "learning_rate": 4.7800700661736085e-05, "loss": 0.9782, "step": 1961 }, { "epoch": 0.7624599242203439, "grad_norm": 0.20410674810409546, "learning_rate": 4.7722849357726744e-05, "loss": 1.0186, "step": 1962 }, { "epoch": 0.7628485378412513, "grad_norm": 0.25312289595603943, "learning_rate": 4.76449980537174e-05, "loss": 1.0103, "step": 1963 }, { "epoch": 0.7632371514621588, "grad_norm": 0.20648318529129028, "learning_rate": 4.7567146749708055e-05, "loss": 1.0314, "step": 1964 }, { "epoch": 0.7636257650830661, "grad_norm": 0.20513702929019928, "learning_rate": 4.7489295445698714e-05, "loss": 0.981, "step": 1965 }, { "epoch": 0.7640143787039736, "grad_norm": 0.20063039660453796, "learning_rate": 4.741144414168938e-05, "loss": 1.0218, "step": 1966 }, { "epoch": 0.764402992324881, "grad_norm": 0.20328521728515625, "learning_rate": 4.733359283768004e-05, "loss": 1.0614, "step": 1967 }, { "epoch": 0.7647916059457884, "grad_norm": 0.2209623008966446, "learning_rate": 4.725574153367069e-05, "loss": 1.0478, "step": 1968 }, { "epoch": 0.7651802195666958, "grad_norm": 0.2023559957742691, "learning_rate": 4.717789022966135e-05, "loss": 1.0455, "step": 1969 }, { "epoch": 0.7655688331876033, "grad_norm": 0.20461297035217285, "learning_rate": 4.710003892565201e-05, "loss": 0.9427, "step": 1970 }, { "epoch": 0.7659574468085106, "grad_norm": 0.2108335793018341, "learning_rate": 4.702218762164267e-05, "loss": 1.0344, "step": 1971 }, { "epoch": 0.766346060429418, "grad_norm": 0.20883473753929138, "learning_rate": 4.6944336317633326e-05, "loss": 1.0336, "step": 1972 }, { "epoch": 0.7667346740503255, "grad_norm": 0.20144741237163544, "learning_rate": 4.686648501362398e-05, "loss": 1.0101, "step": 1973 }, { "epoch": 0.7671232876712328, "grad_norm": 0.21269328892230988, "learning_rate": 4.6788633709614637e-05, "loss": 0.9989, "step": 1974 }, { "epoch": 0.7675119012921403, "grad_norm": 0.20673738420009613, "learning_rate": 4.6710782405605295e-05, "loss": 1.0235, "step": 1975 }, { "epoch": 0.7679005149130477, "grad_norm": 0.1966594159603119, "learning_rate": 4.6632931101595954e-05, "loss": 1.0081, "step": 1976 }, { "epoch": 0.7682891285339551, "grad_norm": 0.22186829149723053, "learning_rate": 4.6555079797586606e-05, "loss": 1.0081, "step": 1977 }, { "epoch": 0.7686777421548625, "grad_norm": 0.20602557063102722, "learning_rate": 4.6477228493577265e-05, "loss": 1.0381, "step": 1978 }, { "epoch": 0.76906635577577, "grad_norm": 0.19581305980682373, "learning_rate": 4.639937718956793e-05, "loss": 1.0196, "step": 1979 }, { "epoch": 0.7694549693966773, "grad_norm": 0.20162086188793182, "learning_rate": 4.632152588555859e-05, "loss": 1.0168, "step": 1980 }, { "epoch": 0.7698435830175848, "grad_norm": 0.21967145800590515, "learning_rate": 4.624367458154925e-05, "loss": 1.0339, "step": 1981 }, { "epoch": 0.7702321966384922, "grad_norm": 0.20245851576328278, "learning_rate": 4.61658232775399e-05, "loss": 1.0349, "step": 1982 }, { "epoch": 0.7706208102593995, "grad_norm": 0.20409934222698212, "learning_rate": 4.608797197353056e-05, "loss": 1.0296, "step": 1983 }, { "epoch": 0.771009423880307, "grad_norm": 0.19757163524627686, "learning_rate": 4.601012066952122e-05, "loss": 1.0443, "step": 1984 }, { "epoch": 0.7713980375012144, "grad_norm": 0.20038221776485443, "learning_rate": 4.593226936551188e-05, "loss": 1.0431, "step": 1985 }, { "epoch": 0.7717866511221219, "grad_norm": 0.2112458199262619, "learning_rate": 4.585441806150253e-05, "loss": 1.0553, "step": 1986 }, { "epoch": 0.7721752647430292, "grad_norm": 0.21868042647838593, "learning_rate": 4.577656675749319e-05, "loss": 1.0061, "step": 1987 }, { "epoch": 0.7725638783639367, "grad_norm": 0.22484582662582397, "learning_rate": 4.5698715453483846e-05, "loss": 1.0831, "step": 1988 }, { "epoch": 0.7729524919848441, "grad_norm": 0.20265011489391327, "learning_rate": 4.5620864149474505e-05, "loss": 1.0206, "step": 1989 }, { "epoch": 0.7733411056057515, "grad_norm": 0.2052810937166214, "learning_rate": 4.5543012845465164e-05, "loss": 1.0366, "step": 1990 }, { "epoch": 0.7737297192266589, "grad_norm": 0.21016088128089905, "learning_rate": 4.546516154145582e-05, "loss": 0.9963, "step": 1991 }, { "epoch": 0.7741183328475663, "grad_norm": 0.19719412922859192, "learning_rate": 4.538731023744648e-05, "loss": 0.9853, "step": 1992 }, { "epoch": 0.7745069464684737, "grad_norm": 0.20447245240211487, "learning_rate": 4.530945893343714e-05, "loss": 0.9977, "step": 1993 }, { "epoch": 0.7748955600893811, "grad_norm": 0.21796588599681854, "learning_rate": 4.52316076294278e-05, "loss": 1.0949, "step": 1994 }, { "epoch": 0.7752841737102886, "grad_norm": 0.2041284590959549, "learning_rate": 4.515375632541845e-05, "loss": 1.0034, "step": 1995 }, { "epoch": 0.7756727873311959, "grad_norm": 0.21134726703166962, "learning_rate": 4.507590502140911e-05, "loss": 1.0076, "step": 1996 }, { "epoch": 0.7760614009521034, "grad_norm": 0.20730996131896973, "learning_rate": 4.499805371739977e-05, "loss": 1.0456, "step": 1997 }, { "epoch": 0.7764500145730108, "grad_norm": 0.22316931188106537, "learning_rate": 4.492020241339043e-05, "loss": 0.9418, "step": 1998 }, { "epoch": 0.7768386281939182, "grad_norm": 0.21494819223880768, "learning_rate": 4.484235110938109e-05, "loss": 1.0597, "step": 1999 }, { "epoch": 0.7772272418148256, "grad_norm": 0.20344491302967072, "learning_rate": 4.476449980537174e-05, "loss": 0.9749, "step": 2000 } ], "logging_steps": 1, "max_steps": 2574, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.7079644445254222e+19, "train_batch_size": 4, "trial_name": null, "trial_params": null }