{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 0.09026900162484203, "eval_steps": 500, "global_step": 10000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 9.026900162484202e-06, "grad_norm": 7.8362650871276855, "learning_rate": 5e-05, "loss": 5.5826, "step": 1 }, { "epoch": 1.8053800324968405e-05, "grad_norm": 12.520882606506348, "learning_rate": 4.999954865499188e-05, "loss": 5.0028, "step": 2 }, { "epoch": 2.7080700487452607e-05, "grad_norm": 9.94706916809082, "learning_rate": 4.9999097309983754e-05, "loss": 5.4074, "step": 3 }, { "epoch": 3.610760064993681e-05, "grad_norm": 8.310630798339844, "learning_rate": 4.999864596497563e-05, "loss": 5.3248, "step": 4 }, { "epoch": 4.513450081242101e-05, "grad_norm": 7.081541061401367, "learning_rate": 4.9998194619967506e-05, "loss": 5.6277, "step": 5 }, { "epoch": 5.4161400974905215e-05, "grad_norm": 8.578364372253418, "learning_rate": 4.999774327495938e-05, "loss": 5.2087, "step": 6 }, { "epoch": 6.318830113738942e-05, "grad_norm": 8.673748970031738, "learning_rate": 4.999729192995126e-05, "loss": 5.0192, "step": 7 }, { "epoch": 7.221520129987362e-05, "grad_norm": 10.252384185791016, "learning_rate": 4.999684058494313e-05, "loss": 5.1054, "step": 8 }, { "epoch": 8.124210146235782e-05, "grad_norm": 9.161073684692383, "learning_rate": 4.999638923993501e-05, "loss": 5.5159, "step": 9 }, { "epoch": 9.026900162484202e-05, "grad_norm": 11.713934898376465, "learning_rate": 4.999593789492689e-05, "loss": 4.8721, "step": 10 }, { "epoch": 9.929590178732623e-05, "grad_norm": 10.410552024841309, "learning_rate": 4.999548654991876e-05, "loss": 5.0056, "step": 11 }, { "epoch": 0.00010832280194981043, "grad_norm": 12.303479194641113, "learning_rate": 4.9995035204910633e-05, "loss": 5.287, "step": 12 }, { "epoch": 0.00011734970211229463, "grad_norm": 6.178354740142822, "learning_rate": 4.999458385990251e-05, "loss": 5.3856, "step": 13 }, { "epoch": 0.00012637660227477883, "grad_norm": 8.602326393127441, "learning_rate": 4.9994132514894385e-05, "loss": 4.5771, "step": 14 }, { "epoch": 0.00013540350243726305, "grad_norm": 7.028525352478027, "learning_rate": 4.9993681169886264e-05, "loss": 5.7803, "step": 15 }, { "epoch": 0.00014443040259974724, "grad_norm": 7.141914367675781, "learning_rate": 4.9993229824878144e-05, "loss": 5.545, "step": 16 }, { "epoch": 0.00015345730276223146, "grad_norm": 9.87399959564209, "learning_rate": 4.9992778479870016e-05, "loss": 5.2503, "step": 17 }, { "epoch": 0.00016248420292471564, "grad_norm": 9.393143653869629, "learning_rate": 4.999232713486189e-05, "loss": 5.2458, "step": 18 }, { "epoch": 0.00017151110308719986, "grad_norm": 9.297143936157227, "learning_rate": 4.999187578985377e-05, "loss": 5.0702, "step": 19 }, { "epoch": 0.00018053800324968405, "grad_norm": 9.123749732971191, "learning_rate": 4.999142444484564e-05, "loss": 4.5812, "step": 20 }, { "epoch": 0.00018956490341216826, "grad_norm": 7.802827835083008, "learning_rate": 4.999097309983752e-05, "loss": 5.6097, "step": 21 }, { "epoch": 0.00019859180357465245, "grad_norm": 15.07286262512207, "learning_rate": 4.999052175482939e-05, "loss": 3.8721, "step": 22 }, { "epoch": 0.00020761870373713667, "grad_norm": 5.826138019561768, "learning_rate": 4.999007040982127e-05, "loss": 5.4309, "step": 23 }, { "epoch": 0.00021664560389962086, "grad_norm": 7.086408615112305, "learning_rate": 4.998961906481315e-05, "loss": 5.557, "step": 24 }, { "epoch": 0.00022567250406210507, "grad_norm": 7.229325294494629, "learning_rate": 4.9989167719805016e-05, "loss": 5.8656, "step": 25 }, { "epoch": 0.00023469940422458926, "grad_norm": 10.291181564331055, "learning_rate": 4.9988716374796895e-05, "loss": 5.3134, "step": 26 }, { "epoch": 0.00024372630438707348, "grad_norm": 9.72523307800293, "learning_rate": 4.9988265029788775e-05, "loss": 5.5786, "step": 27 }, { "epoch": 0.00025275320454955767, "grad_norm": 8.139504432678223, "learning_rate": 4.998781368478065e-05, "loss": 5.6484, "step": 28 }, { "epoch": 0.0002617801047120419, "grad_norm": 7.9897332191467285, "learning_rate": 4.9987362339772526e-05, "loss": 5.399, "step": 29 }, { "epoch": 0.0002708070048745261, "grad_norm": 9.907605171203613, "learning_rate": 4.99869109947644e-05, "loss": 5.2487, "step": 30 }, { "epoch": 0.0002798339050370103, "grad_norm": 8.913695335388184, "learning_rate": 4.998645964975628e-05, "loss": 4.4659, "step": 31 }, { "epoch": 0.0002888608051994945, "grad_norm": 9.919137954711914, "learning_rate": 4.998600830474815e-05, "loss": 5.1025, "step": 32 }, { "epoch": 0.0002978877053619787, "grad_norm": 8.053698539733887, "learning_rate": 4.998555695974003e-05, "loss": 5.3673, "step": 33 }, { "epoch": 0.0003069146055244629, "grad_norm": 6.723404884338379, "learning_rate": 4.99851056147319e-05, "loss": 4.6599, "step": 34 }, { "epoch": 0.0003159415056869471, "grad_norm": 7.670530319213867, "learning_rate": 4.998465426972378e-05, "loss": 5.3265, "step": 35 }, { "epoch": 0.0003249684058494313, "grad_norm": 9.265135765075684, "learning_rate": 4.9984202924715654e-05, "loss": 5.4839, "step": 36 }, { "epoch": 0.00033399530601191553, "grad_norm": 7.694606304168701, "learning_rate": 4.998375157970753e-05, "loss": 5.2777, "step": 37 }, { "epoch": 0.0003430222061743997, "grad_norm": 9.841596603393555, "learning_rate": 4.9983300234699406e-05, "loss": 5.0609, "step": 38 }, { "epoch": 0.0003520491063368839, "grad_norm": 7.007106304168701, "learning_rate": 4.998284888969128e-05, "loss": 5.5848, "step": 39 }, { "epoch": 0.0003610760064993681, "grad_norm": 10.379240989685059, "learning_rate": 4.998239754468316e-05, "loss": 5.7563, "step": 40 }, { "epoch": 0.00037010290666185234, "grad_norm": 7.990135669708252, "learning_rate": 4.9981946199675037e-05, "loss": 5.3767, "step": 41 }, { "epoch": 0.00037912980682433653, "grad_norm": 10.290926933288574, "learning_rate": 4.998149485466691e-05, "loss": 4.4516, "step": 42 }, { "epoch": 0.0003881567069868207, "grad_norm": 10.700989723205566, "learning_rate": 4.998104350965879e-05, "loss": 5.8797, "step": 43 }, { "epoch": 0.0003971836071493049, "grad_norm": 7.0850653648376465, "learning_rate": 4.998059216465066e-05, "loss": 5.6911, "step": 44 }, { "epoch": 0.00040621050731178915, "grad_norm": 6.3839545249938965, "learning_rate": 4.998014081964253e-05, "loss": 5.1487, "step": 45 }, { "epoch": 0.00041523740747427334, "grad_norm": 7.327271938323975, "learning_rate": 4.997968947463441e-05, "loss": 4.8008, "step": 46 }, { "epoch": 0.00042426430763675753, "grad_norm": 9.126873016357422, "learning_rate": 4.9979238129626285e-05, "loss": 5.6461, "step": 47 }, { "epoch": 0.0004332912077992417, "grad_norm": 8.91388988494873, "learning_rate": 4.9978786784618164e-05, "loss": 4.01, "step": 48 }, { "epoch": 0.00044231810796172596, "grad_norm": 9.177200317382812, "learning_rate": 4.997833543961004e-05, "loss": 5.0464, "step": 49 }, { "epoch": 0.00045134500812421015, "grad_norm": 9.812573432922363, "learning_rate": 4.9977884094601916e-05, "loss": 4.3276, "step": 50 }, { "epoch": 0.00046037190828669434, "grad_norm": 7.8470587730407715, "learning_rate": 4.9977432749593795e-05, "loss": 5.6353, "step": 51 }, { "epoch": 0.00046939880844917853, "grad_norm": 7.9620866775512695, "learning_rate": 4.997698140458567e-05, "loss": 5.6812, "step": 52 }, { "epoch": 0.00047842570861166277, "grad_norm": 7.1675705909729, "learning_rate": 4.997653005957754e-05, "loss": 4.8318, "step": 53 }, { "epoch": 0.00048745260877414696, "grad_norm": 6.16520881652832, "learning_rate": 4.997607871456942e-05, "loss": 5.1025, "step": 54 }, { "epoch": 0.0004964795089366312, "grad_norm": 10.292895317077637, "learning_rate": 4.99756273695613e-05, "loss": 5.2905, "step": 55 }, { "epoch": 0.0005055064090991153, "grad_norm": 7.775153160095215, "learning_rate": 4.997517602455317e-05, "loss": 5.1239, "step": 56 }, { "epoch": 0.0005145333092615996, "grad_norm": 9.960454940795898, "learning_rate": 4.997472467954505e-05, "loss": 5.557, "step": 57 }, { "epoch": 0.0005235602094240838, "grad_norm": 5.567709922790527, "learning_rate": 4.997427333453692e-05, "loss": 5.7414, "step": 58 }, { "epoch": 0.000532587109586568, "grad_norm": 8.935986518859863, "learning_rate": 4.9973821989528795e-05, "loss": 5.2345, "step": 59 }, { "epoch": 0.0005416140097490522, "grad_norm": 9.151023864746094, "learning_rate": 4.9973370644520674e-05, "loss": 5.3218, "step": 60 }, { "epoch": 0.0005506409099115363, "grad_norm": 9.755552291870117, "learning_rate": 4.997291929951255e-05, "loss": 5.7142, "step": 61 }, { "epoch": 0.0005596678100740206, "grad_norm": 11.234132766723633, "learning_rate": 4.9972467954504426e-05, "loss": 5.4018, "step": 62 }, { "epoch": 0.0005686947102365048, "grad_norm": 6.105953693389893, "learning_rate": 4.9972016609496305e-05, "loss": 5.7931, "step": 63 }, { "epoch": 0.000577721610398989, "grad_norm": 7.325258731842041, "learning_rate": 4.997156526448818e-05, "loss": 5.3888, "step": 64 }, { "epoch": 0.0005867485105614732, "grad_norm": 10.60838508605957, "learning_rate": 4.997111391948005e-05, "loss": 4.1586, "step": 65 }, { "epoch": 0.0005957754107239574, "grad_norm": 10.369292259216309, "learning_rate": 4.997066257447193e-05, "loss": 5.4708, "step": 66 }, { "epoch": 0.0006048023108864416, "grad_norm": 9.129615783691406, "learning_rate": 4.99702112294638e-05, "loss": 4.0621, "step": 67 }, { "epoch": 0.0006138292110489258, "grad_norm": 8.594255447387695, "learning_rate": 4.996975988445568e-05, "loss": 5.4067, "step": 68 }, { "epoch": 0.00062285611121141, "grad_norm": 8.711878776550293, "learning_rate": 4.9969308539447554e-05, "loss": 4.953, "step": 69 }, { "epoch": 0.0006318830113738942, "grad_norm": 7.299989700317383, "learning_rate": 4.996885719443943e-05, "loss": 3.8801, "step": 70 }, { "epoch": 0.0006409099115363784, "grad_norm": 6.883722305297852, "learning_rate": 4.996840584943131e-05, "loss": 5.117, "step": 71 }, { "epoch": 0.0006499368116988626, "grad_norm": 9.112191200256348, "learning_rate": 4.996795450442318e-05, "loss": 5.5968, "step": 72 }, { "epoch": 0.0006589637118613468, "grad_norm": 7.247189998626709, "learning_rate": 4.996750315941506e-05, "loss": 5.293, "step": 73 }, { "epoch": 0.0006679906120238311, "grad_norm": 7.378231525421143, "learning_rate": 4.9967051814406936e-05, "loss": 5.7267, "step": 74 }, { "epoch": 0.0006770175121863152, "grad_norm": 9.147416114807129, "learning_rate": 4.996660046939881e-05, "loss": 5.2046, "step": 75 }, { "epoch": 0.0006860444123487994, "grad_norm": 7.845174312591553, "learning_rate": 4.996614912439069e-05, "loss": 5.1388, "step": 76 }, { "epoch": 0.0006950713125112836, "grad_norm": 7.671471118927002, "learning_rate": 4.996569777938257e-05, "loss": 4.5635, "step": 77 }, { "epoch": 0.0007040982126737678, "grad_norm": 6.576324462890625, "learning_rate": 4.996524643437444e-05, "loss": 5.1607, "step": 78 }, { "epoch": 0.0007131251128362521, "grad_norm": 5.673051357269287, "learning_rate": 4.996479508936631e-05, "loss": 5.549, "step": 79 }, { "epoch": 0.0007221520129987362, "grad_norm": 6.638739585876465, "learning_rate": 4.996434374435819e-05, "loss": 5.3269, "step": 80 }, { "epoch": 0.0007311789131612204, "grad_norm": 7.621270656585693, "learning_rate": 4.9963892399350064e-05, "loss": 5.1484, "step": 81 }, { "epoch": 0.0007402058133237047, "grad_norm": 7.230222225189209, "learning_rate": 4.996344105434194e-05, "loss": 5.6015, "step": 82 }, { "epoch": 0.0007492327134861888, "grad_norm": 7.890937328338623, "learning_rate": 4.9962989709333816e-05, "loss": 5.5522, "step": 83 }, { "epoch": 0.0007582596136486731, "grad_norm": 5.965391635894775, "learning_rate": 4.9962538364325695e-05, "loss": 5.0855, "step": 84 }, { "epoch": 0.0007672865138111573, "grad_norm": 7.960149765014648, "learning_rate": 4.996208701931757e-05, "loss": 5.5938, "step": 85 }, { "epoch": 0.0007763134139736414, "grad_norm": 7.232391357421875, "learning_rate": 4.996163567430944e-05, "loss": 5.9587, "step": 86 }, { "epoch": 0.0007853403141361257, "grad_norm": 8.46837043762207, "learning_rate": 4.996118432930132e-05, "loss": 5.4748, "step": 87 }, { "epoch": 0.0007943672142986098, "grad_norm": 5.856415748596191, "learning_rate": 4.99607329842932e-05, "loss": 4.3613, "step": 88 }, { "epoch": 0.0008033941144610941, "grad_norm": 8.095649719238281, "learning_rate": 4.996028163928507e-05, "loss": 5.3048, "step": 89 }, { "epoch": 0.0008124210146235783, "grad_norm": 5.506533622741699, "learning_rate": 4.995983029427695e-05, "loss": 4.9516, "step": 90 }, { "epoch": 0.0008214479147860624, "grad_norm": 7.977200031280518, "learning_rate": 4.995937894926883e-05, "loss": 5.8175, "step": 91 }, { "epoch": 0.0008304748149485467, "grad_norm": 5.766261577606201, "learning_rate": 4.9958927604260695e-05, "loss": 5.2154, "step": 92 }, { "epoch": 0.0008395017151110309, "grad_norm": 5.719219207763672, "learning_rate": 4.9958476259252574e-05, "loss": 4.9758, "step": 93 }, { "epoch": 0.0008485286152735151, "grad_norm": 6.498884677886963, "learning_rate": 4.995802491424445e-05, "loss": 4.7987, "step": 94 }, { "epoch": 0.0008575555154359993, "grad_norm": 5.459726810455322, "learning_rate": 4.9957573569236326e-05, "loss": 5.1129, "step": 95 }, { "epoch": 0.0008665824155984834, "grad_norm": 4.959202289581299, "learning_rate": 4.9957122224228205e-05, "loss": 5.2255, "step": 96 }, { "epoch": 0.0008756093157609677, "grad_norm": 4.617876052856445, "learning_rate": 4.995667087922008e-05, "loss": 5.5582, "step": 97 }, { "epoch": 0.0008846362159234519, "grad_norm": 5.88981294631958, "learning_rate": 4.995621953421196e-05, "loss": 5.2846, "step": 98 }, { "epoch": 0.0008936631160859361, "grad_norm": 6.1855010986328125, "learning_rate": 4.995576818920383e-05, "loss": 4.8301, "step": 99 }, { "epoch": 0.0009026900162484203, "grad_norm": 4.977812767028809, "learning_rate": 4.99553168441957e-05, "loss": 5.663, "step": 100 }, { "epoch": 0.0009117169164109045, "grad_norm": 4.26069450378418, "learning_rate": 4.995486549918758e-05, "loss": 5.6895, "step": 101 }, { "epoch": 0.0009207438165733887, "grad_norm": 5.3216776847839355, "learning_rate": 4.995441415417946e-05, "loss": 4.9248, "step": 102 }, { "epoch": 0.0009297707167358729, "grad_norm": 5.444666385650635, "learning_rate": 4.995396280917133e-05, "loss": 5.174, "step": 103 }, { "epoch": 0.0009387976168983571, "grad_norm": 6.172536849975586, "learning_rate": 4.995351146416321e-05, "loss": 4.8518, "step": 104 }, { "epoch": 0.0009478245170608413, "grad_norm": 3.551116943359375, "learning_rate": 4.9953060119155084e-05, "loss": 4.9411, "step": 105 }, { "epoch": 0.0009568514172233255, "grad_norm": 3.7002832889556885, "learning_rate": 4.995260877414696e-05, "loss": 5.2712, "step": 106 }, { "epoch": 0.0009658783173858097, "grad_norm": 5.500906944274902, "learning_rate": 4.9952157429138836e-05, "loss": 5.3829, "step": 107 }, { "epoch": 0.0009749052175482939, "grad_norm": 5.951501369476318, "learning_rate": 4.995170608413071e-05, "loss": 4.7374, "step": 108 }, { "epoch": 0.000983932117710778, "grad_norm": 4.012919902801514, "learning_rate": 4.995125473912259e-05, "loss": 5.1752, "step": 109 }, { "epoch": 0.0009929590178732624, "grad_norm": 4.635753154754639, "learning_rate": 4.995080339411447e-05, "loss": 4.8061, "step": 110 }, { "epoch": 0.0010019859180357465, "grad_norm": 6.280446529388428, "learning_rate": 4.995035204910634e-05, "loss": 5.5487, "step": 111 }, { "epoch": 0.0010110128181982307, "grad_norm": 5.604403495788574, "learning_rate": 4.994990070409821e-05, "loss": 4.4835, "step": 112 }, { "epoch": 0.001020039718360715, "grad_norm": 11.297250747680664, "learning_rate": 4.994944935909009e-05, "loss": 4.7518, "step": 113 }, { "epoch": 0.0010290666185231992, "grad_norm": 6.652409076690674, "learning_rate": 4.9948998014081964e-05, "loss": 5.0442, "step": 114 }, { "epoch": 0.0010380935186856833, "grad_norm": 5.184350490570068, "learning_rate": 4.994854666907384e-05, "loss": 4.5676, "step": 115 }, { "epoch": 0.0010471204188481676, "grad_norm": 6.231746196746826, "learning_rate": 4.994809532406572e-05, "loss": 5.2614, "step": 116 }, { "epoch": 0.0010561473190106518, "grad_norm": 8.688654899597168, "learning_rate": 4.9947643979057595e-05, "loss": 5.3005, "step": 117 }, { "epoch": 0.001065174219173136, "grad_norm": 4.623736381530762, "learning_rate": 4.9947192634049474e-05, "loss": 4.9939, "step": 118 }, { "epoch": 0.00107420111933562, "grad_norm": 6.4367780685424805, "learning_rate": 4.9946741289041346e-05, "loss": 5.2078, "step": 119 }, { "epoch": 0.0010832280194981044, "grad_norm": 4.237805366516113, "learning_rate": 4.994628994403322e-05, "loss": 5.3386, "step": 120 }, { "epoch": 0.0010922549196605885, "grad_norm": 5.620296955108643, "learning_rate": 4.99458385990251e-05, "loss": 5.201, "step": 121 }, { "epoch": 0.0011012818198230727, "grad_norm": 4.342684268951416, "learning_rate": 4.994538725401697e-05, "loss": 4.7002, "step": 122 }, { "epoch": 0.001110308719985557, "grad_norm": 5.919836044311523, "learning_rate": 4.994493590900885e-05, "loss": 4.7795, "step": 123 }, { "epoch": 0.0011193356201480412, "grad_norm": 5.67348575592041, "learning_rate": 4.994448456400073e-05, "loss": 5.3631, "step": 124 }, { "epoch": 0.0011283625203105253, "grad_norm": 5.7495269775390625, "learning_rate": 4.99440332189926e-05, "loss": 5.6839, "step": 125 }, { "epoch": 0.0011373894204730096, "grad_norm": 5.579268455505371, "learning_rate": 4.9943581873984474e-05, "loss": 4.4984, "step": 126 }, { "epoch": 0.0011464163206354938, "grad_norm": 3.6236391067504883, "learning_rate": 4.994313052897635e-05, "loss": 4.8727, "step": 127 }, { "epoch": 0.001155443220797978, "grad_norm": 5.383500099182129, "learning_rate": 4.9942679183968226e-05, "loss": 5.1841, "step": 128 }, { "epoch": 0.0011644701209604623, "grad_norm": 8.279353141784668, "learning_rate": 4.9942227838960105e-05, "loss": 4.0918, "step": 129 }, { "epoch": 0.0011734970211229464, "grad_norm": 3.4767377376556396, "learning_rate": 4.994177649395198e-05, "loss": 5.2044, "step": 130 }, { "epoch": 0.0011825239212854305, "grad_norm": 4.3666605949401855, "learning_rate": 4.9941325148943856e-05, "loss": 5.1911, "step": 131 }, { "epoch": 0.0011915508214479149, "grad_norm": 7.244105815887451, "learning_rate": 4.994087380393573e-05, "loss": 4.1227, "step": 132 }, { "epoch": 0.001200577721610399, "grad_norm": 6.466646671295166, "learning_rate": 4.994042245892761e-05, "loss": 4.7663, "step": 133 }, { "epoch": 0.0012096046217728832, "grad_norm": 5.164224624633789, "learning_rate": 4.993997111391948e-05, "loss": 4.9552, "step": 134 }, { "epoch": 0.0012186315219353673, "grad_norm": 6.225243091583252, "learning_rate": 4.993951976891136e-05, "loss": 4.5013, "step": 135 }, { "epoch": 0.0012276584220978516, "grad_norm": 6.678073883056641, "learning_rate": 4.993906842390323e-05, "loss": 4.9601, "step": 136 }, { "epoch": 0.0012366853222603358, "grad_norm": 4.274768352508545, "learning_rate": 4.993861707889511e-05, "loss": 5.2563, "step": 137 }, { "epoch": 0.00124571222242282, "grad_norm": 3.771730899810791, "learning_rate": 4.993816573388699e-05, "loss": 5.6016, "step": 138 }, { "epoch": 0.0012547391225853043, "grad_norm": 5.381021022796631, "learning_rate": 4.9937714388878857e-05, "loss": 5.1383, "step": 139 }, { "epoch": 0.0012637660227477884, "grad_norm": 4.534156799316406, "learning_rate": 4.9937263043870736e-05, "loss": 5.3151, "step": 140 }, { "epoch": 0.0012727929229102725, "grad_norm": 4.036935806274414, "learning_rate": 4.9936811698862615e-05, "loss": 4.4303, "step": 141 }, { "epoch": 0.0012818198230727569, "grad_norm": 5.927755832672119, "learning_rate": 4.993636035385449e-05, "loss": 5.0507, "step": 142 }, { "epoch": 0.001290846723235241, "grad_norm": 7.061839580535889, "learning_rate": 4.993590900884637e-05, "loss": 4.9659, "step": 143 }, { "epoch": 0.0012998736233977252, "grad_norm": 5.401264667510986, "learning_rate": 4.993545766383824e-05, "loss": 5.0074, "step": 144 }, { "epoch": 0.0013089005235602095, "grad_norm": 4.211348056793213, "learning_rate": 4.993500631883012e-05, "loss": 5.572, "step": 145 }, { "epoch": 0.0013179274237226936, "grad_norm": 3.4184935092926025, "learning_rate": 4.993455497382199e-05, "loss": 5.4112, "step": 146 }, { "epoch": 0.0013269543238851778, "grad_norm": 6.418872356414795, "learning_rate": 4.993410362881386e-05, "loss": 5.6788, "step": 147 }, { "epoch": 0.0013359812240476621, "grad_norm": 6.580074787139893, "learning_rate": 4.993365228380574e-05, "loss": 4.3538, "step": 148 }, { "epoch": 0.0013450081242101463, "grad_norm": 3.6988143920898438, "learning_rate": 4.993320093879762e-05, "loss": 4.6329, "step": 149 }, { "epoch": 0.0013540350243726304, "grad_norm": 4.4106597900390625, "learning_rate": 4.9932749593789494e-05, "loss": 5.384, "step": 150 }, { "epoch": 0.0013630619245351147, "grad_norm": 5.291019439697266, "learning_rate": 4.9932298248781374e-05, "loss": 5.24, "step": 151 }, { "epoch": 0.0013720888246975989, "grad_norm": 4.243605613708496, "learning_rate": 4.9931846903773246e-05, "loss": 5.1562, "step": 152 }, { "epoch": 0.001381115724860083, "grad_norm": 5.044171333312988, "learning_rate": 4.993139555876512e-05, "loss": 5.7931, "step": 153 }, { "epoch": 0.0013901426250225671, "grad_norm": 8.121780395507812, "learning_rate": 4.9930944213757e-05, "loss": 5.278, "step": 154 }, { "epoch": 0.0013991695251850515, "grad_norm": 4.859671115875244, "learning_rate": 4.993049286874888e-05, "loss": 4.9149, "step": 155 }, { "epoch": 0.0014081964253475356, "grad_norm": 4.731335639953613, "learning_rate": 4.993004152374075e-05, "loss": 4.7643, "step": 156 }, { "epoch": 0.0014172233255100198, "grad_norm": 3.6084158420562744, "learning_rate": 4.992959017873263e-05, "loss": 5.3715, "step": 157 }, { "epoch": 0.0014262502256725041, "grad_norm": 4.524412631988525, "learning_rate": 4.99291388337245e-05, "loss": 5.6071, "step": 158 }, { "epoch": 0.0014352771258349883, "grad_norm": 5.226177215576172, "learning_rate": 4.9928687488716374e-05, "loss": 5.1696, "step": 159 }, { "epoch": 0.0014443040259974724, "grad_norm": 4.614068508148193, "learning_rate": 4.992823614370825e-05, "loss": 5.8498, "step": 160 }, { "epoch": 0.0014533309261599567, "grad_norm": 5.204494476318359, "learning_rate": 4.9927784798700125e-05, "loss": 5.3602, "step": 161 }, { "epoch": 0.0014623578263224409, "grad_norm": 5.086613178253174, "learning_rate": 4.9927333453692005e-05, "loss": 5.4153, "step": 162 }, { "epoch": 0.001471384726484925, "grad_norm": 4.032634258270264, "learning_rate": 4.9926882108683884e-05, "loss": 5.4468, "step": 163 }, { "epoch": 0.0014804116266474094, "grad_norm": 4.255616664886475, "learning_rate": 4.9926430763675756e-05, "loss": 5.3606, "step": 164 }, { "epoch": 0.0014894385268098935, "grad_norm": 3.3045125007629395, "learning_rate": 4.9925979418667635e-05, "loss": 5.0117, "step": 165 }, { "epoch": 0.0014984654269723776, "grad_norm": 4.533328533172607, "learning_rate": 4.992552807365951e-05, "loss": 5.3746, "step": 166 }, { "epoch": 0.001507492327134862, "grad_norm": 4.983652591705322, "learning_rate": 4.992507672865138e-05, "loss": 5.1269, "step": 167 }, { "epoch": 0.0015165192272973461, "grad_norm": 5.368861675262451, "learning_rate": 4.992462538364326e-05, "loss": 5.3937, "step": 168 }, { "epoch": 0.0015255461274598303, "grad_norm": 5.124269962310791, "learning_rate": 4.992417403863513e-05, "loss": 5.1702, "step": 169 }, { "epoch": 0.0015345730276223146, "grad_norm": 4.3724212646484375, "learning_rate": 4.992372269362701e-05, "loss": 5.1766, "step": 170 }, { "epoch": 0.0015435999277847987, "grad_norm": 3.8447368144989014, "learning_rate": 4.992327134861889e-05, "loss": 5.2422, "step": 171 }, { "epoch": 0.0015526268279472829, "grad_norm": 3.932455539703369, "learning_rate": 4.992282000361076e-05, "loss": 5.3255, "step": 172 }, { "epoch": 0.001561653728109767, "grad_norm": 5.289744853973389, "learning_rate": 4.9922368658602636e-05, "loss": 5.3903, "step": 173 }, { "epoch": 0.0015706806282722514, "grad_norm": 5.054441928863525, "learning_rate": 4.9921917313594515e-05, "loss": 5.0518, "step": 174 }, { "epoch": 0.0015797075284347355, "grad_norm": 5.426041126251221, "learning_rate": 4.992146596858639e-05, "loss": 4.7158, "step": 175 }, { "epoch": 0.0015887344285972196, "grad_norm": 4.537737846374512, "learning_rate": 4.9921014623578266e-05, "loss": 5.4756, "step": 176 }, { "epoch": 0.001597761328759704, "grad_norm": 6.328954219818115, "learning_rate": 4.9920563278570146e-05, "loss": 5.3957, "step": 177 }, { "epoch": 0.0016067882289221881, "grad_norm": 6.441391468048096, "learning_rate": 4.992011193356202e-05, "loss": 5.3615, "step": 178 }, { "epoch": 0.0016158151290846723, "grad_norm": 4.938143730163574, "learning_rate": 4.991966058855389e-05, "loss": 5.0225, "step": 179 }, { "epoch": 0.0016248420292471566, "grad_norm": 5.432797431945801, "learning_rate": 4.991920924354577e-05, "loss": 5.0782, "step": 180 }, { "epoch": 0.0016338689294096407, "grad_norm": 4.134559154510498, "learning_rate": 4.991875789853764e-05, "loss": 5.3375, "step": 181 }, { "epoch": 0.0016428958295721249, "grad_norm": 4.189735412597656, "learning_rate": 4.991830655352952e-05, "loss": 5.1214, "step": 182 }, { "epoch": 0.0016519227297346092, "grad_norm": 6.9606804847717285, "learning_rate": 4.9917855208521394e-05, "loss": 5.0121, "step": 183 }, { "epoch": 0.0016609496298970934, "grad_norm": 4.842367649078369, "learning_rate": 4.991740386351327e-05, "loss": 5.5876, "step": 184 }, { "epoch": 0.0016699765300595775, "grad_norm": 5.3689470291137695, "learning_rate": 4.991695251850515e-05, "loss": 5.1329, "step": 185 }, { "epoch": 0.0016790034302220618, "grad_norm": 6.760075092315674, "learning_rate": 4.991650117349702e-05, "loss": 4.7815, "step": 186 }, { "epoch": 0.001688030330384546, "grad_norm": 5.678772926330566, "learning_rate": 4.99160498284889e-05, "loss": 4.2575, "step": 187 }, { "epoch": 0.0016970572305470301, "grad_norm": 4.833435535430908, "learning_rate": 4.991559848348078e-05, "loss": 4.5224, "step": 188 }, { "epoch": 0.0017060841307095142, "grad_norm": 5.299649238586426, "learning_rate": 4.991514713847265e-05, "loss": 5.2351, "step": 189 }, { "epoch": 0.0017151110308719986, "grad_norm": 4.814604759216309, "learning_rate": 4.991469579346453e-05, "loss": 5.2694, "step": 190 }, { "epoch": 0.0017241379310344827, "grad_norm": 5.359004497528076, "learning_rate": 4.99142444484564e-05, "loss": 4.4169, "step": 191 }, { "epoch": 0.0017331648311969669, "grad_norm": 4.964508533477783, "learning_rate": 4.991379310344828e-05, "loss": 5.0466, "step": 192 }, { "epoch": 0.0017421917313594512, "grad_norm": 7.191763877868652, "learning_rate": 4.991334175844015e-05, "loss": 4.679, "step": 193 }, { "epoch": 0.0017512186315219354, "grad_norm": 4.597863674163818, "learning_rate": 4.991289041343203e-05, "loss": 5.237, "step": 194 }, { "epoch": 0.0017602455316844195, "grad_norm": 5.166416168212891, "learning_rate": 4.9912439068423904e-05, "loss": 5.1566, "step": 195 }, { "epoch": 0.0017692724318469038, "grad_norm": 4.666167736053467, "learning_rate": 4.9911987723415784e-05, "loss": 5.1123, "step": 196 }, { "epoch": 0.001778299332009388, "grad_norm": 6.197031021118164, "learning_rate": 4.9911536378407656e-05, "loss": 5.1252, "step": 197 }, { "epoch": 0.0017873262321718721, "grad_norm": 6.630825042724609, "learning_rate": 4.9911085033399535e-05, "loss": 4.8518, "step": 198 }, { "epoch": 0.0017963531323343565, "grad_norm": 4.648182392120361, "learning_rate": 4.991063368839141e-05, "loss": 4.6524, "step": 199 }, { "epoch": 0.0018053800324968406, "grad_norm": 5.033623695373535, "learning_rate": 4.991018234338328e-05, "loss": 5.1568, "step": 200 }, { "epoch": 0.0018144069326593247, "grad_norm": 5.699010372161865, "learning_rate": 4.990973099837516e-05, "loss": 5.0339, "step": 201 }, { "epoch": 0.001823433832821809, "grad_norm": 5.0051350593566895, "learning_rate": 4.990927965336704e-05, "loss": 5.2634, "step": 202 }, { "epoch": 0.0018324607329842932, "grad_norm": 4.5059733390808105, "learning_rate": 4.990882830835891e-05, "loss": 5.0728, "step": 203 }, { "epoch": 0.0018414876331467774, "grad_norm": 4.895959854125977, "learning_rate": 4.990837696335079e-05, "loss": 4.7002, "step": 204 }, { "epoch": 0.0018505145333092617, "grad_norm": 4.848843097686768, "learning_rate": 4.990792561834266e-05, "loss": 4.4567, "step": 205 }, { "epoch": 0.0018595414334717458, "grad_norm": 5.591897964477539, "learning_rate": 4.9907474273334535e-05, "loss": 5.5476, "step": 206 }, { "epoch": 0.00186856833363423, "grad_norm": 7.7116780281066895, "learning_rate": 4.9907022928326415e-05, "loss": 4.6912, "step": 207 }, { "epoch": 0.0018775952337967141, "grad_norm": 4.162778854370117, "learning_rate": 4.990657158331829e-05, "loss": 5.3512, "step": 208 }, { "epoch": 0.0018866221339591985, "grad_norm": 4.6462907791137695, "learning_rate": 4.9906120238310166e-05, "loss": 4.4307, "step": 209 }, { "epoch": 0.0018956490341216826, "grad_norm": 4.718513488769531, "learning_rate": 4.9905668893302045e-05, "loss": 5.0457, "step": 210 }, { "epoch": 0.0019046759342841667, "grad_norm": 4.313071250915527, "learning_rate": 4.990521754829392e-05, "loss": 5.6131, "step": 211 }, { "epoch": 0.001913702834446651, "grad_norm": 4.233699321746826, "learning_rate": 4.99047662032858e-05, "loss": 4.9254, "step": 212 }, { "epoch": 0.0019227297346091352, "grad_norm": 6.903225898742676, "learning_rate": 4.990431485827767e-05, "loss": 5.6492, "step": 213 }, { "epoch": 0.0019317566347716194, "grad_norm": 4.87800931930542, "learning_rate": 4.990386351326954e-05, "loss": 5.7241, "step": 214 }, { "epoch": 0.0019407835349341037, "grad_norm": 5.207213878631592, "learning_rate": 4.990341216826142e-05, "loss": 5.3806, "step": 215 }, { "epoch": 0.0019498104350965878, "grad_norm": 5.957947731018066, "learning_rate": 4.99029608232533e-05, "loss": 5.3012, "step": 216 }, { "epoch": 0.001958837335259072, "grad_norm": 5.87772274017334, "learning_rate": 4.990250947824517e-05, "loss": 5.5109, "step": 217 }, { "epoch": 0.001967864235421556, "grad_norm": 5.206866264343262, "learning_rate": 4.990205813323705e-05, "loss": 5.3038, "step": 218 }, { "epoch": 0.0019768911355840402, "grad_norm": 4.107205390930176, "learning_rate": 4.9901606788228925e-05, "loss": 4.9246, "step": 219 }, { "epoch": 0.001985918035746525, "grad_norm": 5.8855767250061035, "learning_rate": 4.99011554432208e-05, "loss": 5.1761, "step": 220 }, { "epoch": 0.001994944935909009, "grad_norm": 5.266073226928711, "learning_rate": 4.9900704098212676e-05, "loss": 4.4998, "step": 221 }, { "epoch": 0.002003971836071493, "grad_norm": 4.896139621734619, "learning_rate": 4.990025275320455e-05, "loss": 4.8993, "step": 222 }, { "epoch": 0.002012998736233977, "grad_norm": 4.206170082092285, "learning_rate": 4.989980140819643e-05, "loss": 5.0812, "step": 223 }, { "epoch": 0.0020220256363964613, "grad_norm": 4.135659217834473, "learning_rate": 4.989935006318831e-05, "loss": 5.3647, "step": 224 }, { "epoch": 0.0020310525365589455, "grad_norm": 3.6154143810272217, "learning_rate": 4.989889871818018e-05, "loss": 5.2117, "step": 225 }, { "epoch": 0.00204007943672143, "grad_norm": 5.031739234924316, "learning_rate": 4.989844737317205e-05, "loss": 4.8563, "step": 226 }, { "epoch": 0.002049106336883914, "grad_norm": 5.777970790863037, "learning_rate": 4.989799602816393e-05, "loss": 5.7669, "step": 227 }, { "epoch": 0.0020581332370463983, "grad_norm": 8.82309627532959, "learning_rate": 4.9897544683155804e-05, "loss": 5.5327, "step": 228 }, { "epoch": 0.0020671601372088825, "grad_norm": 4.682310104370117, "learning_rate": 4.989709333814768e-05, "loss": 5.2309, "step": 229 }, { "epoch": 0.0020761870373713666, "grad_norm": 6.983668327331543, "learning_rate": 4.9896641993139556e-05, "loss": 4.9515, "step": 230 }, { "epoch": 0.0020852139375338507, "grad_norm": 3.723271131515503, "learning_rate": 4.9896190648131435e-05, "loss": 4.0205, "step": 231 }, { "epoch": 0.0020942408376963353, "grad_norm": 5.102370262145996, "learning_rate": 4.9895739303123314e-05, "loss": 5.2686, "step": 232 }, { "epoch": 0.0021032677378588194, "grad_norm": 5.955885410308838, "learning_rate": 4.989528795811518e-05, "loss": 5.3478, "step": 233 }, { "epoch": 0.0021122946380213036, "grad_norm": 4.965644836425781, "learning_rate": 4.989483661310706e-05, "loss": 5.1006, "step": 234 }, { "epoch": 0.0021213215381837877, "grad_norm": 6.386919975280762, "learning_rate": 4.989438526809894e-05, "loss": 5.4725, "step": 235 }, { "epoch": 0.002130348438346272, "grad_norm": 7.445868015289307, "learning_rate": 4.989393392309081e-05, "loss": 4.7433, "step": 236 }, { "epoch": 0.002139375338508756, "grad_norm": 4.041948318481445, "learning_rate": 4.989348257808269e-05, "loss": 4.9314, "step": 237 }, { "epoch": 0.00214840223867124, "grad_norm": 5.829143047332764, "learning_rate": 4.989303123307457e-05, "loss": 4.6102, "step": 238 }, { "epoch": 0.0021574291388337247, "grad_norm": 5.984886646270752, "learning_rate": 4.989257988806644e-05, "loss": 4.5307, "step": 239 }, { "epoch": 0.002166456038996209, "grad_norm": 5.256242275238037, "learning_rate": 4.9892128543058314e-05, "loss": 5.1402, "step": 240 }, { "epoch": 0.002175482939158693, "grad_norm": 6.780630111694336, "learning_rate": 4.9891677198050194e-05, "loss": 5.5201, "step": 241 }, { "epoch": 0.002184509839321177, "grad_norm": 5.493387222290039, "learning_rate": 4.9891225853042066e-05, "loss": 5.3286, "step": 242 }, { "epoch": 0.002193536739483661, "grad_norm": 5.084599494934082, "learning_rate": 4.9890774508033945e-05, "loss": 4.7176, "step": 243 }, { "epoch": 0.0022025636396461453, "grad_norm": 6.372427940368652, "learning_rate": 4.989032316302582e-05, "loss": 4.9766, "step": 244 }, { "epoch": 0.00221159053980863, "grad_norm": 5.0189528465271, "learning_rate": 4.98898718180177e-05, "loss": 5.0168, "step": 245 }, { "epoch": 0.002220617439971114, "grad_norm": 4.203639030456543, "learning_rate": 4.988942047300957e-05, "loss": 5.5675, "step": 246 }, { "epoch": 0.002229644340133598, "grad_norm": 7.10724401473999, "learning_rate": 4.988896912800144e-05, "loss": 5.0758, "step": 247 }, { "epoch": 0.0022386712402960823, "grad_norm": 6.740031719207764, "learning_rate": 4.988851778299332e-05, "loss": 4.6001, "step": 248 }, { "epoch": 0.0022476981404585664, "grad_norm": 8.371086120605469, "learning_rate": 4.98880664379852e-05, "loss": 4.8638, "step": 249 }, { "epoch": 0.0022567250406210506, "grad_norm": 9.349472045898438, "learning_rate": 4.988761509297707e-05, "loss": 5.623, "step": 250 }, { "epoch": 0.0022657519407835347, "grad_norm": 6.028239727020264, "learning_rate": 4.988716374796895e-05, "loss": 5.0068, "step": 251 }, { "epoch": 0.0022747788409460193, "grad_norm": 4.7141337394714355, "learning_rate": 4.988671240296083e-05, "loss": 4.3406, "step": 252 }, { "epoch": 0.0022838057411085034, "grad_norm": 5.936439037322998, "learning_rate": 4.98862610579527e-05, "loss": 5.5637, "step": 253 }, { "epoch": 0.0022928326412709876, "grad_norm": 6.865289211273193, "learning_rate": 4.9885809712944576e-05, "loss": 5.3204, "step": 254 }, { "epoch": 0.0023018595414334717, "grad_norm": 4.12852668762207, "learning_rate": 4.9885358367936455e-05, "loss": 6.035, "step": 255 }, { "epoch": 0.002310886441595956, "grad_norm": 5.463346481323242, "learning_rate": 4.988490702292833e-05, "loss": 5.2713, "step": 256 }, { "epoch": 0.00231991334175844, "grad_norm": 5.132201194763184, "learning_rate": 4.988445567792021e-05, "loss": 4.9379, "step": 257 }, { "epoch": 0.0023289402419209245, "grad_norm": 4.85060453414917, "learning_rate": 4.988400433291208e-05, "loss": 5.7636, "step": 258 }, { "epoch": 0.0023379671420834087, "grad_norm": 3.8125264644622803, "learning_rate": 4.988355298790396e-05, "loss": 5.1456, "step": 259 }, { "epoch": 0.002346994042245893, "grad_norm": 7.0089335441589355, "learning_rate": 4.988310164289583e-05, "loss": 4.9799, "step": 260 }, { "epoch": 0.002356020942408377, "grad_norm": 4.324965000152588, "learning_rate": 4.9882650297887704e-05, "loss": 4.6482, "step": 261 }, { "epoch": 0.002365047842570861, "grad_norm": 5.952146530151367, "learning_rate": 4.988219895287958e-05, "loss": 5.2477, "step": 262 }, { "epoch": 0.002374074742733345, "grad_norm": 6.586727619171143, "learning_rate": 4.988174760787146e-05, "loss": 5.1303, "step": 263 }, { "epoch": 0.0023831016428958298, "grad_norm": 4.347113132476807, "learning_rate": 4.9881296262863335e-05, "loss": 5.1221, "step": 264 }, { "epoch": 0.002392128543058314, "grad_norm": 6.202446460723877, "learning_rate": 4.9880844917855214e-05, "loss": 5.3809, "step": 265 }, { "epoch": 0.002401155443220798, "grad_norm": 3.8619096279144287, "learning_rate": 4.9880393572847086e-05, "loss": 4.6983, "step": 266 }, { "epoch": 0.002410182343383282, "grad_norm": 4.1692705154418945, "learning_rate": 4.987994222783896e-05, "loss": 5.131, "step": 267 }, { "epoch": 0.0024192092435457663, "grad_norm": 4.773333549499512, "learning_rate": 4.987949088283084e-05, "loss": 5.1419, "step": 268 }, { "epoch": 0.0024282361437082504, "grad_norm": 5.046990871429443, "learning_rate": 4.987903953782271e-05, "loss": 5.008, "step": 269 }, { "epoch": 0.0024372630438707346, "grad_norm": 3.751389741897583, "learning_rate": 4.987858819281459e-05, "loss": 5.1689, "step": 270 }, { "epoch": 0.002446289944033219, "grad_norm": 3.5672385692596436, "learning_rate": 4.987813684780647e-05, "loss": 4.9196, "step": 271 }, { "epoch": 0.0024553168441957033, "grad_norm": 5.348905563354492, "learning_rate": 4.987768550279834e-05, "loss": 5.4949, "step": 272 }, { "epoch": 0.0024643437443581874, "grad_norm": 7.349914073944092, "learning_rate": 4.9877234157790214e-05, "loss": 5.5561, "step": 273 }, { "epoch": 0.0024733706445206716, "grad_norm": 4.628376483917236, "learning_rate": 4.987678281278209e-05, "loss": 4.3766, "step": 274 }, { "epoch": 0.0024823975446831557, "grad_norm": 5.1620283126831055, "learning_rate": 4.9876331467773966e-05, "loss": 5.3654, "step": 275 }, { "epoch": 0.00249142444484564, "grad_norm": 3.6543118953704834, "learning_rate": 4.9875880122765845e-05, "loss": 5.204, "step": 276 }, { "epoch": 0.0025004513450081244, "grad_norm": 5.941288948059082, "learning_rate": 4.9875428777757724e-05, "loss": 4.8077, "step": 277 }, { "epoch": 0.0025094782451706085, "grad_norm": 4.309896945953369, "learning_rate": 4.98749774327496e-05, "loss": 5.5435, "step": 278 }, { "epoch": 0.0025185051453330927, "grad_norm": 3.895845413208008, "learning_rate": 4.9874526087741476e-05, "loss": 5.5989, "step": 279 }, { "epoch": 0.002527532045495577, "grad_norm": 5.729174613952637, "learning_rate": 4.987407474273335e-05, "loss": 5.5623, "step": 280 }, { "epoch": 0.002536558945658061, "grad_norm": 3.8973653316497803, "learning_rate": 4.987362339772522e-05, "loss": 5.1828, "step": 281 }, { "epoch": 0.002545585845820545, "grad_norm": 4.521162986755371, "learning_rate": 4.98731720527171e-05, "loss": 4.8001, "step": 282 }, { "epoch": 0.0025546127459830296, "grad_norm": 3.4418630599975586, "learning_rate": 4.987272070770897e-05, "loss": 4.8134, "step": 283 }, { "epoch": 0.0025636396461455138, "grad_norm": 3.954406976699829, "learning_rate": 4.987226936270085e-05, "loss": 5.7543, "step": 284 }, { "epoch": 0.002572666546307998, "grad_norm": 4.527085304260254, "learning_rate": 4.987181801769273e-05, "loss": 5.7082, "step": 285 }, { "epoch": 0.002581693446470482, "grad_norm": 3.800293207168579, "learning_rate": 4.9871366672684604e-05, "loss": 4.855, "step": 286 }, { "epoch": 0.002590720346632966, "grad_norm": 4.596002578735352, "learning_rate": 4.9870915327676476e-05, "loss": 5.2094, "step": 287 }, { "epoch": 0.0025997472467954503, "grad_norm": 4.786351203918457, "learning_rate": 4.9870463982668355e-05, "loss": 4.9286, "step": 288 }, { "epoch": 0.0026087741469579344, "grad_norm": 3.4853193759918213, "learning_rate": 4.987001263766023e-05, "loss": 5.1279, "step": 289 }, { "epoch": 0.002617801047120419, "grad_norm": 4.352386951446533, "learning_rate": 4.986956129265211e-05, "loss": 5.4416, "step": 290 }, { "epoch": 0.002626827947282903, "grad_norm": 3.8838651180267334, "learning_rate": 4.986910994764398e-05, "loss": 4.5435, "step": 291 }, { "epoch": 0.0026358548474453873, "grad_norm": 4.763785362243652, "learning_rate": 4.986865860263586e-05, "loss": 5.4167, "step": 292 }, { "epoch": 0.0026448817476078714, "grad_norm": 5.028802871704102, "learning_rate": 4.986820725762773e-05, "loss": 4.382, "step": 293 }, { "epoch": 0.0026539086477703555, "grad_norm": 5.35697078704834, "learning_rate": 4.986775591261961e-05, "loss": 4.9176, "step": 294 }, { "epoch": 0.0026629355479328397, "grad_norm": 4.571488380432129, "learning_rate": 4.986730456761148e-05, "loss": 5.1021, "step": 295 }, { "epoch": 0.0026719624480953242, "grad_norm": 5.157553672790527, "learning_rate": 4.986685322260336e-05, "loss": 4.9431, "step": 296 }, { "epoch": 0.0026809893482578084, "grad_norm": 5.322408199310303, "learning_rate": 4.9866401877595235e-05, "loss": 4.7446, "step": 297 }, { "epoch": 0.0026900162484202925, "grad_norm": 6.436573028564453, "learning_rate": 4.9865950532587114e-05, "loss": 5.1444, "step": 298 }, { "epoch": 0.0026990431485827767, "grad_norm": 4.7643961906433105, "learning_rate": 4.986549918757899e-05, "loss": 4.692, "step": 299 }, { "epoch": 0.002708070048745261, "grad_norm": 5.269461631774902, "learning_rate": 4.986504784257086e-05, "loss": 4.7377, "step": 300 }, { "epoch": 0.002717096948907745, "grad_norm": 5.381867408752441, "learning_rate": 4.986459649756274e-05, "loss": 5.6669, "step": 301 }, { "epoch": 0.0027261238490702295, "grad_norm": 4.315028190612793, "learning_rate": 4.986414515255462e-05, "loss": 5.0525, "step": 302 }, { "epoch": 0.0027351507492327136, "grad_norm": 5.419351577758789, "learning_rate": 4.986369380754649e-05, "loss": 5.2817, "step": 303 }, { "epoch": 0.0027441776493951978, "grad_norm": 4.22694206237793, "learning_rate": 4.986324246253837e-05, "loss": 4.164, "step": 304 }, { "epoch": 0.002753204549557682, "grad_norm": 3.644566774368286, "learning_rate": 4.986279111753024e-05, "loss": 5.3227, "step": 305 }, { "epoch": 0.002762231449720166, "grad_norm": 5.783388137817383, "learning_rate": 4.986233977252212e-05, "loss": 5.2458, "step": 306 }, { "epoch": 0.00277125834988265, "grad_norm": 7.966733932495117, "learning_rate": 4.986188842751399e-05, "loss": 5.5294, "step": 307 }, { "epoch": 0.0027802852500451343, "grad_norm": 4.107776641845703, "learning_rate": 4.9861437082505866e-05, "loss": 5.5116, "step": 308 }, { "epoch": 0.002789312150207619, "grad_norm": 5.162723064422607, "learning_rate": 4.9860985737497745e-05, "loss": 3.9278, "step": 309 }, { "epoch": 0.002798339050370103, "grad_norm": 4.497034549713135, "learning_rate": 4.9860534392489624e-05, "loss": 5.4271, "step": 310 }, { "epoch": 0.002807365950532587, "grad_norm": 5.236740589141846, "learning_rate": 4.9860083047481496e-05, "loss": 5.178, "step": 311 }, { "epoch": 0.0028163928506950713, "grad_norm": 5.157679080963135, "learning_rate": 4.9859631702473376e-05, "loss": 5.3678, "step": 312 }, { "epoch": 0.0028254197508575554, "grad_norm": 5.667177677154541, "learning_rate": 4.985918035746525e-05, "loss": 5.0268, "step": 313 }, { "epoch": 0.0028344466510200395, "grad_norm": 4.737194061279297, "learning_rate": 4.985872901245712e-05, "loss": 4.9878, "step": 314 }, { "epoch": 0.002843473551182524, "grad_norm": 4.255739212036133, "learning_rate": 4.9858277667449e-05, "loss": 3.9726, "step": 315 }, { "epoch": 0.0028525004513450082, "grad_norm": 4.2768988609313965, "learning_rate": 4.985782632244088e-05, "loss": 4.865, "step": 316 }, { "epoch": 0.0028615273515074924, "grad_norm": 4.024101734161377, "learning_rate": 4.985737497743275e-05, "loss": 5.0769, "step": 317 }, { "epoch": 0.0028705542516699765, "grad_norm": 5.525566101074219, "learning_rate": 4.985692363242463e-05, "loss": 5.5647, "step": 318 }, { "epoch": 0.0028795811518324606, "grad_norm": 4.418797016143799, "learning_rate": 4.98564722874165e-05, "loss": 4.9818, "step": 319 }, { "epoch": 0.0028886080519949448, "grad_norm": 5.249420166015625, "learning_rate": 4.9856020942408376e-05, "loss": 5.5778, "step": 320 }, { "epoch": 0.0028976349521574294, "grad_norm": 4.717482089996338, "learning_rate": 4.9855569597400255e-05, "loss": 5.294, "step": 321 }, { "epoch": 0.0029066618523199135, "grad_norm": 5.7397780418396, "learning_rate": 4.985511825239213e-05, "loss": 5.6553, "step": 322 }, { "epoch": 0.0029156887524823976, "grad_norm": 5.574460029602051, "learning_rate": 4.985466690738401e-05, "loss": 4.9629, "step": 323 }, { "epoch": 0.0029247156526448818, "grad_norm": 3.934701919555664, "learning_rate": 4.9854215562375886e-05, "loss": 5.4118, "step": 324 }, { "epoch": 0.002933742552807366, "grad_norm": 4.577787399291992, "learning_rate": 4.985376421736776e-05, "loss": 4.9114, "step": 325 }, { "epoch": 0.00294276945296985, "grad_norm": 7.418753623962402, "learning_rate": 4.985331287235964e-05, "loss": 5.2624, "step": 326 }, { "epoch": 0.002951796353132334, "grad_norm": 4.691434383392334, "learning_rate": 4.985286152735151e-05, "loss": 5.4305, "step": 327 }, { "epoch": 0.0029608232532948187, "grad_norm": 4.489525318145752, "learning_rate": 4.985241018234338e-05, "loss": 5.5044, "step": 328 }, { "epoch": 0.002969850153457303, "grad_norm": 3.9950616359710693, "learning_rate": 4.985195883733526e-05, "loss": 5.1463, "step": 329 }, { "epoch": 0.002978877053619787, "grad_norm": 5.53897762298584, "learning_rate": 4.9851507492327134e-05, "loss": 4.6632, "step": 330 }, { "epoch": 0.002987903953782271, "grad_norm": 5.401761054992676, "learning_rate": 4.9851056147319014e-05, "loss": 5.3295, "step": 331 }, { "epoch": 0.0029969308539447553, "grad_norm": 5.14931058883667, "learning_rate": 4.985060480231089e-05, "loss": 5.531, "step": 332 }, { "epoch": 0.0030059577541072394, "grad_norm": 6.243673324584961, "learning_rate": 4.9850153457302765e-05, "loss": 4.6234, "step": 333 }, { "epoch": 0.003014984654269724, "grad_norm": 4.365715026855469, "learning_rate": 4.984970211229464e-05, "loss": 3.9061, "step": 334 }, { "epoch": 0.003024011554432208, "grad_norm": 3.888871431350708, "learning_rate": 4.984925076728652e-05, "loss": 5.126, "step": 335 }, { "epoch": 0.0030330384545946922, "grad_norm": 3.5373406410217285, "learning_rate": 4.984879942227839e-05, "loss": 5.0292, "step": 336 }, { "epoch": 0.0030420653547571764, "grad_norm": 5.9421563148498535, "learning_rate": 4.984834807727027e-05, "loss": 5.0872, "step": 337 }, { "epoch": 0.0030510922549196605, "grad_norm": 6.237949371337891, "learning_rate": 4.984789673226215e-05, "loss": 5.1613, "step": 338 }, { "epoch": 0.0030601191550821446, "grad_norm": 5.865431308746338, "learning_rate": 4.984744538725402e-05, "loss": 5.23, "step": 339 }, { "epoch": 0.003069146055244629, "grad_norm": 4.015934944152832, "learning_rate": 4.984699404224589e-05, "loss": 5.2995, "step": 340 }, { "epoch": 0.0030781729554071133, "grad_norm": 3.8161776065826416, "learning_rate": 4.984654269723777e-05, "loss": 4.5557, "step": 341 }, { "epoch": 0.0030871998555695975, "grad_norm": 7.657299041748047, "learning_rate": 4.9846091352229645e-05, "loss": 5.268, "step": 342 }, { "epoch": 0.0030962267557320816, "grad_norm": 4.965270519256592, "learning_rate": 4.9845640007221524e-05, "loss": 5.6208, "step": 343 }, { "epoch": 0.0031052536558945657, "grad_norm": 6.486219882965088, "learning_rate": 4.9845188662213396e-05, "loss": 5.2925, "step": 344 }, { "epoch": 0.00311428055605705, "grad_norm": 3.1749067306518555, "learning_rate": 4.9844737317205275e-05, "loss": 5.5416, "step": 345 }, { "epoch": 0.003123307456219534, "grad_norm": 4.717384338378906, "learning_rate": 4.9844285972197155e-05, "loss": 5.7808, "step": 346 }, { "epoch": 0.0031323343563820186, "grad_norm": 4.224466323852539, "learning_rate": 4.984383462718902e-05, "loss": 5.323, "step": 347 }, { "epoch": 0.0031413612565445027, "grad_norm": 6.081058979034424, "learning_rate": 4.98433832821809e-05, "loss": 5.467, "step": 348 }, { "epoch": 0.003150388156706987, "grad_norm": 4.749838352203369, "learning_rate": 4.984293193717278e-05, "loss": 5.2395, "step": 349 }, { "epoch": 0.003159415056869471, "grad_norm": 3.4874496459960938, "learning_rate": 4.984248059216465e-05, "loss": 4.762, "step": 350 }, { "epoch": 0.003168441957031955, "grad_norm": 3.8021240234375, "learning_rate": 4.984202924715653e-05, "loss": 4.7239, "step": 351 }, { "epoch": 0.0031774688571944393, "grad_norm": 4.484318733215332, "learning_rate": 4.98415779021484e-05, "loss": 5.205, "step": 352 }, { "epoch": 0.003186495757356924, "grad_norm": 3.6944148540496826, "learning_rate": 4.984112655714028e-05, "loss": 4.6681, "step": 353 }, { "epoch": 0.003195522657519408, "grad_norm": 4.382192134857178, "learning_rate": 4.9840675212132155e-05, "loss": 5.2391, "step": 354 }, { "epoch": 0.003204549557681892, "grad_norm": 3.708012104034424, "learning_rate": 4.9840223867124034e-05, "loss": 4.992, "step": 355 }, { "epoch": 0.0032135764578443762, "grad_norm": 3.93467116355896, "learning_rate": 4.9839772522115906e-05, "loss": 5.0233, "step": 356 }, { "epoch": 0.0032226033580068604, "grad_norm": 4.558797359466553, "learning_rate": 4.9839321177107786e-05, "loss": 4.8794, "step": 357 }, { "epoch": 0.0032316302581693445, "grad_norm": 5.2230753898620605, "learning_rate": 4.983886983209966e-05, "loss": 4.5605, "step": 358 }, { "epoch": 0.0032406571583318286, "grad_norm": 4.740249156951904, "learning_rate": 4.983841848709154e-05, "loss": 4.9643, "step": 359 }, { "epoch": 0.003249684058494313, "grad_norm": 3.807795524597168, "learning_rate": 4.983796714208341e-05, "loss": 5.1832, "step": 360 }, { "epoch": 0.0032587109586567973, "grad_norm": 6.207362651824951, "learning_rate": 4.983751579707528e-05, "loss": 4.8661, "step": 361 }, { "epoch": 0.0032677378588192815, "grad_norm": 4.379618167877197, "learning_rate": 4.983706445206716e-05, "loss": 4.9094, "step": 362 }, { "epoch": 0.0032767647589817656, "grad_norm": 4.191213130950928, "learning_rate": 4.983661310705904e-05, "loss": 5.3344, "step": 363 }, { "epoch": 0.0032857916591442497, "grad_norm": 3.7250406742095947, "learning_rate": 4.983616176205091e-05, "loss": 5.0466, "step": 364 }, { "epoch": 0.003294818559306734, "grad_norm": 4.924612045288086, "learning_rate": 4.983571041704279e-05, "loss": 5.1507, "step": 365 }, { "epoch": 0.0033038454594692184, "grad_norm": 6.322413921356201, "learning_rate": 4.9835259072034665e-05, "loss": 4.8008, "step": 366 }, { "epoch": 0.0033128723596317026, "grad_norm": 5.71626091003418, "learning_rate": 4.983480772702654e-05, "loss": 4.1709, "step": 367 }, { "epoch": 0.0033218992597941867, "grad_norm": 5.02719259262085, "learning_rate": 4.983435638201842e-05, "loss": 4.1318, "step": 368 }, { "epoch": 0.003330926159956671, "grad_norm": 4.562554836273193, "learning_rate": 4.983390503701029e-05, "loss": 5.4321, "step": 369 }, { "epoch": 0.003339953060119155, "grad_norm": 3.8800840377807617, "learning_rate": 4.983345369200217e-05, "loss": 4.5245, "step": 370 }, { "epoch": 0.003348979960281639, "grad_norm": 5.474725246429443, "learning_rate": 4.983300234699405e-05, "loss": 4.149, "step": 371 }, { "epoch": 0.0033580068604441237, "grad_norm": 4.693332672119141, "learning_rate": 4.983255100198592e-05, "loss": 5.5079, "step": 372 }, { "epoch": 0.003367033760606608, "grad_norm": 4.1861162185668945, "learning_rate": 4.98320996569778e-05, "loss": 5.157, "step": 373 }, { "epoch": 0.003376060660769092, "grad_norm": 4.066372871398926, "learning_rate": 4.983164831196967e-05, "loss": 5.2813, "step": 374 }, { "epoch": 0.003385087560931576, "grad_norm": 4.761430740356445, "learning_rate": 4.9831196966961544e-05, "loss": 5.5197, "step": 375 }, { "epoch": 0.0033941144610940602, "grad_norm": 4.395769119262695, "learning_rate": 4.9830745621953424e-05, "loss": 4.6381, "step": 376 }, { "epoch": 0.0034031413612565444, "grad_norm": 4.972754955291748, "learning_rate": 4.98302942769453e-05, "loss": 4.7258, "step": 377 }, { "epoch": 0.0034121682614190285, "grad_norm": 4.7737226486206055, "learning_rate": 4.9829842931937175e-05, "loss": 5.3336, "step": 378 }, { "epoch": 0.003421195161581513, "grad_norm": 4.4836297035217285, "learning_rate": 4.9829391586929054e-05, "loss": 5.393, "step": 379 }, { "epoch": 0.003430222061743997, "grad_norm": 3.5221686363220215, "learning_rate": 4.982894024192093e-05, "loss": 4.8024, "step": 380 }, { "epoch": 0.0034392489619064813, "grad_norm": 4.760165691375732, "learning_rate": 4.98284888969128e-05, "loss": 5.6444, "step": 381 }, { "epoch": 0.0034482758620689655, "grad_norm": 3.5919387340545654, "learning_rate": 4.982803755190468e-05, "loss": 4.9364, "step": 382 }, { "epoch": 0.0034573027622314496, "grad_norm": 4.425902366638184, "learning_rate": 4.982758620689655e-05, "loss": 5.0533, "step": 383 }, { "epoch": 0.0034663296623939337, "grad_norm": 4.7572221755981445, "learning_rate": 4.982713486188843e-05, "loss": 4.7505, "step": 384 }, { "epoch": 0.0034753565625564183, "grad_norm": 5.934632778167725, "learning_rate": 4.982668351688031e-05, "loss": 5.3276, "step": 385 }, { "epoch": 0.0034843834627189024, "grad_norm": 6.517950057983398, "learning_rate": 4.982623217187218e-05, "loss": 5.0857, "step": 386 }, { "epoch": 0.0034934103628813866, "grad_norm": 3.2945923805236816, "learning_rate": 4.9825780826864055e-05, "loss": 4.6072, "step": 387 }, { "epoch": 0.0035024372630438707, "grad_norm": 4.76303243637085, "learning_rate": 4.9825329481855934e-05, "loss": 5.4871, "step": 388 }, { "epoch": 0.003511464163206355, "grad_norm": 5.41910982131958, "learning_rate": 4.9824878136847806e-05, "loss": 4.8293, "step": 389 }, { "epoch": 0.003520491063368839, "grad_norm": 3.2078590393066406, "learning_rate": 4.9824426791839685e-05, "loss": 5.1824, "step": 390 }, { "epoch": 0.0035295179635313235, "grad_norm": 4.290473937988281, "learning_rate": 4.982397544683156e-05, "loss": 5.1793, "step": 391 }, { "epoch": 0.0035385448636938077, "grad_norm": 4.214196681976318, "learning_rate": 4.982352410182344e-05, "loss": 5.1251, "step": 392 }, { "epoch": 0.003547571763856292, "grad_norm": 5.052911281585693, "learning_rate": 4.9823072756815316e-05, "loss": 5.0911, "step": 393 }, { "epoch": 0.003556598664018776, "grad_norm": 4.725828647613525, "learning_rate": 4.982262141180718e-05, "loss": 5.3266, "step": 394 }, { "epoch": 0.00356562556418126, "grad_norm": 6.397512912750244, "learning_rate": 4.982217006679906e-05, "loss": 5.551, "step": 395 }, { "epoch": 0.0035746524643437442, "grad_norm": 5.240971088409424, "learning_rate": 4.982171872179094e-05, "loss": 5.2658, "step": 396 }, { "epoch": 0.0035836793645062284, "grad_norm": 4.525820255279541, "learning_rate": 4.982126737678281e-05, "loss": 5.0969, "step": 397 }, { "epoch": 0.003592706264668713, "grad_norm": 3.831834077835083, "learning_rate": 4.982081603177469e-05, "loss": 4.4978, "step": 398 }, { "epoch": 0.003601733164831197, "grad_norm": 4.098292350769043, "learning_rate": 4.982036468676657e-05, "loss": 5.4549, "step": 399 }, { "epoch": 0.003610760064993681, "grad_norm": 5.219669818878174, "learning_rate": 4.9819913341758444e-05, "loss": 5.5582, "step": 400 }, { "epoch": 0.0036197869651561653, "grad_norm": 5.8313212394714355, "learning_rate": 4.9819461996750316e-05, "loss": 5.7174, "step": 401 }, { "epoch": 0.0036288138653186495, "grad_norm": 4.371018409729004, "learning_rate": 4.9819010651742196e-05, "loss": 4.9919, "step": 402 }, { "epoch": 0.0036378407654811336, "grad_norm": 4.11134672164917, "learning_rate": 4.981855930673407e-05, "loss": 4.5267, "step": 403 }, { "epoch": 0.003646867665643618, "grad_norm": 7.486363410949707, "learning_rate": 4.981810796172595e-05, "loss": 5.3033, "step": 404 }, { "epoch": 0.0036558945658061023, "grad_norm": 4.528141498565674, "learning_rate": 4.981765661671782e-05, "loss": 5.2635, "step": 405 }, { "epoch": 0.0036649214659685864, "grad_norm": 6.950198650360107, "learning_rate": 4.98172052717097e-05, "loss": 4.8618, "step": 406 }, { "epoch": 0.0036739483661310706, "grad_norm": 7.496078968048096, "learning_rate": 4.981675392670157e-05, "loss": 4.6297, "step": 407 }, { "epoch": 0.0036829752662935547, "grad_norm": 6.058298587799072, "learning_rate": 4.9816302581693444e-05, "loss": 5.7335, "step": 408 }, { "epoch": 0.003692002166456039, "grad_norm": 4.63859224319458, "learning_rate": 4.981585123668532e-05, "loss": 5.5525, "step": 409 }, { "epoch": 0.0037010290666185234, "grad_norm": 5.666573524475098, "learning_rate": 4.98153998916772e-05, "loss": 4.8676, "step": 410 }, { "epoch": 0.0037100559667810075, "grad_norm": 5.700063705444336, "learning_rate": 4.9814948546669075e-05, "loss": 5.6006, "step": 411 }, { "epoch": 0.0037190828669434917, "grad_norm": 4.8149919509887695, "learning_rate": 4.9814497201660954e-05, "loss": 5.1078, "step": 412 }, { "epoch": 0.003728109767105976, "grad_norm": 7.192021369934082, "learning_rate": 4.9814045856652833e-05, "loss": 5.3307, "step": 413 }, { "epoch": 0.00373713666726846, "grad_norm": 3.43644380569458, "learning_rate": 4.98135945116447e-05, "loss": 5.5721, "step": 414 }, { "epoch": 0.003746163567430944, "grad_norm": 4.625597953796387, "learning_rate": 4.981314316663658e-05, "loss": 4.5042, "step": 415 }, { "epoch": 0.0037551904675934282, "grad_norm": 3.802258253097534, "learning_rate": 4.981269182162846e-05, "loss": 4.8623, "step": 416 }, { "epoch": 0.003764217367755913, "grad_norm": 3.6568260192871094, "learning_rate": 4.981224047662033e-05, "loss": 5.5506, "step": 417 }, { "epoch": 0.003773244267918397, "grad_norm": 4.443288326263428, "learning_rate": 4.981178913161221e-05, "loss": 4.3245, "step": 418 }, { "epoch": 0.003782271168080881, "grad_norm": 5.023831367492676, "learning_rate": 4.981133778660408e-05, "loss": 5.1164, "step": 419 }, { "epoch": 0.003791298068243365, "grad_norm": 5.405661106109619, "learning_rate": 4.981088644159596e-05, "loss": 5.141, "step": 420 }, { "epoch": 0.0038003249684058493, "grad_norm": 5.703847885131836, "learning_rate": 4.9810435096587834e-05, "loss": 3.6915, "step": 421 }, { "epoch": 0.0038093518685683335, "grad_norm": 3.5316848754882812, "learning_rate": 4.9809983751579706e-05, "loss": 5.4638, "step": 422 }, { "epoch": 0.003818378768730818, "grad_norm": 4.303348541259766, "learning_rate": 4.9809532406571585e-05, "loss": 4.7188, "step": 423 }, { "epoch": 0.003827405668893302, "grad_norm": 4.498260974884033, "learning_rate": 4.9809081061563464e-05, "loss": 4.8793, "step": 424 }, { "epoch": 0.0038364325690557863, "grad_norm": 4.554632186889648, "learning_rate": 4.980862971655534e-05, "loss": 4.8127, "step": 425 }, { "epoch": 0.0038454594692182704, "grad_norm": 7.275662422180176, "learning_rate": 4.9808178371547216e-05, "loss": 5.387, "step": 426 }, { "epoch": 0.0038544863693807546, "grad_norm": 6.1712822914123535, "learning_rate": 4.980772702653909e-05, "loss": 5.5465, "step": 427 }, { "epoch": 0.0038635132695432387, "grad_norm": 3.8665707111358643, "learning_rate": 4.980727568153096e-05, "loss": 5.789, "step": 428 }, { "epoch": 0.0038725401697057233, "grad_norm": 4.5868940353393555, "learning_rate": 4.980682433652284e-05, "loss": 5.2352, "step": 429 }, { "epoch": 0.0038815670698682074, "grad_norm": 4.74959659576416, "learning_rate": 4.980637299151471e-05, "loss": 3.467, "step": 430 }, { "epoch": 0.0038905939700306915, "grad_norm": 6.5316548347473145, "learning_rate": 4.980592164650659e-05, "loss": 4.8695, "step": 431 }, { "epoch": 0.0038996208701931757, "grad_norm": 6.091310977935791, "learning_rate": 4.980547030149847e-05, "loss": 5.627, "step": 432 }, { "epoch": 0.00390864777035566, "grad_norm": 4.465155601501465, "learning_rate": 4.9805018956490344e-05, "loss": 5.4661, "step": 433 }, { "epoch": 0.003917674670518144, "grad_norm": 4.367483615875244, "learning_rate": 4.9804567611482216e-05, "loss": 5.3764, "step": 434 }, { "epoch": 0.003926701570680628, "grad_norm": 7.906503677368164, "learning_rate": 4.9804116266474095e-05, "loss": 5.6269, "step": 435 }, { "epoch": 0.003935728470843112, "grad_norm": 4.422842979431152, "learning_rate": 4.980366492146597e-05, "loss": 4.7318, "step": 436 }, { "epoch": 0.003944755371005596, "grad_norm": 4.111903190612793, "learning_rate": 4.980321357645785e-05, "loss": 5.3229, "step": 437 }, { "epoch": 0.0039537822711680805, "grad_norm": 4.6188273429870605, "learning_rate": 4.9802762231449726e-05, "loss": 5.0774, "step": 438 }, { "epoch": 0.0039628091713305655, "grad_norm": 4.446236610412598, "learning_rate": 4.98023108864416e-05, "loss": 5.2993, "step": 439 }, { "epoch": 0.00397183607149305, "grad_norm": 4.033705234527588, "learning_rate": 4.980185954143348e-05, "loss": 4.5316, "step": 440 }, { "epoch": 0.003980862971655534, "grad_norm": 4.103837490081787, "learning_rate": 4.980140819642535e-05, "loss": 5.5363, "step": 441 }, { "epoch": 0.003989889871818018, "grad_norm": 4.516569137573242, "learning_rate": 4.980095685141722e-05, "loss": 4.6371, "step": 442 }, { "epoch": 0.003998916771980502, "grad_norm": 4.687666893005371, "learning_rate": 4.98005055064091e-05, "loss": 5.4001, "step": 443 }, { "epoch": 0.004007943672142986, "grad_norm": 5.3420867919921875, "learning_rate": 4.9800054161400975e-05, "loss": 4.9662, "step": 444 }, { "epoch": 0.00401697057230547, "grad_norm": 4.10528039932251, "learning_rate": 4.9799602816392854e-05, "loss": 3.7953, "step": 445 }, { "epoch": 0.004025997472467954, "grad_norm": 5.086775302886963, "learning_rate": 4.979915147138473e-05, "loss": 5.1009, "step": 446 }, { "epoch": 0.0040350243726304386, "grad_norm": 5.88107442855835, "learning_rate": 4.9798700126376606e-05, "loss": 4.6162, "step": 447 }, { "epoch": 0.004044051272792923, "grad_norm": 5.50982666015625, "learning_rate": 4.979824878136848e-05, "loss": 5.1845, "step": 448 }, { "epoch": 0.004053078172955407, "grad_norm": 3.8533332347869873, "learning_rate": 4.979779743636036e-05, "loss": 5.1999, "step": 449 }, { "epoch": 0.004062105073117891, "grad_norm": 4.919205188751221, "learning_rate": 4.979734609135223e-05, "loss": 5.5796, "step": 450 }, { "epoch": 0.004071131973280375, "grad_norm": 5.3799967765808105, "learning_rate": 4.979689474634411e-05, "loss": 5.3964, "step": 451 }, { "epoch": 0.00408015887344286, "grad_norm": 4.098440170288086, "learning_rate": 4.979644340133598e-05, "loss": 5.0508, "step": 452 }, { "epoch": 0.004089185773605344, "grad_norm": 6.5534868240356445, "learning_rate": 4.979599205632786e-05, "loss": 5.1232, "step": 453 }, { "epoch": 0.004098212673767828, "grad_norm": 4.338656902313232, "learning_rate": 4.979554071131973e-05, "loss": 5.0561, "step": 454 }, { "epoch": 0.0041072395739303125, "grad_norm": 6.794662952423096, "learning_rate": 4.979508936631161e-05, "loss": 4.5414, "step": 455 }, { "epoch": 0.004116266474092797, "grad_norm": 8.440375328063965, "learning_rate": 4.9794638021303485e-05, "loss": 4.8606, "step": 456 }, { "epoch": 0.004125293374255281, "grad_norm": 4.284940242767334, "learning_rate": 4.9794186676295364e-05, "loss": 5.3124, "step": 457 }, { "epoch": 0.004134320274417765, "grad_norm": 4.635278701782227, "learning_rate": 4.979373533128724e-05, "loss": 5.2824, "step": 458 }, { "epoch": 0.004143347174580249, "grad_norm": 4.140478610992432, "learning_rate": 4.9793283986279116e-05, "loss": 5.5934, "step": 459 }, { "epoch": 0.004152374074742733, "grad_norm": 4.449178218841553, "learning_rate": 4.9792832641270995e-05, "loss": 4.6429, "step": 460 }, { "epoch": 0.004161400974905217, "grad_norm": 5.124914169311523, "learning_rate": 4.979238129626286e-05, "loss": 4.6898, "step": 461 }, { "epoch": 0.0041704278750677014, "grad_norm": 4.736512660980225, "learning_rate": 4.979192995125474e-05, "loss": 4.1514, "step": 462 }, { "epoch": 0.004179454775230186, "grad_norm": 3.75803804397583, "learning_rate": 4.979147860624662e-05, "loss": 4.9236, "step": 463 }, { "epoch": 0.004188481675392671, "grad_norm": 8.251612663269043, "learning_rate": 4.979102726123849e-05, "loss": 4.8834, "step": 464 }, { "epoch": 0.004197508575555155, "grad_norm": 3.8943066596984863, "learning_rate": 4.979057591623037e-05, "loss": 5.0097, "step": 465 }, { "epoch": 0.004206535475717639, "grad_norm": 6.691514492034912, "learning_rate": 4.9790124571222244e-05, "loss": 5.0866, "step": 466 }, { "epoch": 0.004215562375880123, "grad_norm": 6.288057327270508, "learning_rate": 4.978967322621412e-05, "loss": 5.1816, "step": 467 }, { "epoch": 0.004224589276042607, "grad_norm": 4.202695369720459, "learning_rate": 4.9789221881205995e-05, "loss": 4.9953, "step": 468 }, { "epoch": 0.004233616176205091, "grad_norm": 4.616065979003906, "learning_rate": 4.978877053619787e-05, "loss": 4.3972, "step": 469 }, { "epoch": 0.004242643076367575, "grad_norm": 5.451911449432373, "learning_rate": 4.978831919118975e-05, "loss": 5.4691, "step": 470 }, { "epoch": 0.0042516699765300595, "grad_norm": 6.316236972808838, "learning_rate": 4.9787867846181626e-05, "loss": 5.4561, "step": 471 }, { "epoch": 0.004260696876692544, "grad_norm": 5.914906978607178, "learning_rate": 4.97874165011735e-05, "loss": 4.5183, "step": 472 }, { "epoch": 0.004269723776855028, "grad_norm": 3.8779654502868652, "learning_rate": 4.978696515616538e-05, "loss": 5.8229, "step": 473 }, { "epoch": 0.004278750677017512, "grad_norm": 3.7703235149383545, "learning_rate": 4.978651381115725e-05, "loss": 5.1507, "step": 474 }, { "epoch": 0.004287777577179996, "grad_norm": 4.313806056976318, "learning_rate": 4.978606246614912e-05, "loss": 5.2388, "step": 475 }, { "epoch": 0.00429680447734248, "grad_norm": 5.320243835449219, "learning_rate": 4.9785611121141e-05, "loss": 5.1719, "step": 476 }, { "epoch": 0.004305831377504965, "grad_norm": 3.3057007789611816, "learning_rate": 4.978515977613288e-05, "loss": 5.0132, "step": 477 }, { "epoch": 0.004314858277667449, "grad_norm": 3.6631019115448, "learning_rate": 4.9784708431124754e-05, "loss": 5.1234, "step": 478 }, { "epoch": 0.0043238851778299335, "grad_norm": 3.7721569538116455, "learning_rate": 4.978425708611663e-05, "loss": 4.1837, "step": 479 }, { "epoch": 0.004332912077992418, "grad_norm": 4.2821364402771, "learning_rate": 4.9783805741108505e-05, "loss": 3.913, "step": 480 }, { "epoch": 0.004341938978154902, "grad_norm": 5.1245598793029785, "learning_rate": 4.978335439610038e-05, "loss": 5.6562, "step": 481 }, { "epoch": 0.004350965878317386, "grad_norm": 3.9139137268066406, "learning_rate": 4.978290305109226e-05, "loss": 5.1161, "step": 482 }, { "epoch": 0.00435999277847987, "grad_norm": 3.714146137237549, "learning_rate": 4.978245170608413e-05, "loss": 4.6567, "step": 483 }, { "epoch": 0.004369019678642354, "grad_norm": 4.875441074371338, "learning_rate": 4.978200036107601e-05, "loss": 5.038, "step": 484 }, { "epoch": 0.004378046578804838, "grad_norm": 5.037322521209717, "learning_rate": 4.978154901606789e-05, "loss": 5.4423, "step": 485 }, { "epoch": 0.004387073478967322, "grad_norm": 4.726399898529053, "learning_rate": 4.978109767105976e-05, "loss": 5.0756, "step": 486 }, { "epoch": 0.0043961003791298065, "grad_norm": 4.796133995056152, "learning_rate": 4.978064632605164e-05, "loss": 4.6266, "step": 487 }, { "epoch": 0.004405127279292291, "grad_norm": 5.79962158203125, "learning_rate": 4.978019498104351e-05, "loss": 3.8762, "step": 488 }, { "epoch": 0.004414154179454775, "grad_norm": 3.4478869438171387, "learning_rate": 4.9779743636035385e-05, "loss": 5.189, "step": 489 }, { "epoch": 0.00442318107961726, "grad_norm": 4.47123384475708, "learning_rate": 4.9779292291027264e-05, "loss": 4.8787, "step": 490 }, { "epoch": 0.004432207979779744, "grad_norm": 3.6668589115142822, "learning_rate": 4.9778840946019136e-05, "loss": 5.0212, "step": 491 }, { "epoch": 0.004441234879942228, "grad_norm": 6.017575740814209, "learning_rate": 4.9778389601011016e-05, "loss": 5.3932, "step": 492 }, { "epoch": 0.004450261780104712, "grad_norm": 4.627450942993164, "learning_rate": 4.9777938256002895e-05, "loss": 5.1537, "step": 493 }, { "epoch": 0.004459288680267196, "grad_norm": 4.104442119598389, "learning_rate": 4.977748691099477e-05, "loss": 5.5997, "step": 494 }, { "epoch": 0.0044683155804296805, "grad_norm": 3.882295846939087, "learning_rate": 4.977703556598664e-05, "loss": 5.3063, "step": 495 }, { "epoch": 0.004477342480592165, "grad_norm": 5.108547687530518, "learning_rate": 4.977658422097852e-05, "loss": 5.2353, "step": 496 }, { "epoch": 0.004486369380754649, "grad_norm": 4.615786075592041, "learning_rate": 4.977613287597039e-05, "loss": 4.7243, "step": 497 }, { "epoch": 0.004495396280917133, "grad_norm": 3.8889360427856445, "learning_rate": 4.977568153096227e-05, "loss": 5.5478, "step": 498 }, { "epoch": 0.004504423181079617, "grad_norm": 4.672197341918945, "learning_rate": 4.977523018595415e-05, "loss": 5.0678, "step": 499 }, { "epoch": 0.004513450081242101, "grad_norm": 3.692472457885742, "learning_rate": 4.977477884094602e-05, "loss": 4.6479, "step": 500 }, { "epoch": 0.004522476981404585, "grad_norm": 4.0645599365234375, "learning_rate": 4.9774327495937895e-05, "loss": 5.2946, "step": 501 }, { "epoch": 0.0045315038815670694, "grad_norm": 4.748575687408447, "learning_rate": 4.9773876150929774e-05, "loss": 4.7368, "step": 502 }, { "epoch": 0.0045405307817295544, "grad_norm": 4.782742500305176, "learning_rate": 4.977342480592165e-05, "loss": 5.3433, "step": 503 }, { "epoch": 0.004549557681892039, "grad_norm": 5.095293998718262, "learning_rate": 4.9772973460913526e-05, "loss": 5.6053, "step": 504 }, { "epoch": 0.004558584582054523, "grad_norm": 3.755499839782715, "learning_rate": 4.97725221159054e-05, "loss": 5.3312, "step": 505 }, { "epoch": 0.004567611482217007, "grad_norm": 5.836657524108887, "learning_rate": 4.977207077089728e-05, "loss": 4.9453, "step": 506 }, { "epoch": 0.004576638382379491, "grad_norm": 4.7052788734436035, "learning_rate": 4.977161942588916e-05, "loss": 4.2011, "step": 507 }, { "epoch": 0.004585665282541975, "grad_norm": 5.432224750518799, "learning_rate": 4.977116808088102e-05, "loss": 5.2616, "step": 508 }, { "epoch": 0.004594692182704459, "grad_norm": 4.421679973602295, "learning_rate": 4.97707167358729e-05, "loss": 5.4967, "step": 509 }, { "epoch": 0.004603719082866943, "grad_norm": 4.374495983123779, "learning_rate": 4.977026539086478e-05, "loss": 5.3736, "step": 510 }, { "epoch": 0.0046127459830294275, "grad_norm": 4.062595367431641, "learning_rate": 4.9769814045856653e-05, "loss": 4.9874, "step": 511 }, { "epoch": 0.004621772883191912, "grad_norm": 4.457193851470947, "learning_rate": 4.976936270084853e-05, "loss": 4.9379, "step": 512 }, { "epoch": 0.004630799783354396, "grad_norm": 4.17273473739624, "learning_rate": 4.9768911355840405e-05, "loss": 5.2996, "step": 513 }, { "epoch": 0.00463982668351688, "grad_norm": 3.009300708770752, "learning_rate": 4.9768460010832284e-05, "loss": 5.2866, "step": 514 }, { "epoch": 0.004648853583679365, "grad_norm": 4.124131202697754, "learning_rate": 4.976800866582416e-05, "loss": 4.6521, "step": 515 }, { "epoch": 0.004657880483841849, "grad_norm": 4.44077205657959, "learning_rate": 4.9767557320816036e-05, "loss": 4.8261, "step": 516 }, { "epoch": 0.004666907384004333, "grad_norm": 3.6857316493988037, "learning_rate": 4.976710597580791e-05, "loss": 5.2483, "step": 517 }, { "epoch": 0.004675934284166817, "grad_norm": 4.320807933807373, "learning_rate": 4.976665463079979e-05, "loss": 4.7202, "step": 518 }, { "epoch": 0.0046849611843293015, "grad_norm": 4.260260105133057, "learning_rate": 4.976620328579166e-05, "loss": 5.109, "step": 519 }, { "epoch": 0.004693988084491786, "grad_norm": 10.7841157913208, "learning_rate": 4.976575194078354e-05, "loss": 5.015, "step": 520 }, { "epoch": 0.00470301498465427, "grad_norm": 5.414941310882568, "learning_rate": 4.976530059577541e-05, "loss": 4.9631, "step": 521 }, { "epoch": 0.004712041884816754, "grad_norm": 4.7589497566223145, "learning_rate": 4.9764849250767284e-05, "loss": 4.8916, "step": 522 }, { "epoch": 0.004721068784979238, "grad_norm": 5.033052921295166, "learning_rate": 4.9764397905759164e-05, "loss": 4.6693, "step": 523 }, { "epoch": 0.004730095685141722, "grad_norm": 4.67963171005249, "learning_rate": 4.976394656075104e-05, "loss": 4.901, "step": 524 }, { "epoch": 0.004739122585304206, "grad_norm": 6.182500839233398, "learning_rate": 4.9763495215742915e-05, "loss": 5.5538, "step": 525 }, { "epoch": 0.00474814948546669, "grad_norm": 3.9078755378723145, "learning_rate": 4.9763043870734795e-05, "loss": 4.6602, "step": 526 }, { "epoch": 0.0047571763856291745, "grad_norm": 4.114792346954346, "learning_rate": 4.976259252572667e-05, "loss": 5.1687, "step": 527 }, { "epoch": 0.0047662032857916595, "grad_norm": 6.065999507904053, "learning_rate": 4.976214118071854e-05, "loss": 4.8042, "step": 528 }, { "epoch": 0.004775230185954144, "grad_norm": 4.480125904083252, "learning_rate": 4.976168983571042e-05, "loss": 4.686, "step": 529 }, { "epoch": 0.004784257086116628, "grad_norm": 3.6080777645111084, "learning_rate": 4.976123849070229e-05, "loss": 5.0082, "step": 530 }, { "epoch": 0.004793283986279112, "grad_norm": 2.7340734004974365, "learning_rate": 4.976078714569417e-05, "loss": 5.0481, "step": 531 }, { "epoch": 0.004802310886441596, "grad_norm": 6.441117763519287, "learning_rate": 4.976033580068605e-05, "loss": 4.4959, "step": 532 }, { "epoch": 0.00481133778660408, "grad_norm": 4.0671234130859375, "learning_rate": 4.975988445567792e-05, "loss": 4.4381, "step": 533 }, { "epoch": 0.004820364686766564, "grad_norm": 5.670238971710205, "learning_rate": 4.97594331106698e-05, "loss": 3.6814, "step": 534 }, { "epoch": 0.0048293915869290485, "grad_norm": 4.818734645843506, "learning_rate": 4.9758981765661674e-05, "loss": 5.0674, "step": 535 }, { "epoch": 0.004838418487091533, "grad_norm": 5.502256393432617, "learning_rate": 4.9758530420653546e-05, "loss": 5.2853, "step": 536 }, { "epoch": 0.004847445387254017, "grad_norm": 5.687674045562744, "learning_rate": 4.9758079075645426e-05, "loss": 5.3215, "step": 537 }, { "epoch": 0.004856472287416501, "grad_norm": 4.3550496101379395, "learning_rate": 4.9757627730637305e-05, "loss": 5.3277, "step": 538 }, { "epoch": 0.004865499187578985, "grad_norm": 3.9432897567749023, "learning_rate": 4.975717638562918e-05, "loss": 5.3455, "step": 539 }, { "epoch": 0.004874526087741469, "grad_norm": 3.9839696884155273, "learning_rate": 4.975672504062106e-05, "loss": 5.2441, "step": 540 }, { "epoch": 0.004883552987903954, "grad_norm": 4.151726245880127, "learning_rate": 4.975627369561293e-05, "loss": 4.9794, "step": 541 }, { "epoch": 0.004892579888066438, "grad_norm": 4.067487716674805, "learning_rate": 4.97558223506048e-05, "loss": 5.5334, "step": 542 }, { "epoch": 0.004901606788228922, "grad_norm": 4.388070106506348, "learning_rate": 4.975537100559668e-05, "loss": 5.3499, "step": 543 }, { "epoch": 0.0049106336883914066, "grad_norm": 15.404187202453613, "learning_rate": 4.975491966058855e-05, "loss": 5.3839, "step": 544 }, { "epoch": 0.004919660588553891, "grad_norm": 6.11668062210083, "learning_rate": 4.975446831558043e-05, "loss": 4.3841, "step": 545 }, { "epoch": 0.004928687488716375, "grad_norm": 4.036369800567627, "learning_rate": 4.975401697057231e-05, "loss": 5.2605, "step": 546 }, { "epoch": 0.004937714388878859, "grad_norm": 4.996504783630371, "learning_rate": 4.9753565625564184e-05, "loss": 4.9125, "step": 547 }, { "epoch": 0.004946741289041343, "grad_norm": 4.536583423614502, "learning_rate": 4.975311428055606e-05, "loss": 4.8613, "step": 548 }, { "epoch": 0.004955768189203827, "grad_norm": 3.8143222332000732, "learning_rate": 4.9752662935547936e-05, "loss": 4.9157, "step": 549 }, { "epoch": 0.004964795089366311, "grad_norm": 3.0844266414642334, "learning_rate": 4.975221159053981e-05, "loss": 4.6415, "step": 550 }, { "epoch": 0.0049738219895287955, "grad_norm": 4.853735446929932, "learning_rate": 4.975176024553169e-05, "loss": 5.2576, "step": 551 }, { "epoch": 0.00498284888969128, "grad_norm": 4.95042085647583, "learning_rate": 4.975130890052356e-05, "loss": 5.29, "step": 552 }, { "epoch": 0.004991875789853765, "grad_norm": 3.789947032928467, "learning_rate": 4.975085755551544e-05, "loss": 5.2218, "step": 553 }, { "epoch": 0.005000902690016249, "grad_norm": 4.229473114013672, "learning_rate": 4.975040621050732e-05, "loss": 4.7305, "step": 554 }, { "epoch": 0.005009929590178733, "grad_norm": 3.6703059673309326, "learning_rate": 4.9749954865499184e-05, "loss": 5.1329, "step": 555 }, { "epoch": 0.005018956490341217, "grad_norm": 4.352149486541748, "learning_rate": 4.9749503520491063e-05, "loss": 4.921, "step": 556 }, { "epoch": 0.005027983390503701, "grad_norm": 3.9177727699279785, "learning_rate": 4.974905217548294e-05, "loss": 4.6863, "step": 557 }, { "epoch": 0.005037010290666185, "grad_norm": 7.922440528869629, "learning_rate": 4.9748600830474815e-05, "loss": 5.5767, "step": 558 }, { "epoch": 0.0050460371908286695, "grad_norm": 7.0821452140808105, "learning_rate": 4.9748149485466694e-05, "loss": 4.6192, "step": 559 }, { "epoch": 0.005055064090991154, "grad_norm": 5.447481155395508, "learning_rate": 4.9747698140458574e-05, "loss": 5.5641, "step": 560 }, { "epoch": 0.005064090991153638, "grad_norm": 5.350694179534912, "learning_rate": 4.9747246795450446e-05, "loss": 5.4713, "step": 561 }, { "epoch": 0.005073117891316122, "grad_norm": 5.778390884399414, "learning_rate": 4.974679545044232e-05, "loss": 5.1052, "step": 562 }, { "epoch": 0.005082144791478606, "grad_norm": 8.906133651733398, "learning_rate": 4.97463441054342e-05, "loss": 5.178, "step": 563 }, { "epoch": 0.00509117169164109, "grad_norm": 3.600684881210327, "learning_rate": 4.974589276042607e-05, "loss": 4.5028, "step": 564 }, { "epoch": 0.005100198591803574, "grad_norm": 4.6290388107299805, "learning_rate": 4.974544141541795e-05, "loss": 5.0557, "step": 565 }, { "epoch": 0.005109225491966059, "grad_norm": 4.240969657897949, "learning_rate": 4.974499007040982e-05, "loss": 4.8862, "step": 566 }, { "epoch": 0.005118252392128543, "grad_norm": 4.235538959503174, "learning_rate": 4.97445387254017e-05, "loss": 5.0071, "step": 567 }, { "epoch": 0.0051272792922910275, "grad_norm": 3.869985342025757, "learning_rate": 4.9744087380393574e-05, "loss": 5.0128, "step": 568 }, { "epoch": 0.005136306192453512, "grad_norm": 4.873660564422607, "learning_rate": 4.9743636035385446e-05, "loss": 4.8014, "step": 569 }, { "epoch": 0.005145333092615996, "grad_norm": 2.8679840564727783, "learning_rate": 4.9743184690377325e-05, "loss": 4.3027, "step": 570 }, { "epoch": 0.00515435999277848, "grad_norm": 3.9443540573120117, "learning_rate": 4.9742733345369205e-05, "loss": 4.9393, "step": 571 }, { "epoch": 0.005163386892940964, "grad_norm": 4.239754676818848, "learning_rate": 4.974228200036108e-05, "loss": 5.3883, "step": 572 }, { "epoch": 0.005172413793103448, "grad_norm": 4.684035778045654, "learning_rate": 4.9741830655352956e-05, "loss": 4.4068, "step": 573 }, { "epoch": 0.005181440693265932, "grad_norm": 6.2101898193359375, "learning_rate": 4.9741379310344836e-05, "loss": 5.0126, "step": 574 }, { "epoch": 0.0051904675934284165, "grad_norm": 5.692546367645264, "learning_rate": 4.97409279653367e-05, "loss": 4.6631, "step": 575 }, { "epoch": 0.005199494493590901, "grad_norm": 3.2562050819396973, "learning_rate": 4.974047662032858e-05, "loss": 4.6258, "step": 576 }, { "epoch": 0.005208521393753385, "grad_norm": 4.145088195800781, "learning_rate": 4.974002527532046e-05, "loss": 5.2915, "step": 577 }, { "epoch": 0.005217548293915869, "grad_norm": 4.073465347290039, "learning_rate": 4.973957393031233e-05, "loss": 5.4031, "step": 578 }, { "epoch": 0.005226575194078354, "grad_norm": 4.712744235992432, "learning_rate": 4.973912258530421e-05, "loss": 5.2819, "step": 579 }, { "epoch": 0.005235602094240838, "grad_norm": 4.431913375854492, "learning_rate": 4.9738671240296084e-05, "loss": 5.5518, "step": 580 }, { "epoch": 0.005244628994403322, "grad_norm": 4.845635414123535, "learning_rate": 4.973821989528796e-05, "loss": 5.4822, "step": 581 }, { "epoch": 0.005253655894565806, "grad_norm": 4.462169170379639, "learning_rate": 4.9737768550279836e-05, "loss": 5.1756, "step": 582 }, { "epoch": 0.00526268279472829, "grad_norm": 4.003049373626709, "learning_rate": 4.973731720527171e-05, "loss": 4.6684, "step": 583 }, { "epoch": 0.0052717096948907746, "grad_norm": 4.827277183532715, "learning_rate": 4.973686586026359e-05, "loss": 4.917, "step": 584 }, { "epoch": 0.005280736595053259, "grad_norm": 3.5793256759643555, "learning_rate": 4.9736414515255467e-05, "loss": 4.9107, "step": 585 }, { "epoch": 0.005289763495215743, "grad_norm": 6.659903526306152, "learning_rate": 4.973596317024734e-05, "loss": 5.5367, "step": 586 }, { "epoch": 0.005298790395378227, "grad_norm": 3.553875684738159, "learning_rate": 4.973551182523922e-05, "loss": 4.9975, "step": 587 }, { "epoch": 0.005307817295540711, "grad_norm": 7.168386936187744, "learning_rate": 4.973506048023109e-05, "loss": 5.0935, "step": 588 }, { "epoch": 0.005316844195703195, "grad_norm": 6.157546043395996, "learning_rate": 4.973460913522296e-05, "loss": 5.0208, "step": 589 }, { "epoch": 0.005325871095865679, "grad_norm": 4.707059383392334, "learning_rate": 4.973415779021484e-05, "loss": 5.8686, "step": 590 }, { "epoch": 0.0053348979960281635, "grad_norm": 3.6856613159179688, "learning_rate": 4.9733706445206715e-05, "loss": 4.8117, "step": 591 }, { "epoch": 0.0053439248961906485, "grad_norm": 4.508827209472656, "learning_rate": 4.9733255100198594e-05, "loss": 5.152, "step": 592 }, { "epoch": 0.005352951796353133, "grad_norm": 4.751213550567627, "learning_rate": 4.9732803755190473e-05, "loss": 3.6716, "step": 593 }, { "epoch": 0.005361978696515617, "grad_norm": 7.489810466766357, "learning_rate": 4.9732352410182346e-05, "loss": 4.1036, "step": 594 }, { "epoch": 0.005371005596678101, "grad_norm": 5.1193766593933105, "learning_rate": 4.973190106517422e-05, "loss": 5.4849, "step": 595 }, { "epoch": 0.005380032496840585, "grad_norm": 3.623011589050293, "learning_rate": 4.97314497201661e-05, "loss": 4.8448, "step": 596 }, { "epoch": 0.005389059397003069, "grad_norm": 6.721268653869629, "learning_rate": 4.973099837515797e-05, "loss": 5.4434, "step": 597 }, { "epoch": 0.005398086297165553, "grad_norm": 4.074625015258789, "learning_rate": 4.973054703014985e-05, "loss": 5.2166, "step": 598 }, { "epoch": 0.0054071131973280374, "grad_norm": 3.9862937927246094, "learning_rate": 4.973009568514173e-05, "loss": 4.5557, "step": 599 }, { "epoch": 0.005416140097490522, "grad_norm": 5.367059230804443, "learning_rate": 4.97296443401336e-05, "loss": 5.2215, "step": 600 }, { "epoch": 0.005425166997653006, "grad_norm": 3.5886390209198, "learning_rate": 4.972919299512548e-05, "loss": 4.7601, "step": 601 }, { "epoch": 0.00543419389781549, "grad_norm": 5.050940036773682, "learning_rate": 4.972874165011735e-05, "loss": 5.2349, "step": 602 }, { "epoch": 0.005443220797977974, "grad_norm": 4.378365993499756, "learning_rate": 4.9728290305109225e-05, "loss": 5.2813, "step": 603 }, { "epoch": 0.005452247698140459, "grad_norm": 4.275803565979004, "learning_rate": 4.9727838960101104e-05, "loss": 4.4222, "step": 604 }, { "epoch": 0.005461274598302943, "grad_norm": 4.310400009155273, "learning_rate": 4.972738761509298e-05, "loss": 3.9875, "step": 605 }, { "epoch": 0.005470301498465427, "grad_norm": 5.995329856872559, "learning_rate": 4.9726936270084856e-05, "loss": 4.7611, "step": 606 }, { "epoch": 0.005479328398627911, "grad_norm": 3.9013166427612305, "learning_rate": 4.9726484925076735e-05, "loss": 4.9312, "step": 607 }, { "epoch": 0.0054883552987903955, "grad_norm": 4.7266082763671875, "learning_rate": 4.972603358006861e-05, "loss": 5.5214, "step": 608 }, { "epoch": 0.00549738219895288, "grad_norm": 7.650891304016113, "learning_rate": 4.972558223506048e-05, "loss": 5.1727, "step": 609 }, { "epoch": 0.005506409099115364, "grad_norm": 4.53589391708374, "learning_rate": 4.972513089005236e-05, "loss": 4.8541, "step": 610 }, { "epoch": 0.005515435999277848, "grad_norm": 4.910439491271973, "learning_rate": 4.972467954504423e-05, "loss": 4.7788, "step": 611 }, { "epoch": 0.005524462899440332, "grad_norm": 3.8983795642852783, "learning_rate": 4.972422820003611e-05, "loss": 4.9884, "step": 612 }, { "epoch": 0.005533489799602816, "grad_norm": 5.057757377624512, "learning_rate": 4.9723776855027984e-05, "loss": 4.9584, "step": 613 }, { "epoch": 0.0055425166997653, "grad_norm": 3.9963746070861816, "learning_rate": 4.972332551001986e-05, "loss": 5.0223, "step": 614 }, { "epoch": 0.0055515435999277845, "grad_norm": 3.726250648498535, "learning_rate": 4.9722874165011735e-05, "loss": 4.883, "step": 615 }, { "epoch": 0.005560570500090269, "grad_norm": 3.726716995239258, "learning_rate": 4.9722422820003615e-05, "loss": 5.0352, "step": 616 }, { "epoch": 0.005569597400252754, "grad_norm": 4.219272136688232, "learning_rate": 4.972197147499549e-05, "loss": 4.6631, "step": 617 }, { "epoch": 0.005578624300415238, "grad_norm": 2.801262140274048, "learning_rate": 4.9721520129987366e-05, "loss": 5.3553, "step": 618 }, { "epoch": 0.005587651200577722, "grad_norm": 4.912095546722412, "learning_rate": 4.972106878497924e-05, "loss": 5.304, "step": 619 }, { "epoch": 0.005596678100740206, "grad_norm": 5.653271198272705, "learning_rate": 4.972061743997112e-05, "loss": 5.0458, "step": 620 }, { "epoch": 0.00560570500090269, "grad_norm": 3.6995558738708496, "learning_rate": 4.9720166094963e-05, "loss": 5.3779, "step": 621 }, { "epoch": 0.005614731901065174, "grad_norm": 4.9777750968933105, "learning_rate": 4.971971474995486e-05, "loss": 4.8795, "step": 622 }, { "epoch": 0.005623758801227658, "grad_norm": 3.7769620418548584, "learning_rate": 4.971926340494674e-05, "loss": 5.1352, "step": 623 }, { "epoch": 0.0056327857013901425, "grad_norm": 4.8127970695495605, "learning_rate": 4.971881205993862e-05, "loss": 4.5883, "step": 624 }, { "epoch": 0.005641812601552627, "grad_norm": 4.982100486755371, "learning_rate": 4.9718360714930494e-05, "loss": 4.5588, "step": 625 }, { "epoch": 0.005650839501715111, "grad_norm": 3.5655105113983154, "learning_rate": 4.971790936992237e-05, "loss": 4.5715, "step": 626 }, { "epoch": 0.005659866401877595, "grad_norm": 4.178361892700195, "learning_rate": 4.9717458024914246e-05, "loss": 4.8761, "step": 627 }, { "epoch": 0.005668893302040079, "grad_norm": 3.8141636848449707, "learning_rate": 4.9717006679906125e-05, "loss": 5.2857, "step": 628 }, { "epoch": 0.005677920202202563, "grad_norm": 3.77724027633667, "learning_rate": 4.9716555334898e-05, "loss": 5.3382, "step": 629 }, { "epoch": 0.005686947102365048, "grad_norm": 5.152548789978027, "learning_rate": 4.971610398988987e-05, "loss": 5.2749, "step": 630 }, { "epoch": 0.005695974002527532, "grad_norm": 3.0649962425231934, "learning_rate": 4.971565264488175e-05, "loss": 4.7611, "step": 631 }, { "epoch": 0.0057050009026900165, "grad_norm": 4.574838638305664, "learning_rate": 4.971520129987363e-05, "loss": 6.2329, "step": 632 }, { "epoch": 0.005714027802852501, "grad_norm": 3.828866720199585, "learning_rate": 4.97147499548655e-05, "loss": 5.44, "step": 633 }, { "epoch": 0.005723054703014985, "grad_norm": 5.882302761077881, "learning_rate": 4.971429860985738e-05, "loss": 3.8751, "step": 634 }, { "epoch": 0.005732081603177469, "grad_norm": 3.5405776500701904, "learning_rate": 4.971384726484926e-05, "loss": 5.0982, "step": 635 }, { "epoch": 0.005741108503339953, "grad_norm": 4.009702682495117, "learning_rate": 4.9713395919841125e-05, "loss": 4.3846, "step": 636 }, { "epoch": 0.005750135403502437, "grad_norm": 4.212607383728027, "learning_rate": 4.9712944574833004e-05, "loss": 4.1156, "step": 637 }, { "epoch": 0.005759162303664921, "grad_norm": 4.18304443359375, "learning_rate": 4.9712493229824883e-05, "loss": 4.9334, "step": 638 }, { "epoch": 0.005768189203827405, "grad_norm": 3.639580726623535, "learning_rate": 4.9712041884816756e-05, "loss": 5.2927, "step": 639 }, { "epoch": 0.0057772161039898896, "grad_norm": 5.908946990966797, "learning_rate": 4.9711590539808635e-05, "loss": 5.5033, "step": 640 }, { "epoch": 0.005786243004152374, "grad_norm": 3.5894827842712402, "learning_rate": 4.971113919480051e-05, "loss": 4.3765, "step": 641 }, { "epoch": 0.005795269904314859, "grad_norm": 3.3261542320251465, "learning_rate": 4.971068784979238e-05, "loss": 4.8429, "step": 642 }, { "epoch": 0.005804296804477343, "grad_norm": 4.8894195556640625, "learning_rate": 4.971023650478426e-05, "loss": 4.8187, "step": 643 }, { "epoch": 0.005813323704639827, "grad_norm": 4.816640853881836, "learning_rate": 4.970978515977613e-05, "loss": 4.9516, "step": 644 }, { "epoch": 0.005822350604802311, "grad_norm": 4.663883209228516, "learning_rate": 4.970933381476801e-05, "loss": 4.9038, "step": 645 }, { "epoch": 0.005831377504964795, "grad_norm": 5.012569904327393, "learning_rate": 4.970888246975989e-05, "loss": 5.6828, "step": 646 }, { "epoch": 0.005840404405127279, "grad_norm": 5.530849933624268, "learning_rate": 4.970843112475176e-05, "loss": 4.7766, "step": 647 }, { "epoch": 0.0058494313052897635, "grad_norm": 4.538765907287598, "learning_rate": 4.970797977974364e-05, "loss": 5.1513, "step": 648 }, { "epoch": 0.005858458205452248, "grad_norm": 5.278341293334961, "learning_rate": 4.9707528434735514e-05, "loss": 5.304, "step": 649 }, { "epoch": 0.005867485105614732, "grad_norm": 3.220534086227417, "learning_rate": 4.970707708972739e-05, "loss": 4.9016, "step": 650 }, { "epoch": 0.005876512005777216, "grad_norm": 5.035581111907959, "learning_rate": 4.9706625744719266e-05, "loss": 5.0657, "step": 651 }, { "epoch": 0.0058855389059397, "grad_norm": 4.781709671020508, "learning_rate": 4.970617439971114e-05, "loss": 5.212, "step": 652 }, { "epoch": 0.005894565806102184, "grad_norm": 3.6574089527130127, "learning_rate": 4.970572305470302e-05, "loss": 4.9981, "step": 653 }, { "epoch": 0.005903592706264668, "grad_norm": 5.495360374450684, "learning_rate": 4.97052717096949e-05, "loss": 4.5557, "step": 654 }, { "epoch": 0.005912619606427153, "grad_norm": 3.553037166595459, "learning_rate": 4.970482036468677e-05, "loss": 4.7675, "step": 655 }, { "epoch": 0.0059216465065896375, "grad_norm": 4.6427741050720215, "learning_rate": 4.970436901967864e-05, "loss": 4.2292, "step": 656 }, { "epoch": 0.005930673406752122, "grad_norm": 4.094669818878174, "learning_rate": 4.970391767467052e-05, "loss": 5.0918, "step": 657 }, { "epoch": 0.005939700306914606, "grad_norm": 5.003294467926025, "learning_rate": 4.9703466329662394e-05, "loss": 5.4896, "step": 658 }, { "epoch": 0.00594872720707709, "grad_norm": 5.327884197235107, "learning_rate": 4.970301498465427e-05, "loss": 5.3706, "step": 659 }, { "epoch": 0.005957754107239574, "grad_norm": 4.236532211303711, "learning_rate": 4.970256363964615e-05, "loss": 5.0307, "step": 660 }, { "epoch": 0.005966781007402058, "grad_norm": 4.748040199279785, "learning_rate": 4.9702112294638025e-05, "loss": 5.1814, "step": 661 }, { "epoch": 0.005975807907564542, "grad_norm": 4.852215766906738, "learning_rate": 4.97016609496299e-05, "loss": 5.1812, "step": 662 }, { "epoch": 0.005984834807727026, "grad_norm": 5.843736171722412, "learning_rate": 4.9701209604621776e-05, "loss": 4.9125, "step": 663 }, { "epoch": 0.0059938617078895105, "grad_norm": 4.040783405303955, "learning_rate": 4.970075825961365e-05, "loss": 5.1056, "step": 664 }, { "epoch": 0.006002888608051995, "grad_norm": 3.774702787399292, "learning_rate": 4.970030691460553e-05, "loss": 5.1202, "step": 665 }, { "epoch": 0.006011915508214479, "grad_norm": 4.5960917472839355, "learning_rate": 4.96998555695974e-05, "loss": 4.9092, "step": 666 }, { "epoch": 0.006020942408376963, "grad_norm": 4.058343887329102, "learning_rate": 4.969940422458928e-05, "loss": 5.2667, "step": 667 }, { "epoch": 0.006029969308539448, "grad_norm": 2.8151602745056152, "learning_rate": 4.969895287958116e-05, "loss": 5.0651, "step": 668 }, { "epoch": 0.006038996208701932, "grad_norm": 4.584304332733154, "learning_rate": 4.9698501534573025e-05, "loss": 5.7119, "step": 669 }, { "epoch": 0.006048023108864416, "grad_norm": 5.722696781158447, "learning_rate": 4.9698050189564904e-05, "loss": 4.998, "step": 670 }, { "epoch": 0.0060570500090269, "grad_norm": 3.2472171783447266, "learning_rate": 4.969759884455678e-05, "loss": 4.9041, "step": 671 }, { "epoch": 0.0060660769091893845, "grad_norm": 4.743481159210205, "learning_rate": 4.9697147499548656e-05, "loss": 5.0267, "step": 672 }, { "epoch": 0.006075103809351869, "grad_norm": 3.892082691192627, "learning_rate": 4.9696696154540535e-05, "loss": 4.7188, "step": 673 }, { "epoch": 0.006084130709514353, "grad_norm": 3.3495614528656006, "learning_rate": 4.969624480953241e-05, "loss": 5.176, "step": 674 }, { "epoch": 0.006093157609676837, "grad_norm": 6.192718029022217, "learning_rate": 4.9695793464524287e-05, "loss": 5.4086, "step": 675 }, { "epoch": 0.006102184509839321, "grad_norm": 4.838039875030518, "learning_rate": 4.969534211951616e-05, "loss": 5.3309, "step": 676 }, { "epoch": 0.006111211410001805, "grad_norm": 3.4746978282928467, "learning_rate": 4.969489077450804e-05, "loss": 5.2772, "step": 677 }, { "epoch": 0.006120238310164289, "grad_norm": 5.396434307098389, "learning_rate": 4.969443942949991e-05, "loss": 4.4676, "step": 678 }, { "epoch": 0.006129265210326773, "grad_norm": 5.370175361633301, "learning_rate": 4.969398808449179e-05, "loss": 4.6858, "step": 679 }, { "epoch": 0.006138292110489258, "grad_norm": 3.6738431453704834, "learning_rate": 4.969353673948366e-05, "loss": 4.0808, "step": 680 }, { "epoch": 0.0061473190106517426, "grad_norm": 4.543558120727539, "learning_rate": 4.969308539447554e-05, "loss": 4.9911, "step": 681 }, { "epoch": 0.006156345910814227, "grad_norm": 5.266188621520996, "learning_rate": 4.9692634049467414e-05, "loss": 5.0809, "step": 682 }, { "epoch": 0.006165372810976711, "grad_norm": 4.623319149017334, "learning_rate": 4.969218270445929e-05, "loss": 5.2824, "step": 683 }, { "epoch": 0.006174399711139195, "grad_norm": 4.165961742401123, "learning_rate": 4.9691731359451166e-05, "loss": 5.2246, "step": 684 }, { "epoch": 0.006183426611301679, "grad_norm": 4.027560710906982, "learning_rate": 4.9691280014443045e-05, "loss": 5.1045, "step": 685 }, { "epoch": 0.006192453511464163, "grad_norm": 3.5906286239624023, "learning_rate": 4.969082866943492e-05, "loss": 4.5314, "step": 686 }, { "epoch": 0.006201480411626647, "grad_norm": 3.7294461727142334, "learning_rate": 4.96903773244268e-05, "loss": 3.9557, "step": 687 }, { "epoch": 0.0062105073117891315, "grad_norm": 2.981731414794922, "learning_rate": 4.968992597941867e-05, "loss": 3.7643, "step": 688 }, { "epoch": 0.006219534211951616, "grad_norm": 5.166871547698975, "learning_rate": 4.968947463441054e-05, "loss": 4.9729, "step": 689 }, { "epoch": 0.0062285611121141, "grad_norm": 3.0953001976013184, "learning_rate": 4.968902328940242e-05, "loss": 4.4837, "step": 690 }, { "epoch": 0.006237588012276584, "grad_norm": 5.974477767944336, "learning_rate": 4.9688571944394293e-05, "loss": 5.6716, "step": 691 }, { "epoch": 0.006246614912439068, "grad_norm": 3.984257936477661, "learning_rate": 4.968812059938617e-05, "loss": 5.207, "step": 692 }, { "epoch": 0.006255641812601553, "grad_norm": 5.672134876251221, "learning_rate": 4.968766925437805e-05, "loss": 3.9192, "step": 693 }, { "epoch": 0.006264668712764037, "grad_norm": 5.588447093963623, "learning_rate": 4.9687217909369924e-05, "loss": 4.5119, "step": 694 }, { "epoch": 0.006273695612926521, "grad_norm": 5.473875045776367, "learning_rate": 4.9686766564361804e-05, "loss": 4.2276, "step": 695 }, { "epoch": 0.0062827225130890054, "grad_norm": 3.619205951690674, "learning_rate": 4.9686315219353676e-05, "loss": 5.1997, "step": 696 }, { "epoch": 0.00629174941325149, "grad_norm": 3.394047498703003, "learning_rate": 4.968586387434555e-05, "loss": 4.3187, "step": 697 }, { "epoch": 0.006300776313413974, "grad_norm": 4.6200785636901855, "learning_rate": 4.968541252933743e-05, "loss": 5.3715, "step": 698 }, { "epoch": 0.006309803213576458, "grad_norm": 4.13812255859375, "learning_rate": 4.968496118432931e-05, "loss": 5.3658, "step": 699 }, { "epoch": 0.006318830113738942, "grad_norm": 5.048144340515137, "learning_rate": 4.968450983932118e-05, "loss": 4.786, "step": 700 }, { "epoch": 0.006327857013901426, "grad_norm": 4.257413387298584, "learning_rate": 4.968405849431306e-05, "loss": 4.8218, "step": 701 }, { "epoch": 0.00633688391406391, "grad_norm": 4.52841329574585, "learning_rate": 4.968360714930493e-05, "loss": 5.0322, "step": 702 }, { "epoch": 0.006345910814226394, "grad_norm": 4.800861835479736, "learning_rate": 4.9683155804296804e-05, "loss": 4.1429, "step": 703 }, { "epoch": 0.0063549377143888785, "grad_norm": 4.361137866973877, "learning_rate": 4.968270445928868e-05, "loss": 5.3007, "step": 704 }, { "epoch": 0.006363964614551363, "grad_norm": 3.554116725921631, "learning_rate": 4.9682253114280555e-05, "loss": 4.2159, "step": 705 }, { "epoch": 0.006372991514713848, "grad_norm": 5.836273670196533, "learning_rate": 4.9681801769272435e-05, "loss": 5.0183, "step": 706 }, { "epoch": 0.006382018414876332, "grad_norm": 5.769260883331299, "learning_rate": 4.9681350424264314e-05, "loss": 4.7746, "step": 707 }, { "epoch": 0.006391045315038816, "grad_norm": 4.288827419281006, "learning_rate": 4.9680899079256186e-05, "loss": 4.3247, "step": 708 }, { "epoch": 0.0064000722152013, "grad_norm": 6.4901204109191895, "learning_rate": 4.968044773424806e-05, "loss": 5.292, "step": 709 }, { "epoch": 0.006409099115363784, "grad_norm": 2.9850857257843018, "learning_rate": 4.967999638923994e-05, "loss": 5.2463, "step": 710 }, { "epoch": 0.006418126015526268, "grad_norm": 4.6910576820373535, "learning_rate": 4.967954504423181e-05, "loss": 3.6634, "step": 711 }, { "epoch": 0.0064271529156887525, "grad_norm": 4.392322540283203, "learning_rate": 4.967909369922369e-05, "loss": 4.1039, "step": 712 }, { "epoch": 0.006436179815851237, "grad_norm": 4.515756130218506, "learning_rate": 4.967864235421556e-05, "loss": 4.9638, "step": 713 }, { "epoch": 0.006445206716013721, "grad_norm": 5.008366107940674, "learning_rate": 4.967819100920744e-05, "loss": 5.2422, "step": 714 }, { "epoch": 0.006454233616176205, "grad_norm": 4.816945552825928, "learning_rate": 4.967773966419932e-05, "loss": 5.5652, "step": 715 }, { "epoch": 0.006463260516338689, "grad_norm": 3.8850462436676025, "learning_rate": 4.9677288319191186e-05, "loss": 5.0833, "step": 716 }, { "epoch": 0.006472287416501173, "grad_norm": 3.491128444671631, "learning_rate": 4.9676836974183066e-05, "loss": 5.4066, "step": 717 }, { "epoch": 0.006481314316663657, "grad_norm": 5.038502216339111, "learning_rate": 4.9676385629174945e-05, "loss": 4.9717, "step": 718 }, { "epoch": 0.006490341216826142, "grad_norm": 5.829882621765137, "learning_rate": 4.967593428416682e-05, "loss": 4.8755, "step": 719 }, { "epoch": 0.006499368116988626, "grad_norm": 6.641509532928467, "learning_rate": 4.9675482939158697e-05, "loss": 3.9906, "step": 720 }, { "epoch": 0.0065083950171511105, "grad_norm": 4.065728187561035, "learning_rate": 4.9675031594150576e-05, "loss": 5.0598, "step": 721 }, { "epoch": 0.006517421917313595, "grad_norm": 3.3008382320404053, "learning_rate": 4.967458024914245e-05, "loss": 5.0681, "step": 722 }, { "epoch": 0.006526448817476079, "grad_norm": 4.039988994598389, "learning_rate": 4.967412890413432e-05, "loss": 4.8057, "step": 723 }, { "epoch": 0.006535475717638563, "grad_norm": 5.34893274307251, "learning_rate": 4.96736775591262e-05, "loss": 5.5164, "step": 724 }, { "epoch": 0.006544502617801047, "grad_norm": 4.716335773468018, "learning_rate": 4.967322621411807e-05, "loss": 5.1435, "step": 725 }, { "epoch": 0.006553529517963531, "grad_norm": 5.328979969024658, "learning_rate": 4.967277486910995e-05, "loss": 5.2964, "step": 726 }, { "epoch": 0.006562556418126015, "grad_norm": 4.035988807678223, "learning_rate": 4.9672323524101824e-05, "loss": 4.8249, "step": 727 }, { "epoch": 0.0065715833182884995, "grad_norm": 4.515258312225342, "learning_rate": 4.9671872179093703e-05, "loss": 4.7637, "step": 728 }, { "epoch": 0.006580610218450984, "grad_norm": 4.202758312225342, "learning_rate": 4.9671420834085576e-05, "loss": 5.6595, "step": 729 }, { "epoch": 0.006589637118613468, "grad_norm": 3.720872640609741, "learning_rate": 4.967096948907745e-05, "loss": 4.78, "step": 730 }, { "epoch": 0.006598664018775953, "grad_norm": 4.128564834594727, "learning_rate": 4.967051814406933e-05, "loss": 4.5626, "step": 731 }, { "epoch": 0.006607690918938437, "grad_norm": 5.659574508666992, "learning_rate": 4.967006679906121e-05, "loss": 4.4746, "step": 732 }, { "epoch": 0.006616717819100921, "grad_norm": 5.18796443939209, "learning_rate": 4.966961545405308e-05, "loss": 5.2802, "step": 733 }, { "epoch": 0.006625744719263405, "grad_norm": 5.25480842590332, "learning_rate": 4.966916410904496e-05, "loss": 4.8655, "step": 734 }, { "epoch": 0.006634771619425889, "grad_norm": 4.588565826416016, "learning_rate": 4.966871276403684e-05, "loss": 5.0257, "step": 735 }, { "epoch": 0.006643798519588373, "grad_norm": 4.837741374969482, "learning_rate": 4.9668261419028703e-05, "loss": 4.6953, "step": 736 }, { "epoch": 0.006652825419750858, "grad_norm": 3.101762294769287, "learning_rate": 4.966781007402058e-05, "loss": 5.1313, "step": 737 }, { "epoch": 0.006661852319913342, "grad_norm": 3.455875873565674, "learning_rate": 4.966735872901246e-05, "loss": 4.7865, "step": 738 }, { "epoch": 0.006670879220075826, "grad_norm": 4.796215534210205, "learning_rate": 4.9666907384004334e-05, "loss": 5.8221, "step": 739 }, { "epoch": 0.00667990612023831, "grad_norm": 5.07595157623291, "learning_rate": 4.9666456038996214e-05, "loss": 5.6006, "step": 740 }, { "epoch": 0.006688933020400794, "grad_norm": 6.460494518280029, "learning_rate": 4.9666004693988086e-05, "loss": 3.573, "step": 741 }, { "epoch": 0.006697959920563278, "grad_norm": 4.665312767028809, "learning_rate": 4.9665553348979965e-05, "loss": 5.1509, "step": 742 }, { "epoch": 0.006706986820725762, "grad_norm": 6.533115386962891, "learning_rate": 4.966510200397184e-05, "loss": 5.6204, "step": 743 }, { "epoch": 0.006716013720888247, "grad_norm": 5.694360256195068, "learning_rate": 4.966465065896371e-05, "loss": 5.0989, "step": 744 }, { "epoch": 0.0067250406210507315, "grad_norm": 6.104094982147217, "learning_rate": 4.966419931395559e-05, "loss": 4.8204, "step": 745 }, { "epoch": 0.006734067521213216, "grad_norm": 3.3278491497039795, "learning_rate": 4.966374796894747e-05, "loss": 5.1428, "step": 746 }, { "epoch": 0.0067430944213757, "grad_norm": 4.113167762756348, "learning_rate": 4.966329662393934e-05, "loss": 4.5746, "step": 747 }, { "epoch": 0.006752121321538184, "grad_norm": 3.656660556793213, "learning_rate": 4.966284527893122e-05, "loss": 5.2439, "step": 748 }, { "epoch": 0.006761148221700668, "grad_norm": 4.547497272491455, "learning_rate": 4.966239393392309e-05, "loss": 4.8503, "step": 749 }, { "epoch": 0.006770175121863152, "grad_norm": 4.75850248336792, "learning_rate": 4.9661942588914965e-05, "loss": 4.6544, "step": 750 }, { "epoch": 0.006779202022025636, "grad_norm": 3.432523250579834, "learning_rate": 4.9661491243906845e-05, "loss": 5.2455, "step": 751 }, { "epoch": 0.0067882289221881205, "grad_norm": 3.913602113723755, "learning_rate": 4.966103989889872e-05, "loss": 4.8935, "step": 752 }, { "epoch": 0.006797255822350605, "grad_norm": 3.8646185398101807, "learning_rate": 4.9660588553890596e-05, "loss": 5.2503, "step": 753 }, { "epoch": 0.006806282722513089, "grad_norm": 7.878702640533447, "learning_rate": 4.9660137208882476e-05, "loss": 5.3649, "step": 754 }, { "epoch": 0.006815309622675573, "grad_norm": 4.913766384124756, "learning_rate": 4.965968586387435e-05, "loss": 3.6759, "step": 755 }, { "epoch": 0.006824336522838057, "grad_norm": 4.109209060668945, "learning_rate": 4.965923451886622e-05, "loss": 4.7741, "step": 756 }, { "epoch": 0.006833363423000542, "grad_norm": 4.078432559967041, "learning_rate": 4.96587831738581e-05, "loss": 5.8103, "step": 757 }, { "epoch": 0.006842390323163026, "grad_norm": 4.173638343811035, "learning_rate": 4.965833182884997e-05, "loss": 5.0898, "step": 758 }, { "epoch": 0.00685141722332551, "grad_norm": 3.3371999263763428, "learning_rate": 4.965788048384185e-05, "loss": 5.1516, "step": 759 }, { "epoch": 0.006860444123487994, "grad_norm": 5.852723121643066, "learning_rate": 4.965742913883373e-05, "loss": 4.5769, "step": 760 }, { "epoch": 0.0068694710236504785, "grad_norm": 5.346510887145996, "learning_rate": 4.96569777938256e-05, "loss": 4.7928, "step": 761 }, { "epoch": 0.006878497923812963, "grad_norm": 4.3426737785339355, "learning_rate": 4.965652644881748e-05, "loss": 4.1163, "step": 762 }, { "epoch": 0.006887524823975447, "grad_norm": 5.027246475219727, "learning_rate": 4.9656075103809355e-05, "loss": 4.8348, "step": 763 }, { "epoch": 0.006896551724137931, "grad_norm": 3.5052082538604736, "learning_rate": 4.965562375880123e-05, "loss": 4.8242, "step": 764 }, { "epoch": 0.006905578624300415, "grad_norm": 4.610098838806152, "learning_rate": 4.9655172413793107e-05, "loss": 5.5243, "step": 765 }, { "epoch": 0.006914605524462899, "grad_norm": 4.358154773712158, "learning_rate": 4.965472106878498e-05, "loss": 5.043, "step": 766 }, { "epoch": 0.006923632424625383, "grad_norm": 4.335635185241699, "learning_rate": 4.965426972377686e-05, "loss": 4.7211, "step": 767 }, { "epoch": 0.0069326593247878675, "grad_norm": 6.823611736297607, "learning_rate": 4.965381837876874e-05, "loss": 4.9041, "step": 768 }, { "epoch": 0.0069416862249503525, "grad_norm": 4.351554870605469, "learning_rate": 4.965336703376061e-05, "loss": 4.9713, "step": 769 }, { "epoch": 0.006950713125112837, "grad_norm": 7.732394695281982, "learning_rate": 4.965291568875248e-05, "loss": 4.7393, "step": 770 }, { "epoch": 0.006959740025275321, "grad_norm": 4.753767967224121, "learning_rate": 4.965246434374436e-05, "loss": 5.1951, "step": 771 }, { "epoch": 0.006968766925437805, "grad_norm": 6.6605682373046875, "learning_rate": 4.9652012998736234e-05, "loss": 4.7665, "step": 772 }, { "epoch": 0.006977793825600289, "grad_norm": 4.182164669036865, "learning_rate": 4.9651561653728113e-05, "loss": 4.629, "step": 773 }, { "epoch": 0.006986820725762773, "grad_norm": 6.4998064041137695, "learning_rate": 4.9651110308719986e-05, "loss": 5.0117, "step": 774 }, { "epoch": 0.006995847625925257, "grad_norm": 5.213385105133057, "learning_rate": 4.9650658963711865e-05, "loss": 4.84, "step": 775 }, { "epoch": 0.007004874526087741, "grad_norm": 5.014222621917725, "learning_rate": 4.965020761870374e-05, "loss": 5.0025, "step": 776 }, { "epoch": 0.0070139014262502256, "grad_norm": 3.1822030544281006, "learning_rate": 4.964975627369562e-05, "loss": 5.0502, "step": 777 }, { "epoch": 0.00702292832641271, "grad_norm": 5.3389973640441895, "learning_rate": 4.964930492868749e-05, "loss": 5.0912, "step": 778 }, { "epoch": 0.007031955226575194, "grad_norm": 4.911290168762207, "learning_rate": 4.964885358367937e-05, "loss": 5.0192, "step": 779 }, { "epoch": 0.007040982126737678, "grad_norm": 5.816397190093994, "learning_rate": 4.964840223867124e-05, "loss": 5.2657, "step": 780 }, { "epoch": 0.007050009026900162, "grad_norm": 5.3271660804748535, "learning_rate": 4.964795089366312e-05, "loss": 4.189, "step": 781 }, { "epoch": 0.007059035927062647, "grad_norm": 3.4733617305755615, "learning_rate": 4.9647499548655e-05, "loss": 5.0547, "step": 782 }, { "epoch": 0.007068062827225131, "grad_norm": 4.7114481925964355, "learning_rate": 4.9647048203646865e-05, "loss": 4.5159, "step": 783 }, { "epoch": 0.007077089727387615, "grad_norm": 4.121319770812988, "learning_rate": 4.9646596858638744e-05, "loss": 4.6858, "step": 784 }, { "epoch": 0.0070861166275500995, "grad_norm": 4.022204875946045, "learning_rate": 4.9646145513630624e-05, "loss": 4.3668, "step": 785 }, { "epoch": 0.007095143527712584, "grad_norm": 5.095218181610107, "learning_rate": 4.9645694168622496e-05, "loss": 4.7621, "step": 786 }, { "epoch": 0.007104170427875068, "grad_norm": 4.107555866241455, "learning_rate": 4.9645242823614375e-05, "loss": 5.0946, "step": 787 }, { "epoch": 0.007113197328037552, "grad_norm": 4.432422161102295, "learning_rate": 4.964479147860625e-05, "loss": 4.6285, "step": 788 }, { "epoch": 0.007122224228200036, "grad_norm": 5.398739337921143, "learning_rate": 4.964434013359813e-05, "loss": 4.6845, "step": 789 }, { "epoch": 0.00713125112836252, "grad_norm": 3.9171090126037598, "learning_rate": 4.964388878859e-05, "loss": 5.5755, "step": 790 }, { "epoch": 0.007140278028525004, "grad_norm": 3.7898004055023193, "learning_rate": 4.964343744358187e-05, "loss": 4.6983, "step": 791 }, { "epoch": 0.0071493049286874884, "grad_norm": 7.445708751678467, "learning_rate": 4.964298609857375e-05, "loss": 5.2806, "step": 792 }, { "epoch": 0.007158331828849973, "grad_norm": 5.137912750244141, "learning_rate": 4.964253475356563e-05, "loss": 4.4631, "step": 793 }, { "epoch": 0.007167358729012457, "grad_norm": 3.732832908630371, "learning_rate": 4.96420834085575e-05, "loss": 4.6277, "step": 794 }, { "epoch": 0.007176385629174942, "grad_norm": 4.19340705871582, "learning_rate": 4.964163206354938e-05, "loss": 5.0272, "step": 795 }, { "epoch": 0.007185412529337426, "grad_norm": 2.609771728515625, "learning_rate": 4.964118071854126e-05, "loss": 5.0312, "step": 796 }, { "epoch": 0.00719443942949991, "grad_norm": 4.145847320556641, "learning_rate": 4.964072937353313e-05, "loss": 5.1234, "step": 797 }, { "epoch": 0.007203466329662394, "grad_norm": 4.913527488708496, "learning_rate": 4.9640278028525006e-05, "loss": 5.2155, "step": 798 }, { "epoch": 0.007212493229824878, "grad_norm": 4.664303779602051, "learning_rate": 4.9639826683516886e-05, "loss": 5.239, "step": 799 }, { "epoch": 0.007221520129987362, "grad_norm": 3.7170166969299316, "learning_rate": 4.963937533850876e-05, "loss": 5.1564, "step": 800 }, { "epoch": 0.0072305470301498465, "grad_norm": 4.625992774963379, "learning_rate": 4.963892399350064e-05, "loss": 5.15, "step": 801 }, { "epoch": 0.007239573930312331, "grad_norm": 4.889558792114258, "learning_rate": 4.963847264849251e-05, "loss": 5.1401, "step": 802 }, { "epoch": 0.007248600830474815, "grad_norm": 4.150251865386963, "learning_rate": 4.963802130348438e-05, "loss": 5.2592, "step": 803 }, { "epoch": 0.007257627730637299, "grad_norm": 6.314458847045898, "learning_rate": 4.963756995847626e-05, "loss": 4.8734, "step": 804 }, { "epoch": 0.007266654630799783, "grad_norm": 4.529664039611816, "learning_rate": 4.9637118613468134e-05, "loss": 5.0885, "step": 805 }, { "epoch": 0.007275681530962267, "grad_norm": 5.599713325500488, "learning_rate": 4.963666726846001e-05, "loss": 4.8865, "step": 806 }, { "epoch": 0.007284708431124751, "grad_norm": 5.988199234008789, "learning_rate": 4.963621592345189e-05, "loss": 5.3837, "step": 807 }, { "epoch": 0.007293735331287236, "grad_norm": 4.478553295135498, "learning_rate": 4.9635764578443765e-05, "loss": 4.4777, "step": 808 }, { "epoch": 0.0073027622314497205, "grad_norm": 3.6604278087615967, "learning_rate": 4.9635313233435644e-05, "loss": 4.48, "step": 809 }, { "epoch": 0.007311789131612205, "grad_norm": 5.3583550453186035, "learning_rate": 4.9634861888427517e-05, "loss": 4.4917, "step": 810 }, { "epoch": 0.007320816031774689, "grad_norm": 6.4993815422058105, "learning_rate": 4.963441054341939e-05, "loss": 4.7135, "step": 811 }, { "epoch": 0.007329842931937173, "grad_norm": 4.508959770202637, "learning_rate": 4.963395919841127e-05, "loss": 5.4079, "step": 812 }, { "epoch": 0.007338869832099657, "grad_norm": 4.955261707305908, "learning_rate": 4.963350785340314e-05, "loss": 5.0809, "step": 813 }, { "epoch": 0.007347896732262141, "grad_norm": 2.9802796840667725, "learning_rate": 4.963305650839502e-05, "loss": 5.298, "step": 814 }, { "epoch": 0.007356923632424625, "grad_norm": 4.5007452964782715, "learning_rate": 4.96326051633869e-05, "loss": 5.6817, "step": 815 }, { "epoch": 0.007365950532587109, "grad_norm": 5.110789775848389, "learning_rate": 4.963215381837877e-05, "loss": 4.9531, "step": 816 }, { "epoch": 0.0073749774327495935, "grad_norm": 6.552303791046143, "learning_rate": 4.9631702473370644e-05, "loss": 5.7267, "step": 817 }, { "epoch": 0.007384004332912078, "grad_norm": 4.146636009216309, "learning_rate": 4.9631251128362523e-05, "loss": 5.0484, "step": 818 }, { "epoch": 0.007393031233074562, "grad_norm": 4.1692585945129395, "learning_rate": 4.9630799783354396e-05, "loss": 5.0142, "step": 819 }, { "epoch": 0.007402058133237047, "grad_norm": 7.123874664306641, "learning_rate": 4.9630348438346275e-05, "loss": 5.1341, "step": 820 }, { "epoch": 0.007411085033399531, "grad_norm": 3.895420789718628, "learning_rate": 4.9629897093338154e-05, "loss": 4.5959, "step": 821 }, { "epoch": 0.007420111933562015, "grad_norm": 5.255703449249268, "learning_rate": 4.962944574833003e-05, "loss": 4.8995, "step": 822 }, { "epoch": 0.007429138833724499, "grad_norm": 3.3405182361602783, "learning_rate": 4.96289944033219e-05, "loss": 4.8859, "step": 823 }, { "epoch": 0.007438165733886983, "grad_norm": 3.900172710418701, "learning_rate": 4.962854305831378e-05, "loss": 4.8131, "step": 824 }, { "epoch": 0.0074471926340494675, "grad_norm": 5.193853378295898, "learning_rate": 4.962809171330565e-05, "loss": 4.6909, "step": 825 }, { "epoch": 0.007456219534211952, "grad_norm": 4.154839992523193, "learning_rate": 4.962764036829753e-05, "loss": 5.1181, "step": 826 }, { "epoch": 0.007465246434374436, "grad_norm": 3.8087897300720215, "learning_rate": 4.96271890232894e-05, "loss": 3.8123, "step": 827 }, { "epoch": 0.00747427333453692, "grad_norm": 4.675388336181641, "learning_rate": 4.962673767828128e-05, "loss": 4.6278, "step": 828 }, { "epoch": 0.007483300234699404, "grad_norm": 4.675389766693115, "learning_rate": 4.962628633327316e-05, "loss": 4.9741, "step": 829 }, { "epoch": 0.007492327134861888, "grad_norm": 4.074936866760254, "learning_rate": 4.962583498826503e-05, "loss": 4.7933, "step": 830 }, { "epoch": 0.007501354035024372, "grad_norm": 4.208528995513916, "learning_rate": 4.9625383643256906e-05, "loss": 5.3687, "step": 831 }, { "epoch": 0.0075103809351868564, "grad_norm": 6.875718593597412, "learning_rate": 4.9624932298248785e-05, "loss": 4.3347, "step": 832 }, { "epoch": 0.0075194078353493414, "grad_norm": 4.950725078582764, "learning_rate": 4.962448095324066e-05, "loss": 4.5282, "step": 833 }, { "epoch": 0.007528434735511826, "grad_norm": 2.740273952484131, "learning_rate": 4.962402960823254e-05, "loss": 5.049, "step": 834 }, { "epoch": 0.00753746163567431, "grad_norm": 6.285519123077393, "learning_rate": 4.962357826322441e-05, "loss": 5.2493, "step": 835 }, { "epoch": 0.007546488535836794, "grad_norm": 4.407564640045166, "learning_rate": 4.962312691821629e-05, "loss": 4.8408, "step": 836 }, { "epoch": 0.007555515435999278, "grad_norm": 6.434680461883545, "learning_rate": 4.962267557320816e-05, "loss": 4.5521, "step": 837 }, { "epoch": 0.007564542336161762, "grad_norm": 3.581092119216919, "learning_rate": 4.962222422820004e-05, "loss": 4.9595, "step": 838 }, { "epoch": 0.007573569236324246, "grad_norm": 4.43980598449707, "learning_rate": 4.962177288319191e-05, "loss": 4.6591, "step": 839 }, { "epoch": 0.00758259613648673, "grad_norm": 3.8812263011932373, "learning_rate": 4.962132153818379e-05, "loss": 5.1453, "step": 840 }, { "epoch": 0.0075916230366492145, "grad_norm": 4.17105770111084, "learning_rate": 4.9620870193175665e-05, "loss": 5.3047, "step": 841 }, { "epoch": 0.007600649936811699, "grad_norm": 3.916118860244751, "learning_rate": 4.9620418848167544e-05, "loss": 4.7646, "step": 842 }, { "epoch": 0.007609676836974183, "grad_norm": 5.6656317710876465, "learning_rate": 4.9619967503159416e-05, "loss": 4.668, "step": 843 }, { "epoch": 0.007618703737136667, "grad_norm": 3.7318124771118164, "learning_rate": 4.961951615815129e-05, "loss": 4.9093, "step": 844 }, { "epoch": 0.007627730637299151, "grad_norm": 5.027513027191162, "learning_rate": 4.961906481314317e-05, "loss": 4.5561, "step": 845 }, { "epoch": 0.007636757537461636, "grad_norm": 4.778604984283447, "learning_rate": 4.961861346813505e-05, "loss": 4.8226, "step": 846 }, { "epoch": 0.00764578443762412, "grad_norm": 4.205298900604248, "learning_rate": 4.961816212312692e-05, "loss": 5.3234, "step": 847 }, { "epoch": 0.007654811337786604, "grad_norm": 3.5978381633758545, "learning_rate": 4.96177107781188e-05, "loss": 4.6875, "step": 848 }, { "epoch": 0.0076638382379490885, "grad_norm": 3.1814024448394775, "learning_rate": 4.961725943311067e-05, "loss": 4.9147, "step": 849 }, { "epoch": 0.007672865138111573, "grad_norm": 4.786133289337158, "learning_rate": 4.9616808088102544e-05, "loss": 4.8764, "step": 850 }, { "epoch": 0.007681892038274057, "grad_norm": 3.3609349727630615, "learning_rate": 4.961635674309442e-05, "loss": 4.943, "step": 851 }, { "epoch": 0.007690918938436541, "grad_norm": 4.33212947845459, "learning_rate": 4.9615905398086296e-05, "loss": 4.507, "step": 852 }, { "epoch": 0.007699945838599025, "grad_norm": 4.761166572570801, "learning_rate": 4.9615454053078175e-05, "loss": 5.3893, "step": 853 }, { "epoch": 0.007708972738761509, "grad_norm": 3.9671149253845215, "learning_rate": 4.9615002708070054e-05, "loss": 4.6927, "step": 854 }, { "epoch": 0.007717999638923993, "grad_norm": 3.335536003112793, "learning_rate": 4.9614551363061927e-05, "loss": 4.4838, "step": 855 }, { "epoch": 0.007727026539086477, "grad_norm": 4.23444938659668, "learning_rate": 4.9614100018053806e-05, "loss": 4.8599, "step": 856 }, { "epoch": 0.0077360534392489615, "grad_norm": 3.8254194259643555, "learning_rate": 4.961364867304568e-05, "loss": 4.4848, "step": 857 }, { "epoch": 0.0077450803394114465, "grad_norm": 3.7598717212677, "learning_rate": 4.961319732803755e-05, "loss": 4.8481, "step": 858 }, { "epoch": 0.007754107239573931, "grad_norm": 5.077148914337158, "learning_rate": 4.961274598302943e-05, "loss": 5.0877, "step": 859 }, { "epoch": 0.007763134139736415, "grad_norm": 5.876866817474365, "learning_rate": 4.961229463802131e-05, "loss": 4.9192, "step": 860 }, { "epoch": 0.007772161039898899, "grad_norm": 4.502058506011963, "learning_rate": 4.961184329301318e-05, "loss": 4.6232, "step": 861 }, { "epoch": 0.007781187940061383, "grad_norm": 4.088423252105713, "learning_rate": 4.961139194800506e-05, "loss": 5.179, "step": 862 }, { "epoch": 0.007790214840223867, "grad_norm": 3.9125564098358154, "learning_rate": 4.961094060299693e-05, "loss": 5.0344, "step": 863 }, { "epoch": 0.007799241740386351, "grad_norm": 3.49788498878479, "learning_rate": 4.9610489257988806e-05, "loss": 5.2236, "step": 864 }, { "epoch": 0.0078082686405488355, "grad_norm": 4.850168228149414, "learning_rate": 4.9610037912980685e-05, "loss": 4.6391, "step": 865 }, { "epoch": 0.00781729554071132, "grad_norm": 5.2009477615356445, "learning_rate": 4.960958656797256e-05, "loss": 4.7698, "step": 866 }, { "epoch": 0.007826322440873805, "grad_norm": 3.5983197689056396, "learning_rate": 4.960913522296444e-05, "loss": 4.7142, "step": 867 }, { "epoch": 0.007835349341036288, "grad_norm": 5.389955043792725, "learning_rate": 4.9608683877956316e-05, "loss": 4.3548, "step": 868 }, { "epoch": 0.007844376241198773, "grad_norm": 4.205588340759277, "learning_rate": 4.960823253294819e-05, "loss": 5.4184, "step": 869 }, { "epoch": 0.007853403141361256, "grad_norm": 3.837185859680176, "learning_rate": 4.960778118794006e-05, "loss": 4.4787, "step": 870 }, { "epoch": 0.007862430041523741, "grad_norm": 3.568202495574951, "learning_rate": 4.960732984293194e-05, "loss": 5.0731, "step": 871 }, { "epoch": 0.007871456941686224, "grad_norm": 5.216371536254883, "learning_rate": 4.960687849792381e-05, "loss": 5.2419, "step": 872 }, { "epoch": 0.00788048384184871, "grad_norm": 3.820770502090454, "learning_rate": 4.960642715291569e-05, "loss": 4.8218, "step": 873 }, { "epoch": 0.007889510742011193, "grad_norm": 3.243699312210083, "learning_rate": 4.9605975807907564e-05, "loss": 4.7756, "step": 874 }, { "epoch": 0.007898537642173678, "grad_norm": 6.015995979309082, "learning_rate": 4.9605524462899444e-05, "loss": 4.6933, "step": 875 }, { "epoch": 0.007907564542336161, "grad_norm": 3.6482553482055664, "learning_rate": 4.960507311789132e-05, "loss": 4.9651, "step": 876 }, { "epoch": 0.007916591442498646, "grad_norm": 3.973867654800415, "learning_rate": 4.960462177288319e-05, "loss": 4.3329, "step": 877 }, { "epoch": 0.007925618342661131, "grad_norm": 4.272880554199219, "learning_rate": 4.960417042787507e-05, "loss": 4.8034, "step": 878 }, { "epoch": 0.007934645242823614, "grad_norm": 3.6880383491516113, "learning_rate": 4.960371908286695e-05, "loss": 5.2037, "step": 879 }, { "epoch": 0.0079436721429861, "grad_norm": 4.710310935974121, "learning_rate": 4.960326773785882e-05, "loss": 4.7799, "step": 880 }, { "epoch": 0.007952699043148583, "grad_norm": 5.159243583679199, "learning_rate": 4.96028163928507e-05, "loss": 4.5786, "step": 881 }, { "epoch": 0.007961725943311068, "grad_norm": 3.683128595352173, "learning_rate": 4.960236504784258e-05, "loss": 5.0896, "step": 882 }, { "epoch": 0.00797075284347355, "grad_norm": 4.634469509124756, "learning_rate": 4.960191370283445e-05, "loss": 4.9059, "step": 883 }, { "epoch": 0.007979779743636036, "grad_norm": 4.024899005889893, "learning_rate": 4.960146235782632e-05, "loss": 5.2301, "step": 884 }, { "epoch": 0.007988806643798519, "grad_norm": 4.304896354675293, "learning_rate": 4.96010110128182e-05, "loss": 5.1434, "step": 885 }, { "epoch": 0.007997833543961004, "grad_norm": 3.9876997470855713, "learning_rate": 4.9600559667810075e-05, "loss": 4.8163, "step": 886 }, { "epoch": 0.008006860444123487, "grad_norm": 4.17072868347168, "learning_rate": 4.9600108322801954e-05, "loss": 5.2152, "step": 887 }, { "epoch": 0.008015887344285972, "grad_norm": 4.332200050354004, "learning_rate": 4.9599656977793826e-05, "loss": 5.2495, "step": 888 }, { "epoch": 0.008024914244448456, "grad_norm": 4.296065330505371, "learning_rate": 4.9599205632785706e-05, "loss": 4.869, "step": 889 }, { "epoch": 0.00803394114461094, "grad_norm": 3.6634716987609863, "learning_rate": 4.959875428777758e-05, "loss": 4.7192, "step": 890 }, { "epoch": 0.008042968044773426, "grad_norm": 4.644104957580566, "learning_rate": 4.959830294276945e-05, "loss": 4.9656, "step": 891 }, { "epoch": 0.008051994944935909, "grad_norm": 3.582808017730713, "learning_rate": 4.959785159776133e-05, "loss": 4.4264, "step": 892 }, { "epoch": 0.008061021845098394, "grad_norm": 4.098066806793213, "learning_rate": 4.959740025275321e-05, "loss": 4.4984, "step": 893 }, { "epoch": 0.008070048745260877, "grad_norm": 5.172050476074219, "learning_rate": 4.959694890774508e-05, "loss": 5.4522, "step": 894 }, { "epoch": 0.008079075645423362, "grad_norm": 3.9049458503723145, "learning_rate": 4.959649756273696e-05, "loss": 4.9626, "step": 895 }, { "epoch": 0.008088102545585845, "grad_norm": 3.6398298740386963, "learning_rate": 4.959604621772883e-05, "loss": 5.1567, "step": 896 }, { "epoch": 0.00809712944574833, "grad_norm": 3.6054434776306152, "learning_rate": 4.9595594872720706e-05, "loss": 5.2544, "step": 897 }, { "epoch": 0.008106156345910814, "grad_norm": 3.9117000102996826, "learning_rate": 4.9595143527712585e-05, "loss": 5.1825, "step": 898 }, { "epoch": 0.008115183246073299, "grad_norm": 4.957103729248047, "learning_rate": 4.9594692182704464e-05, "loss": 5.0699, "step": 899 }, { "epoch": 0.008124210146235782, "grad_norm": 6.244205474853516, "learning_rate": 4.9594240837696337e-05, "loss": 5.1371, "step": 900 }, { "epoch": 0.008133237046398267, "grad_norm": 4.481984615325928, "learning_rate": 4.9593789492688216e-05, "loss": 5.07, "step": 901 }, { "epoch": 0.00814226394656075, "grad_norm": 4.07331657409668, "learning_rate": 4.959333814768009e-05, "loss": 5.2907, "step": 902 }, { "epoch": 0.008151290846723235, "grad_norm": 3.4016764163970947, "learning_rate": 4.959288680267197e-05, "loss": 5.3786, "step": 903 }, { "epoch": 0.00816031774688572, "grad_norm": 4.794071674346924, "learning_rate": 4.959243545766384e-05, "loss": 4.6125, "step": 904 }, { "epoch": 0.008169344647048203, "grad_norm": 4.97048807144165, "learning_rate": 4.959198411265571e-05, "loss": 3.6815, "step": 905 }, { "epoch": 0.008178371547210688, "grad_norm": 7.094961643218994, "learning_rate": 4.959153276764759e-05, "loss": 4.6609, "step": 906 }, { "epoch": 0.008187398447373172, "grad_norm": 4.3128156661987305, "learning_rate": 4.959108142263947e-05, "loss": 4.4625, "step": 907 }, { "epoch": 0.008196425347535657, "grad_norm": 3.7217185497283936, "learning_rate": 4.959063007763134e-05, "loss": 4.7031, "step": 908 }, { "epoch": 0.00820545224769814, "grad_norm": 4.982461929321289, "learning_rate": 4.959017873262322e-05, "loss": 4.8108, "step": 909 }, { "epoch": 0.008214479147860625, "grad_norm": 3.660353183746338, "learning_rate": 4.9589727387615095e-05, "loss": 4.8546, "step": 910 }, { "epoch": 0.008223506048023108, "grad_norm": 5.893610954284668, "learning_rate": 4.958927604260697e-05, "loss": 4.7113, "step": 911 }, { "epoch": 0.008232532948185593, "grad_norm": 4.759519100189209, "learning_rate": 4.958882469759885e-05, "loss": 4.8214, "step": 912 }, { "epoch": 0.008241559848348077, "grad_norm": 5.0641984939575195, "learning_rate": 4.958837335259072e-05, "loss": 5.1468, "step": 913 }, { "epoch": 0.008250586748510562, "grad_norm": 3.938575506210327, "learning_rate": 4.95879220075826e-05, "loss": 4.3602, "step": 914 }, { "epoch": 0.008259613648673045, "grad_norm": 4.2669267654418945, "learning_rate": 4.958747066257448e-05, "loss": 5.2006, "step": 915 }, { "epoch": 0.00826864054883553, "grad_norm": 3.7501020431518555, "learning_rate": 4.958701931756635e-05, "loss": 4.4428, "step": 916 }, { "epoch": 0.008277667448998015, "grad_norm": 4.597887992858887, "learning_rate": 4.958656797255822e-05, "loss": 4.5689, "step": 917 }, { "epoch": 0.008286694349160498, "grad_norm": 4.150274753570557, "learning_rate": 4.95861166275501e-05, "loss": 3.9735, "step": 918 }, { "epoch": 0.008295721249322983, "grad_norm": 4.255503177642822, "learning_rate": 4.9585665282541974e-05, "loss": 5.5563, "step": 919 }, { "epoch": 0.008304748149485466, "grad_norm": 4.376003265380859, "learning_rate": 4.9585213937533854e-05, "loss": 4.9855, "step": 920 }, { "epoch": 0.008313775049647951, "grad_norm": 4.616285800933838, "learning_rate": 4.958476259252573e-05, "loss": 5.0844, "step": 921 }, { "epoch": 0.008322801949810435, "grad_norm": 3.377626419067383, "learning_rate": 4.9584311247517605e-05, "loss": 4.5822, "step": 922 }, { "epoch": 0.00833182884997292, "grad_norm": 4.2388787269592285, "learning_rate": 4.9583859902509485e-05, "loss": 5.0131, "step": 923 }, { "epoch": 0.008340855750135403, "grad_norm": 5.712177276611328, "learning_rate": 4.958340855750136e-05, "loss": 5.1725, "step": 924 }, { "epoch": 0.008349882650297888, "grad_norm": 3.4337239265441895, "learning_rate": 4.958295721249323e-05, "loss": 5.1958, "step": 925 }, { "epoch": 0.008358909550460371, "grad_norm": 3.270965814590454, "learning_rate": 4.958250586748511e-05, "loss": 4.3866, "step": 926 }, { "epoch": 0.008367936450622856, "grad_norm": 4.742428302764893, "learning_rate": 4.958205452247698e-05, "loss": 5.3884, "step": 927 }, { "epoch": 0.008376963350785341, "grad_norm": 3.755423069000244, "learning_rate": 4.958160317746886e-05, "loss": 4.7262, "step": 928 }, { "epoch": 0.008385990250947824, "grad_norm": 4.3868889808654785, "learning_rate": 4.958115183246074e-05, "loss": 5.3672, "step": 929 }, { "epoch": 0.00839501715111031, "grad_norm": 3.83990216255188, "learning_rate": 4.958070048745261e-05, "loss": 4.6667, "step": 930 }, { "epoch": 0.008404044051272793, "grad_norm": 4.5216593742370605, "learning_rate": 4.9580249142444485e-05, "loss": 4.5049, "step": 931 }, { "epoch": 0.008413070951435278, "grad_norm": 4.129252910614014, "learning_rate": 4.9579797797436364e-05, "loss": 5.3518, "step": 932 }, { "epoch": 0.008422097851597761, "grad_norm": 3.3689942359924316, "learning_rate": 4.9579346452428236e-05, "loss": 5.2107, "step": 933 }, { "epoch": 0.008431124751760246, "grad_norm": 4.800335884094238, "learning_rate": 4.9578895107420116e-05, "loss": 4.6488, "step": 934 }, { "epoch": 0.00844015165192273, "grad_norm": 4.479372978210449, "learning_rate": 4.957844376241199e-05, "loss": 4.6509, "step": 935 }, { "epoch": 0.008449178552085214, "grad_norm": 3.1489977836608887, "learning_rate": 4.957799241740387e-05, "loss": 4.8325, "step": 936 }, { "epoch": 0.008458205452247698, "grad_norm": 3.772808313369751, "learning_rate": 4.957754107239574e-05, "loss": 4.8371, "step": 937 }, { "epoch": 0.008467232352410183, "grad_norm": 3.69559645652771, "learning_rate": 4.957708972738762e-05, "loss": 4.8914, "step": 938 }, { "epoch": 0.008476259252572666, "grad_norm": 3.1565663814544678, "learning_rate": 4.957663838237949e-05, "loss": 4.547, "step": 939 }, { "epoch": 0.00848528615273515, "grad_norm": 3.881692409515381, "learning_rate": 4.957618703737137e-05, "loss": 5.2078, "step": 940 }, { "epoch": 0.008494313052897636, "grad_norm": 3.9849953651428223, "learning_rate": 4.957573569236324e-05, "loss": 5.299, "step": 941 }, { "epoch": 0.008503339953060119, "grad_norm": 4.4804768562316895, "learning_rate": 4.957528434735512e-05, "loss": 4.1768, "step": 942 }, { "epoch": 0.008512366853222604, "grad_norm": 4.0193352699279785, "learning_rate": 4.9574833002347e-05, "loss": 4.8362, "step": 943 }, { "epoch": 0.008521393753385087, "grad_norm": 4.352176666259766, "learning_rate": 4.957438165733887e-05, "loss": 4.1899, "step": 944 }, { "epoch": 0.008530420653547572, "grad_norm": 4.471424102783203, "learning_rate": 4.9573930312330747e-05, "loss": 5.0193, "step": 945 }, { "epoch": 0.008539447553710056, "grad_norm": 4.432508945465088, "learning_rate": 4.9573478967322626e-05, "loss": 5.1577, "step": 946 }, { "epoch": 0.00854847445387254, "grad_norm": 3.933807373046875, "learning_rate": 4.95730276223145e-05, "loss": 4.2458, "step": 947 }, { "epoch": 0.008557501354035024, "grad_norm": 3.74588680267334, "learning_rate": 4.957257627730638e-05, "loss": 4.3632, "step": 948 }, { "epoch": 0.008566528254197509, "grad_norm": 7.013723373413086, "learning_rate": 4.957212493229825e-05, "loss": 4.7553, "step": 949 }, { "epoch": 0.008575555154359992, "grad_norm": 5.709097385406494, "learning_rate": 4.957167358729013e-05, "loss": 5.738, "step": 950 }, { "epoch": 0.008584582054522477, "grad_norm": 4.348210334777832, "learning_rate": 4.9571222242282e-05, "loss": 5.2156, "step": 951 }, { "epoch": 0.00859360895468496, "grad_norm": 5.5155930519104, "learning_rate": 4.9570770897273874e-05, "loss": 4.8273, "step": 952 }, { "epoch": 0.008602635854847445, "grad_norm": 4.596023082733154, "learning_rate": 4.957031955226575e-05, "loss": 4.7265, "step": 953 }, { "epoch": 0.00861166275500993, "grad_norm": 5.78956937789917, "learning_rate": 4.956986820725763e-05, "loss": 5.3612, "step": 954 }, { "epoch": 0.008620689655172414, "grad_norm": 4.503453731536865, "learning_rate": 4.9569416862249505e-05, "loss": 4.8224, "step": 955 }, { "epoch": 0.008629716555334899, "grad_norm": 5.545749664306641, "learning_rate": 4.9568965517241384e-05, "loss": 5.3341, "step": 956 }, { "epoch": 0.008638743455497382, "grad_norm": 4.635727882385254, "learning_rate": 4.9568514172233264e-05, "loss": 4.2456, "step": 957 }, { "epoch": 0.008647770355659867, "grad_norm": 4.701042175292969, "learning_rate": 4.956806282722513e-05, "loss": 5.6962, "step": 958 }, { "epoch": 0.00865679725582235, "grad_norm": 4.821402072906494, "learning_rate": 4.956761148221701e-05, "loss": 5.9234, "step": 959 }, { "epoch": 0.008665824155984835, "grad_norm": 5.356356620788574, "learning_rate": 4.956716013720889e-05, "loss": 4.6395, "step": 960 }, { "epoch": 0.008674851056147318, "grad_norm": 3.5075788497924805, "learning_rate": 4.956670879220076e-05, "loss": 4.9471, "step": 961 }, { "epoch": 0.008683877956309803, "grad_norm": 4.61034631729126, "learning_rate": 4.956625744719264e-05, "loss": 4.2211, "step": 962 }, { "epoch": 0.008692904856472287, "grad_norm": 3.7751574516296387, "learning_rate": 4.956580610218451e-05, "loss": 4.8389, "step": 963 }, { "epoch": 0.008701931756634772, "grad_norm": 4.49608039855957, "learning_rate": 4.9565354757176384e-05, "loss": 4.8507, "step": 964 }, { "epoch": 0.008710958656797255, "grad_norm": 5.545968055725098, "learning_rate": 4.9564903412168264e-05, "loss": 5.0883, "step": 965 }, { "epoch": 0.00871998555695974, "grad_norm": 3.569321870803833, "learning_rate": 4.9564452067160136e-05, "loss": 5.1615, "step": 966 }, { "epoch": 0.008729012457122225, "grad_norm": 4.603216648101807, "learning_rate": 4.9564000722152015e-05, "loss": 5.0931, "step": 967 }, { "epoch": 0.008738039357284708, "grad_norm": 4.613211631774902, "learning_rate": 4.9563549377143895e-05, "loss": 4.7772, "step": 968 }, { "epoch": 0.008747066257447193, "grad_norm": 3.8311593532562256, "learning_rate": 4.956309803213577e-05, "loss": 5.1008, "step": 969 }, { "epoch": 0.008756093157609677, "grad_norm": 4.9820637702941895, "learning_rate": 4.9562646687127646e-05, "loss": 5.5126, "step": 970 }, { "epoch": 0.008765120057772162, "grad_norm": 4.616842746734619, "learning_rate": 4.956219534211952e-05, "loss": 4.8862, "step": 971 }, { "epoch": 0.008774146957934645, "grad_norm": 3.829270362854004, "learning_rate": 4.956174399711139e-05, "loss": 3.7184, "step": 972 }, { "epoch": 0.00878317385809713, "grad_norm": 4.190557956695557, "learning_rate": 4.956129265210327e-05, "loss": 4.958, "step": 973 }, { "epoch": 0.008792200758259613, "grad_norm": 3.6733572483062744, "learning_rate": 4.956084130709514e-05, "loss": 4.376, "step": 974 }, { "epoch": 0.008801227658422098, "grad_norm": 6.440627098083496, "learning_rate": 4.956038996208702e-05, "loss": 4.5494, "step": 975 }, { "epoch": 0.008810254558584581, "grad_norm": 4.769351005554199, "learning_rate": 4.95599386170789e-05, "loss": 5.1168, "step": 976 }, { "epoch": 0.008819281458747066, "grad_norm": 2.9914872646331787, "learning_rate": 4.9559487272070774e-05, "loss": 3.7346, "step": 977 }, { "epoch": 0.00882830835890955, "grad_norm": 4.70028829574585, "learning_rate": 4.9559035927062646e-05, "loss": 4.3883, "step": 978 }, { "epoch": 0.008837335259072035, "grad_norm": 4.7128005027771, "learning_rate": 4.9558584582054526e-05, "loss": 4.6515, "step": 979 }, { "epoch": 0.00884636215923452, "grad_norm": 3.814944267272949, "learning_rate": 4.95581332370464e-05, "loss": 4.0496, "step": 980 }, { "epoch": 0.008855389059397003, "grad_norm": 3.569796085357666, "learning_rate": 4.955768189203828e-05, "loss": 5.5766, "step": 981 }, { "epoch": 0.008864415959559488, "grad_norm": 4.7198805809021, "learning_rate": 4.9557230547030156e-05, "loss": 4.9109, "step": 982 }, { "epoch": 0.008873442859721971, "grad_norm": 4.794322490692139, "learning_rate": 4.955677920202203e-05, "loss": 4.703, "step": 983 }, { "epoch": 0.008882469759884456, "grad_norm": 4.9944658279418945, "learning_rate": 4.95563278570139e-05, "loss": 5.5908, "step": 984 }, { "epoch": 0.00889149666004694, "grad_norm": 6.144598484039307, "learning_rate": 4.955587651200578e-05, "loss": 4.8498, "step": 985 }, { "epoch": 0.008900523560209424, "grad_norm": 4.101416110992432, "learning_rate": 4.955542516699765e-05, "loss": 4.7033, "step": 986 }, { "epoch": 0.008909550460371908, "grad_norm": 5.8452229499816895, "learning_rate": 4.955497382198953e-05, "loss": 5.4978, "step": 987 }, { "epoch": 0.008918577360534393, "grad_norm": 6.372506141662598, "learning_rate": 4.9554522476981405e-05, "loss": 5.0315, "step": 988 }, { "epoch": 0.008927604260696876, "grad_norm": 3.7845938205718994, "learning_rate": 4.9554071131973284e-05, "loss": 5.1122, "step": 989 }, { "epoch": 0.008936631160859361, "grad_norm": 4.760998249053955, "learning_rate": 4.955361978696516e-05, "loss": 5.1858, "step": 990 }, { "epoch": 0.008945658061021844, "grad_norm": 4.0142316818237305, "learning_rate": 4.955316844195703e-05, "loss": 5.585, "step": 991 }, { "epoch": 0.00895468496118433, "grad_norm": 3.511352062225342, "learning_rate": 4.955271709694891e-05, "loss": 4.7578, "step": 992 }, { "epoch": 0.008963711861346814, "grad_norm": 4.390182018280029, "learning_rate": 4.955226575194079e-05, "loss": 5.1606, "step": 993 }, { "epoch": 0.008972738761509298, "grad_norm": 3.538532257080078, "learning_rate": 4.955181440693266e-05, "loss": 4.8006, "step": 994 }, { "epoch": 0.008981765661671783, "grad_norm": 6.3022074699401855, "learning_rate": 4.955136306192454e-05, "loss": 4.2814, "step": 995 }, { "epoch": 0.008990792561834266, "grad_norm": 5.180533409118652, "learning_rate": 4.955091171691641e-05, "loss": 4.8719, "step": 996 }, { "epoch": 0.00899981946199675, "grad_norm": 4.759675025939941, "learning_rate": 4.955046037190829e-05, "loss": 5.0188, "step": 997 }, { "epoch": 0.009008846362159234, "grad_norm": 3.545356512069702, "learning_rate": 4.955000902690016e-05, "loss": 4.6979, "step": 998 }, { "epoch": 0.009017873262321719, "grad_norm": 4.674427509307861, "learning_rate": 4.954955768189204e-05, "loss": 5.0364, "step": 999 }, { "epoch": 0.009026900162484202, "grad_norm": 4.241589546203613, "learning_rate": 4.9549106336883915e-05, "loss": 4.542, "step": 1000 }, { "epoch": 0.009035927062646687, "grad_norm": 3.179924249649048, "learning_rate": 4.9548654991875794e-05, "loss": 4.736, "step": 1001 }, { "epoch": 0.00904495396280917, "grad_norm": 4.20405387878418, "learning_rate": 4.954820364686767e-05, "loss": 4.8819, "step": 1002 }, { "epoch": 0.009053980862971656, "grad_norm": 5.236631393432617, "learning_rate": 4.9547752301859546e-05, "loss": 4.6713, "step": 1003 }, { "epoch": 0.009063007763134139, "grad_norm": 5.635905742645264, "learning_rate": 4.9547300956851425e-05, "loss": 5.4535, "step": 1004 }, { "epoch": 0.009072034663296624, "grad_norm": 3.631531000137329, "learning_rate": 4.954684961184329e-05, "loss": 4.7612, "step": 1005 }, { "epoch": 0.009081061563459109, "grad_norm": 5.635919094085693, "learning_rate": 4.954639826683517e-05, "loss": 5.3485, "step": 1006 }, { "epoch": 0.009090088463621592, "grad_norm": 5.184910297393799, "learning_rate": 4.954594692182705e-05, "loss": 5.0676, "step": 1007 }, { "epoch": 0.009099115363784077, "grad_norm": 3.323617935180664, "learning_rate": 4.954549557681892e-05, "loss": 4.0581, "step": 1008 }, { "epoch": 0.00910814226394656, "grad_norm": 6.605253219604492, "learning_rate": 4.95450442318108e-05, "loss": 4.6846, "step": 1009 }, { "epoch": 0.009117169164109045, "grad_norm": 3.519116163253784, "learning_rate": 4.9544592886802674e-05, "loss": 4.4104, "step": 1010 }, { "epoch": 0.009126196064271529, "grad_norm": 4.303508758544922, "learning_rate": 4.9544141541794546e-05, "loss": 5.304, "step": 1011 }, { "epoch": 0.009135222964434014, "grad_norm": 5.556924343109131, "learning_rate": 4.9543690196786425e-05, "loss": 5.3201, "step": 1012 }, { "epoch": 0.009144249864596497, "grad_norm": 4.186751365661621, "learning_rate": 4.95432388517783e-05, "loss": 5.4194, "step": 1013 }, { "epoch": 0.009153276764758982, "grad_norm": 3.6302361488342285, "learning_rate": 4.954278750677018e-05, "loss": 4.64, "step": 1014 }, { "epoch": 0.009162303664921465, "grad_norm": 4.197775840759277, "learning_rate": 4.9542336161762056e-05, "loss": 5.2358, "step": 1015 }, { "epoch": 0.00917133056508395, "grad_norm": 3.5932276248931885, "learning_rate": 4.954188481675393e-05, "loss": 4.6844, "step": 1016 }, { "epoch": 0.009180357465246435, "grad_norm": 4.5740509033203125, "learning_rate": 4.954143347174581e-05, "loss": 4.9477, "step": 1017 }, { "epoch": 0.009189384365408918, "grad_norm": 5.406645774841309, "learning_rate": 4.954098212673768e-05, "loss": 4.9336, "step": 1018 }, { "epoch": 0.009198411265571403, "grad_norm": 2.74259090423584, "learning_rate": 4.954053078172955e-05, "loss": 4.6197, "step": 1019 }, { "epoch": 0.009207438165733887, "grad_norm": 4.4355292320251465, "learning_rate": 4.954007943672143e-05, "loss": 5.3913, "step": 1020 }, { "epoch": 0.009216465065896372, "grad_norm": 3.842796564102173, "learning_rate": 4.953962809171331e-05, "loss": 4.7412, "step": 1021 }, { "epoch": 0.009225491966058855, "grad_norm": 3.781888484954834, "learning_rate": 4.9539176746705184e-05, "loss": 5.4227, "step": 1022 }, { "epoch": 0.00923451886622134, "grad_norm": 2.5644359588623047, "learning_rate": 4.953872540169706e-05, "loss": 5.0604, "step": 1023 }, { "epoch": 0.009243545766383823, "grad_norm": 3.633495330810547, "learning_rate": 4.9538274056688936e-05, "loss": 5.5179, "step": 1024 }, { "epoch": 0.009252572666546308, "grad_norm": 4.7162604331970215, "learning_rate": 4.953782271168081e-05, "loss": 3.4885, "step": 1025 }, { "epoch": 0.009261599566708792, "grad_norm": 4.346991062164307, "learning_rate": 4.953737136667269e-05, "loss": 5.2172, "step": 1026 }, { "epoch": 0.009270626466871277, "grad_norm": 3.6756293773651123, "learning_rate": 4.953692002166456e-05, "loss": 3.7051, "step": 1027 }, { "epoch": 0.00927965336703376, "grad_norm": 4.25848388671875, "learning_rate": 4.953646867665644e-05, "loss": 5.5522, "step": 1028 }, { "epoch": 0.009288680267196245, "grad_norm": 3.732581377029419, "learning_rate": 4.953601733164832e-05, "loss": 4.8318, "step": 1029 }, { "epoch": 0.00929770716735873, "grad_norm": 5.624423503875732, "learning_rate": 4.953556598664019e-05, "loss": 4.6889, "step": 1030 }, { "epoch": 0.009306734067521213, "grad_norm": 4.599568843841553, "learning_rate": 4.953511464163206e-05, "loss": 5.1194, "step": 1031 }, { "epoch": 0.009315760967683698, "grad_norm": 4.010011196136475, "learning_rate": 4.953466329662394e-05, "loss": 5.0235, "step": 1032 }, { "epoch": 0.009324787867846181, "grad_norm": 5.3375654220581055, "learning_rate": 4.9534211951615815e-05, "loss": 5.0056, "step": 1033 }, { "epoch": 0.009333814768008666, "grad_norm": 3.645550012588501, "learning_rate": 4.9533760606607694e-05, "loss": 5.296, "step": 1034 }, { "epoch": 0.00934284166817115, "grad_norm": 4.036087512969971, "learning_rate": 4.9533309261599567e-05, "loss": 4.8975, "step": 1035 }, { "epoch": 0.009351868568333635, "grad_norm": 3.9397237300872803, "learning_rate": 4.9532857916591446e-05, "loss": 5.0299, "step": 1036 }, { "epoch": 0.009360895468496118, "grad_norm": 2.8856875896453857, "learning_rate": 4.9532406571583325e-05, "loss": 4.7593, "step": 1037 }, { "epoch": 0.009369922368658603, "grad_norm": 4.164451599121094, "learning_rate": 4.953195522657519e-05, "loss": 4.5907, "step": 1038 }, { "epoch": 0.009378949268821086, "grad_norm": 5.29353666305542, "learning_rate": 4.953150388156707e-05, "loss": 5.065, "step": 1039 }, { "epoch": 0.009387976168983571, "grad_norm": 3.6999216079711914, "learning_rate": 4.953105253655895e-05, "loss": 5.1236, "step": 1040 }, { "epoch": 0.009397003069146054, "grad_norm": 3.636389970779419, "learning_rate": 4.953060119155082e-05, "loss": 4.8445, "step": 1041 }, { "epoch": 0.00940602996930854, "grad_norm": 4.4510884284973145, "learning_rate": 4.95301498465427e-05, "loss": 5.3563, "step": 1042 }, { "epoch": 0.009415056869471024, "grad_norm": 4.1736321449279785, "learning_rate": 4.952969850153458e-05, "loss": 4.7427, "step": 1043 }, { "epoch": 0.009424083769633508, "grad_norm": 4.081650733947754, "learning_rate": 4.952924715652645e-05, "loss": 4.1981, "step": 1044 }, { "epoch": 0.009433110669795993, "grad_norm": 3.819715738296509, "learning_rate": 4.9528795811518325e-05, "loss": 5.1057, "step": 1045 }, { "epoch": 0.009442137569958476, "grad_norm": 3.410074472427368, "learning_rate": 4.9528344466510204e-05, "loss": 3.7751, "step": 1046 }, { "epoch": 0.009451164470120961, "grad_norm": 5.127204418182373, "learning_rate": 4.952789312150208e-05, "loss": 5.2381, "step": 1047 }, { "epoch": 0.009460191370283444, "grad_norm": 5.757767677307129, "learning_rate": 4.9527441776493956e-05, "loss": 4.7366, "step": 1048 }, { "epoch": 0.00946921827044593, "grad_norm": 3.3116114139556885, "learning_rate": 4.952699043148583e-05, "loss": 4.5826, "step": 1049 }, { "epoch": 0.009478245170608413, "grad_norm": 3.657491445541382, "learning_rate": 4.952653908647771e-05, "loss": 4.8627, "step": 1050 }, { "epoch": 0.009487272070770898, "grad_norm": 6.969394207000732, "learning_rate": 4.952608774146958e-05, "loss": 5.3909, "step": 1051 }, { "epoch": 0.00949629897093338, "grad_norm": 4.4469990730285645, "learning_rate": 4.952563639646145e-05, "loss": 4.6648, "step": 1052 }, { "epoch": 0.009505325871095866, "grad_norm": 4.009494781494141, "learning_rate": 4.952518505145333e-05, "loss": 4.7773, "step": 1053 }, { "epoch": 0.009514352771258349, "grad_norm": 3.717228412628174, "learning_rate": 4.952473370644521e-05, "loss": 5.0057, "step": 1054 }, { "epoch": 0.009523379671420834, "grad_norm": 3.5310773849487305, "learning_rate": 4.9524282361437084e-05, "loss": 4.7922, "step": 1055 }, { "epoch": 0.009532406571583319, "grad_norm": 5.146940231323242, "learning_rate": 4.952383101642896e-05, "loss": 4.8719, "step": 1056 }, { "epoch": 0.009541433471745802, "grad_norm": 4.050644397735596, "learning_rate": 4.9523379671420835e-05, "loss": 5.3128, "step": 1057 }, { "epoch": 0.009550460371908287, "grad_norm": 4.084057807922363, "learning_rate": 4.952292832641271e-05, "loss": 4.9236, "step": 1058 }, { "epoch": 0.00955948727207077, "grad_norm": 7.304225444793701, "learning_rate": 4.952247698140459e-05, "loss": 5.1198, "step": 1059 }, { "epoch": 0.009568514172233256, "grad_norm": 3.544642448425293, "learning_rate": 4.9522025636396466e-05, "loss": 5.129, "step": 1060 }, { "epoch": 0.009577541072395739, "grad_norm": 6.216887950897217, "learning_rate": 4.952157429138834e-05, "loss": 4.8601, "step": 1061 }, { "epoch": 0.009586567972558224, "grad_norm": 4.753251075744629, "learning_rate": 4.952112294638022e-05, "loss": 4.6877, "step": 1062 }, { "epoch": 0.009595594872720707, "grad_norm": 4.005147933959961, "learning_rate": 4.952067160137209e-05, "loss": 4.3168, "step": 1063 }, { "epoch": 0.009604621772883192, "grad_norm": 4.673619270324707, "learning_rate": 4.952022025636397e-05, "loss": 5.0247, "step": 1064 }, { "epoch": 0.009613648673045675, "grad_norm": 3.687422513961792, "learning_rate": 4.951976891135584e-05, "loss": 5.0604, "step": 1065 }, { "epoch": 0.00962267557320816, "grad_norm": 5.161110877990723, "learning_rate": 4.9519317566347715e-05, "loss": 4.8708, "step": 1066 }, { "epoch": 0.009631702473370644, "grad_norm": 8.370479583740234, "learning_rate": 4.9518866221339594e-05, "loss": 4.5763, "step": 1067 }, { "epoch": 0.009640729373533129, "grad_norm": 3.7863564491271973, "learning_rate": 4.951841487633147e-05, "loss": 4.8782, "step": 1068 }, { "epoch": 0.009649756273695614, "grad_norm": 5.2165303230285645, "learning_rate": 4.9517963531323346e-05, "loss": 5.1389, "step": 1069 }, { "epoch": 0.009658783173858097, "grad_norm": 4.356987953186035, "learning_rate": 4.9517512186315225e-05, "loss": 4.961, "step": 1070 }, { "epoch": 0.009667810074020582, "grad_norm": 3.646433115005493, "learning_rate": 4.95170608413071e-05, "loss": 4.8782, "step": 1071 }, { "epoch": 0.009676836974183065, "grad_norm": 3.930959701538086, "learning_rate": 4.951660949629897e-05, "loss": 4.9714, "step": 1072 }, { "epoch": 0.00968586387434555, "grad_norm": 4.032708644866943, "learning_rate": 4.951615815129085e-05, "loss": 4.868, "step": 1073 }, { "epoch": 0.009694890774508034, "grad_norm": 3.0648727416992188, "learning_rate": 4.951570680628272e-05, "loss": 4.7812, "step": 1074 }, { "epoch": 0.009703917674670519, "grad_norm": 3.2728991508483887, "learning_rate": 4.95152554612746e-05, "loss": 5.2389, "step": 1075 }, { "epoch": 0.009712944574833002, "grad_norm": 5.194617748260498, "learning_rate": 4.951480411626648e-05, "loss": 4.332, "step": 1076 }, { "epoch": 0.009721971474995487, "grad_norm": 4.577701568603516, "learning_rate": 4.951435277125835e-05, "loss": 5.1031, "step": 1077 }, { "epoch": 0.00973099837515797, "grad_norm": 6.824482440948486, "learning_rate": 4.9513901426250225e-05, "loss": 5.2572, "step": 1078 }, { "epoch": 0.009740025275320455, "grad_norm": 4.786196708679199, "learning_rate": 4.9513450081242104e-05, "loss": 4.7962, "step": 1079 }, { "epoch": 0.009749052175482938, "grad_norm": 4.941115379333496, "learning_rate": 4.9512998736233977e-05, "loss": 4.9506, "step": 1080 }, { "epoch": 0.009758079075645423, "grad_norm": 4.065633296966553, "learning_rate": 4.9512547391225856e-05, "loss": 5.2873, "step": 1081 }, { "epoch": 0.009767105975807908, "grad_norm": 4.50375509262085, "learning_rate": 4.9512096046217735e-05, "loss": 3.7273, "step": 1082 }, { "epoch": 0.009776132875970392, "grad_norm": 4.767701148986816, "learning_rate": 4.951164470120961e-05, "loss": 5.227, "step": 1083 }, { "epoch": 0.009785159776132877, "grad_norm": 4.925496578216553, "learning_rate": 4.951119335620149e-05, "loss": 4.7574, "step": 1084 }, { "epoch": 0.00979418667629536, "grad_norm": 4.956662178039551, "learning_rate": 4.951074201119336e-05, "loss": 4.5131, "step": 1085 }, { "epoch": 0.009803213576457845, "grad_norm": 7.02203893661499, "learning_rate": 4.951029066618523e-05, "loss": 5.0749, "step": 1086 }, { "epoch": 0.009812240476620328, "grad_norm": 3.1948366165161133, "learning_rate": 4.950983932117711e-05, "loss": 4.5971, "step": 1087 }, { "epoch": 0.009821267376782813, "grad_norm": 3.9716248512268066, "learning_rate": 4.950938797616898e-05, "loss": 4.7483, "step": 1088 }, { "epoch": 0.009830294276945296, "grad_norm": 4.9861249923706055, "learning_rate": 4.950893663116086e-05, "loss": 5.431, "step": 1089 }, { "epoch": 0.009839321177107781, "grad_norm": 5.004669666290283, "learning_rate": 4.950848528615274e-05, "loss": 5.5177, "step": 1090 }, { "epoch": 0.009848348077270265, "grad_norm": 4.175624370574951, "learning_rate": 4.9508033941144614e-05, "loss": 5.2532, "step": 1091 }, { "epoch": 0.00985737497743275, "grad_norm": 3.991927146911621, "learning_rate": 4.950758259613649e-05, "loss": 5.0987, "step": 1092 }, { "epoch": 0.009866401877595233, "grad_norm": 9.242430686950684, "learning_rate": 4.9507131251128366e-05, "loss": 4.445, "step": 1093 }, { "epoch": 0.009875428777757718, "grad_norm": 3.700880765914917, "learning_rate": 4.950667990612024e-05, "loss": 4.9746, "step": 1094 }, { "epoch": 0.009884455677920203, "grad_norm": 3.9410033226013184, "learning_rate": 4.950622856111212e-05, "loss": 5.2362, "step": 1095 }, { "epoch": 0.009893482578082686, "grad_norm": 3.301900625228882, "learning_rate": 4.950577721610399e-05, "loss": 5.1794, "step": 1096 }, { "epoch": 0.009902509478245171, "grad_norm": 3.193131685256958, "learning_rate": 4.950532587109587e-05, "loss": 5.2956, "step": 1097 }, { "epoch": 0.009911536378407654, "grad_norm": 3.830988645553589, "learning_rate": 4.950487452608774e-05, "loss": 4.6242, "step": 1098 }, { "epoch": 0.00992056327857014, "grad_norm": 4.066760540008545, "learning_rate": 4.950442318107962e-05, "loss": 5.1263, "step": 1099 }, { "epoch": 0.009929590178732623, "grad_norm": 4.627004623413086, "learning_rate": 4.9503971836071494e-05, "loss": 3.4993, "step": 1100 }, { "epoch": 0.009938617078895108, "grad_norm": 4.073619365692139, "learning_rate": 4.950352049106337e-05, "loss": 5.427, "step": 1101 }, { "epoch": 0.009947643979057591, "grad_norm": 6.7589826583862305, "learning_rate": 4.9503069146055245e-05, "loss": 4.1636, "step": 1102 }, { "epoch": 0.009956670879220076, "grad_norm": 3.9075169563293457, "learning_rate": 4.9502617801047125e-05, "loss": 4.8989, "step": 1103 }, { "epoch": 0.00996569777938256, "grad_norm": 3.534029960632324, "learning_rate": 4.9502166456039004e-05, "loss": 4.6076, "step": 1104 }, { "epoch": 0.009974724679545044, "grad_norm": 3.876556634902954, "learning_rate": 4.950171511103087e-05, "loss": 4.7591, "step": 1105 }, { "epoch": 0.00998375157970753, "grad_norm": 4.437699794769287, "learning_rate": 4.950126376602275e-05, "loss": 5.4961, "step": 1106 }, { "epoch": 0.009992778479870013, "grad_norm": 5.22467041015625, "learning_rate": 4.950081242101463e-05, "loss": 5.1072, "step": 1107 }, { "epoch": 0.010001805380032498, "grad_norm": 3.401338815689087, "learning_rate": 4.95003610760065e-05, "loss": 4.7122, "step": 1108 }, { "epoch": 0.01001083228019498, "grad_norm": 4.733249664306641, "learning_rate": 4.949990973099838e-05, "loss": 4.7969, "step": 1109 }, { "epoch": 0.010019859180357466, "grad_norm": 3.12799072265625, "learning_rate": 4.949945838599025e-05, "loss": 4.4876, "step": 1110 }, { "epoch": 0.010028886080519949, "grad_norm": 3.379387140274048, "learning_rate": 4.949900704098213e-05, "loss": 4.8321, "step": 1111 }, { "epoch": 0.010037912980682434, "grad_norm": 4.041637420654297, "learning_rate": 4.9498555695974004e-05, "loss": 4.2284, "step": 1112 }, { "epoch": 0.010046939880844917, "grad_norm": 4.299101829528809, "learning_rate": 4.9498104350965876e-05, "loss": 4.817, "step": 1113 }, { "epoch": 0.010055966781007402, "grad_norm": 4.136139392852783, "learning_rate": 4.9497653005957756e-05, "loss": 4.4398, "step": 1114 }, { "epoch": 0.010064993681169886, "grad_norm": 4.79197883605957, "learning_rate": 4.9497201660949635e-05, "loss": 5.0914, "step": 1115 }, { "epoch": 0.01007402058133237, "grad_norm": 5.342098712921143, "learning_rate": 4.949675031594151e-05, "loss": 5.2557, "step": 1116 }, { "epoch": 0.010083047481494854, "grad_norm": 2.851348876953125, "learning_rate": 4.9496298970933386e-05, "loss": 4.8117, "step": 1117 }, { "epoch": 0.010092074381657339, "grad_norm": 3.497001886367798, "learning_rate": 4.9495847625925266e-05, "loss": 4.7945, "step": 1118 }, { "epoch": 0.010101101281819824, "grad_norm": 11.366806030273438, "learning_rate": 4.949539628091713e-05, "loss": 4.9235, "step": 1119 }, { "epoch": 0.010110128181982307, "grad_norm": 4.711489200592041, "learning_rate": 4.949494493590901e-05, "loss": 4.8526, "step": 1120 }, { "epoch": 0.010119155082144792, "grad_norm": 4.573590278625488, "learning_rate": 4.949449359090089e-05, "loss": 5.3268, "step": 1121 }, { "epoch": 0.010128181982307275, "grad_norm": 5.749955654144287, "learning_rate": 4.949404224589276e-05, "loss": 4.4518, "step": 1122 }, { "epoch": 0.01013720888246976, "grad_norm": 4.809941291809082, "learning_rate": 4.949359090088464e-05, "loss": 4.2096, "step": 1123 }, { "epoch": 0.010146235782632244, "grad_norm": 4.491140842437744, "learning_rate": 4.9493139555876514e-05, "loss": 5.3175, "step": 1124 }, { "epoch": 0.010155262682794729, "grad_norm": 5.436829566955566, "learning_rate": 4.9492688210868387e-05, "loss": 3.8455, "step": 1125 }, { "epoch": 0.010164289582957212, "grad_norm": 3.2138547897338867, "learning_rate": 4.9492236865860266e-05, "loss": 4.9725, "step": 1126 }, { "epoch": 0.010173316483119697, "grad_norm": 8.022100448608398, "learning_rate": 4.949178552085214e-05, "loss": 5.0588, "step": 1127 }, { "epoch": 0.01018234338328218, "grad_norm": 5.210722923278809, "learning_rate": 4.949133417584402e-05, "loss": 5.0523, "step": 1128 }, { "epoch": 0.010191370283444665, "grad_norm": 7.391053676605225, "learning_rate": 4.94908828308359e-05, "loss": 4.841, "step": 1129 }, { "epoch": 0.010200397183607149, "grad_norm": 6.0455756187438965, "learning_rate": 4.949043148582777e-05, "loss": 4.784, "step": 1130 }, { "epoch": 0.010209424083769634, "grad_norm": 3.7393057346343994, "learning_rate": 4.948998014081965e-05, "loss": 5.1981, "step": 1131 }, { "epoch": 0.010218450983932119, "grad_norm": 4.109616279602051, "learning_rate": 4.948952879581152e-05, "loss": 4.4079, "step": 1132 }, { "epoch": 0.010227477884094602, "grad_norm": 4.238804817199707, "learning_rate": 4.948907745080339e-05, "loss": 5.2915, "step": 1133 }, { "epoch": 0.010236504784257087, "grad_norm": 3.6382980346679688, "learning_rate": 4.948862610579527e-05, "loss": 4.5025, "step": 1134 }, { "epoch": 0.01024553168441957, "grad_norm": 4.870605945587158, "learning_rate": 4.9488174760787145e-05, "loss": 4.8834, "step": 1135 }, { "epoch": 0.010254558584582055, "grad_norm": 4.308560371398926, "learning_rate": 4.9487723415779024e-05, "loss": 5.024, "step": 1136 }, { "epoch": 0.010263585484744538, "grad_norm": 6.30046272277832, "learning_rate": 4.9487272070770904e-05, "loss": 4.686, "step": 1137 }, { "epoch": 0.010272612384907023, "grad_norm": 7.405848979949951, "learning_rate": 4.9486820725762776e-05, "loss": 5.6571, "step": 1138 }, { "epoch": 0.010281639285069507, "grad_norm": 3.8572804927825928, "learning_rate": 4.948636938075465e-05, "loss": 4.8973, "step": 1139 }, { "epoch": 0.010290666185231992, "grad_norm": 5.874499797821045, "learning_rate": 4.948591803574653e-05, "loss": 4.3849, "step": 1140 }, { "epoch": 0.010299693085394475, "grad_norm": 7.280343055725098, "learning_rate": 4.94854666907384e-05, "loss": 5.0178, "step": 1141 }, { "epoch": 0.01030871998555696, "grad_norm": 4.309957981109619, "learning_rate": 4.948501534573028e-05, "loss": 4.9574, "step": 1142 }, { "epoch": 0.010317746885719443, "grad_norm": 3.839399814605713, "learning_rate": 4.948456400072216e-05, "loss": 4.734, "step": 1143 }, { "epoch": 0.010326773785881928, "grad_norm": 5.2196855545043945, "learning_rate": 4.948411265571403e-05, "loss": 4.3078, "step": 1144 }, { "epoch": 0.010335800686044413, "grad_norm": 4.747811794281006, "learning_rate": 4.9483661310705904e-05, "loss": 3.9921, "step": 1145 }, { "epoch": 0.010344827586206896, "grad_norm": 2.800800085067749, "learning_rate": 4.948320996569778e-05, "loss": 4.5315, "step": 1146 }, { "epoch": 0.010353854486369381, "grad_norm": 3.4969534873962402, "learning_rate": 4.9482758620689655e-05, "loss": 4.9441, "step": 1147 }, { "epoch": 0.010362881386531865, "grad_norm": 3.368704319000244, "learning_rate": 4.9482307275681535e-05, "loss": 4.9111, "step": 1148 }, { "epoch": 0.01037190828669435, "grad_norm": 5.392389297485352, "learning_rate": 4.948185593067341e-05, "loss": 5.3889, "step": 1149 }, { "epoch": 0.010380935186856833, "grad_norm": 3.636711359024048, "learning_rate": 4.9481404585665286e-05, "loss": 3.6839, "step": 1150 }, { "epoch": 0.010389962087019318, "grad_norm": 4.713812351226807, "learning_rate": 4.9480953240657165e-05, "loss": 5.5928, "step": 1151 }, { "epoch": 0.010398988987181801, "grad_norm": 6.133978366851807, "learning_rate": 4.948050189564903e-05, "loss": 4.5473, "step": 1152 }, { "epoch": 0.010408015887344286, "grad_norm": 3.159846544265747, "learning_rate": 4.948005055064091e-05, "loss": 4.9756, "step": 1153 }, { "epoch": 0.01041704278750677, "grad_norm": 4.915037155151367, "learning_rate": 4.947959920563279e-05, "loss": 5.0691, "step": 1154 }, { "epoch": 0.010426069687669254, "grad_norm": 4.408469200134277, "learning_rate": 4.947914786062466e-05, "loss": 5.6792, "step": 1155 }, { "epoch": 0.010435096587831738, "grad_norm": 3.780221700668335, "learning_rate": 4.947869651561654e-05, "loss": 5.2238, "step": 1156 }, { "epoch": 0.010444123487994223, "grad_norm": 4.047801494598389, "learning_rate": 4.9478245170608414e-05, "loss": 4.7828, "step": 1157 }, { "epoch": 0.010453150388156708, "grad_norm": 4.184853553771973, "learning_rate": 4.947779382560029e-05, "loss": 4.722, "step": 1158 }, { "epoch": 0.010462177288319191, "grad_norm": 3.948432683944702, "learning_rate": 4.9477342480592166e-05, "loss": 4.7414, "step": 1159 }, { "epoch": 0.010471204188481676, "grad_norm": 4.373456001281738, "learning_rate": 4.9476891135584045e-05, "loss": 4.6465, "step": 1160 }, { "epoch": 0.01048023108864416, "grad_norm": 4.6548357009887695, "learning_rate": 4.947643979057592e-05, "loss": 4.4314, "step": 1161 }, { "epoch": 0.010489257988806644, "grad_norm": 4.666328430175781, "learning_rate": 4.9475988445567796e-05, "loss": 4.8473, "step": 1162 }, { "epoch": 0.010498284888969128, "grad_norm": 2.914553642272949, "learning_rate": 4.947553710055967e-05, "loss": 5.2311, "step": 1163 }, { "epoch": 0.010507311789131613, "grad_norm": 3.2870359420776367, "learning_rate": 4.947508575555155e-05, "loss": 4.7651, "step": 1164 }, { "epoch": 0.010516338689294096, "grad_norm": 4.027866363525391, "learning_rate": 4.947463441054343e-05, "loss": 5.0565, "step": 1165 }, { "epoch": 0.01052536558945658, "grad_norm": 3.322584629058838, "learning_rate": 4.947418306553529e-05, "loss": 4.7529, "step": 1166 }, { "epoch": 0.010534392489619064, "grad_norm": 3.6487607955932617, "learning_rate": 4.947373172052717e-05, "loss": 4.5969, "step": 1167 }, { "epoch": 0.010543419389781549, "grad_norm": 4.4692206382751465, "learning_rate": 4.947328037551905e-05, "loss": 4.2718, "step": 1168 }, { "epoch": 0.010552446289944032, "grad_norm": 4.1409220695495605, "learning_rate": 4.9472829030510924e-05, "loss": 4.7652, "step": 1169 }, { "epoch": 0.010561473190106517, "grad_norm": 4.199126720428467, "learning_rate": 4.94723776855028e-05, "loss": 4.8892, "step": 1170 }, { "epoch": 0.010570500090269002, "grad_norm": 3.9079673290252686, "learning_rate": 4.9471926340494676e-05, "loss": 4.7309, "step": 1171 }, { "epoch": 0.010579526990431486, "grad_norm": 4.78713846206665, "learning_rate": 4.947147499548655e-05, "loss": 5.1829, "step": 1172 }, { "epoch": 0.01058855389059397, "grad_norm": 4.123274326324463, "learning_rate": 4.947102365047843e-05, "loss": 4.932, "step": 1173 }, { "epoch": 0.010597580790756454, "grad_norm": 3.427090883255005, "learning_rate": 4.94705723054703e-05, "loss": 4.987, "step": 1174 }, { "epoch": 0.010606607690918939, "grad_norm": 3.949970006942749, "learning_rate": 4.947012096046218e-05, "loss": 5.0733, "step": 1175 }, { "epoch": 0.010615634591081422, "grad_norm": 4.019908905029297, "learning_rate": 4.946966961545406e-05, "loss": 4.6883, "step": 1176 }, { "epoch": 0.010624661491243907, "grad_norm": 3.110323905944824, "learning_rate": 4.946921827044593e-05, "loss": 4.5292, "step": 1177 }, { "epoch": 0.01063368839140639, "grad_norm": 3.945462942123413, "learning_rate": 4.946876692543781e-05, "loss": 5.3017, "step": 1178 }, { "epoch": 0.010642715291568875, "grad_norm": 2.925663948059082, "learning_rate": 4.946831558042968e-05, "loss": 4.8393, "step": 1179 }, { "epoch": 0.010651742191731359, "grad_norm": 10.694543838500977, "learning_rate": 4.9467864235421555e-05, "loss": 4.6322, "step": 1180 }, { "epoch": 0.010660769091893844, "grad_norm": 4.643951416015625, "learning_rate": 4.9467412890413434e-05, "loss": 5.3932, "step": 1181 }, { "epoch": 0.010669795992056327, "grad_norm": 4.411935806274414, "learning_rate": 4.9466961545405314e-05, "loss": 4.4433, "step": 1182 }, { "epoch": 0.010678822892218812, "grad_norm": 4.001672744750977, "learning_rate": 4.9466510200397186e-05, "loss": 4.4925, "step": 1183 }, { "epoch": 0.010687849792381297, "grad_norm": 5.422354698181152, "learning_rate": 4.9466058855389065e-05, "loss": 4.8772, "step": 1184 }, { "epoch": 0.01069687669254378, "grad_norm": 3.2686214447021484, "learning_rate": 4.946560751038094e-05, "loss": 5.1871, "step": 1185 }, { "epoch": 0.010705903592706265, "grad_norm": 5.4335784912109375, "learning_rate": 4.946515616537281e-05, "loss": 3.876, "step": 1186 }, { "epoch": 0.010714930492868749, "grad_norm": 3.3351783752441406, "learning_rate": 4.946470482036469e-05, "loss": 5.4007, "step": 1187 }, { "epoch": 0.010723957393031234, "grad_norm": 3.8724610805511475, "learning_rate": 4.946425347535656e-05, "loss": 5.0985, "step": 1188 }, { "epoch": 0.010732984293193717, "grad_norm": 3.1578147411346436, "learning_rate": 4.946380213034844e-05, "loss": 5.2693, "step": 1189 }, { "epoch": 0.010742011193356202, "grad_norm": 2.9926488399505615, "learning_rate": 4.946335078534032e-05, "loss": 4.7078, "step": 1190 }, { "epoch": 0.010751038093518685, "grad_norm": 4.732287883758545, "learning_rate": 4.946289944033219e-05, "loss": 4.7984, "step": 1191 }, { "epoch": 0.01076006499368117, "grad_norm": 3.8915669918060303, "learning_rate": 4.9462448095324065e-05, "loss": 4.891, "step": 1192 }, { "epoch": 0.010769091893843653, "grad_norm": 3.6625001430511475, "learning_rate": 4.9461996750315945e-05, "loss": 5.3138, "step": 1193 }, { "epoch": 0.010778118794006138, "grad_norm": 4.696578502655029, "learning_rate": 4.946154540530782e-05, "loss": 3.1049, "step": 1194 }, { "epoch": 0.010787145694168623, "grad_norm": 5.847769260406494, "learning_rate": 4.9461094060299696e-05, "loss": 6.1099, "step": 1195 }, { "epoch": 0.010796172594331107, "grad_norm": 3.633331775665283, "learning_rate": 4.946064271529157e-05, "loss": 5.1996, "step": 1196 }, { "epoch": 0.010805199494493592, "grad_norm": 6.743048667907715, "learning_rate": 4.946019137028345e-05, "loss": 4.6327, "step": 1197 }, { "epoch": 0.010814226394656075, "grad_norm": 4.1112236976623535, "learning_rate": 4.945974002527533e-05, "loss": 4.3707, "step": 1198 }, { "epoch": 0.01082325329481856, "grad_norm": 6.041843414306641, "learning_rate": 4.945928868026719e-05, "loss": 5.198, "step": 1199 }, { "epoch": 0.010832280194981043, "grad_norm": 3.695202350616455, "learning_rate": 4.945883733525907e-05, "loss": 5.2344, "step": 1200 }, { "epoch": 0.010841307095143528, "grad_norm": 4.572292804718018, "learning_rate": 4.945838599025095e-05, "loss": 4.5732, "step": 1201 }, { "epoch": 0.010850333995306011, "grad_norm": 4.123940944671631, "learning_rate": 4.9457934645242824e-05, "loss": 4.7484, "step": 1202 }, { "epoch": 0.010859360895468496, "grad_norm": 7.8217453956604, "learning_rate": 4.94574833002347e-05, "loss": 4.1368, "step": 1203 }, { "epoch": 0.01086838779563098, "grad_norm": 5.898826599121094, "learning_rate": 4.945703195522658e-05, "loss": 4.9545, "step": 1204 }, { "epoch": 0.010877414695793465, "grad_norm": 5.148050785064697, "learning_rate": 4.9456580610218455e-05, "loss": 4.7676, "step": 1205 }, { "epoch": 0.010886441595955948, "grad_norm": 3.7611122131347656, "learning_rate": 4.945612926521033e-05, "loss": 4.7209, "step": 1206 }, { "epoch": 0.010895468496118433, "grad_norm": 3.402618646621704, "learning_rate": 4.9455677920202206e-05, "loss": 4.9084, "step": 1207 }, { "epoch": 0.010904495396280918, "grad_norm": 3.901594400405884, "learning_rate": 4.945522657519408e-05, "loss": 5.2406, "step": 1208 }, { "epoch": 0.010913522296443401, "grad_norm": 3.780928373336792, "learning_rate": 4.945477523018596e-05, "loss": 4.9767, "step": 1209 }, { "epoch": 0.010922549196605886, "grad_norm": 4.990306377410889, "learning_rate": 4.945432388517783e-05, "loss": 4.6577, "step": 1210 }, { "epoch": 0.01093157609676837, "grad_norm": 3.692810535430908, "learning_rate": 4.945387254016971e-05, "loss": 4.5722, "step": 1211 }, { "epoch": 0.010940602996930855, "grad_norm": 4.045557022094727, "learning_rate": 4.945342119516158e-05, "loss": 4.9832, "step": 1212 }, { "epoch": 0.010949629897093338, "grad_norm": 4.697025299072266, "learning_rate": 4.9452969850153455e-05, "loss": 4.992, "step": 1213 }, { "epoch": 0.010958656797255823, "grad_norm": 3.6151978969573975, "learning_rate": 4.9452518505145334e-05, "loss": 4.659, "step": 1214 }, { "epoch": 0.010967683697418306, "grad_norm": 4.913609981536865, "learning_rate": 4.945206716013721e-05, "loss": 4.5092, "step": 1215 }, { "epoch": 0.010976710597580791, "grad_norm": 3.0239739418029785, "learning_rate": 4.9451615815129086e-05, "loss": 5.1423, "step": 1216 }, { "epoch": 0.010985737497743274, "grad_norm": 4.770012855529785, "learning_rate": 4.9451164470120965e-05, "loss": 4.949, "step": 1217 }, { "epoch": 0.01099476439790576, "grad_norm": 4.891515731811523, "learning_rate": 4.945071312511284e-05, "loss": 4.3585, "step": 1218 }, { "epoch": 0.011003791298068243, "grad_norm": 4.485438346862793, "learning_rate": 4.945026178010471e-05, "loss": 4.2085, "step": 1219 }, { "epoch": 0.011012818198230728, "grad_norm": 5.661469459533691, "learning_rate": 4.944981043509659e-05, "loss": 4.7357, "step": 1220 }, { "epoch": 0.011021845098393213, "grad_norm": 3.3951914310455322, "learning_rate": 4.944935909008847e-05, "loss": 5.0616, "step": 1221 }, { "epoch": 0.011030871998555696, "grad_norm": 3.077972650527954, "learning_rate": 4.944890774508034e-05, "loss": 5.2292, "step": 1222 }, { "epoch": 0.01103989889871818, "grad_norm": 2.8831310272216797, "learning_rate": 4.944845640007222e-05, "loss": 4.9129, "step": 1223 }, { "epoch": 0.011048925798880664, "grad_norm": 5.509570121765137, "learning_rate": 4.944800505506409e-05, "loss": 4.9685, "step": 1224 }, { "epoch": 0.011057952699043149, "grad_norm": 3.154984951019287, "learning_rate": 4.944755371005597e-05, "loss": 4.8803, "step": 1225 }, { "epoch": 0.011066979599205632, "grad_norm": 3.894683361053467, "learning_rate": 4.9447102365047844e-05, "loss": 5.3389, "step": 1226 }, { "epoch": 0.011076006499368117, "grad_norm": 4.701049327850342, "learning_rate": 4.944665102003972e-05, "loss": 5.2931, "step": 1227 }, { "epoch": 0.0110850333995306, "grad_norm": 4.935464859008789, "learning_rate": 4.9446199675031596e-05, "loss": 4.2189, "step": 1228 }, { "epoch": 0.011094060299693086, "grad_norm": 4.986292362213135, "learning_rate": 4.9445748330023475e-05, "loss": 3.9696, "step": 1229 }, { "epoch": 0.011103087199855569, "grad_norm": 3.3993144035339355, "learning_rate": 4.944529698501535e-05, "loss": 3.9762, "step": 1230 }, { "epoch": 0.011112114100018054, "grad_norm": 3.3631033897399902, "learning_rate": 4.944484564000723e-05, "loss": 5.3541, "step": 1231 }, { "epoch": 0.011121141000180537, "grad_norm": 3.9366633892059326, "learning_rate": 4.94443942949991e-05, "loss": 4.7618, "step": 1232 }, { "epoch": 0.011130167900343022, "grad_norm": 4.138396263122559, "learning_rate": 4.944394294999097e-05, "loss": 4.6122, "step": 1233 }, { "epoch": 0.011139194800505507, "grad_norm": 4.89894962310791, "learning_rate": 4.944349160498285e-05, "loss": 4.6142, "step": 1234 }, { "epoch": 0.01114822170066799, "grad_norm": 5.140895843505859, "learning_rate": 4.9443040259974724e-05, "loss": 5.4725, "step": 1235 }, { "epoch": 0.011157248600830475, "grad_norm": 4.64768648147583, "learning_rate": 4.94425889149666e-05, "loss": 5.1875, "step": 1236 }, { "epoch": 0.011166275500992959, "grad_norm": 5.045102596282959, "learning_rate": 4.944213756995848e-05, "loss": 5.4592, "step": 1237 }, { "epoch": 0.011175302401155444, "grad_norm": 4.258023262023926, "learning_rate": 4.9441686224950355e-05, "loss": 5.0773, "step": 1238 }, { "epoch": 0.011184329301317927, "grad_norm": 4.74140739440918, "learning_rate": 4.944123487994223e-05, "loss": 4.9444, "step": 1239 }, { "epoch": 0.011193356201480412, "grad_norm": 7.220585346221924, "learning_rate": 4.9440783534934106e-05, "loss": 5.2443, "step": 1240 }, { "epoch": 0.011202383101642895, "grad_norm": 4.126645088195801, "learning_rate": 4.944033218992598e-05, "loss": 4.9135, "step": 1241 }, { "epoch": 0.01121141000180538, "grad_norm": 3.3155601024627686, "learning_rate": 4.943988084491786e-05, "loss": 4.9416, "step": 1242 }, { "epoch": 0.011220436901967864, "grad_norm": 4.028341770172119, "learning_rate": 4.943942949990974e-05, "loss": 4.9172, "step": 1243 }, { "epoch": 0.011229463802130349, "grad_norm": 3.0666720867156982, "learning_rate": 4.943897815490161e-05, "loss": 4.5952, "step": 1244 }, { "epoch": 0.011238490702292832, "grad_norm": 5.152514457702637, "learning_rate": 4.943852680989349e-05, "loss": 5.3474, "step": 1245 }, { "epoch": 0.011247517602455317, "grad_norm": 5.1277313232421875, "learning_rate": 4.943807546488536e-05, "loss": 4.165, "step": 1246 }, { "epoch": 0.011256544502617802, "grad_norm": 4.634603977203369, "learning_rate": 4.9437624119877234e-05, "loss": 5.4904, "step": 1247 }, { "epoch": 0.011265571402780285, "grad_norm": 6.874091625213623, "learning_rate": 4.943717277486911e-05, "loss": 3.7285, "step": 1248 }, { "epoch": 0.01127459830294277, "grad_norm": 3.8778436183929443, "learning_rate": 4.9436721429860985e-05, "loss": 5.5821, "step": 1249 }, { "epoch": 0.011283625203105253, "grad_norm": 7.267788887023926, "learning_rate": 4.9436270084852865e-05, "loss": 5.1273, "step": 1250 }, { "epoch": 0.011292652103267738, "grad_norm": 3.447422504425049, "learning_rate": 4.9435818739844744e-05, "loss": 5.1709, "step": 1251 }, { "epoch": 0.011301679003430222, "grad_norm": 3.83552622795105, "learning_rate": 4.9435367394836616e-05, "loss": 5.3138, "step": 1252 }, { "epoch": 0.011310705903592707, "grad_norm": 4.056714057922363, "learning_rate": 4.943491604982849e-05, "loss": 4.9417, "step": 1253 }, { "epoch": 0.01131973280375519, "grad_norm": 4.77383279800415, "learning_rate": 4.943446470482037e-05, "loss": 5.5251, "step": 1254 }, { "epoch": 0.011328759703917675, "grad_norm": 3.9806268215179443, "learning_rate": 4.943401335981224e-05, "loss": 4.2021, "step": 1255 }, { "epoch": 0.011337786604080158, "grad_norm": 3.909907341003418, "learning_rate": 4.943356201480412e-05, "loss": 5.0001, "step": 1256 }, { "epoch": 0.011346813504242643, "grad_norm": 4.038882732391357, "learning_rate": 4.943311066979599e-05, "loss": 4.2038, "step": 1257 }, { "epoch": 0.011355840404405126, "grad_norm": 3.455800771713257, "learning_rate": 4.943265932478787e-05, "loss": 5.002, "step": 1258 }, { "epoch": 0.011364867304567611, "grad_norm": 3.3631999492645264, "learning_rate": 4.9432207979779744e-05, "loss": 4.7174, "step": 1259 }, { "epoch": 0.011373894204730096, "grad_norm": 4.57729434967041, "learning_rate": 4.943175663477162e-05, "loss": 5.2376, "step": 1260 }, { "epoch": 0.01138292110489258, "grad_norm": 4.251590251922607, "learning_rate": 4.9431305289763496e-05, "loss": 4.7766, "step": 1261 }, { "epoch": 0.011391948005055065, "grad_norm": 4.891411781311035, "learning_rate": 4.9430853944755375e-05, "loss": 4.4203, "step": 1262 }, { "epoch": 0.011400974905217548, "grad_norm": 4.165139675140381, "learning_rate": 4.943040259974725e-05, "loss": 4.3767, "step": 1263 }, { "epoch": 0.011410001805380033, "grad_norm": 3.5337469577789307, "learning_rate": 4.942995125473913e-05, "loss": 4.5989, "step": 1264 }, { "epoch": 0.011419028705542516, "grad_norm": 3.447878837585449, "learning_rate": 4.9429499909731006e-05, "loss": 4.5253, "step": 1265 }, { "epoch": 0.011428055605705001, "grad_norm": 4.471466541290283, "learning_rate": 4.942904856472287e-05, "loss": 4.1279, "step": 1266 }, { "epoch": 0.011437082505867485, "grad_norm": 4.166001319885254, "learning_rate": 4.942859721971475e-05, "loss": 4.5314, "step": 1267 }, { "epoch": 0.01144610940602997, "grad_norm": 5.529746055603027, "learning_rate": 4.942814587470663e-05, "loss": 5.1517, "step": 1268 }, { "epoch": 0.011455136306192453, "grad_norm": 4.783323764801025, "learning_rate": 4.94276945296985e-05, "loss": 5.1598, "step": 1269 }, { "epoch": 0.011464163206354938, "grad_norm": 4.339188098907471, "learning_rate": 4.942724318469038e-05, "loss": 5.1531, "step": 1270 }, { "epoch": 0.011473190106517423, "grad_norm": 3.6588656902313232, "learning_rate": 4.9426791839682254e-05, "loss": 5.0153, "step": 1271 }, { "epoch": 0.011482217006679906, "grad_norm": 4.870579719543457, "learning_rate": 4.9426340494674134e-05, "loss": 5.0592, "step": 1272 }, { "epoch": 0.011491243906842391, "grad_norm": 4.258254051208496, "learning_rate": 4.9425889149666006e-05, "loss": 4.8547, "step": 1273 }, { "epoch": 0.011500270807004874, "grad_norm": 4.319006443023682, "learning_rate": 4.942543780465788e-05, "loss": 4.2247, "step": 1274 }, { "epoch": 0.01150929770716736, "grad_norm": 3.460817575454712, "learning_rate": 4.942498645964976e-05, "loss": 5.5578, "step": 1275 }, { "epoch": 0.011518324607329843, "grad_norm": 3.848689317703247, "learning_rate": 4.942453511464164e-05, "loss": 5.2437, "step": 1276 }, { "epoch": 0.011527351507492328, "grad_norm": 2.9206349849700928, "learning_rate": 4.942408376963351e-05, "loss": 5.0078, "step": 1277 }, { "epoch": 0.01153637840765481, "grad_norm": 4.793228626251221, "learning_rate": 4.942363242462539e-05, "loss": 4.9723, "step": 1278 }, { "epoch": 0.011545405307817296, "grad_norm": 4.7605881690979, "learning_rate": 4.942318107961727e-05, "loss": 4.7136, "step": 1279 }, { "epoch": 0.011554432207979779, "grad_norm": 3.738159418106079, "learning_rate": 4.9422729734609134e-05, "loss": 4.7332, "step": 1280 }, { "epoch": 0.011563459108142264, "grad_norm": 3.148641586303711, "learning_rate": 4.942227838960101e-05, "loss": 4.955, "step": 1281 }, { "epoch": 0.011572486008304747, "grad_norm": 4.373576641082764, "learning_rate": 4.942182704459289e-05, "loss": 4.8391, "step": 1282 }, { "epoch": 0.011581512908467232, "grad_norm": 3.621273994445801, "learning_rate": 4.9421375699584764e-05, "loss": 5.4192, "step": 1283 }, { "epoch": 0.011590539808629717, "grad_norm": 3.187734365463257, "learning_rate": 4.9420924354576644e-05, "loss": 4.5468, "step": 1284 }, { "epoch": 0.0115995667087922, "grad_norm": 4.02885627746582, "learning_rate": 4.9420473009568516e-05, "loss": 4.6609, "step": 1285 }, { "epoch": 0.011608593608954686, "grad_norm": 4.49727201461792, "learning_rate": 4.942002166456039e-05, "loss": 5.1093, "step": 1286 }, { "epoch": 0.011617620509117169, "grad_norm": 5.074577331542969, "learning_rate": 4.941957031955227e-05, "loss": 5.2476, "step": 1287 }, { "epoch": 0.011626647409279654, "grad_norm": 4.546145915985107, "learning_rate": 4.941911897454414e-05, "loss": 4.7859, "step": 1288 }, { "epoch": 0.011635674309442137, "grad_norm": 3.1325221061706543, "learning_rate": 4.941866762953602e-05, "loss": 4.4161, "step": 1289 }, { "epoch": 0.011644701209604622, "grad_norm": 4.230278015136719, "learning_rate": 4.94182162845279e-05, "loss": 5.1161, "step": 1290 }, { "epoch": 0.011653728109767105, "grad_norm": 4.198645114898682, "learning_rate": 4.941776493951977e-05, "loss": 4.6063, "step": 1291 }, { "epoch": 0.01166275500992959, "grad_norm": 6.146627426147461, "learning_rate": 4.941731359451165e-05, "loss": 5.4595, "step": 1292 }, { "epoch": 0.011671781910092074, "grad_norm": 4.982247829437256, "learning_rate": 4.941686224950352e-05, "loss": 4.7807, "step": 1293 }, { "epoch": 0.011680808810254559, "grad_norm": 4.012475967407227, "learning_rate": 4.9416410904495395e-05, "loss": 4.4526, "step": 1294 }, { "epoch": 0.011689835710417042, "grad_norm": 4.5518317222595215, "learning_rate": 4.9415959559487275e-05, "loss": 4.7748, "step": 1295 }, { "epoch": 0.011698862610579527, "grad_norm": 3.2414920330047607, "learning_rate": 4.941550821447915e-05, "loss": 4.3564, "step": 1296 }, { "epoch": 0.011707889510742012, "grad_norm": 4.560291767120361, "learning_rate": 4.9415056869471026e-05, "loss": 5.2679, "step": 1297 }, { "epoch": 0.011716916410904495, "grad_norm": 3.090766191482544, "learning_rate": 4.9414605524462906e-05, "loss": 4.6628, "step": 1298 }, { "epoch": 0.01172594331106698, "grad_norm": 3.8495724201202393, "learning_rate": 4.941415417945478e-05, "loss": 4.3706, "step": 1299 }, { "epoch": 0.011734970211229464, "grad_norm": 5.527219772338867, "learning_rate": 4.941370283444665e-05, "loss": 5.0488, "step": 1300 }, { "epoch": 0.011743997111391949, "grad_norm": 4.136092185974121, "learning_rate": 4.941325148943853e-05, "loss": 4.4, "step": 1301 }, { "epoch": 0.011753024011554432, "grad_norm": 4.704952239990234, "learning_rate": 4.94128001444304e-05, "loss": 5.1633, "step": 1302 }, { "epoch": 0.011762050911716917, "grad_norm": 3.700702667236328, "learning_rate": 4.941234879942228e-05, "loss": 3.6489, "step": 1303 }, { "epoch": 0.0117710778118794, "grad_norm": 4.030555248260498, "learning_rate": 4.941189745441416e-05, "loss": 4.8639, "step": 1304 }, { "epoch": 0.011780104712041885, "grad_norm": 7.316988468170166, "learning_rate": 4.941144610940603e-05, "loss": 5.197, "step": 1305 }, { "epoch": 0.011789131612204368, "grad_norm": 3.598806858062744, "learning_rate": 4.9410994764397906e-05, "loss": 4.0759, "step": 1306 }, { "epoch": 0.011798158512366853, "grad_norm": 3.466377019882202, "learning_rate": 4.9410543419389785e-05, "loss": 4.9337, "step": 1307 }, { "epoch": 0.011807185412529337, "grad_norm": 4.252083778381348, "learning_rate": 4.941009207438166e-05, "loss": 5.1858, "step": 1308 }, { "epoch": 0.011816212312691822, "grad_norm": 5.51515007019043, "learning_rate": 4.940964072937354e-05, "loss": 4.0298, "step": 1309 }, { "epoch": 0.011825239212854307, "grad_norm": 3.4082882404327393, "learning_rate": 4.940918938436541e-05, "loss": 4.877, "step": 1310 }, { "epoch": 0.01183426611301679, "grad_norm": 5.754518032073975, "learning_rate": 4.940873803935729e-05, "loss": 3.61, "step": 1311 }, { "epoch": 0.011843293013179275, "grad_norm": 3.9678311347961426, "learning_rate": 4.940828669434917e-05, "loss": 5.3857, "step": 1312 }, { "epoch": 0.011852319913341758, "grad_norm": 4.4502339363098145, "learning_rate": 4.940783534934103e-05, "loss": 4.9324, "step": 1313 }, { "epoch": 0.011861346813504243, "grad_norm": 5.422971248626709, "learning_rate": 4.940738400433291e-05, "loss": 4.6671, "step": 1314 }, { "epoch": 0.011870373713666726, "grad_norm": 3.841655731201172, "learning_rate": 4.940693265932479e-05, "loss": 4.9189, "step": 1315 }, { "epoch": 0.011879400613829211, "grad_norm": 4.392859935760498, "learning_rate": 4.9406481314316664e-05, "loss": 5.1593, "step": 1316 }, { "epoch": 0.011888427513991695, "grad_norm": 4.925874710083008, "learning_rate": 4.9406029969308543e-05, "loss": 4.7543, "step": 1317 }, { "epoch": 0.01189745441415418, "grad_norm": 8.151869773864746, "learning_rate": 4.9405578624300416e-05, "loss": 4.9978, "step": 1318 }, { "epoch": 0.011906481314316663, "grad_norm": 3.4940600395202637, "learning_rate": 4.9405127279292295e-05, "loss": 5.2132, "step": 1319 }, { "epoch": 0.011915508214479148, "grad_norm": 3.487550973892212, "learning_rate": 4.940467593428417e-05, "loss": 5.1655, "step": 1320 }, { "epoch": 0.011924535114641631, "grad_norm": 4.664997577667236, "learning_rate": 4.940422458927605e-05, "loss": 3.941, "step": 1321 }, { "epoch": 0.011933562014804116, "grad_norm": 4.3085832595825195, "learning_rate": 4.940377324426792e-05, "loss": 4.5673, "step": 1322 }, { "epoch": 0.011942588914966601, "grad_norm": 4.840452671051025, "learning_rate": 4.94033218992598e-05, "loss": 4.6072, "step": 1323 }, { "epoch": 0.011951615815129085, "grad_norm": 3.576341152191162, "learning_rate": 4.940287055425167e-05, "loss": 4.6948, "step": 1324 }, { "epoch": 0.01196064271529157, "grad_norm": 4.8745293617248535, "learning_rate": 4.940241920924355e-05, "loss": 5.3677, "step": 1325 }, { "epoch": 0.011969669615454053, "grad_norm": 3.8922088146209717, "learning_rate": 4.940196786423543e-05, "loss": 5.0348, "step": 1326 }, { "epoch": 0.011978696515616538, "grad_norm": 4.138410568237305, "learning_rate": 4.9401516519227295e-05, "loss": 4.6243, "step": 1327 }, { "epoch": 0.011987723415779021, "grad_norm": 3.270482301712036, "learning_rate": 4.9401065174219174e-05, "loss": 4.516, "step": 1328 }, { "epoch": 0.011996750315941506, "grad_norm": 3.720442771911621, "learning_rate": 4.9400613829211054e-05, "loss": 5.4129, "step": 1329 }, { "epoch": 0.01200577721610399, "grad_norm": 4.635166645050049, "learning_rate": 4.9400162484202926e-05, "loss": 5.3266, "step": 1330 }, { "epoch": 0.012014804116266474, "grad_norm": 3.7957515716552734, "learning_rate": 4.9399711139194805e-05, "loss": 5.0544, "step": 1331 }, { "epoch": 0.012023831016428958, "grad_norm": 5.62252140045166, "learning_rate": 4.939925979418668e-05, "loss": 5.0514, "step": 1332 }, { "epoch": 0.012032857916591443, "grad_norm": 4.788458824157715, "learning_rate": 4.939880844917855e-05, "loss": 4.9484, "step": 1333 }, { "epoch": 0.012041884816753926, "grad_norm": 3.8305349349975586, "learning_rate": 4.939835710417043e-05, "loss": 4.2204, "step": 1334 }, { "epoch": 0.012050911716916411, "grad_norm": 3.687129020690918, "learning_rate": 4.93979057591623e-05, "loss": 4.7019, "step": 1335 }, { "epoch": 0.012059938617078896, "grad_norm": 3.434675931930542, "learning_rate": 4.939745441415418e-05, "loss": 4.9516, "step": 1336 }, { "epoch": 0.01206896551724138, "grad_norm": 7.457440376281738, "learning_rate": 4.939700306914606e-05, "loss": 5.1833, "step": 1337 }, { "epoch": 0.012077992417403864, "grad_norm": 6.273955821990967, "learning_rate": 4.939655172413793e-05, "loss": 4.2881, "step": 1338 }, { "epoch": 0.012087019317566347, "grad_norm": 3.7345755100250244, "learning_rate": 4.939610037912981e-05, "loss": 4.9831, "step": 1339 }, { "epoch": 0.012096046217728832, "grad_norm": 2.753933906555176, "learning_rate": 4.9395649034121685e-05, "loss": 4.8063, "step": 1340 }, { "epoch": 0.012105073117891316, "grad_norm": 3.199284553527832, "learning_rate": 4.939519768911356e-05, "loss": 5.0885, "step": 1341 }, { "epoch": 0.0121141000180538, "grad_norm": 4.977487087249756, "learning_rate": 4.9394746344105436e-05, "loss": 5.0519, "step": 1342 }, { "epoch": 0.012123126918216284, "grad_norm": 3.7824110984802246, "learning_rate": 4.9394294999097316e-05, "loss": 4.883, "step": 1343 }, { "epoch": 0.012132153818378769, "grad_norm": 3.579320192337036, "learning_rate": 4.939384365408919e-05, "loss": 5.3966, "step": 1344 }, { "epoch": 0.012141180718541252, "grad_norm": 4.252265930175781, "learning_rate": 4.939339230908107e-05, "loss": 4.6618, "step": 1345 }, { "epoch": 0.012150207618703737, "grad_norm": 4.96865701675415, "learning_rate": 4.939294096407294e-05, "loss": 3.8568, "step": 1346 }, { "epoch": 0.01215923451886622, "grad_norm": 4.031436920166016, "learning_rate": 4.939248961906481e-05, "loss": 4.985, "step": 1347 }, { "epoch": 0.012168261419028705, "grad_norm": 3.964413642883301, "learning_rate": 4.939203827405669e-05, "loss": 5.3975, "step": 1348 }, { "epoch": 0.01217728831919119, "grad_norm": 3.6391501426696777, "learning_rate": 4.9391586929048564e-05, "loss": 4.6299, "step": 1349 }, { "epoch": 0.012186315219353674, "grad_norm": 3.267834424972534, "learning_rate": 4.939113558404044e-05, "loss": 5.4747, "step": 1350 }, { "epoch": 0.012195342119516159, "grad_norm": 5.0569047927856445, "learning_rate": 4.939068423903232e-05, "loss": 4.9135, "step": 1351 }, { "epoch": 0.012204369019678642, "grad_norm": 4.096376419067383, "learning_rate": 4.9390232894024195e-05, "loss": 5.1408, "step": 1352 }, { "epoch": 0.012213395919841127, "grad_norm": 7.672942161560059, "learning_rate": 4.938978154901607e-05, "loss": 4.0038, "step": 1353 }, { "epoch": 0.01222242282000361, "grad_norm": 3.5208470821380615, "learning_rate": 4.938933020400795e-05, "loss": 4.8572, "step": 1354 }, { "epoch": 0.012231449720166095, "grad_norm": 3.566149950027466, "learning_rate": 4.938887885899982e-05, "loss": 4.5011, "step": 1355 }, { "epoch": 0.012240476620328579, "grad_norm": 3.6635403633117676, "learning_rate": 4.93884275139917e-05, "loss": 5.3045, "step": 1356 }, { "epoch": 0.012249503520491064, "grad_norm": 4.123534679412842, "learning_rate": 4.938797616898357e-05, "loss": 4.6333, "step": 1357 }, { "epoch": 0.012258530420653547, "grad_norm": 3.9883923530578613, "learning_rate": 4.938752482397545e-05, "loss": 4.6864, "step": 1358 }, { "epoch": 0.012267557320816032, "grad_norm": 3.6722042560577393, "learning_rate": 4.938707347896733e-05, "loss": 5.5376, "step": 1359 }, { "epoch": 0.012276584220978517, "grad_norm": 7.189025402069092, "learning_rate": 4.9386622133959195e-05, "loss": 4.721, "step": 1360 }, { "epoch": 0.012285611121141, "grad_norm": 3.9769351482391357, "learning_rate": 4.9386170788951074e-05, "loss": 5.4193, "step": 1361 }, { "epoch": 0.012294638021303485, "grad_norm": 2.8915457725524902, "learning_rate": 4.9385719443942953e-05, "loss": 5.0165, "step": 1362 }, { "epoch": 0.012303664921465968, "grad_norm": 3.8430066108703613, "learning_rate": 4.9385268098934826e-05, "loss": 4.919, "step": 1363 }, { "epoch": 0.012312691821628453, "grad_norm": 3.369408130645752, "learning_rate": 4.9384816753926705e-05, "loss": 4.5925, "step": 1364 }, { "epoch": 0.012321718721790937, "grad_norm": 4.366002082824707, "learning_rate": 4.9384365408918584e-05, "loss": 4.8438, "step": 1365 }, { "epoch": 0.012330745621953422, "grad_norm": 8.521025657653809, "learning_rate": 4.938391406391046e-05, "loss": 3.8649, "step": 1366 }, { "epoch": 0.012339772522115905, "grad_norm": 4.782292366027832, "learning_rate": 4.938346271890233e-05, "loss": 4.8191, "step": 1367 }, { "epoch": 0.01234879942227839, "grad_norm": 3.446610927581787, "learning_rate": 4.938301137389421e-05, "loss": 4.2469, "step": 1368 }, { "epoch": 0.012357826322440873, "grad_norm": 3.3387129306793213, "learning_rate": 4.938256002888608e-05, "loss": 5.0334, "step": 1369 }, { "epoch": 0.012366853222603358, "grad_norm": 4.112967491149902, "learning_rate": 4.938210868387796e-05, "loss": 3.7674, "step": 1370 }, { "epoch": 0.012375880122765841, "grad_norm": 5.0502190589904785, "learning_rate": 4.938165733886983e-05, "loss": 5.0386, "step": 1371 }, { "epoch": 0.012384907022928326, "grad_norm": 5.544337749481201, "learning_rate": 4.938120599386171e-05, "loss": 2.5045, "step": 1372 }, { "epoch": 0.012393933923090811, "grad_norm": 4.261744499206543, "learning_rate": 4.938075464885359e-05, "loss": 5.6056, "step": 1373 }, { "epoch": 0.012402960823253295, "grad_norm": 3.468864917755127, "learning_rate": 4.938030330384546e-05, "loss": 5.285, "step": 1374 }, { "epoch": 0.01241198772341578, "grad_norm": 5.491814613342285, "learning_rate": 4.9379851958837336e-05, "loss": 5.7399, "step": 1375 }, { "epoch": 0.012421014623578263, "grad_norm": 3.871513605117798, "learning_rate": 4.9379400613829215e-05, "loss": 4.3828, "step": 1376 }, { "epoch": 0.012430041523740748, "grad_norm": 3.3360133171081543, "learning_rate": 4.937894926882109e-05, "loss": 4.7314, "step": 1377 }, { "epoch": 0.012439068423903231, "grad_norm": 3.768447160720825, "learning_rate": 4.937849792381297e-05, "loss": 5.2603, "step": 1378 }, { "epoch": 0.012448095324065716, "grad_norm": 4.326568126678467, "learning_rate": 4.937804657880484e-05, "loss": 4.472, "step": 1379 }, { "epoch": 0.0124571222242282, "grad_norm": 3.968127727508545, "learning_rate": 4.937759523379671e-05, "loss": 4.8549, "step": 1380 }, { "epoch": 0.012466149124390685, "grad_norm": 4.690174102783203, "learning_rate": 4.937714388878859e-05, "loss": 5.0361, "step": 1381 }, { "epoch": 0.012475176024553168, "grad_norm": 3.793745279312134, "learning_rate": 4.937669254378047e-05, "loss": 5.1321, "step": 1382 }, { "epoch": 0.012484202924715653, "grad_norm": 3.604121446609497, "learning_rate": 4.937624119877234e-05, "loss": 4.8401, "step": 1383 }, { "epoch": 0.012493229824878136, "grad_norm": 3.3972017765045166, "learning_rate": 4.937578985376422e-05, "loss": 4.8264, "step": 1384 }, { "epoch": 0.012502256725040621, "grad_norm": 3.390611410140991, "learning_rate": 4.9375338508756095e-05, "loss": 4.5791, "step": 1385 }, { "epoch": 0.012511283625203106, "grad_norm": 6.1628804206848145, "learning_rate": 4.9374887163747974e-05, "loss": 4.7591, "step": 1386 }, { "epoch": 0.01252031052536559, "grad_norm": 2.3633921146392822, "learning_rate": 4.9374435818739846e-05, "loss": 4.7753, "step": 1387 }, { "epoch": 0.012529337425528074, "grad_norm": 3.6960184574127197, "learning_rate": 4.937398447373172e-05, "loss": 5.2975, "step": 1388 }, { "epoch": 0.012538364325690558, "grad_norm": 3.7768938541412354, "learning_rate": 4.93735331287236e-05, "loss": 4.9337, "step": 1389 }, { "epoch": 0.012547391225853043, "grad_norm": 3.89339280128479, "learning_rate": 4.937308178371548e-05, "loss": 4.8236, "step": 1390 }, { "epoch": 0.012556418126015526, "grad_norm": 4.510386943817139, "learning_rate": 4.937263043870735e-05, "loss": 4.8363, "step": 1391 }, { "epoch": 0.012565445026178011, "grad_norm": 4.869132041931152, "learning_rate": 4.937217909369923e-05, "loss": 4.4601, "step": 1392 }, { "epoch": 0.012574471926340494, "grad_norm": 5.526404857635498, "learning_rate": 4.93717277486911e-05, "loss": 4.8499, "step": 1393 }, { "epoch": 0.01258349882650298, "grad_norm": 5.022368907928467, "learning_rate": 4.9371276403682974e-05, "loss": 5.0476, "step": 1394 }, { "epoch": 0.012592525726665462, "grad_norm": 4.110621929168701, "learning_rate": 4.937082505867485e-05, "loss": 4.7084, "step": 1395 }, { "epoch": 0.012601552626827947, "grad_norm": 4.396614074707031, "learning_rate": 4.9370373713666726e-05, "loss": 4.6319, "step": 1396 }, { "epoch": 0.01261057952699043, "grad_norm": 3.1149730682373047, "learning_rate": 4.9369922368658605e-05, "loss": 4.5711, "step": 1397 }, { "epoch": 0.012619606427152916, "grad_norm": 3.922621726989746, "learning_rate": 4.9369471023650484e-05, "loss": 4.9463, "step": 1398 }, { "epoch": 0.0126286333273154, "grad_norm": 4.954376220703125, "learning_rate": 4.936901967864236e-05, "loss": 5.1011, "step": 1399 }, { "epoch": 0.012637660227477884, "grad_norm": 3.551724672317505, "learning_rate": 4.936856833363423e-05, "loss": 3.2803, "step": 1400 }, { "epoch": 0.012646687127640369, "grad_norm": 4.608685493469238, "learning_rate": 4.936811698862611e-05, "loss": 4.9919, "step": 1401 }, { "epoch": 0.012655714027802852, "grad_norm": 3.713625192642212, "learning_rate": 4.936766564361798e-05, "loss": 4.8783, "step": 1402 }, { "epoch": 0.012664740927965337, "grad_norm": 4.168898105621338, "learning_rate": 4.936721429860986e-05, "loss": 4.9148, "step": 1403 }, { "epoch": 0.01267376782812782, "grad_norm": 4.351160526275635, "learning_rate": 4.936676295360174e-05, "loss": 4.9435, "step": 1404 }, { "epoch": 0.012682794728290306, "grad_norm": 4.139448165893555, "learning_rate": 4.936631160859361e-05, "loss": 4.4634, "step": 1405 }, { "epoch": 0.012691821628452789, "grad_norm": 7.223930835723877, "learning_rate": 4.936586026358549e-05, "loss": 5.1586, "step": 1406 }, { "epoch": 0.012700848528615274, "grad_norm": 6.1633524894714355, "learning_rate": 4.9365408918577363e-05, "loss": 4.6613, "step": 1407 }, { "epoch": 0.012709875428777757, "grad_norm": 3.3306429386138916, "learning_rate": 4.9364957573569236e-05, "loss": 4.6351, "step": 1408 }, { "epoch": 0.012718902328940242, "grad_norm": 3.9830353260040283, "learning_rate": 4.9364506228561115e-05, "loss": 4.2434, "step": 1409 }, { "epoch": 0.012727929229102725, "grad_norm": 3.349914789199829, "learning_rate": 4.936405488355299e-05, "loss": 5.1597, "step": 1410 }, { "epoch": 0.01273695612926521, "grad_norm": 7.813857555389404, "learning_rate": 4.936360353854487e-05, "loss": 5.3041, "step": 1411 }, { "epoch": 0.012745983029427695, "grad_norm": 3.7567498683929443, "learning_rate": 4.9363152193536746e-05, "loss": 4.1682, "step": 1412 }, { "epoch": 0.012755009929590179, "grad_norm": 4.482232570648193, "learning_rate": 4.936270084852862e-05, "loss": 4.9119, "step": 1413 }, { "epoch": 0.012764036829752664, "grad_norm": 4.099105358123779, "learning_rate": 4.936224950352049e-05, "loss": 4.7804, "step": 1414 }, { "epoch": 0.012773063729915147, "grad_norm": 4.593034267425537, "learning_rate": 4.936179815851237e-05, "loss": 4.9423, "step": 1415 }, { "epoch": 0.012782090630077632, "grad_norm": 4.290309429168701, "learning_rate": 4.936134681350424e-05, "loss": 4.882, "step": 1416 }, { "epoch": 0.012791117530240115, "grad_norm": 6.926157474517822, "learning_rate": 4.936089546849612e-05, "loss": 5.0467, "step": 1417 }, { "epoch": 0.0128001444304026, "grad_norm": 3.126817464828491, "learning_rate": 4.9360444123487994e-05, "loss": 4.6831, "step": 1418 }, { "epoch": 0.012809171330565083, "grad_norm": 3.724536895751953, "learning_rate": 4.9359992778479874e-05, "loss": 4.9243, "step": 1419 }, { "epoch": 0.012818198230727568, "grad_norm": 4.690654754638672, "learning_rate": 4.9359541433471746e-05, "loss": 4.9939, "step": 1420 }, { "epoch": 0.012827225130890052, "grad_norm": 5.637835502624512, "learning_rate": 4.9359090088463625e-05, "loss": 5.0738, "step": 1421 }, { "epoch": 0.012836252031052537, "grad_norm": 3.972559928894043, "learning_rate": 4.93586387434555e-05, "loss": 5.4975, "step": 1422 }, { "epoch": 0.01284527893121502, "grad_norm": 3.862428665161133, "learning_rate": 4.935818739844738e-05, "loss": 3.8513, "step": 1423 }, { "epoch": 0.012854305831377505, "grad_norm": 4.781521797180176, "learning_rate": 4.935773605343925e-05, "loss": 5.2014, "step": 1424 }, { "epoch": 0.01286333273153999, "grad_norm": 4.287476539611816, "learning_rate": 4.935728470843113e-05, "loss": 5.0946, "step": 1425 }, { "epoch": 0.012872359631702473, "grad_norm": 5.879729747772217, "learning_rate": 4.935683336342301e-05, "loss": 4.6801, "step": 1426 }, { "epoch": 0.012881386531864958, "grad_norm": 3.7419052124023438, "learning_rate": 4.9356382018414874e-05, "loss": 3.6659, "step": 1427 }, { "epoch": 0.012890413432027441, "grad_norm": 3.600729465484619, "learning_rate": 4.935593067340675e-05, "loss": 4.8862, "step": 1428 }, { "epoch": 0.012899440332189926, "grad_norm": 3.758721113204956, "learning_rate": 4.935547932839863e-05, "loss": 4.089, "step": 1429 }, { "epoch": 0.01290846723235241, "grad_norm": 4.568583965301514, "learning_rate": 4.9355027983390505e-05, "loss": 5.5553, "step": 1430 }, { "epoch": 0.012917494132514895, "grad_norm": 3.931347608566284, "learning_rate": 4.9354576638382384e-05, "loss": 5.2779, "step": 1431 }, { "epoch": 0.012926521032677378, "grad_norm": 4.558419227600098, "learning_rate": 4.9354125293374256e-05, "loss": 5.1051, "step": 1432 }, { "epoch": 0.012935547932839863, "grad_norm": 3.5476858615875244, "learning_rate": 4.9353673948366136e-05, "loss": 4.4463, "step": 1433 }, { "epoch": 0.012944574833002346, "grad_norm": 3.809781789779663, "learning_rate": 4.935322260335801e-05, "loss": 4.3536, "step": 1434 }, { "epoch": 0.012953601733164831, "grad_norm": 2.54215407371521, "learning_rate": 4.935277125834988e-05, "loss": 5.2429, "step": 1435 }, { "epoch": 0.012962628633327315, "grad_norm": 5.27992057800293, "learning_rate": 4.935231991334176e-05, "loss": 5.3106, "step": 1436 }, { "epoch": 0.0129716555334898, "grad_norm": 3.916983127593994, "learning_rate": 4.935186856833364e-05, "loss": 4.4255, "step": 1437 }, { "epoch": 0.012980682433652285, "grad_norm": 4.63434362411499, "learning_rate": 4.935141722332551e-05, "loss": 4.8466, "step": 1438 }, { "epoch": 0.012989709333814768, "grad_norm": 5.6624979972839355, "learning_rate": 4.935096587831739e-05, "loss": 4.7763, "step": 1439 }, { "epoch": 0.012998736233977253, "grad_norm": 3.3070170879364014, "learning_rate": 4.935051453330927e-05, "loss": 5.1426, "step": 1440 }, { "epoch": 0.013007763134139736, "grad_norm": 4.079953670501709, "learning_rate": 4.9350063188301136e-05, "loss": 4.8713, "step": 1441 }, { "epoch": 0.013016790034302221, "grad_norm": 5.500600814819336, "learning_rate": 4.9349611843293015e-05, "loss": 4.5865, "step": 1442 }, { "epoch": 0.013025816934464704, "grad_norm": 3.360496997833252, "learning_rate": 4.9349160498284894e-05, "loss": 5.1072, "step": 1443 }, { "epoch": 0.01303484383462719, "grad_norm": 3.7629809379577637, "learning_rate": 4.934870915327677e-05, "loss": 5.3586, "step": 1444 }, { "epoch": 0.013043870734789673, "grad_norm": 3.3974359035491943, "learning_rate": 4.9348257808268646e-05, "loss": 4.7946, "step": 1445 }, { "epoch": 0.013052897634952158, "grad_norm": 3.7959439754486084, "learning_rate": 4.934780646326052e-05, "loss": 3.9947, "step": 1446 }, { "epoch": 0.013061924535114641, "grad_norm": 5.848896026611328, "learning_rate": 4.934735511825239e-05, "loss": 3.4734, "step": 1447 }, { "epoch": 0.013070951435277126, "grad_norm": 3.83923077583313, "learning_rate": 4.934690377324427e-05, "loss": 4.6449, "step": 1448 }, { "epoch": 0.013079978335439611, "grad_norm": 4.257209777832031, "learning_rate": 4.934645242823614e-05, "loss": 3.9445, "step": 1449 }, { "epoch": 0.013089005235602094, "grad_norm": 3.9191277027130127, "learning_rate": 4.934600108322802e-05, "loss": 5.4235, "step": 1450 }, { "epoch": 0.01309803213576458, "grad_norm": 4.668932914733887, "learning_rate": 4.93455497382199e-05, "loss": 5.0517, "step": 1451 }, { "epoch": 0.013107059035927062, "grad_norm": 3.297539472579956, "learning_rate": 4.9345098393211773e-05, "loss": 4.7155, "step": 1452 }, { "epoch": 0.013116085936089547, "grad_norm": 3.494335651397705, "learning_rate": 4.934464704820365e-05, "loss": 5.3166, "step": 1453 }, { "epoch": 0.01312511283625203, "grad_norm": 6.243860721588135, "learning_rate": 4.9344195703195525e-05, "loss": 4.5306, "step": 1454 }, { "epoch": 0.013134139736414516, "grad_norm": 5.086581230163574, "learning_rate": 4.93437443581874e-05, "loss": 5.5449, "step": 1455 }, { "epoch": 0.013143166636576999, "grad_norm": 3.1534128189086914, "learning_rate": 4.934329301317928e-05, "loss": 4.6944, "step": 1456 }, { "epoch": 0.013152193536739484, "grad_norm": 5.610918998718262, "learning_rate": 4.934284166817115e-05, "loss": 4.8988, "step": 1457 }, { "epoch": 0.013161220436901967, "grad_norm": 3.9026894569396973, "learning_rate": 4.934239032316303e-05, "loss": 4.7639, "step": 1458 }, { "epoch": 0.013170247337064452, "grad_norm": 4.149236679077148, "learning_rate": 4.934193897815491e-05, "loss": 5.3009, "step": 1459 }, { "epoch": 0.013179274237226936, "grad_norm": 4.084713459014893, "learning_rate": 4.934148763314678e-05, "loss": 5.3906, "step": 1460 }, { "epoch": 0.01318830113738942, "grad_norm": 3.5079619884490967, "learning_rate": 4.934103628813865e-05, "loss": 4.6339, "step": 1461 }, { "epoch": 0.013197328037551906, "grad_norm": 4.5606842041015625, "learning_rate": 4.934058494313053e-05, "loss": 4.8973, "step": 1462 }, { "epoch": 0.013206354937714389, "grad_norm": 4.453874111175537, "learning_rate": 4.9340133598122404e-05, "loss": 5.1189, "step": 1463 }, { "epoch": 0.013215381837876874, "grad_norm": 4.040550708770752, "learning_rate": 4.9339682253114284e-05, "loss": 3.9325, "step": 1464 }, { "epoch": 0.013224408738039357, "grad_norm": 4.506809234619141, "learning_rate": 4.933923090810616e-05, "loss": 4.9267, "step": 1465 }, { "epoch": 0.013233435638201842, "grad_norm": 3.4074501991271973, "learning_rate": 4.9338779563098035e-05, "loss": 4.9108, "step": 1466 }, { "epoch": 0.013242462538364325, "grad_norm": 5.863563060760498, "learning_rate": 4.933832821808991e-05, "loss": 4.3142, "step": 1467 }, { "epoch": 0.01325148943852681, "grad_norm": 4.140857696533203, "learning_rate": 4.933787687308179e-05, "loss": 4.9558, "step": 1468 }, { "epoch": 0.013260516338689294, "grad_norm": 3.7835793495178223, "learning_rate": 4.933742552807366e-05, "loss": 5.0221, "step": 1469 }, { "epoch": 0.013269543238851779, "grad_norm": 3.9014766216278076, "learning_rate": 4.933697418306554e-05, "loss": 4.7045, "step": 1470 }, { "epoch": 0.013278570139014262, "grad_norm": 4.894827365875244, "learning_rate": 4.933652283805741e-05, "loss": 5.5225, "step": 1471 }, { "epoch": 0.013287597039176747, "grad_norm": 4.622293472290039, "learning_rate": 4.933607149304929e-05, "loss": 4.6318, "step": 1472 }, { "epoch": 0.01329662393933923, "grad_norm": 3.6974809169769287, "learning_rate": 4.933562014804117e-05, "loss": 5.327, "step": 1473 }, { "epoch": 0.013305650839501715, "grad_norm": 5.996678352355957, "learning_rate": 4.9335168803033035e-05, "loss": 5.2217, "step": 1474 }, { "epoch": 0.0133146777396642, "grad_norm": 4.2731523513793945, "learning_rate": 4.9334717458024915e-05, "loss": 4.4294, "step": 1475 }, { "epoch": 0.013323704639826683, "grad_norm": 3.8386688232421875, "learning_rate": 4.9334266113016794e-05, "loss": 4.3619, "step": 1476 }, { "epoch": 0.013332731539989168, "grad_norm": 3.1611568927764893, "learning_rate": 4.9333814768008666e-05, "loss": 4.8613, "step": 1477 }, { "epoch": 0.013341758440151652, "grad_norm": 5.0734405517578125, "learning_rate": 4.9333363423000546e-05, "loss": 4.9205, "step": 1478 }, { "epoch": 0.013350785340314137, "grad_norm": 5.37762975692749, "learning_rate": 4.933291207799242e-05, "loss": 5.5653, "step": 1479 }, { "epoch": 0.01335981224047662, "grad_norm": 3.386352300643921, "learning_rate": 4.93324607329843e-05, "loss": 4.8914, "step": 1480 }, { "epoch": 0.013368839140639105, "grad_norm": 3.6827638149261475, "learning_rate": 4.933200938797617e-05, "loss": 4.0947, "step": 1481 }, { "epoch": 0.013377866040801588, "grad_norm": 4.194892406463623, "learning_rate": 4.933155804296805e-05, "loss": 4.5173, "step": 1482 }, { "epoch": 0.013386892940964073, "grad_norm": 3.927053689956665, "learning_rate": 4.933110669795992e-05, "loss": 4.9634, "step": 1483 }, { "epoch": 0.013395919841126556, "grad_norm": 4.946029186248779, "learning_rate": 4.93306553529518e-05, "loss": 5.1301, "step": 1484 }, { "epoch": 0.013404946741289041, "grad_norm": 4.601600170135498, "learning_rate": 4.933020400794367e-05, "loss": 4.7826, "step": 1485 }, { "epoch": 0.013413973641451525, "grad_norm": 3.258819580078125, "learning_rate": 4.932975266293555e-05, "loss": 4.9863, "step": 1486 }, { "epoch": 0.01342300054161401, "grad_norm": 3.411864757537842, "learning_rate": 4.932930131792743e-05, "loss": 5.0066, "step": 1487 }, { "epoch": 0.013432027441776495, "grad_norm": 3.6426591873168945, "learning_rate": 4.93288499729193e-05, "loss": 4.3939, "step": 1488 }, { "epoch": 0.013441054341938978, "grad_norm": 5.10517692565918, "learning_rate": 4.932839862791118e-05, "loss": 4.6859, "step": 1489 }, { "epoch": 0.013450081242101463, "grad_norm": 6.436878204345703, "learning_rate": 4.9327947282903056e-05, "loss": 4.3581, "step": 1490 }, { "epoch": 0.013459108142263946, "grad_norm": 3.5496132373809814, "learning_rate": 4.932749593789493e-05, "loss": 5.037, "step": 1491 }, { "epoch": 0.013468135042426431, "grad_norm": 3.0907680988311768, "learning_rate": 4.932704459288681e-05, "loss": 5.0903, "step": 1492 }, { "epoch": 0.013477161942588915, "grad_norm": 4.611671447753906, "learning_rate": 4.932659324787868e-05, "loss": 4.8664, "step": 1493 }, { "epoch": 0.0134861888427514, "grad_norm": 4.102431297302246, "learning_rate": 4.932614190287055e-05, "loss": 4.547, "step": 1494 }, { "epoch": 0.013495215742913883, "grad_norm": 3.1096200942993164, "learning_rate": 4.932569055786243e-05, "loss": 4.8197, "step": 1495 }, { "epoch": 0.013504242643076368, "grad_norm": 4.319351673126221, "learning_rate": 4.9325239212854304e-05, "loss": 4.6836, "step": 1496 }, { "epoch": 0.013513269543238851, "grad_norm": 4.230674743652344, "learning_rate": 4.9324787867846183e-05, "loss": 4.7826, "step": 1497 }, { "epoch": 0.013522296443401336, "grad_norm": 3.6756064891815186, "learning_rate": 4.932433652283806e-05, "loss": 5.4355, "step": 1498 }, { "epoch": 0.01353132334356382, "grad_norm": 3.6152775287628174, "learning_rate": 4.9323885177829935e-05, "loss": 4.9167, "step": 1499 }, { "epoch": 0.013540350243726304, "grad_norm": 3.559957981109619, "learning_rate": 4.9323433832821814e-05, "loss": 5.3513, "step": 1500 }, { "epoch": 0.01354937714388879, "grad_norm": 6.238595962524414, "learning_rate": 4.932298248781369e-05, "loss": 4.5257, "step": 1501 }, { "epoch": 0.013558404044051273, "grad_norm": 3.487497091293335, "learning_rate": 4.932253114280556e-05, "loss": 5.1147, "step": 1502 }, { "epoch": 0.013567430944213758, "grad_norm": 3.6665139198303223, "learning_rate": 4.932207979779744e-05, "loss": 5.0757, "step": 1503 }, { "epoch": 0.013576457844376241, "grad_norm": 5.237826347351074, "learning_rate": 4.932162845278932e-05, "loss": 5.2194, "step": 1504 }, { "epoch": 0.013585484744538726, "grad_norm": 5.284128665924072, "learning_rate": 4.932117710778119e-05, "loss": 5.1328, "step": 1505 }, { "epoch": 0.01359451164470121, "grad_norm": 4.595756530761719, "learning_rate": 4.932072576277307e-05, "loss": 4.7629, "step": 1506 }, { "epoch": 0.013603538544863694, "grad_norm": 3.964712142944336, "learning_rate": 4.932027441776494e-05, "loss": 4.5531, "step": 1507 }, { "epoch": 0.013612565445026177, "grad_norm": 4.605990886688232, "learning_rate": 4.9319823072756814e-05, "loss": 4.4169, "step": 1508 }, { "epoch": 0.013621592345188662, "grad_norm": 3.6736342906951904, "learning_rate": 4.9319371727748694e-05, "loss": 4.5739, "step": 1509 }, { "epoch": 0.013630619245351146, "grad_norm": 5.705549240112305, "learning_rate": 4.9318920382740566e-05, "loss": 4.6093, "step": 1510 }, { "epoch": 0.01363964614551363, "grad_norm": 4.076338768005371, "learning_rate": 4.9318469037732445e-05, "loss": 5.1099, "step": 1511 }, { "epoch": 0.013648673045676114, "grad_norm": 5.3485798835754395, "learning_rate": 4.9318017692724325e-05, "loss": 4.6165, "step": 1512 }, { "epoch": 0.013657699945838599, "grad_norm": 2.9795949459075928, "learning_rate": 4.93175663477162e-05, "loss": 4.8525, "step": 1513 }, { "epoch": 0.013666726846001084, "grad_norm": 3.9624276161193848, "learning_rate": 4.931711500270807e-05, "loss": 4.7349, "step": 1514 }, { "epoch": 0.013675753746163567, "grad_norm": 4.985337257385254, "learning_rate": 4.931666365769995e-05, "loss": 5.2065, "step": 1515 }, { "epoch": 0.013684780646326052, "grad_norm": 6.562350273132324, "learning_rate": 4.931621231269182e-05, "loss": 4.6644, "step": 1516 }, { "epoch": 0.013693807546488536, "grad_norm": 3.6337192058563232, "learning_rate": 4.93157609676837e-05, "loss": 4.7673, "step": 1517 }, { "epoch": 0.01370283444665102, "grad_norm": 4.221133232116699, "learning_rate": 4.931530962267557e-05, "loss": 4.8551, "step": 1518 }, { "epoch": 0.013711861346813504, "grad_norm": 3.689579725265503, "learning_rate": 4.931485827766745e-05, "loss": 4.4425, "step": 1519 }, { "epoch": 0.013720888246975989, "grad_norm": 4.628645896911621, "learning_rate": 4.931440693265933e-05, "loss": 5.2852, "step": 1520 }, { "epoch": 0.013729915147138472, "grad_norm": 4.517068386077881, "learning_rate": 4.93139555876512e-05, "loss": 4.8514, "step": 1521 }, { "epoch": 0.013738942047300957, "grad_norm": 3.7355153560638428, "learning_rate": 4.9313504242643076e-05, "loss": 4.7305, "step": 1522 }, { "epoch": 0.01374796894746344, "grad_norm": 3.854743242263794, "learning_rate": 4.9313052897634956e-05, "loss": 4.8651, "step": 1523 }, { "epoch": 0.013756995847625925, "grad_norm": 3.5922584533691406, "learning_rate": 4.931260155262683e-05, "loss": 5.2146, "step": 1524 }, { "epoch": 0.013766022747788409, "grad_norm": 6.935342311859131, "learning_rate": 4.931215020761871e-05, "loss": 5.0945, "step": 1525 }, { "epoch": 0.013775049647950894, "grad_norm": 3.827483654022217, "learning_rate": 4.9311698862610587e-05, "loss": 4.7581, "step": 1526 }, { "epoch": 0.013784076548113379, "grad_norm": 4.040938854217529, "learning_rate": 4.931124751760246e-05, "loss": 5.2643, "step": 1527 }, { "epoch": 0.013793103448275862, "grad_norm": 3.8043830394744873, "learning_rate": 4.931079617259433e-05, "loss": 4.6844, "step": 1528 }, { "epoch": 0.013802130348438347, "grad_norm": 3.9452526569366455, "learning_rate": 4.931034482758621e-05, "loss": 5.2445, "step": 1529 }, { "epoch": 0.01381115724860083, "grad_norm": 4.238149642944336, "learning_rate": 4.930989348257808e-05, "loss": 4.984, "step": 1530 }, { "epoch": 0.013820184148763315, "grad_norm": 3.414966344833374, "learning_rate": 4.930944213756996e-05, "loss": 4.6005, "step": 1531 }, { "epoch": 0.013829211048925798, "grad_norm": 4.071689605712891, "learning_rate": 4.9308990792561835e-05, "loss": 3.9859, "step": 1532 }, { "epoch": 0.013838237949088283, "grad_norm": 3.5656449794769287, "learning_rate": 4.9308539447553714e-05, "loss": 4.8292, "step": 1533 }, { "epoch": 0.013847264849250767, "grad_norm": 5.966017723083496, "learning_rate": 4.9308088102545593e-05, "loss": 3.9653, "step": 1534 }, { "epoch": 0.013856291749413252, "grad_norm": 4.572762966156006, "learning_rate": 4.930763675753746e-05, "loss": 4.9438, "step": 1535 }, { "epoch": 0.013865318649575735, "grad_norm": 4.789292335510254, "learning_rate": 4.930718541252934e-05, "loss": 4.1961, "step": 1536 }, { "epoch": 0.01387434554973822, "grad_norm": 3.8886280059814453, "learning_rate": 4.930673406752122e-05, "loss": 5.2617, "step": 1537 }, { "epoch": 0.013883372449900705, "grad_norm": 3.4717330932617188, "learning_rate": 4.930628272251309e-05, "loss": 4.5262, "step": 1538 }, { "epoch": 0.013892399350063188, "grad_norm": 3.178502082824707, "learning_rate": 4.930583137750497e-05, "loss": 4.9128, "step": 1539 }, { "epoch": 0.013901426250225673, "grad_norm": 6.669200420379639, "learning_rate": 4.930538003249684e-05, "loss": 5.1456, "step": 1540 }, { "epoch": 0.013910453150388156, "grad_norm": 4.04219913482666, "learning_rate": 4.9304928687488714e-05, "loss": 5.3104, "step": 1541 }, { "epoch": 0.013919480050550641, "grad_norm": 5.233236789703369, "learning_rate": 4.9304477342480593e-05, "loss": 4.0333, "step": 1542 }, { "epoch": 0.013928506950713125, "grad_norm": 3.967617988586426, "learning_rate": 4.930402599747247e-05, "loss": 4.2492, "step": 1543 }, { "epoch": 0.01393753385087561, "grad_norm": 4.899129867553711, "learning_rate": 4.9303574652464345e-05, "loss": 4.8181, "step": 1544 }, { "epoch": 0.013946560751038093, "grad_norm": 4.174698352813721, "learning_rate": 4.9303123307456224e-05, "loss": 5.1043, "step": 1545 }, { "epoch": 0.013955587651200578, "grad_norm": 5.905158996582031, "learning_rate": 4.93026719624481e-05, "loss": 4.7808, "step": 1546 }, { "epoch": 0.013964614551363061, "grad_norm": 4.980924606323242, "learning_rate": 4.9302220617439976e-05, "loss": 4.9455, "step": 1547 }, { "epoch": 0.013973641451525546, "grad_norm": 2.9616074562072754, "learning_rate": 4.930176927243185e-05, "loss": 5.0982, "step": 1548 }, { "epoch": 0.01398266835168803, "grad_norm": 4.049345970153809, "learning_rate": 4.930131792742372e-05, "loss": 4.9243, "step": 1549 }, { "epoch": 0.013991695251850515, "grad_norm": 3.951158046722412, "learning_rate": 4.93008665824156e-05, "loss": 4.5996, "step": 1550 }, { "epoch": 0.014000722152013, "grad_norm": 3.7015981674194336, "learning_rate": 4.930041523740748e-05, "loss": 5.0885, "step": 1551 }, { "epoch": 0.014009749052175483, "grad_norm": 4.217982292175293, "learning_rate": 4.929996389239935e-05, "loss": 4.7562, "step": 1552 }, { "epoch": 0.014018775952337968, "grad_norm": 4.437270641326904, "learning_rate": 4.929951254739123e-05, "loss": 5.3207, "step": 1553 }, { "epoch": 0.014027802852500451, "grad_norm": 6.1495041847229, "learning_rate": 4.9299061202383104e-05, "loss": 4.8811, "step": 1554 }, { "epoch": 0.014036829752662936, "grad_norm": 3.4800007343292236, "learning_rate": 4.9298609857374976e-05, "loss": 5.0159, "step": 1555 }, { "epoch": 0.01404585665282542, "grad_norm": 5.917176723480225, "learning_rate": 4.9298158512366855e-05, "loss": 3.8093, "step": 1556 }, { "epoch": 0.014054883552987904, "grad_norm": 4.0104546546936035, "learning_rate": 4.929770716735873e-05, "loss": 4.9763, "step": 1557 }, { "epoch": 0.014063910453150388, "grad_norm": 3.6930887699127197, "learning_rate": 4.929725582235061e-05, "loss": 4.5264, "step": 1558 }, { "epoch": 0.014072937353312873, "grad_norm": 7.369104385375977, "learning_rate": 4.9296804477342486e-05, "loss": 5.3171, "step": 1559 }, { "epoch": 0.014081964253475356, "grad_norm": 4.864550590515137, "learning_rate": 4.929635313233436e-05, "loss": 4.7357, "step": 1560 }, { "epoch": 0.014090991153637841, "grad_norm": 3.7003393173217773, "learning_rate": 4.929590178732623e-05, "loss": 5.0629, "step": 1561 }, { "epoch": 0.014100018053800324, "grad_norm": 3.7286269664764404, "learning_rate": 4.929545044231811e-05, "loss": 4.6867, "step": 1562 }, { "epoch": 0.01410904495396281, "grad_norm": 5.935271263122559, "learning_rate": 4.929499909730998e-05, "loss": 4.6568, "step": 1563 }, { "epoch": 0.014118071854125294, "grad_norm": 3.0559022426605225, "learning_rate": 4.929454775230186e-05, "loss": 5.0752, "step": 1564 }, { "epoch": 0.014127098754287777, "grad_norm": 5.41775369644165, "learning_rate": 4.929409640729374e-05, "loss": 5.054, "step": 1565 }, { "epoch": 0.014136125654450262, "grad_norm": 4.409483909606934, "learning_rate": 4.9293645062285614e-05, "loss": 5.3353, "step": 1566 }, { "epoch": 0.014145152554612746, "grad_norm": 4.399726867675781, "learning_rate": 4.929319371727749e-05, "loss": 5.1904, "step": 1567 }, { "epoch": 0.01415417945477523, "grad_norm": 3.3765032291412354, "learning_rate": 4.9292742372269366e-05, "loss": 4.5178, "step": 1568 }, { "epoch": 0.014163206354937714, "grad_norm": 3.3873281478881836, "learning_rate": 4.929229102726124e-05, "loss": 5.1146, "step": 1569 }, { "epoch": 0.014172233255100199, "grad_norm": 3.480980634689331, "learning_rate": 4.929183968225312e-05, "loss": 5.3715, "step": 1570 }, { "epoch": 0.014181260155262682, "grad_norm": 5.082538604736328, "learning_rate": 4.929138833724499e-05, "loss": 5.0719, "step": 1571 }, { "epoch": 0.014190287055425167, "grad_norm": 4.2504472732543945, "learning_rate": 4.929093699223687e-05, "loss": 4.5478, "step": 1572 }, { "epoch": 0.01419931395558765, "grad_norm": 3.255114793777466, "learning_rate": 4.929048564722875e-05, "loss": 4.4482, "step": 1573 }, { "epoch": 0.014208340855750136, "grad_norm": 4.348743438720703, "learning_rate": 4.929003430222062e-05, "loss": 5.6543, "step": 1574 }, { "epoch": 0.014217367755912619, "grad_norm": 4.074105739593506, "learning_rate": 4.928958295721249e-05, "loss": 4.4231, "step": 1575 }, { "epoch": 0.014226394656075104, "grad_norm": 3.2648065090179443, "learning_rate": 4.928913161220437e-05, "loss": 5.5942, "step": 1576 }, { "epoch": 0.014235421556237589, "grad_norm": 3.710505723953247, "learning_rate": 4.9288680267196245e-05, "loss": 4.7494, "step": 1577 }, { "epoch": 0.014244448456400072, "grad_norm": 3.7545828819274902, "learning_rate": 4.9288228922188124e-05, "loss": 5.1895, "step": 1578 }, { "epoch": 0.014253475356562557, "grad_norm": 3.851830005645752, "learning_rate": 4.928777757718e-05, "loss": 4.4873, "step": 1579 }, { "epoch": 0.01426250225672504, "grad_norm": 4.267833709716797, "learning_rate": 4.9287326232171876e-05, "loss": 4.5891, "step": 1580 }, { "epoch": 0.014271529156887525, "grad_norm": 4.191375255584717, "learning_rate": 4.928687488716375e-05, "loss": 4.9379, "step": 1581 }, { "epoch": 0.014280556057050009, "grad_norm": 6.440689563751221, "learning_rate": 4.928642354215563e-05, "loss": 4.7171, "step": 1582 }, { "epoch": 0.014289582957212494, "grad_norm": 4.3634772300720215, "learning_rate": 4.92859721971475e-05, "loss": 5.2938, "step": 1583 }, { "epoch": 0.014298609857374977, "grad_norm": 3.445173740386963, "learning_rate": 4.928552085213938e-05, "loss": 4.1932, "step": 1584 }, { "epoch": 0.014307636757537462, "grad_norm": 3.8372693061828613, "learning_rate": 4.928506950713125e-05, "loss": 5.1586, "step": 1585 }, { "epoch": 0.014316663657699945, "grad_norm": 3.6476683616638184, "learning_rate": 4.928461816212313e-05, "loss": 4.6563, "step": 1586 }, { "epoch": 0.01432569055786243, "grad_norm": 3.5369181632995605, "learning_rate": 4.928416681711501e-05, "loss": 5.1539, "step": 1587 }, { "epoch": 0.014334717458024913, "grad_norm": 5.079349517822266, "learning_rate": 4.9283715472106876e-05, "loss": 5.1644, "step": 1588 }, { "epoch": 0.014343744358187398, "grad_norm": 6.081517696380615, "learning_rate": 4.9283264127098755e-05, "loss": 5.6564, "step": 1589 }, { "epoch": 0.014352771258349883, "grad_norm": 3.755855083465576, "learning_rate": 4.9282812782090634e-05, "loss": 4.9103, "step": 1590 }, { "epoch": 0.014361798158512367, "grad_norm": 7.585200309753418, "learning_rate": 4.928236143708251e-05, "loss": 4.9409, "step": 1591 }, { "epoch": 0.014370825058674852, "grad_norm": 5.112760066986084, "learning_rate": 4.9281910092074386e-05, "loss": 3.2529, "step": 1592 }, { "epoch": 0.014379851958837335, "grad_norm": 6.983741760253906, "learning_rate": 4.928145874706626e-05, "loss": 5.1308, "step": 1593 }, { "epoch": 0.01438887885899982, "grad_norm": 3.6358208656311035, "learning_rate": 4.928100740205814e-05, "loss": 4.6955, "step": 1594 }, { "epoch": 0.014397905759162303, "grad_norm": 4.744623184204102, "learning_rate": 4.928055605705001e-05, "loss": 4.5754, "step": 1595 }, { "epoch": 0.014406932659324788, "grad_norm": 3.8720030784606934, "learning_rate": 4.928010471204188e-05, "loss": 4.8142, "step": 1596 }, { "epoch": 0.014415959559487272, "grad_norm": 4.955869197845459, "learning_rate": 4.927965336703376e-05, "loss": 5.0968, "step": 1597 }, { "epoch": 0.014424986459649757, "grad_norm": 2.7726728916168213, "learning_rate": 4.927920202202564e-05, "loss": 5.2859, "step": 1598 }, { "epoch": 0.01443401335981224, "grad_norm": 3.3350300788879395, "learning_rate": 4.9278750677017514e-05, "loss": 4.917, "step": 1599 }, { "epoch": 0.014443040259974725, "grad_norm": 3.813864231109619, "learning_rate": 4.927829933200939e-05, "loss": 5.2683, "step": 1600 }, { "epoch": 0.014452067160137208, "grad_norm": 3.881674289703369, "learning_rate": 4.927784798700127e-05, "loss": 4.7843, "step": 1601 }, { "epoch": 0.014461094060299693, "grad_norm": 3.437152624130249, "learning_rate": 4.927739664199314e-05, "loss": 5.2218, "step": 1602 }, { "epoch": 0.014470120960462178, "grad_norm": 4.544203281402588, "learning_rate": 4.927694529698502e-05, "loss": 5.114, "step": 1603 }, { "epoch": 0.014479147860624661, "grad_norm": 4.115044116973877, "learning_rate": 4.9276493951976896e-05, "loss": 4.9577, "step": 1604 }, { "epoch": 0.014488174760787146, "grad_norm": 4.95378303527832, "learning_rate": 4.927604260696877e-05, "loss": 4.2244, "step": 1605 }, { "epoch": 0.01449720166094963, "grad_norm": 3.5381393432617188, "learning_rate": 4.927559126196065e-05, "loss": 5.0955, "step": 1606 }, { "epoch": 0.014506228561112115, "grad_norm": 5.649842262268066, "learning_rate": 4.927513991695252e-05, "loss": 4.3767, "step": 1607 }, { "epoch": 0.014515255461274598, "grad_norm": 3.55305814743042, "learning_rate": 4.927468857194439e-05, "loss": 4.7349, "step": 1608 }, { "epoch": 0.014524282361437083, "grad_norm": 7.78743839263916, "learning_rate": 4.927423722693627e-05, "loss": 4.7609, "step": 1609 }, { "epoch": 0.014533309261599566, "grad_norm": 3.8738129138946533, "learning_rate": 4.9273785881928145e-05, "loss": 4.6744, "step": 1610 }, { "epoch": 0.014542336161762051, "grad_norm": 3.4039740562438965, "learning_rate": 4.9273334536920024e-05, "loss": 4.7081, "step": 1611 }, { "epoch": 0.014551363061924534, "grad_norm": 3.5395667552948, "learning_rate": 4.92728831919119e-05, "loss": 4.5242, "step": 1612 }, { "epoch": 0.01456038996208702, "grad_norm": 4.0613203048706055, "learning_rate": 4.9272431846903776e-05, "loss": 5.1817, "step": 1613 }, { "epoch": 0.014569416862249503, "grad_norm": 7.7040534019470215, "learning_rate": 4.9271980501895655e-05, "loss": 4.032, "step": 1614 }, { "epoch": 0.014578443762411988, "grad_norm": 3.3153302669525146, "learning_rate": 4.927152915688753e-05, "loss": 4.906, "step": 1615 }, { "epoch": 0.014587470662574473, "grad_norm": 4.018549919128418, "learning_rate": 4.92710778118794e-05, "loss": 4.5674, "step": 1616 }, { "epoch": 0.014596497562736956, "grad_norm": 3.3511099815368652, "learning_rate": 4.927062646687128e-05, "loss": 4.93, "step": 1617 }, { "epoch": 0.014605524462899441, "grad_norm": 4.019015312194824, "learning_rate": 4.927017512186315e-05, "loss": 4.3467, "step": 1618 }, { "epoch": 0.014614551363061924, "grad_norm": 5.960495471954346, "learning_rate": 4.926972377685503e-05, "loss": 4.256, "step": 1619 }, { "epoch": 0.01462357826322441, "grad_norm": 3.8855462074279785, "learning_rate": 4.926927243184691e-05, "loss": 5.1287, "step": 1620 }, { "epoch": 0.014632605163386892, "grad_norm": 3.4594225883483887, "learning_rate": 4.926882108683878e-05, "loss": 5.7795, "step": 1621 }, { "epoch": 0.014641632063549377, "grad_norm": 3.673543930053711, "learning_rate": 4.9268369741830655e-05, "loss": 4.5799, "step": 1622 }, { "epoch": 0.01465065896371186, "grad_norm": 3.1146504878997803, "learning_rate": 4.9267918396822534e-05, "loss": 4.4654, "step": 1623 }, { "epoch": 0.014659685863874346, "grad_norm": 7.148996829986572, "learning_rate": 4.926746705181441e-05, "loss": 5.0238, "step": 1624 }, { "epoch": 0.014668712764036829, "grad_norm": 5.0376152992248535, "learning_rate": 4.9267015706806286e-05, "loss": 5.1154, "step": 1625 }, { "epoch": 0.014677739664199314, "grad_norm": 3.201141119003296, "learning_rate": 4.9266564361798165e-05, "loss": 4.5175, "step": 1626 }, { "epoch": 0.014686766564361799, "grad_norm": 3.8499948978424072, "learning_rate": 4.926611301679004e-05, "loss": 4.9532, "step": 1627 }, { "epoch": 0.014695793464524282, "grad_norm": 5.581600666046143, "learning_rate": 4.926566167178191e-05, "loss": 4.6603, "step": 1628 }, { "epoch": 0.014704820364686767, "grad_norm": 3.88535737991333, "learning_rate": 4.926521032677379e-05, "loss": 4.9257, "step": 1629 }, { "epoch": 0.01471384726484925, "grad_norm": 3.813821792602539, "learning_rate": 4.926475898176566e-05, "loss": 5.2824, "step": 1630 }, { "epoch": 0.014722874165011736, "grad_norm": 3.898993730545044, "learning_rate": 4.926430763675754e-05, "loss": 5.1395, "step": 1631 }, { "epoch": 0.014731901065174219, "grad_norm": 3.5987181663513184, "learning_rate": 4.9263856291749413e-05, "loss": 4.9956, "step": 1632 }, { "epoch": 0.014740927965336704, "grad_norm": 2.9940884113311768, "learning_rate": 4.926340494674129e-05, "loss": 4.9036, "step": 1633 }, { "epoch": 0.014749954865499187, "grad_norm": 3.8521225452423096, "learning_rate": 4.926295360173317e-05, "loss": 4.6611, "step": 1634 }, { "epoch": 0.014758981765661672, "grad_norm": 4.068364143371582, "learning_rate": 4.926250225672504e-05, "loss": 4.8143, "step": 1635 }, { "epoch": 0.014768008665824155, "grad_norm": 3.8699405193328857, "learning_rate": 4.926205091171692e-05, "loss": 5.019, "step": 1636 }, { "epoch": 0.01477703556598664, "grad_norm": 2.7434680461883545, "learning_rate": 4.9261599566708796e-05, "loss": 5.0919, "step": 1637 }, { "epoch": 0.014786062466149124, "grad_norm": 5.184078216552734, "learning_rate": 4.926114822170067e-05, "loss": 4.5066, "step": 1638 }, { "epoch": 0.014795089366311609, "grad_norm": 3.674030065536499, "learning_rate": 4.926069687669255e-05, "loss": 4.6378, "step": 1639 }, { "epoch": 0.014804116266474094, "grad_norm": 3.379956007003784, "learning_rate": 4.926024553168442e-05, "loss": 4.0626, "step": 1640 }, { "epoch": 0.014813143166636577, "grad_norm": 3.147420883178711, "learning_rate": 4.92597941866763e-05, "loss": 4.8649, "step": 1641 }, { "epoch": 0.014822170066799062, "grad_norm": 4.269362926483154, "learning_rate": 4.925934284166817e-05, "loss": 4.0058, "step": 1642 }, { "epoch": 0.014831196966961545, "grad_norm": 4.430767059326172, "learning_rate": 4.925889149666005e-05, "loss": 4.5736, "step": 1643 }, { "epoch": 0.01484022386712403, "grad_norm": 4.566793441772461, "learning_rate": 4.9258440151651924e-05, "loss": 4.8103, "step": 1644 }, { "epoch": 0.014849250767286513, "grad_norm": 2.680988311767578, "learning_rate": 4.92579888066438e-05, "loss": 4.8929, "step": 1645 }, { "epoch": 0.014858277667448998, "grad_norm": 3.974734306335449, "learning_rate": 4.9257537461635675e-05, "loss": 5.1019, "step": 1646 }, { "epoch": 0.014867304567611482, "grad_norm": 2.734255790710449, "learning_rate": 4.9257086116627555e-05, "loss": 5.2122, "step": 1647 }, { "epoch": 0.014876331467773967, "grad_norm": 4.833951473236084, "learning_rate": 4.9256634771619434e-05, "loss": 4.9909, "step": 1648 }, { "epoch": 0.01488535836793645, "grad_norm": 4.384193420410156, "learning_rate": 4.92561834266113e-05, "loss": 5.2962, "step": 1649 }, { "epoch": 0.014894385268098935, "grad_norm": 3.840407133102417, "learning_rate": 4.925573208160318e-05, "loss": 4.797, "step": 1650 }, { "epoch": 0.014903412168261418, "grad_norm": 4.044092655181885, "learning_rate": 4.925528073659506e-05, "loss": 4.6705, "step": 1651 }, { "epoch": 0.014912439068423903, "grad_norm": 3.5107204914093018, "learning_rate": 4.925482939158693e-05, "loss": 4.9588, "step": 1652 }, { "epoch": 0.014921465968586388, "grad_norm": 2.6928012371063232, "learning_rate": 4.925437804657881e-05, "loss": 5.3396, "step": 1653 }, { "epoch": 0.014930492868748872, "grad_norm": 3.561983346939087, "learning_rate": 4.925392670157068e-05, "loss": 4.9272, "step": 1654 }, { "epoch": 0.014939519768911357, "grad_norm": 4.0832133293151855, "learning_rate": 4.9253475356562555e-05, "loss": 4.9494, "step": 1655 }, { "epoch": 0.01494854666907384, "grad_norm": 3.683666229248047, "learning_rate": 4.9253024011554434e-05, "loss": 4.864, "step": 1656 }, { "epoch": 0.014957573569236325, "grad_norm": 2.932175636291504, "learning_rate": 4.9252572666546306e-05, "loss": 4.7386, "step": 1657 }, { "epoch": 0.014966600469398808, "grad_norm": 5.8611741065979, "learning_rate": 4.9252121321538186e-05, "loss": 5.287, "step": 1658 }, { "epoch": 0.014975627369561293, "grad_norm": 4.026108264923096, "learning_rate": 4.9251669976530065e-05, "loss": 5.4458, "step": 1659 }, { "epoch": 0.014984654269723776, "grad_norm": 3.306638479232788, "learning_rate": 4.925121863152194e-05, "loss": 5.4448, "step": 1660 }, { "epoch": 0.014993681169886261, "grad_norm": 4.207057476043701, "learning_rate": 4.9250767286513817e-05, "loss": 4.635, "step": 1661 }, { "epoch": 0.015002708070048745, "grad_norm": 3.766463041305542, "learning_rate": 4.925031594150569e-05, "loss": 5.0507, "step": 1662 }, { "epoch": 0.01501173497021123, "grad_norm": 5.948224067687988, "learning_rate": 4.924986459649756e-05, "loss": 5.1186, "step": 1663 }, { "epoch": 0.015020761870373713, "grad_norm": 3.766111135482788, "learning_rate": 4.924941325148944e-05, "loss": 4.7963, "step": 1664 }, { "epoch": 0.015029788770536198, "grad_norm": 4.184983730316162, "learning_rate": 4.924896190648132e-05, "loss": 5.2557, "step": 1665 }, { "epoch": 0.015038815670698683, "grad_norm": 3.639707326889038, "learning_rate": 4.924851056147319e-05, "loss": 4.2376, "step": 1666 }, { "epoch": 0.015047842570861166, "grad_norm": 3.5733702182769775, "learning_rate": 4.924805921646507e-05, "loss": 5.3367, "step": 1667 }, { "epoch": 0.015056869471023651, "grad_norm": 3.942835569381714, "learning_rate": 4.9247607871456944e-05, "loss": 4.7414, "step": 1668 }, { "epoch": 0.015065896371186134, "grad_norm": 5.041374206542969, "learning_rate": 4.9247156526448817e-05, "loss": 4.4282, "step": 1669 }, { "epoch": 0.01507492327134862, "grad_norm": 4.18682861328125, "learning_rate": 4.9246705181440696e-05, "loss": 4.9848, "step": 1670 }, { "epoch": 0.015083950171511103, "grad_norm": 4.242360591888428, "learning_rate": 4.924625383643257e-05, "loss": 3.8649, "step": 1671 }, { "epoch": 0.015092977071673588, "grad_norm": 3.0979795455932617, "learning_rate": 4.924580249142445e-05, "loss": 5.2381, "step": 1672 }, { "epoch": 0.015102003971836071, "grad_norm": 4.929898262023926, "learning_rate": 4.924535114641633e-05, "loss": 5.2141, "step": 1673 }, { "epoch": 0.015111030871998556, "grad_norm": 5.436182498931885, "learning_rate": 4.92448998014082e-05, "loss": 5.194, "step": 1674 }, { "epoch": 0.01512005777216104, "grad_norm": 3.2264721393585205, "learning_rate": 4.924444845640007e-05, "loss": 4.7702, "step": 1675 }, { "epoch": 0.015129084672323524, "grad_norm": 3.8467299938201904, "learning_rate": 4.924399711139195e-05, "loss": 4.9522, "step": 1676 }, { "epoch": 0.015138111572486007, "grad_norm": 4.129514217376709, "learning_rate": 4.9243545766383823e-05, "loss": 5.1554, "step": 1677 }, { "epoch": 0.015147138472648492, "grad_norm": 4.0183491706848145, "learning_rate": 4.92430944213757e-05, "loss": 5.1329, "step": 1678 }, { "epoch": 0.015156165372810977, "grad_norm": 5.047775745391846, "learning_rate": 4.9242643076367575e-05, "loss": 4.9014, "step": 1679 }, { "epoch": 0.01516519227297346, "grad_norm": 4.685017108917236, "learning_rate": 4.9242191731359454e-05, "loss": 5.4508, "step": 1680 }, { "epoch": 0.015174219173135946, "grad_norm": 4.583311080932617, "learning_rate": 4.9241740386351334e-05, "loss": 3.9415, "step": 1681 }, { "epoch": 0.015183246073298429, "grad_norm": 4.855068206787109, "learning_rate": 4.92412890413432e-05, "loss": 4.7018, "step": 1682 }, { "epoch": 0.015192272973460914, "grad_norm": 3.093632698059082, "learning_rate": 4.924083769633508e-05, "loss": 5.262, "step": 1683 }, { "epoch": 0.015201299873623397, "grad_norm": 6.092831134796143, "learning_rate": 4.924038635132696e-05, "loss": 5.0488, "step": 1684 }, { "epoch": 0.015210326773785882, "grad_norm": 3.7471466064453125, "learning_rate": 4.923993500631883e-05, "loss": 5.2965, "step": 1685 }, { "epoch": 0.015219353673948366, "grad_norm": 3.584824562072754, "learning_rate": 4.923948366131071e-05, "loss": 4.6917, "step": 1686 }, { "epoch": 0.01522838057411085, "grad_norm": 11.129680633544922, "learning_rate": 4.923903231630259e-05, "loss": 5.0166, "step": 1687 }, { "epoch": 0.015237407474273334, "grad_norm": 3.6155030727386475, "learning_rate": 4.923858097129446e-05, "loss": 5.3051, "step": 1688 }, { "epoch": 0.015246434374435819, "grad_norm": 3.3137543201446533, "learning_rate": 4.9238129626286334e-05, "loss": 4.8403, "step": 1689 }, { "epoch": 0.015255461274598302, "grad_norm": 3.683126449584961, "learning_rate": 4.923767828127821e-05, "loss": 4.7344, "step": 1690 }, { "epoch": 0.015264488174760787, "grad_norm": 4.637896537780762, "learning_rate": 4.9237226936270085e-05, "loss": 4.9031, "step": 1691 }, { "epoch": 0.015273515074923272, "grad_norm": 3.9292538166046143, "learning_rate": 4.9236775591261965e-05, "loss": 4.2757, "step": 1692 }, { "epoch": 0.015282541975085755, "grad_norm": 4.4167914390563965, "learning_rate": 4.923632424625384e-05, "loss": 5.0135, "step": 1693 }, { "epoch": 0.01529156887524824, "grad_norm": 3.790126323699951, "learning_rate": 4.9235872901245716e-05, "loss": 4.3076, "step": 1694 }, { "epoch": 0.015300595775410724, "grad_norm": 4.208259105682373, "learning_rate": 4.9235421556237596e-05, "loss": 4.388, "step": 1695 }, { "epoch": 0.015309622675573209, "grad_norm": 3.8210840225219727, "learning_rate": 4.923497021122946e-05, "loss": 5.0247, "step": 1696 }, { "epoch": 0.015318649575735692, "grad_norm": 4.279548168182373, "learning_rate": 4.923451886622134e-05, "loss": 5.0809, "step": 1697 }, { "epoch": 0.015327676475898177, "grad_norm": 13.53091812133789, "learning_rate": 4.923406752121322e-05, "loss": 4.751, "step": 1698 }, { "epoch": 0.01533670337606066, "grad_norm": 4.288211345672607, "learning_rate": 4.923361617620509e-05, "loss": 5.395, "step": 1699 }, { "epoch": 0.015345730276223145, "grad_norm": 4.257620811462402, "learning_rate": 4.923316483119697e-05, "loss": 4.3405, "step": 1700 }, { "epoch": 0.015354757176385628, "grad_norm": 2.5632340908050537, "learning_rate": 4.9232713486188844e-05, "loss": 4.5974, "step": 1701 }, { "epoch": 0.015363784076548113, "grad_norm": 3.3721678256988525, "learning_rate": 4.9232262141180716e-05, "loss": 3.8514, "step": 1702 }, { "epoch": 0.015372810976710597, "grad_norm": 4.499845504760742, "learning_rate": 4.9231810796172596e-05, "loss": 5.2953, "step": 1703 }, { "epoch": 0.015381837876873082, "grad_norm": 4.232858180999756, "learning_rate": 4.9231359451164475e-05, "loss": 4.9419, "step": 1704 }, { "epoch": 0.015390864777035567, "grad_norm": 3.464974880218506, "learning_rate": 4.923090810615635e-05, "loss": 5.2918, "step": 1705 }, { "epoch": 0.01539989167719805, "grad_norm": 3.4812703132629395, "learning_rate": 4.9230456761148227e-05, "loss": 4.1683, "step": 1706 }, { "epoch": 0.015408918577360535, "grad_norm": 4.095772743225098, "learning_rate": 4.92300054161401e-05, "loss": 4.7672, "step": 1707 }, { "epoch": 0.015417945477523018, "grad_norm": 5.528397560119629, "learning_rate": 4.922955407113198e-05, "loss": 5.0492, "step": 1708 }, { "epoch": 0.015426972377685503, "grad_norm": 4.037591457366943, "learning_rate": 4.922910272612385e-05, "loss": 4.4834, "step": 1709 }, { "epoch": 0.015435999277847987, "grad_norm": 3.574317693710327, "learning_rate": 4.922865138111572e-05, "loss": 4.5976, "step": 1710 }, { "epoch": 0.015445026178010472, "grad_norm": 3.678480386734009, "learning_rate": 4.92282000361076e-05, "loss": 4.1969, "step": 1711 }, { "epoch": 0.015454053078172955, "grad_norm": 3.786205291748047, "learning_rate": 4.922774869109948e-05, "loss": 4.7833, "step": 1712 }, { "epoch": 0.01546307997833544, "grad_norm": 3.828003406524658, "learning_rate": 4.9227297346091354e-05, "loss": 4.8009, "step": 1713 }, { "epoch": 0.015472106878497923, "grad_norm": 4.009328842163086, "learning_rate": 4.922684600108323e-05, "loss": 4.9739, "step": 1714 }, { "epoch": 0.015481133778660408, "grad_norm": 4.4761481285095215, "learning_rate": 4.9226394656075106e-05, "loss": 5.0125, "step": 1715 }, { "epoch": 0.015490160678822893, "grad_norm": 3.513389825820923, "learning_rate": 4.922594331106698e-05, "loss": 4.8983, "step": 1716 }, { "epoch": 0.015499187578985376, "grad_norm": 5.144423961639404, "learning_rate": 4.922549196605886e-05, "loss": 4.7339, "step": 1717 }, { "epoch": 0.015508214479147861, "grad_norm": 4.275842666625977, "learning_rate": 4.922504062105073e-05, "loss": 4.7993, "step": 1718 }, { "epoch": 0.015517241379310345, "grad_norm": 2.60394024848938, "learning_rate": 4.922458927604261e-05, "loss": 4.8392, "step": 1719 }, { "epoch": 0.01552626827947283, "grad_norm": 4.7242655754089355, "learning_rate": 4.922413793103449e-05, "loss": 4.8125, "step": 1720 }, { "epoch": 0.015535295179635313, "grad_norm": 3.886718273162842, "learning_rate": 4.922368658602636e-05, "loss": 5.2376, "step": 1721 }, { "epoch": 0.015544322079797798, "grad_norm": 4.214581489562988, "learning_rate": 4.9223235241018233e-05, "loss": 4.213, "step": 1722 }, { "epoch": 0.015553348979960281, "grad_norm": 3.401792049407959, "learning_rate": 4.922278389601011e-05, "loss": 4.7715, "step": 1723 }, { "epoch": 0.015562375880122766, "grad_norm": 6.901301383972168, "learning_rate": 4.9222332551001985e-05, "loss": 5.111, "step": 1724 }, { "epoch": 0.01557140278028525, "grad_norm": 4.5419487953186035, "learning_rate": 4.9221881205993864e-05, "loss": 5.2462, "step": 1725 }, { "epoch": 0.015580429680447734, "grad_norm": 3.4458465576171875, "learning_rate": 4.9221429860985744e-05, "loss": 4.8128, "step": 1726 }, { "epoch": 0.015589456580610218, "grad_norm": 3.8580029010772705, "learning_rate": 4.9220978515977616e-05, "loss": 4.5692, "step": 1727 }, { "epoch": 0.015598483480772703, "grad_norm": 3.998676300048828, "learning_rate": 4.9220527170969495e-05, "loss": 4.9847, "step": 1728 }, { "epoch": 0.015607510380935188, "grad_norm": 4.647660255432129, "learning_rate": 4.922007582596137e-05, "loss": 4.7459, "step": 1729 }, { "epoch": 0.015616537281097671, "grad_norm": 6.4043121337890625, "learning_rate": 4.921962448095324e-05, "loss": 5.1447, "step": 1730 }, { "epoch": 0.015625564181260154, "grad_norm": 3.369075059890747, "learning_rate": 4.921917313594512e-05, "loss": 4.897, "step": 1731 }, { "epoch": 0.01563459108142264, "grad_norm": 3.3253207206726074, "learning_rate": 4.921872179093699e-05, "loss": 5.1384, "step": 1732 }, { "epoch": 0.015643617981585124, "grad_norm": 4.379225254058838, "learning_rate": 4.921827044592887e-05, "loss": 4.5441, "step": 1733 }, { "epoch": 0.01565264488174761, "grad_norm": 4.339646339416504, "learning_rate": 4.921781910092075e-05, "loss": 4.4519, "step": 1734 }, { "epoch": 0.01566167178191009, "grad_norm": 3.6397690773010254, "learning_rate": 4.921736775591262e-05, "loss": 4.9879, "step": 1735 }, { "epoch": 0.015670698682072576, "grad_norm": 4.201405048370361, "learning_rate": 4.9216916410904495e-05, "loss": 4.7318, "step": 1736 }, { "epoch": 0.01567972558223506, "grad_norm": 5.681923866271973, "learning_rate": 4.9216465065896375e-05, "loss": 4.3231, "step": 1737 }, { "epoch": 0.015688752482397546, "grad_norm": 4.024292469024658, "learning_rate": 4.921601372088825e-05, "loss": 5.0676, "step": 1738 }, { "epoch": 0.015697779382560027, "grad_norm": 3.874340057373047, "learning_rate": 4.9215562375880126e-05, "loss": 5.0604, "step": 1739 }, { "epoch": 0.015706806282722512, "grad_norm": 4.6610846519470215, "learning_rate": 4.9215111030872e-05, "loss": 5.0265, "step": 1740 }, { "epoch": 0.015715833182884997, "grad_norm": 2.89193058013916, "learning_rate": 4.921465968586388e-05, "loss": 4.9758, "step": 1741 }, { "epoch": 0.015724860083047482, "grad_norm": 6.662283420562744, "learning_rate": 4.921420834085576e-05, "loss": 5.1203, "step": 1742 }, { "epoch": 0.015733886983209967, "grad_norm": 4.2216362953186035, "learning_rate": 4.921375699584763e-05, "loss": 4.8702, "step": 1743 }, { "epoch": 0.01574291388337245, "grad_norm": 2.8689661026000977, "learning_rate": 4.92133056508395e-05, "loss": 4.996, "step": 1744 }, { "epoch": 0.015751940783534934, "grad_norm": 3.17086124420166, "learning_rate": 4.921285430583138e-05, "loss": 4.8698, "step": 1745 }, { "epoch": 0.01576096768369742, "grad_norm": 2.8994789123535156, "learning_rate": 4.9212402960823254e-05, "loss": 5.337, "step": 1746 }, { "epoch": 0.015769994583859904, "grad_norm": 4.198799133300781, "learning_rate": 4.921195161581513e-05, "loss": 3.9434, "step": 1747 }, { "epoch": 0.015779021484022385, "grad_norm": 6.421247482299805, "learning_rate": 4.921150027080701e-05, "loss": 4.7108, "step": 1748 }, { "epoch": 0.01578804838418487, "grad_norm": 4.439505577087402, "learning_rate": 4.921104892579888e-05, "loss": 3.9889, "step": 1749 }, { "epoch": 0.015797075284347355, "grad_norm": 4.579911231994629, "learning_rate": 4.921059758079076e-05, "loss": 3.4153, "step": 1750 }, { "epoch": 0.01580610218450984, "grad_norm": 3.8830008506774902, "learning_rate": 4.9210146235782637e-05, "loss": 4.442, "step": 1751 }, { "epoch": 0.015815129084672322, "grad_norm": 3.5273356437683105, "learning_rate": 4.920969489077451e-05, "loss": 4.8328, "step": 1752 }, { "epoch": 0.015824155984834807, "grad_norm": 4.546022891998291, "learning_rate": 4.920924354576639e-05, "loss": 5.3955, "step": 1753 }, { "epoch": 0.015833182884997292, "grad_norm": 3.843698501586914, "learning_rate": 4.920879220075826e-05, "loss": 4.2821, "step": 1754 }, { "epoch": 0.015842209785159777, "grad_norm": 3.834350824356079, "learning_rate": 4.920834085575014e-05, "loss": 5.5976, "step": 1755 }, { "epoch": 0.015851236685322262, "grad_norm": 6.153573036193848, "learning_rate": 4.920788951074201e-05, "loss": 4.7692, "step": 1756 }, { "epoch": 0.015860263585484743, "grad_norm": 6.631134510040283, "learning_rate": 4.9207438165733885e-05, "loss": 4.0821, "step": 1757 }, { "epoch": 0.01586929048564723, "grad_norm": 8.322803497314453, "learning_rate": 4.9206986820725764e-05, "loss": 5.2219, "step": 1758 }, { "epoch": 0.015878317385809713, "grad_norm": 3.9671971797943115, "learning_rate": 4.920653547571764e-05, "loss": 5.0241, "step": 1759 }, { "epoch": 0.0158873442859722, "grad_norm": 4.155241966247559, "learning_rate": 4.9206084130709516e-05, "loss": 4.7903, "step": 1760 }, { "epoch": 0.01589637118613468, "grad_norm": 3.8261988162994385, "learning_rate": 4.9205632785701395e-05, "loss": 4.6414, "step": 1761 }, { "epoch": 0.015905398086297165, "grad_norm": 3.0021281242370605, "learning_rate": 4.9205181440693274e-05, "loss": 4.7694, "step": 1762 }, { "epoch": 0.01591442498645965, "grad_norm": 3.7718594074249268, "learning_rate": 4.920473009568514e-05, "loss": 4.9786, "step": 1763 }, { "epoch": 0.015923451886622135, "grad_norm": 3.1808881759643555, "learning_rate": 4.920427875067702e-05, "loss": 5.0954, "step": 1764 }, { "epoch": 0.015932478786784617, "grad_norm": 5.77583646774292, "learning_rate": 4.92038274056689e-05, "loss": 4.712, "step": 1765 }, { "epoch": 0.0159415056869471, "grad_norm": 3.1177103519439697, "learning_rate": 4.920337606066077e-05, "loss": 4.8345, "step": 1766 }, { "epoch": 0.015950532587109587, "grad_norm": 4.110717296600342, "learning_rate": 4.920292471565265e-05, "loss": 4.7893, "step": 1767 }, { "epoch": 0.01595955948727207, "grad_norm": 3.5338754653930664, "learning_rate": 4.920247337064452e-05, "loss": 4.9638, "step": 1768 }, { "epoch": 0.015968586387434557, "grad_norm": 5.573867321014404, "learning_rate": 4.9202022025636395e-05, "loss": 4.7229, "step": 1769 }, { "epoch": 0.015977613287597038, "grad_norm": 3.9319326877593994, "learning_rate": 4.9201570680628274e-05, "loss": 4.7551, "step": 1770 }, { "epoch": 0.015986640187759523, "grad_norm": 3.318671226501465, "learning_rate": 4.920111933562015e-05, "loss": 5.2046, "step": 1771 }, { "epoch": 0.015995667087922008, "grad_norm": 3.661059617996216, "learning_rate": 4.9200667990612026e-05, "loss": 5.5002, "step": 1772 }, { "epoch": 0.016004693988084493, "grad_norm": 6.432276248931885, "learning_rate": 4.9200216645603905e-05, "loss": 3.6613, "step": 1773 }, { "epoch": 0.016013720888246975, "grad_norm": 4.4228315353393555, "learning_rate": 4.919976530059578e-05, "loss": 5.1078, "step": 1774 }, { "epoch": 0.01602274778840946, "grad_norm": 3.586118698120117, "learning_rate": 4.919931395558766e-05, "loss": 4.547, "step": 1775 }, { "epoch": 0.016031774688571945, "grad_norm": 3.839111566543579, "learning_rate": 4.919886261057953e-05, "loss": 5.0348, "step": 1776 }, { "epoch": 0.01604080158873443, "grad_norm": 3.573751211166382, "learning_rate": 4.91984112655714e-05, "loss": 5.5951, "step": 1777 }, { "epoch": 0.01604982848889691, "grad_norm": 4.026891231536865, "learning_rate": 4.919795992056328e-05, "loss": 4.6836, "step": 1778 }, { "epoch": 0.016058855389059396, "grad_norm": 3.113450527191162, "learning_rate": 4.9197508575555154e-05, "loss": 5.5603, "step": 1779 }, { "epoch": 0.01606788228922188, "grad_norm": 4.477795600891113, "learning_rate": 4.919705723054703e-05, "loss": 4.6114, "step": 1780 }, { "epoch": 0.016076909189384366, "grad_norm": 3.4712507724761963, "learning_rate": 4.919660588553891e-05, "loss": 4.6142, "step": 1781 }, { "epoch": 0.01608593608954685, "grad_norm": 4.059403896331787, "learning_rate": 4.9196154540530785e-05, "loss": 5.2254, "step": 1782 }, { "epoch": 0.016094962989709333, "grad_norm": 4.179435729980469, "learning_rate": 4.919570319552266e-05, "loss": 4.8793, "step": 1783 }, { "epoch": 0.016103989889871818, "grad_norm": 6.2971272468566895, "learning_rate": 4.9195251850514536e-05, "loss": 4.5534, "step": 1784 }, { "epoch": 0.016113016790034303, "grad_norm": 5.168407917022705, "learning_rate": 4.919480050550641e-05, "loss": 4.8358, "step": 1785 }, { "epoch": 0.016122043690196788, "grad_norm": 4.2104620933532715, "learning_rate": 4.919434916049829e-05, "loss": 4.6805, "step": 1786 }, { "epoch": 0.01613107059035927, "grad_norm": 4.908720970153809, "learning_rate": 4.919389781549017e-05, "loss": 4.1478, "step": 1787 }, { "epoch": 0.016140097490521754, "grad_norm": 5.614633083343506, "learning_rate": 4.919344647048204e-05, "loss": 4.6123, "step": 1788 }, { "epoch": 0.01614912439068424, "grad_norm": 3.923227548599243, "learning_rate": 4.919299512547391e-05, "loss": 4.865, "step": 1789 }, { "epoch": 0.016158151290846724, "grad_norm": 3.4145865440368652, "learning_rate": 4.919254378046579e-05, "loss": 4.4474, "step": 1790 }, { "epoch": 0.016167178191009206, "grad_norm": 4.95242166519165, "learning_rate": 4.9192092435457664e-05, "loss": 5.3277, "step": 1791 }, { "epoch": 0.01617620509117169, "grad_norm": 4.301572799682617, "learning_rate": 4.919164109044954e-05, "loss": 5.4773, "step": 1792 }, { "epoch": 0.016185231991334176, "grad_norm": 5.928084373474121, "learning_rate": 4.9191189745441416e-05, "loss": 4.9113, "step": 1793 }, { "epoch": 0.01619425889149666, "grad_norm": 3.706949234008789, "learning_rate": 4.9190738400433295e-05, "loss": 4.7383, "step": 1794 }, { "epoch": 0.016203285791659146, "grad_norm": 4.239721775054932, "learning_rate": 4.9190287055425174e-05, "loss": 4.8527, "step": 1795 }, { "epoch": 0.016212312691821627, "grad_norm": 3.1242446899414062, "learning_rate": 4.918983571041704e-05, "loss": 4.8432, "step": 1796 }, { "epoch": 0.016221339591984112, "grad_norm": 4.499575138092041, "learning_rate": 4.918938436540892e-05, "loss": 4.7705, "step": 1797 }, { "epoch": 0.016230366492146597, "grad_norm": 3.6329758167266846, "learning_rate": 4.91889330204008e-05, "loss": 4.5339, "step": 1798 }, { "epoch": 0.016239393392309082, "grad_norm": 5.328938961029053, "learning_rate": 4.918848167539267e-05, "loss": 5.3371, "step": 1799 }, { "epoch": 0.016248420292471564, "grad_norm": 5.400532245635986, "learning_rate": 4.918803033038455e-05, "loss": 4.9491, "step": 1800 }, { "epoch": 0.01625744719263405, "grad_norm": 4.3582682609558105, "learning_rate": 4.918757898537642e-05, "loss": 4.8443, "step": 1801 }, { "epoch": 0.016266474092796534, "grad_norm": 4.584572792053223, "learning_rate": 4.91871276403683e-05, "loss": 5.1058, "step": 1802 }, { "epoch": 0.01627550099295902, "grad_norm": 3.30522084236145, "learning_rate": 4.9186676295360174e-05, "loss": 4.9148, "step": 1803 }, { "epoch": 0.0162845278931215, "grad_norm": 9.719292640686035, "learning_rate": 4.918622495035205e-05, "loss": 5.2926, "step": 1804 }, { "epoch": 0.016293554793283985, "grad_norm": 4.520979404449463, "learning_rate": 4.9185773605343926e-05, "loss": 4.6362, "step": 1805 }, { "epoch": 0.01630258169344647, "grad_norm": 4.979296684265137, "learning_rate": 4.9185322260335805e-05, "loss": 4.7699, "step": 1806 }, { "epoch": 0.016311608593608955, "grad_norm": 4.792953014373779, "learning_rate": 4.918487091532768e-05, "loss": 4.6448, "step": 1807 }, { "epoch": 0.01632063549377144, "grad_norm": 4.750773906707764, "learning_rate": 4.918441957031956e-05, "loss": 4.8482, "step": 1808 }, { "epoch": 0.016329662393933922, "grad_norm": 3.6871888637542725, "learning_rate": 4.9183968225311436e-05, "loss": 4.6505, "step": 1809 }, { "epoch": 0.016338689294096407, "grad_norm": 5.063503265380859, "learning_rate": 4.91835168803033e-05, "loss": 5.1011, "step": 1810 }, { "epoch": 0.016347716194258892, "grad_norm": 4.255102634429932, "learning_rate": 4.918306553529518e-05, "loss": 4.9478, "step": 1811 }, { "epoch": 0.016356743094421377, "grad_norm": 3.1925048828125, "learning_rate": 4.918261419028706e-05, "loss": 4.6736, "step": 1812 }, { "epoch": 0.01636576999458386, "grad_norm": 3.016477108001709, "learning_rate": 4.918216284527893e-05, "loss": 4.6546, "step": 1813 }, { "epoch": 0.016374796894746343, "grad_norm": 3.4416186809539795, "learning_rate": 4.918171150027081e-05, "loss": 5.4215, "step": 1814 }, { "epoch": 0.01638382379490883, "grad_norm": 3.6347386837005615, "learning_rate": 4.9181260155262684e-05, "loss": 4.7117, "step": 1815 }, { "epoch": 0.016392850695071313, "grad_norm": 4.04032564163208, "learning_rate": 4.918080881025456e-05, "loss": 4.42, "step": 1816 }, { "epoch": 0.016401877595233795, "grad_norm": 3.8906333446502686, "learning_rate": 4.9180357465246436e-05, "loss": 5.1477, "step": 1817 }, { "epoch": 0.01641090449539628, "grad_norm": 3.693121910095215, "learning_rate": 4.917990612023831e-05, "loss": 5.018, "step": 1818 }, { "epoch": 0.016419931395558765, "grad_norm": 3.8439767360687256, "learning_rate": 4.917945477523019e-05, "loss": 4.0225, "step": 1819 }, { "epoch": 0.01642895829572125, "grad_norm": 3.747131586074829, "learning_rate": 4.917900343022207e-05, "loss": 4.6521, "step": 1820 }, { "epoch": 0.016437985195883735, "grad_norm": 3.4809951782226562, "learning_rate": 4.917855208521394e-05, "loss": 4.4704, "step": 1821 }, { "epoch": 0.016447012096046217, "grad_norm": 4.508114337921143, "learning_rate": 4.917810074020582e-05, "loss": 4.9137, "step": 1822 }, { "epoch": 0.0164560389962087, "grad_norm": 3.953552484512329, "learning_rate": 4.917764939519769e-05, "loss": 4.6667, "step": 1823 }, { "epoch": 0.016465065896371187, "grad_norm": 4.534947395324707, "learning_rate": 4.9177198050189564e-05, "loss": 4.2261, "step": 1824 }, { "epoch": 0.01647409279653367, "grad_norm": 5.441312789916992, "learning_rate": 4.917674670518144e-05, "loss": 4.3778, "step": 1825 }, { "epoch": 0.016483119696696153, "grad_norm": 4.143200874328613, "learning_rate": 4.917629536017332e-05, "loss": 5.0743, "step": 1826 }, { "epoch": 0.016492146596858638, "grad_norm": 4.156723976135254, "learning_rate": 4.9175844015165195e-05, "loss": 5.1398, "step": 1827 }, { "epoch": 0.016501173497021123, "grad_norm": 3.463364839553833, "learning_rate": 4.9175392670157074e-05, "loss": 5.4797, "step": 1828 }, { "epoch": 0.016510200397183608, "grad_norm": 3.698495388031006, "learning_rate": 4.9174941325148946e-05, "loss": 5.1459, "step": 1829 }, { "epoch": 0.01651922729734609, "grad_norm": 4.96151876449585, "learning_rate": 4.917448998014082e-05, "loss": 4.4382, "step": 1830 }, { "epoch": 0.016528254197508575, "grad_norm": 5.081183910369873, "learning_rate": 4.91740386351327e-05, "loss": 5.2055, "step": 1831 }, { "epoch": 0.01653728109767106, "grad_norm": 4.051260948181152, "learning_rate": 4.917358729012457e-05, "loss": 4.7259, "step": 1832 }, { "epoch": 0.016546307997833545, "grad_norm": 3.8583147525787354, "learning_rate": 4.917313594511645e-05, "loss": 4.5761, "step": 1833 }, { "epoch": 0.01655533489799603, "grad_norm": 5.454555988311768, "learning_rate": 4.917268460010833e-05, "loss": 4.7394, "step": 1834 }, { "epoch": 0.01656436179815851, "grad_norm": 3.4157538414001465, "learning_rate": 4.91722332551002e-05, "loss": 4.8384, "step": 1835 }, { "epoch": 0.016573388698320996, "grad_norm": 2.7650716304779053, "learning_rate": 4.9171781910092074e-05, "loss": 4.8601, "step": 1836 }, { "epoch": 0.01658241559848348, "grad_norm": 3.7848339080810547, "learning_rate": 4.917133056508395e-05, "loss": 5.1983, "step": 1837 }, { "epoch": 0.016591442498645966, "grad_norm": 5.551590919494629, "learning_rate": 4.9170879220075826e-05, "loss": 4.5946, "step": 1838 }, { "epoch": 0.016600469398808448, "grad_norm": 3.111004114151001, "learning_rate": 4.9170427875067705e-05, "loss": 4.7771, "step": 1839 }, { "epoch": 0.016609496298970933, "grad_norm": 5.049638748168945, "learning_rate": 4.916997653005958e-05, "loss": 4.5322, "step": 1840 }, { "epoch": 0.016618523199133418, "grad_norm": 3.8108720779418945, "learning_rate": 4.9169525185051457e-05, "loss": 4.3913, "step": 1841 }, { "epoch": 0.016627550099295903, "grad_norm": 4.28173303604126, "learning_rate": 4.9169073840043336e-05, "loss": 5.5165, "step": 1842 }, { "epoch": 0.016636576999458388, "grad_norm": 4.424776077270508, "learning_rate": 4.91686224950352e-05, "loss": 5.2079, "step": 1843 }, { "epoch": 0.01664560389962087, "grad_norm": 5.29373025894165, "learning_rate": 4.916817115002708e-05, "loss": 4.8403, "step": 1844 }, { "epoch": 0.016654630799783354, "grad_norm": 3.704237461090088, "learning_rate": 4.916771980501896e-05, "loss": 4.6561, "step": 1845 }, { "epoch": 0.01666365769994584, "grad_norm": 9.594893455505371, "learning_rate": 4.916726846001083e-05, "loss": 5.2489, "step": 1846 }, { "epoch": 0.016672684600108324, "grad_norm": 3.6998236179351807, "learning_rate": 4.916681711500271e-05, "loss": 5.2674, "step": 1847 }, { "epoch": 0.016681711500270806, "grad_norm": 6.059239864349365, "learning_rate": 4.916636576999459e-05, "loss": 4.7823, "step": 1848 }, { "epoch": 0.01669073840043329, "grad_norm": 3.3541455268859863, "learning_rate": 4.916591442498646e-05, "loss": 5.3879, "step": 1849 }, { "epoch": 0.016699765300595776, "grad_norm": 2.7224321365356445, "learning_rate": 4.9165463079978336e-05, "loss": 5.3743, "step": 1850 }, { "epoch": 0.01670879220075826, "grad_norm": 3.407416343688965, "learning_rate": 4.9165011734970215e-05, "loss": 4.8328, "step": 1851 }, { "epoch": 0.016717819100920742, "grad_norm": 6.209898948669434, "learning_rate": 4.916456038996209e-05, "loss": 4.8184, "step": 1852 }, { "epoch": 0.016726846001083227, "grad_norm": 4.133207321166992, "learning_rate": 4.916410904495397e-05, "loss": 4.6775, "step": 1853 }, { "epoch": 0.016735872901245712, "grad_norm": 3.7884905338287354, "learning_rate": 4.916365769994584e-05, "loss": 4.9833, "step": 1854 }, { "epoch": 0.016744899801408197, "grad_norm": 5.50409460067749, "learning_rate": 4.916320635493772e-05, "loss": 4.8883, "step": 1855 }, { "epoch": 0.016753926701570682, "grad_norm": 3.9700236320495605, "learning_rate": 4.91627550099296e-05, "loss": 5.0441, "step": 1856 }, { "epoch": 0.016762953601733164, "grad_norm": 3.412889003753662, "learning_rate": 4.9162303664921463e-05, "loss": 5.3685, "step": 1857 }, { "epoch": 0.01677198050189565, "grad_norm": 4.635268211364746, "learning_rate": 4.916185231991334e-05, "loss": 5.0815, "step": 1858 }, { "epoch": 0.016781007402058134, "grad_norm": 3.7899229526519775, "learning_rate": 4.916140097490522e-05, "loss": 4.7048, "step": 1859 }, { "epoch": 0.01679003430222062, "grad_norm": 4.044137954711914, "learning_rate": 4.9160949629897094e-05, "loss": 4.777, "step": 1860 }, { "epoch": 0.0167990612023831, "grad_norm": 5.055237770080566, "learning_rate": 4.9160498284888974e-05, "loss": 4.9185, "step": 1861 }, { "epoch": 0.016808088102545585, "grad_norm": 4.7524189949035645, "learning_rate": 4.9160046939880846e-05, "loss": 5.0489, "step": 1862 }, { "epoch": 0.01681711500270807, "grad_norm": 3.6071970462799072, "learning_rate": 4.915959559487272e-05, "loss": 4.6085, "step": 1863 }, { "epoch": 0.016826141902870555, "grad_norm": 6.86301326751709, "learning_rate": 4.91591442498646e-05, "loss": 4.5401, "step": 1864 }, { "epoch": 0.016835168803033037, "grad_norm": 3.1789426803588867, "learning_rate": 4.915869290485648e-05, "loss": 4.7119, "step": 1865 }, { "epoch": 0.016844195703195522, "grad_norm": 3.5957159996032715, "learning_rate": 4.915824155984835e-05, "loss": 5.2866, "step": 1866 }, { "epoch": 0.016853222603358007, "grad_norm": 2.971107006072998, "learning_rate": 4.915779021484023e-05, "loss": 4.5132, "step": 1867 }, { "epoch": 0.016862249503520492, "grad_norm": 3.4838576316833496, "learning_rate": 4.91573388698321e-05, "loss": 5.0605, "step": 1868 }, { "epoch": 0.016871276403682977, "grad_norm": 5.368085861206055, "learning_rate": 4.915688752482398e-05, "loss": 4.5031, "step": 1869 }, { "epoch": 0.01688030330384546, "grad_norm": 4.499110221862793, "learning_rate": 4.915643617981585e-05, "loss": 5.1454, "step": 1870 }, { "epoch": 0.016889330204007943, "grad_norm": 3.5230846405029297, "learning_rate": 4.9155984834807725e-05, "loss": 4.5997, "step": 1871 }, { "epoch": 0.01689835710417043, "grad_norm": 3.770979404449463, "learning_rate": 4.9155533489799605e-05, "loss": 4.722, "step": 1872 }, { "epoch": 0.016907384004332913, "grad_norm": 4.066607475280762, "learning_rate": 4.9155082144791484e-05, "loss": 4.5275, "step": 1873 }, { "epoch": 0.016916410904495395, "grad_norm": 3.519540548324585, "learning_rate": 4.9154630799783356e-05, "loss": 4.6973, "step": 1874 }, { "epoch": 0.01692543780465788, "grad_norm": 4.65676736831665, "learning_rate": 4.9154179454775236e-05, "loss": 4.7059, "step": 1875 }, { "epoch": 0.016934464704820365, "grad_norm": 3.144810199737549, "learning_rate": 4.915372810976711e-05, "loss": 4.9699, "step": 1876 }, { "epoch": 0.01694349160498285, "grad_norm": 3.9383790493011475, "learning_rate": 4.915327676475898e-05, "loss": 4.8026, "step": 1877 }, { "epoch": 0.01695251850514533, "grad_norm": 4.870471954345703, "learning_rate": 4.915282541975086e-05, "loss": 4.0751, "step": 1878 }, { "epoch": 0.016961545405307817, "grad_norm": 5.672579288482666, "learning_rate": 4.915237407474273e-05, "loss": 4.9151, "step": 1879 }, { "epoch": 0.0169705723054703, "grad_norm": 5.042507648468018, "learning_rate": 4.915192272973461e-05, "loss": 4.7617, "step": 1880 }, { "epoch": 0.016979599205632787, "grad_norm": 3.3492770195007324, "learning_rate": 4.915147138472649e-05, "loss": 5.1088, "step": 1881 }, { "epoch": 0.01698862610579527, "grad_norm": 3.6079165935516357, "learning_rate": 4.915102003971836e-05, "loss": 4.9435, "step": 1882 }, { "epoch": 0.016997653005957753, "grad_norm": 4.090900421142578, "learning_rate": 4.9150568694710236e-05, "loss": 4.4235, "step": 1883 }, { "epoch": 0.017006679906120238, "grad_norm": 5.456752300262451, "learning_rate": 4.9150117349702115e-05, "loss": 4.6473, "step": 1884 }, { "epoch": 0.017015706806282723, "grad_norm": 4.292070388793945, "learning_rate": 4.914966600469399e-05, "loss": 4.7985, "step": 1885 }, { "epoch": 0.017024733706445208, "grad_norm": 4.011219024658203, "learning_rate": 4.9149214659685867e-05, "loss": 4.8255, "step": 1886 }, { "epoch": 0.01703376060660769, "grad_norm": 4.596063137054443, "learning_rate": 4.9148763314677746e-05, "loss": 4.7491, "step": 1887 }, { "epoch": 0.017042787506770175, "grad_norm": 2.9699759483337402, "learning_rate": 4.914831196966962e-05, "loss": 4.9882, "step": 1888 }, { "epoch": 0.01705181440693266, "grad_norm": 4.116121768951416, "learning_rate": 4.91478606246615e-05, "loss": 4.4961, "step": 1889 }, { "epoch": 0.017060841307095145, "grad_norm": 3.8625292778015137, "learning_rate": 4.914740927965337e-05, "loss": 4.7011, "step": 1890 }, { "epoch": 0.017069868207257626, "grad_norm": 4.449184417724609, "learning_rate": 4.914695793464524e-05, "loss": 5.2609, "step": 1891 }, { "epoch": 0.01707889510742011, "grad_norm": 2.5072808265686035, "learning_rate": 4.914650658963712e-05, "loss": 5.3998, "step": 1892 }, { "epoch": 0.017087922007582596, "grad_norm": 3.4111382961273193, "learning_rate": 4.9146055244628994e-05, "loss": 5.8211, "step": 1893 }, { "epoch": 0.01709694890774508, "grad_norm": 4.057845115661621, "learning_rate": 4.914560389962087e-05, "loss": 5.293, "step": 1894 }, { "epoch": 0.017105975807907566, "grad_norm": 4.489046096801758, "learning_rate": 4.914515255461275e-05, "loss": 5.2063, "step": 1895 }, { "epoch": 0.017115002708070048, "grad_norm": 3.9272005558013916, "learning_rate": 4.9144701209604625e-05, "loss": 5.1218, "step": 1896 }, { "epoch": 0.017124029608232533, "grad_norm": 4.598461151123047, "learning_rate": 4.91442498645965e-05, "loss": 5.0261, "step": 1897 }, { "epoch": 0.017133056508395018, "grad_norm": 4.550034523010254, "learning_rate": 4.914379851958838e-05, "loss": 5.2846, "step": 1898 }, { "epoch": 0.017142083408557503, "grad_norm": 4.340222358703613, "learning_rate": 4.914334717458025e-05, "loss": 5.0423, "step": 1899 }, { "epoch": 0.017151110308719984, "grad_norm": 4.989524841308594, "learning_rate": 4.914289582957213e-05, "loss": 3.9411, "step": 1900 }, { "epoch": 0.01716013720888247, "grad_norm": 3.4924027919769287, "learning_rate": 4.9142444484564e-05, "loss": 5.3224, "step": 1901 }, { "epoch": 0.017169164109044954, "grad_norm": 6.347515106201172, "learning_rate": 4.914199313955588e-05, "loss": 4.698, "step": 1902 }, { "epoch": 0.01717819100920744, "grad_norm": 3.2502381801605225, "learning_rate": 4.914154179454776e-05, "loss": 4.7802, "step": 1903 }, { "epoch": 0.01718721790936992, "grad_norm": 3.823234796524048, "learning_rate": 4.9141090449539625e-05, "loss": 4.5638, "step": 1904 }, { "epoch": 0.017196244809532406, "grad_norm": 2.7915256023406982, "learning_rate": 4.9140639104531504e-05, "loss": 5.0294, "step": 1905 }, { "epoch": 0.01720527170969489, "grad_norm": 4.124701976776123, "learning_rate": 4.9140187759523384e-05, "loss": 4.364, "step": 1906 }, { "epoch": 0.017214298609857376, "grad_norm": 3.5605292320251465, "learning_rate": 4.9139736414515256e-05, "loss": 4.7983, "step": 1907 }, { "epoch": 0.01722332551001986, "grad_norm": 3.5490052700042725, "learning_rate": 4.9139285069507135e-05, "loss": 4.7168, "step": 1908 }, { "epoch": 0.017232352410182342, "grad_norm": 3.4034955501556396, "learning_rate": 4.9138833724499015e-05, "loss": 4.4529, "step": 1909 }, { "epoch": 0.017241379310344827, "grad_norm": 3.632934331893921, "learning_rate": 4.913838237949088e-05, "loss": 4.794, "step": 1910 }, { "epoch": 0.017250406210507312, "grad_norm": 5.379853248596191, "learning_rate": 4.913793103448276e-05, "loss": 4.9341, "step": 1911 }, { "epoch": 0.017259433110669797, "grad_norm": 4.689905166625977, "learning_rate": 4.913747968947464e-05, "loss": 4.7589, "step": 1912 }, { "epoch": 0.01726846001083228, "grad_norm": 4.339552402496338, "learning_rate": 4.913702834446651e-05, "loss": 4.6118, "step": 1913 }, { "epoch": 0.017277486910994764, "grad_norm": 7.172119617462158, "learning_rate": 4.913657699945839e-05, "loss": 5.1262, "step": 1914 }, { "epoch": 0.01728651381115725, "grad_norm": 3.8520748615264893, "learning_rate": 4.913612565445026e-05, "loss": 4.754, "step": 1915 }, { "epoch": 0.017295540711319734, "grad_norm": 3.2003092765808105, "learning_rate": 4.913567430944214e-05, "loss": 5.0624, "step": 1916 }, { "epoch": 0.017304567611482215, "grad_norm": 3.9096412658691406, "learning_rate": 4.9135222964434015e-05, "loss": 5.0168, "step": 1917 }, { "epoch": 0.0173135945116447, "grad_norm": 3.177353620529175, "learning_rate": 4.913477161942589e-05, "loss": 5.1903, "step": 1918 }, { "epoch": 0.017322621411807185, "grad_norm": 4.232254505157471, "learning_rate": 4.9134320274417766e-05, "loss": 4.4632, "step": 1919 }, { "epoch": 0.01733164831196967, "grad_norm": 3.1533420085906982, "learning_rate": 4.9133868929409646e-05, "loss": 5.1116, "step": 1920 }, { "epoch": 0.017340675212132155, "grad_norm": 4.955524921417236, "learning_rate": 4.913341758440152e-05, "loss": 5.2238, "step": 1921 }, { "epoch": 0.017349702112294637, "grad_norm": 3.8890233039855957, "learning_rate": 4.91329662393934e-05, "loss": 4.4114, "step": 1922 }, { "epoch": 0.017358729012457122, "grad_norm": 4.556668281555176, "learning_rate": 4.9132514894385276e-05, "loss": 5.1388, "step": 1923 }, { "epoch": 0.017367755912619607, "grad_norm": 3.8395397663116455, "learning_rate": 4.913206354937714e-05, "loss": 4.4434, "step": 1924 }, { "epoch": 0.017376782812782092, "grad_norm": 4.573689937591553, "learning_rate": 4.913161220436902e-05, "loss": 4.3666, "step": 1925 }, { "epoch": 0.017385809712944574, "grad_norm": 4.189978122711182, "learning_rate": 4.91311608593609e-05, "loss": 4.4489, "step": 1926 }, { "epoch": 0.01739483661310706, "grad_norm": 3.928588628768921, "learning_rate": 4.913070951435277e-05, "loss": 4.3414, "step": 1927 }, { "epoch": 0.017403863513269544, "grad_norm": 3.530383348464966, "learning_rate": 4.913025816934465e-05, "loss": 4.7995, "step": 1928 }, { "epoch": 0.01741289041343203, "grad_norm": 3.407148838043213, "learning_rate": 4.9129806824336525e-05, "loss": 5.286, "step": 1929 }, { "epoch": 0.01742191731359451, "grad_norm": 3.730221748352051, "learning_rate": 4.91293554793284e-05, "loss": 5.3794, "step": 1930 }, { "epoch": 0.017430944213756995, "grad_norm": 3.8698601722717285, "learning_rate": 4.9128904134320277e-05, "loss": 4.8523, "step": 1931 }, { "epoch": 0.01743997111391948, "grad_norm": 5.423507213592529, "learning_rate": 4.912845278931215e-05, "loss": 5.3583, "step": 1932 }, { "epoch": 0.017448998014081965, "grad_norm": 4.267336368560791, "learning_rate": 4.912800144430403e-05, "loss": 5.2043, "step": 1933 }, { "epoch": 0.01745802491424445, "grad_norm": 3.789365530014038, "learning_rate": 4.912755009929591e-05, "loss": 4.7652, "step": 1934 }, { "epoch": 0.01746705181440693, "grad_norm": 4.108798503875732, "learning_rate": 4.912709875428778e-05, "loss": 5.22, "step": 1935 }, { "epoch": 0.017476078714569417, "grad_norm": 4.976925849914551, "learning_rate": 4.912664740927966e-05, "loss": 4.2272, "step": 1936 }, { "epoch": 0.0174851056147319, "grad_norm": 6.873584747314453, "learning_rate": 4.912619606427153e-05, "loss": 5.0638, "step": 1937 }, { "epoch": 0.017494132514894387, "grad_norm": 4.0368571281433105, "learning_rate": 4.9125744719263404e-05, "loss": 5.2252, "step": 1938 }, { "epoch": 0.017503159415056868, "grad_norm": 3.9685518741607666, "learning_rate": 4.912529337425528e-05, "loss": 5.4466, "step": 1939 }, { "epoch": 0.017512186315219353, "grad_norm": 3.7405521869659424, "learning_rate": 4.9124842029247156e-05, "loss": 5.174, "step": 1940 }, { "epoch": 0.017521213215381838, "grad_norm": 3.745330572128296, "learning_rate": 4.9124390684239035e-05, "loss": 4.4228, "step": 1941 }, { "epoch": 0.017530240115544323, "grad_norm": 3.3374183177948, "learning_rate": 4.9123939339230914e-05, "loss": 4.8476, "step": 1942 }, { "epoch": 0.017539267015706805, "grad_norm": 3.461564540863037, "learning_rate": 4.912348799422279e-05, "loss": 4.2604, "step": 1943 }, { "epoch": 0.01754829391586929, "grad_norm": 3.758622646331787, "learning_rate": 4.912303664921466e-05, "loss": 5.6533, "step": 1944 }, { "epoch": 0.017557320816031775, "grad_norm": 3.852255344390869, "learning_rate": 4.912258530420654e-05, "loss": 5.1701, "step": 1945 }, { "epoch": 0.01756634771619426, "grad_norm": 3.8005082607269287, "learning_rate": 4.912213395919841e-05, "loss": 5.3204, "step": 1946 }, { "epoch": 0.017575374616356745, "grad_norm": 3.6428146362304688, "learning_rate": 4.912168261419029e-05, "loss": 5.0313, "step": 1947 }, { "epoch": 0.017584401516519226, "grad_norm": 4.537120342254639, "learning_rate": 4.912123126918217e-05, "loss": 4.6334, "step": 1948 }, { "epoch": 0.01759342841668171, "grad_norm": 3.069803476333618, "learning_rate": 4.912077992417404e-05, "loss": 4.9073, "step": 1949 }, { "epoch": 0.017602455316844196, "grad_norm": 3.3770511150360107, "learning_rate": 4.9120328579165914e-05, "loss": 4.8743, "step": 1950 }, { "epoch": 0.01761148221700668, "grad_norm": 3.573185682296753, "learning_rate": 4.9119877234157794e-05, "loss": 4.7376, "step": 1951 }, { "epoch": 0.017620509117169163, "grad_norm": 3.5945539474487305, "learning_rate": 4.9119425889149666e-05, "loss": 5.2117, "step": 1952 }, { "epoch": 0.017629536017331648, "grad_norm": 4.039890289306641, "learning_rate": 4.9118974544141545e-05, "loss": 4.9994, "step": 1953 }, { "epoch": 0.017638562917494133, "grad_norm": 4.83317232131958, "learning_rate": 4.911852319913342e-05, "loss": 5.3371, "step": 1954 }, { "epoch": 0.017647589817656618, "grad_norm": 3.517991065979004, "learning_rate": 4.91180718541253e-05, "loss": 4.7057, "step": 1955 }, { "epoch": 0.0176566167178191, "grad_norm": 6.5581865310668945, "learning_rate": 4.9117620509117176e-05, "loss": 5.285, "step": 1956 }, { "epoch": 0.017665643617981584, "grad_norm": 6.120394706726074, "learning_rate": 4.911716916410904e-05, "loss": 4.4128, "step": 1957 }, { "epoch": 0.01767467051814407, "grad_norm": 3.495138645172119, "learning_rate": 4.911671781910092e-05, "loss": 4.6099, "step": 1958 }, { "epoch": 0.017683697418306554, "grad_norm": 5.017294406890869, "learning_rate": 4.91162664740928e-05, "loss": 4.8876, "step": 1959 }, { "epoch": 0.01769272431846904, "grad_norm": 3.2117562294006348, "learning_rate": 4.911581512908467e-05, "loss": 4.5027, "step": 1960 }, { "epoch": 0.01770175121863152, "grad_norm": 4.378935813903809, "learning_rate": 4.911536378407655e-05, "loss": 4.9412, "step": 1961 }, { "epoch": 0.017710778118794006, "grad_norm": 3.6196935176849365, "learning_rate": 4.9114912439068425e-05, "loss": 4.4929, "step": 1962 }, { "epoch": 0.01771980501895649, "grad_norm": 7.025827884674072, "learning_rate": 4.9114461094060304e-05, "loss": 4.7595, "step": 1963 }, { "epoch": 0.017728831919118976, "grad_norm": 3.4888381958007812, "learning_rate": 4.9114009749052176e-05, "loss": 5.3216, "step": 1964 }, { "epoch": 0.017737858819281457, "grad_norm": 3.989227533340454, "learning_rate": 4.9113558404044056e-05, "loss": 4.6432, "step": 1965 }, { "epoch": 0.017746885719443942, "grad_norm": 3.953238010406494, "learning_rate": 4.911310705903593e-05, "loss": 4.7512, "step": 1966 }, { "epoch": 0.017755912619606427, "grad_norm": 4.141457557678223, "learning_rate": 4.911265571402781e-05, "loss": 4.8276, "step": 1967 }, { "epoch": 0.017764939519768912, "grad_norm": 6.774924278259277, "learning_rate": 4.911220436901968e-05, "loss": 4.7056, "step": 1968 }, { "epoch": 0.017773966419931394, "grad_norm": 5.081954479217529, "learning_rate": 4.911175302401156e-05, "loss": 4.8952, "step": 1969 }, { "epoch": 0.01778299332009388, "grad_norm": 3.27681565284729, "learning_rate": 4.911130167900344e-05, "loss": 4.1559, "step": 1970 }, { "epoch": 0.017792020220256364, "grad_norm": 4.218776226043701, "learning_rate": 4.9110850333995304e-05, "loss": 4.87, "step": 1971 }, { "epoch": 0.01780104712041885, "grad_norm": 3.9476523399353027, "learning_rate": 4.911039898898718e-05, "loss": 4.9991, "step": 1972 }, { "epoch": 0.017810074020581334, "grad_norm": 3.9725089073181152, "learning_rate": 4.910994764397906e-05, "loss": 4.3687, "step": 1973 }, { "epoch": 0.017819100920743815, "grad_norm": 3.69500994682312, "learning_rate": 4.9109496298970935e-05, "loss": 4.7324, "step": 1974 }, { "epoch": 0.0178281278209063, "grad_norm": 3.087998628616333, "learning_rate": 4.9109044953962814e-05, "loss": 5.0789, "step": 1975 }, { "epoch": 0.017837154721068785, "grad_norm": 4.942154407501221, "learning_rate": 4.9108593608954687e-05, "loss": 5.0091, "step": 1976 }, { "epoch": 0.01784618162123127, "grad_norm": 3.851125478744507, "learning_rate": 4.910814226394656e-05, "loss": 5.3061, "step": 1977 }, { "epoch": 0.017855208521393752, "grad_norm": 4.202939510345459, "learning_rate": 4.910769091893844e-05, "loss": 4.6882, "step": 1978 }, { "epoch": 0.017864235421556237, "grad_norm": 3.375303268432617, "learning_rate": 4.910723957393031e-05, "loss": 4.5607, "step": 1979 }, { "epoch": 0.017873262321718722, "grad_norm": 3.6031219959259033, "learning_rate": 4.910678822892219e-05, "loss": 4.492, "step": 1980 }, { "epoch": 0.017882289221881207, "grad_norm": 4.165266990661621, "learning_rate": 4.910633688391407e-05, "loss": 5.0026, "step": 1981 }, { "epoch": 0.01789131612204369, "grad_norm": 3.625606060028076, "learning_rate": 4.910588553890594e-05, "loss": 4.6712, "step": 1982 }, { "epoch": 0.017900343022206174, "grad_norm": 4.262526988983154, "learning_rate": 4.910543419389782e-05, "loss": 5.0462, "step": 1983 }, { "epoch": 0.01790936992236866, "grad_norm": 3.80904483795166, "learning_rate": 4.910498284888969e-05, "loss": 5.0132, "step": 1984 }, { "epoch": 0.017918396822531144, "grad_norm": 4.920229911804199, "learning_rate": 4.9104531503881566e-05, "loss": 5.2093, "step": 1985 }, { "epoch": 0.01792742372269363, "grad_norm": 4.900843143463135, "learning_rate": 4.9104080158873445e-05, "loss": 4.9347, "step": 1986 }, { "epoch": 0.01793645062285611, "grad_norm": 4.428861141204834, "learning_rate": 4.9103628813865324e-05, "loss": 4.9618, "step": 1987 }, { "epoch": 0.017945477523018595, "grad_norm": 3.895920515060425, "learning_rate": 4.91031774688572e-05, "loss": 5.2857, "step": 1988 }, { "epoch": 0.01795450442318108, "grad_norm": 4.836122512817383, "learning_rate": 4.9102726123849076e-05, "loss": 5.0172, "step": 1989 }, { "epoch": 0.017963531323343565, "grad_norm": 3.8130030632019043, "learning_rate": 4.910227477884095e-05, "loss": 5.0527, "step": 1990 }, { "epoch": 0.017972558223506047, "grad_norm": 3.6434719562530518, "learning_rate": 4.910182343383282e-05, "loss": 4.7133, "step": 1991 }, { "epoch": 0.01798158512366853, "grad_norm": 3.361246109008789, "learning_rate": 4.91013720888247e-05, "loss": 5.6864, "step": 1992 }, { "epoch": 0.017990612023831017, "grad_norm": 4.927320957183838, "learning_rate": 4.910092074381657e-05, "loss": 4.9824, "step": 1993 }, { "epoch": 0.0179996389239935, "grad_norm": 2.973278522491455, "learning_rate": 4.910046939880845e-05, "loss": 4.8451, "step": 1994 }, { "epoch": 0.018008665824155983, "grad_norm": 3.294184684753418, "learning_rate": 4.910001805380033e-05, "loss": 4.5604, "step": 1995 }, { "epoch": 0.018017692724318468, "grad_norm": 3.7731926441192627, "learning_rate": 4.9099566708792204e-05, "loss": 4.5347, "step": 1996 }, { "epoch": 0.018026719624480953, "grad_norm": 3.433797836303711, "learning_rate": 4.9099115363784076e-05, "loss": 5.1589, "step": 1997 }, { "epoch": 0.018035746524643438, "grad_norm": 3.2849111557006836, "learning_rate": 4.9098664018775955e-05, "loss": 5.1825, "step": 1998 }, { "epoch": 0.018044773424805923, "grad_norm": 4.833727836608887, "learning_rate": 4.909821267376783e-05, "loss": 5.0437, "step": 1999 }, { "epoch": 0.018053800324968405, "grad_norm": 4.919244766235352, "learning_rate": 4.909776132875971e-05, "loss": 5.2316, "step": 2000 }, { "epoch": 0.01806282722513089, "grad_norm": 3.8142380714416504, "learning_rate": 4.909730998375158e-05, "loss": 4.5313, "step": 2001 }, { "epoch": 0.018071854125293375, "grad_norm": 3.5523605346679688, "learning_rate": 4.909685863874346e-05, "loss": 4.97, "step": 2002 }, { "epoch": 0.01808088102545586, "grad_norm": 4.814506530761719, "learning_rate": 4.909640729373534e-05, "loss": 5.0481, "step": 2003 }, { "epoch": 0.01808990792561834, "grad_norm": 3.649346351623535, "learning_rate": 4.9095955948727204e-05, "loss": 4.7433, "step": 2004 }, { "epoch": 0.018098934825780826, "grad_norm": 3.3611414432525635, "learning_rate": 4.909550460371908e-05, "loss": 4.759, "step": 2005 }, { "epoch": 0.01810796172594331, "grad_norm": 2.896547317504883, "learning_rate": 4.909505325871096e-05, "loss": 4.8341, "step": 2006 }, { "epoch": 0.018116988626105796, "grad_norm": 4.987926483154297, "learning_rate": 4.9094601913702835e-05, "loss": 5.2566, "step": 2007 }, { "epoch": 0.018126015526268278, "grad_norm": 3.4138083457946777, "learning_rate": 4.9094150568694714e-05, "loss": 5.0116, "step": 2008 }, { "epoch": 0.018135042426430763, "grad_norm": 4.30384635925293, "learning_rate": 4.909369922368659e-05, "loss": 5.2906, "step": 2009 }, { "epoch": 0.018144069326593248, "grad_norm": 5.17842960357666, "learning_rate": 4.9093247878678466e-05, "loss": 5.2798, "step": 2010 }, { "epoch": 0.018153096226755733, "grad_norm": 4.266278266906738, "learning_rate": 4.909279653367034e-05, "loss": 4.4522, "step": 2011 }, { "epoch": 0.018162123126918218, "grad_norm": 3.8494575023651123, "learning_rate": 4.909234518866222e-05, "loss": 4.9902, "step": 2012 }, { "epoch": 0.0181711500270807, "grad_norm": 3.7175533771514893, "learning_rate": 4.909189384365409e-05, "loss": 5.1695, "step": 2013 }, { "epoch": 0.018180176927243184, "grad_norm": 3.912518262863159, "learning_rate": 4.909144249864597e-05, "loss": 4.9862, "step": 2014 }, { "epoch": 0.01818920382740567, "grad_norm": 4.574335098266602, "learning_rate": 4.909099115363784e-05, "loss": 4.5044, "step": 2015 }, { "epoch": 0.018198230727568154, "grad_norm": 4.1077399253845215, "learning_rate": 4.909053980862972e-05, "loss": 4.0751, "step": 2016 }, { "epoch": 0.018207257627730636, "grad_norm": 3.735466241836548, "learning_rate": 4.90900884636216e-05, "loss": 4.5819, "step": 2017 }, { "epoch": 0.01821628452789312, "grad_norm": 4.04451847076416, "learning_rate": 4.9089637118613466e-05, "loss": 4.9861, "step": 2018 }, { "epoch": 0.018225311428055606, "grad_norm": 5.723209381103516, "learning_rate": 4.9089185773605345e-05, "loss": 3.8351, "step": 2019 }, { "epoch": 0.01823433832821809, "grad_norm": 4.774405002593994, "learning_rate": 4.9088734428597224e-05, "loss": 4.7012, "step": 2020 }, { "epoch": 0.018243365228380576, "grad_norm": 3.206845998764038, "learning_rate": 4.9088283083589096e-05, "loss": 5.729, "step": 2021 }, { "epoch": 0.018252392128543057, "grad_norm": 3.4504342079162598, "learning_rate": 4.9087831738580976e-05, "loss": 5.0202, "step": 2022 }, { "epoch": 0.018261419028705542, "grad_norm": 4.029533863067627, "learning_rate": 4.908738039357285e-05, "loss": 4.8475, "step": 2023 }, { "epoch": 0.018270445928868027, "grad_norm": 4.4906511306762695, "learning_rate": 4.908692904856472e-05, "loss": 5.4188, "step": 2024 }, { "epoch": 0.018279472829030512, "grad_norm": 3.6973440647125244, "learning_rate": 4.90864777035566e-05, "loss": 5.1103, "step": 2025 }, { "epoch": 0.018288499729192994, "grad_norm": 3.50223445892334, "learning_rate": 4.908602635854848e-05, "loss": 5.2689, "step": 2026 }, { "epoch": 0.01829752662935548, "grad_norm": 4.020206451416016, "learning_rate": 4.908557501354035e-05, "loss": 4.6854, "step": 2027 }, { "epoch": 0.018306553529517964, "grad_norm": 3.833003520965576, "learning_rate": 4.908512366853223e-05, "loss": 5.2072, "step": 2028 }, { "epoch": 0.01831558042968045, "grad_norm": 4.538300514221191, "learning_rate": 4.90846723235241e-05, "loss": 3.4669, "step": 2029 }, { "epoch": 0.01832460732984293, "grad_norm": 3.209207057952881, "learning_rate": 4.908422097851598e-05, "loss": 5.0285, "step": 2030 }, { "epoch": 0.018333634230005415, "grad_norm": 4.329929351806641, "learning_rate": 4.9083769633507855e-05, "loss": 4.7476, "step": 2031 }, { "epoch": 0.0183426611301679, "grad_norm": 4.542428493499756, "learning_rate": 4.908331828849973e-05, "loss": 4.8706, "step": 2032 }, { "epoch": 0.018351688030330385, "grad_norm": 3.5688297748565674, "learning_rate": 4.908286694349161e-05, "loss": 5.0487, "step": 2033 }, { "epoch": 0.01836071493049287, "grad_norm": 3.358766794204712, "learning_rate": 4.9082415598483486e-05, "loss": 4.3423, "step": 2034 }, { "epoch": 0.018369741830655352, "grad_norm": 4.417934417724609, "learning_rate": 4.908196425347536e-05, "loss": 4.6853, "step": 2035 }, { "epoch": 0.018378768730817837, "grad_norm": 4.065502166748047, "learning_rate": 4.908151290846724e-05, "loss": 4.8681, "step": 2036 }, { "epoch": 0.018387795630980322, "grad_norm": 4.299354076385498, "learning_rate": 4.908106156345911e-05, "loss": 4.8063, "step": 2037 }, { "epoch": 0.018396822531142807, "grad_norm": 5.042741298675537, "learning_rate": 4.908061021845098e-05, "loss": 4.6737, "step": 2038 }, { "epoch": 0.01840584943130529, "grad_norm": 3.3907978534698486, "learning_rate": 4.908015887344286e-05, "loss": 5.2398, "step": 2039 }, { "epoch": 0.018414876331467774, "grad_norm": 3.9402525424957275, "learning_rate": 4.9079707528434734e-05, "loss": 4.8409, "step": 2040 }, { "epoch": 0.01842390323163026, "grad_norm": 3.6438193321228027, "learning_rate": 4.9079256183426614e-05, "loss": 4.7667, "step": 2041 }, { "epoch": 0.018432930131792744, "grad_norm": 3.233536958694458, "learning_rate": 4.907880483841849e-05, "loss": 4.467, "step": 2042 }, { "epoch": 0.018441957031955225, "grad_norm": 3.585157632827759, "learning_rate": 4.9078353493410365e-05, "loss": 5.2012, "step": 2043 }, { "epoch": 0.01845098393211771, "grad_norm": 3.8369808197021484, "learning_rate": 4.907790214840224e-05, "loss": 4.6409, "step": 2044 }, { "epoch": 0.018460010832280195, "grad_norm": 4.069700241088867, "learning_rate": 4.907745080339412e-05, "loss": 5.0512, "step": 2045 }, { "epoch": 0.01846903773244268, "grad_norm": 4.474863052368164, "learning_rate": 4.907699945838599e-05, "loss": 4.9455, "step": 2046 }, { "epoch": 0.018478064632605165, "grad_norm": 4.420050621032715, "learning_rate": 4.907654811337787e-05, "loss": 5.1284, "step": 2047 }, { "epoch": 0.018487091532767647, "grad_norm": 3.4613006114959717, "learning_rate": 4.907609676836975e-05, "loss": 4.6857, "step": 2048 }, { "epoch": 0.01849611843293013, "grad_norm": 4.355159759521484, "learning_rate": 4.907564542336162e-05, "loss": 5.2624, "step": 2049 }, { "epoch": 0.018505145333092617, "grad_norm": 4.754657745361328, "learning_rate": 4.90751940783535e-05, "loss": 5.0213, "step": 2050 }, { "epoch": 0.0185141722332551, "grad_norm": 2.932971239089966, "learning_rate": 4.907474273334537e-05, "loss": 5.1268, "step": 2051 }, { "epoch": 0.018523199133417583, "grad_norm": 4.445348739624023, "learning_rate": 4.9074291388337245e-05, "loss": 4.6098, "step": 2052 }, { "epoch": 0.018532226033580068, "grad_norm": 4.239518165588379, "learning_rate": 4.9073840043329124e-05, "loss": 5.3127, "step": 2053 }, { "epoch": 0.018541252933742553, "grad_norm": 2.953017473220825, "learning_rate": 4.9073388698320996e-05, "loss": 4.8543, "step": 2054 }, { "epoch": 0.018550279833905038, "grad_norm": 5.640254020690918, "learning_rate": 4.9072937353312875e-05, "loss": 4.2159, "step": 2055 }, { "epoch": 0.01855930673406752, "grad_norm": 3.3130009174346924, "learning_rate": 4.9072486008304755e-05, "loss": 4.291, "step": 2056 }, { "epoch": 0.018568333634230005, "grad_norm": 3.908445119857788, "learning_rate": 4.907203466329663e-05, "loss": 4.596, "step": 2057 }, { "epoch": 0.01857736053439249, "grad_norm": 4.12477970123291, "learning_rate": 4.90715833182885e-05, "loss": 4.2634, "step": 2058 }, { "epoch": 0.018586387434554975, "grad_norm": 3.3236398696899414, "learning_rate": 4.907113197328038e-05, "loss": 4.3865, "step": 2059 }, { "epoch": 0.01859541433471746, "grad_norm": 3.9974277019500732, "learning_rate": 4.907068062827225e-05, "loss": 3.6539, "step": 2060 }, { "epoch": 0.01860444123487994, "grad_norm": 3.3607308864593506, "learning_rate": 4.907022928326413e-05, "loss": 5.1252, "step": 2061 }, { "epoch": 0.018613468135042426, "grad_norm": 4.097604274749756, "learning_rate": 4.9069777938256e-05, "loss": 5.3019, "step": 2062 }, { "epoch": 0.01862249503520491, "grad_norm": 5.521720886230469, "learning_rate": 4.906932659324788e-05, "loss": 5.3407, "step": 2063 }, { "epoch": 0.018631521935367396, "grad_norm": 4.11037540435791, "learning_rate": 4.906887524823976e-05, "loss": 5.2575, "step": 2064 }, { "epoch": 0.018640548835529878, "grad_norm": 4.191189289093018, "learning_rate": 4.906842390323163e-05, "loss": 4.8783, "step": 2065 }, { "epoch": 0.018649575735692363, "grad_norm": 4.110476970672607, "learning_rate": 4.9067972558223506e-05, "loss": 4.926, "step": 2066 }, { "epoch": 0.018658602635854848, "grad_norm": 4.592854022979736, "learning_rate": 4.9067521213215386e-05, "loss": 5.0196, "step": 2067 }, { "epoch": 0.018667629536017333, "grad_norm": 7.287646293640137, "learning_rate": 4.906706986820726e-05, "loss": 3.6738, "step": 2068 }, { "epoch": 0.018676656436179814, "grad_norm": 4.3278326988220215, "learning_rate": 4.906661852319914e-05, "loss": 5.1739, "step": 2069 }, { "epoch": 0.0186856833363423, "grad_norm": 2.9180049896240234, "learning_rate": 4.906616717819102e-05, "loss": 5.0424, "step": 2070 }, { "epoch": 0.018694710236504784, "grad_norm": 3.2438559532165527, "learning_rate": 4.906571583318288e-05, "loss": 4.8307, "step": 2071 }, { "epoch": 0.01870373713666727, "grad_norm": 3.4833881855010986, "learning_rate": 4.906526448817476e-05, "loss": 4.8604, "step": 2072 }, { "epoch": 0.018712764036829754, "grad_norm": 4.032395362854004, "learning_rate": 4.906481314316664e-05, "loss": 5.2806, "step": 2073 }, { "epoch": 0.018721790936992236, "grad_norm": 3.849454879760742, "learning_rate": 4.906436179815851e-05, "loss": 5.2996, "step": 2074 }, { "epoch": 0.01873081783715472, "grad_norm": 5.1531662940979, "learning_rate": 4.906391045315039e-05, "loss": 3.6538, "step": 2075 }, { "epoch": 0.018739844737317206, "grad_norm": 3.533859968185425, "learning_rate": 4.9063459108142265e-05, "loss": 4.844, "step": 2076 }, { "epoch": 0.01874887163747969, "grad_norm": 3.8854591846466064, "learning_rate": 4.9063007763134144e-05, "loss": 4.7172, "step": 2077 }, { "epoch": 0.018757898537642172, "grad_norm": 4.229945182800293, "learning_rate": 4.906255641812602e-05, "loss": 3.9684, "step": 2078 }, { "epoch": 0.018766925437804657, "grad_norm": 6.4450201988220215, "learning_rate": 4.906210507311789e-05, "loss": 5.3818, "step": 2079 }, { "epoch": 0.018775952337967142, "grad_norm": 4.453969478607178, "learning_rate": 4.906165372810977e-05, "loss": 5.0318, "step": 2080 }, { "epoch": 0.018784979238129627, "grad_norm": 5.678875923156738, "learning_rate": 4.906120238310165e-05, "loss": 4.9418, "step": 2081 }, { "epoch": 0.01879400613829211, "grad_norm": 3.58832049369812, "learning_rate": 4.906075103809352e-05, "loss": 4.9032, "step": 2082 }, { "epoch": 0.018803033038454594, "grad_norm": 5.114472389221191, "learning_rate": 4.90602996930854e-05, "loss": 4.8201, "step": 2083 }, { "epoch": 0.01881205993861708, "grad_norm": 2.7819881439208984, "learning_rate": 4.905984834807728e-05, "loss": 5.3769, "step": 2084 }, { "epoch": 0.018821086838779564, "grad_norm": 4.171091079711914, "learning_rate": 4.9059397003069144e-05, "loss": 4.1197, "step": 2085 }, { "epoch": 0.01883011373894205, "grad_norm": 3.7981035709381104, "learning_rate": 4.9058945658061024e-05, "loss": 5.2849, "step": 2086 }, { "epoch": 0.01883914063910453, "grad_norm": 4.328599452972412, "learning_rate": 4.90584943130529e-05, "loss": 5.058, "step": 2087 }, { "epoch": 0.018848167539267015, "grad_norm": 4.718459129333496, "learning_rate": 4.9058042968044775e-05, "loss": 4.9762, "step": 2088 }, { "epoch": 0.0188571944394295, "grad_norm": 4.693949222564697, "learning_rate": 4.9057591623036654e-05, "loss": 4.446, "step": 2089 }, { "epoch": 0.018866221339591985, "grad_norm": 3.5669174194335938, "learning_rate": 4.905714027802853e-05, "loss": 4.8393, "step": 2090 }, { "epoch": 0.018875248239754467, "grad_norm": 4.764926910400391, "learning_rate": 4.90566889330204e-05, "loss": 5.4296, "step": 2091 }, { "epoch": 0.018884275139916952, "grad_norm": 5.352468013763428, "learning_rate": 4.905623758801228e-05, "loss": 4.5579, "step": 2092 }, { "epoch": 0.018893302040079437, "grad_norm": 4.252882957458496, "learning_rate": 4.905578624300415e-05, "loss": 5.0327, "step": 2093 }, { "epoch": 0.018902328940241922, "grad_norm": 5.2285614013671875, "learning_rate": 4.905533489799603e-05, "loss": 4.2769, "step": 2094 }, { "epoch": 0.018911355840404404, "grad_norm": 3.642662286758423, "learning_rate": 4.905488355298791e-05, "loss": 4.1061, "step": 2095 }, { "epoch": 0.01892038274056689, "grad_norm": 3.901705265045166, "learning_rate": 4.905443220797978e-05, "loss": 4.6565, "step": 2096 }, { "epoch": 0.018929409640729374, "grad_norm": 4.978423118591309, "learning_rate": 4.905398086297166e-05, "loss": 4.3401, "step": 2097 }, { "epoch": 0.01893843654089186, "grad_norm": 3.163423538208008, "learning_rate": 4.9053529517963534e-05, "loss": 4.5354, "step": 2098 }, { "epoch": 0.018947463441054344, "grad_norm": 4.5671162605285645, "learning_rate": 4.9053078172955406e-05, "loss": 4.9942, "step": 2099 }, { "epoch": 0.018956490341216825, "grad_norm": 3.3961665630340576, "learning_rate": 4.9052626827947285e-05, "loss": 4.1105, "step": 2100 }, { "epoch": 0.01896551724137931, "grad_norm": 3.297783851623535, "learning_rate": 4.905217548293916e-05, "loss": 4.7109, "step": 2101 }, { "epoch": 0.018974544141541795, "grad_norm": 4.651186943054199, "learning_rate": 4.905172413793104e-05, "loss": 4.8892, "step": 2102 }, { "epoch": 0.01898357104170428, "grad_norm": 3.256810426712036, "learning_rate": 4.9051272792922916e-05, "loss": 4.9802, "step": 2103 }, { "epoch": 0.01899259794186676, "grad_norm": 10.20473575592041, "learning_rate": 4.905082144791479e-05, "loss": 3.1518, "step": 2104 }, { "epoch": 0.019001624842029247, "grad_norm": 4.586795330047607, "learning_rate": 4.905037010290666e-05, "loss": 5.118, "step": 2105 }, { "epoch": 0.01901065174219173, "grad_norm": 3.3626818656921387, "learning_rate": 4.904991875789854e-05, "loss": 4.2169, "step": 2106 }, { "epoch": 0.019019678642354217, "grad_norm": 4.880466461181641, "learning_rate": 4.904946741289041e-05, "loss": 5.0307, "step": 2107 }, { "epoch": 0.019028705542516698, "grad_norm": 4.340914726257324, "learning_rate": 4.904901606788229e-05, "loss": 4.8076, "step": 2108 }, { "epoch": 0.019037732442679183, "grad_norm": 3.642052173614502, "learning_rate": 4.904856472287417e-05, "loss": 5.366, "step": 2109 }, { "epoch": 0.019046759342841668, "grad_norm": 3.710766553878784, "learning_rate": 4.9048113377866044e-05, "loss": 4.3522, "step": 2110 }, { "epoch": 0.019055786243004153, "grad_norm": 4.093945503234863, "learning_rate": 4.9047662032857916e-05, "loss": 4.351, "step": 2111 }, { "epoch": 0.019064813143166638, "grad_norm": 6.330406188964844, "learning_rate": 4.9047210687849796e-05, "loss": 5.0845, "step": 2112 }, { "epoch": 0.01907384004332912, "grad_norm": 3.472285270690918, "learning_rate": 4.904675934284167e-05, "loss": 5.0273, "step": 2113 }, { "epoch": 0.019082866943491605, "grad_norm": 4.034302234649658, "learning_rate": 4.904630799783355e-05, "loss": 5.2933, "step": 2114 }, { "epoch": 0.01909189384365409, "grad_norm": 3.1391336917877197, "learning_rate": 4.904585665282542e-05, "loss": 4.4176, "step": 2115 }, { "epoch": 0.019100920743816575, "grad_norm": 4.355051040649414, "learning_rate": 4.90454053078173e-05, "loss": 4.2858, "step": 2116 }, { "epoch": 0.019109947643979056, "grad_norm": 3.9604485034942627, "learning_rate": 4.904495396280918e-05, "loss": 4.5804, "step": 2117 }, { "epoch": 0.01911897454414154, "grad_norm": 4.724828720092773, "learning_rate": 4.9044502617801044e-05, "loss": 5.4619, "step": 2118 }, { "epoch": 0.019128001444304026, "grad_norm": 3.635542154312134, "learning_rate": 4.904405127279292e-05, "loss": 5.0465, "step": 2119 }, { "epoch": 0.01913702834446651, "grad_norm": 3.3053641319274902, "learning_rate": 4.90435999277848e-05, "loss": 4.2, "step": 2120 }, { "epoch": 0.019146055244628993, "grad_norm": 2.889575242996216, "learning_rate": 4.9043148582776675e-05, "loss": 5.3089, "step": 2121 }, { "epoch": 0.019155082144791478, "grad_norm": 4.55643892288208, "learning_rate": 4.9042697237768554e-05, "loss": 4.9126, "step": 2122 }, { "epoch": 0.019164109044953963, "grad_norm": 4.978509902954102, "learning_rate": 4.904224589276043e-05, "loss": 4.9285, "step": 2123 }, { "epoch": 0.019173135945116448, "grad_norm": 3.7409822940826416, "learning_rate": 4.9041794547752306e-05, "loss": 4.3225, "step": 2124 }, { "epoch": 0.019182162845278933, "grad_norm": 3.7132954597473145, "learning_rate": 4.904134320274418e-05, "loss": 3.9839, "step": 2125 }, { "epoch": 0.019191189745441414, "grad_norm": 4.295714378356934, "learning_rate": 4.904089185773606e-05, "loss": 5.0059, "step": 2126 }, { "epoch": 0.0192002166456039, "grad_norm": 4.382133483886719, "learning_rate": 4.904044051272793e-05, "loss": 4.9389, "step": 2127 }, { "epoch": 0.019209243545766384, "grad_norm": 4.609371662139893, "learning_rate": 4.903998916771981e-05, "loss": 4.441, "step": 2128 }, { "epoch": 0.01921827044592887, "grad_norm": 3.626193046569824, "learning_rate": 4.903953782271168e-05, "loss": 5.1645, "step": 2129 }, { "epoch": 0.01922729734609135, "grad_norm": 3.584031105041504, "learning_rate": 4.903908647770356e-05, "loss": 4.9008, "step": 2130 }, { "epoch": 0.019236324246253836, "grad_norm": 4.39001989364624, "learning_rate": 4.903863513269544e-05, "loss": 4.878, "step": 2131 }, { "epoch": 0.01924535114641632, "grad_norm": 3.301258087158203, "learning_rate": 4.9038183787687306e-05, "loss": 5.4253, "step": 2132 }, { "epoch": 0.019254378046578806, "grad_norm": 3.9147160053253174, "learning_rate": 4.9037732442679185e-05, "loss": 4.4458, "step": 2133 }, { "epoch": 0.019263404946741287, "grad_norm": 5.766073226928711, "learning_rate": 4.9037281097671064e-05, "loss": 5.0131, "step": 2134 }, { "epoch": 0.019272431846903772, "grad_norm": 4.319394588470459, "learning_rate": 4.903682975266294e-05, "loss": 4.2923, "step": 2135 }, { "epoch": 0.019281458747066257, "grad_norm": 4.929179668426514, "learning_rate": 4.9036378407654816e-05, "loss": 5.2424, "step": 2136 }, { "epoch": 0.019290485647228742, "grad_norm": 7.305233001708984, "learning_rate": 4.903592706264669e-05, "loss": 4.6838, "step": 2137 }, { "epoch": 0.019299512547391227, "grad_norm": 3.5842244625091553, "learning_rate": 4.903547571763856e-05, "loss": 3.3435, "step": 2138 }, { "epoch": 0.01930853944755371, "grad_norm": 3.9109065532684326, "learning_rate": 4.903502437263044e-05, "loss": 4.1663, "step": 2139 }, { "epoch": 0.019317566347716194, "grad_norm": 4.023662090301514, "learning_rate": 4.903457302762231e-05, "loss": 5.1603, "step": 2140 }, { "epoch": 0.01932659324787868, "grad_norm": 4.739776134490967, "learning_rate": 4.903412168261419e-05, "loss": 3.4418, "step": 2141 }, { "epoch": 0.019335620148041164, "grad_norm": 3.7356209754943848, "learning_rate": 4.903367033760607e-05, "loss": 4.7005, "step": 2142 }, { "epoch": 0.019344647048203645, "grad_norm": 4.743292808532715, "learning_rate": 4.9033218992597944e-05, "loss": 4.9552, "step": 2143 }, { "epoch": 0.01935367394836613, "grad_norm": 3.9861814975738525, "learning_rate": 4.903276764758982e-05, "loss": 4.8566, "step": 2144 }, { "epoch": 0.019362700848528615, "grad_norm": 3.9462344646453857, "learning_rate": 4.9032316302581695e-05, "loss": 4.7324, "step": 2145 }, { "epoch": 0.0193717277486911, "grad_norm": 3.2092621326446533, "learning_rate": 4.903186495757357e-05, "loss": 4.9508, "step": 2146 }, { "epoch": 0.019380754648853582, "grad_norm": 3.437721014022827, "learning_rate": 4.903141361256545e-05, "loss": 4.6647, "step": 2147 }, { "epoch": 0.019389781549016067, "grad_norm": 5.148375988006592, "learning_rate": 4.9030962267557326e-05, "loss": 4.9653, "step": 2148 }, { "epoch": 0.019398808449178552, "grad_norm": 4.497570514678955, "learning_rate": 4.90305109225492e-05, "loss": 3.4834, "step": 2149 }, { "epoch": 0.019407835349341037, "grad_norm": 3.8016316890716553, "learning_rate": 4.903005957754108e-05, "loss": 5.3355, "step": 2150 }, { "epoch": 0.019416862249503522, "grad_norm": 3.8371691703796387, "learning_rate": 4.902960823253295e-05, "loss": 4.6784, "step": 2151 }, { "epoch": 0.019425889149666004, "grad_norm": 3.4216792583465576, "learning_rate": 4.902915688752482e-05, "loss": 3.6457, "step": 2152 }, { "epoch": 0.01943491604982849, "grad_norm": 3.1374847888946533, "learning_rate": 4.90287055425167e-05, "loss": 4.75, "step": 2153 }, { "epoch": 0.019443942949990974, "grad_norm": 3.6699976921081543, "learning_rate": 4.9028254197508575e-05, "loss": 5.1191, "step": 2154 }, { "epoch": 0.01945296985015346, "grad_norm": 2.9183175563812256, "learning_rate": 4.9027802852500454e-05, "loss": 4.8544, "step": 2155 }, { "epoch": 0.01946199675031594, "grad_norm": 3.7447471618652344, "learning_rate": 4.902735150749233e-05, "loss": 5.1498, "step": 2156 }, { "epoch": 0.019471023650478425, "grad_norm": 3.1225075721740723, "learning_rate": 4.9026900162484206e-05, "loss": 5.03, "step": 2157 }, { "epoch": 0.01948005055064091, "grad_norm": 3.20241641998291, "learning_rate": 4.902644881747608e-05, "loss": 5.2306, "step": 2158 }, { "epoch": 0.019489077450803395, "grad_norm": 3.750896692276001, "learning_rate": 4.902599747246796e-05, "loss": 4.6939, "step": 2159 }, { "epoch": 0.019498104350965877, "grad_norm": 3.2953009605407715, "learning_rate": 4.902554612745983e-05, "loss": 5.0681, "step": 2160 }, { "epoch": 0.01950713125112836, "grad_norm": 3.336872100830078, "learning_rate": 4.902509478245171e-05, "loss": 3.9642, "step": 2161 }, { "epoch": 0.019516158151290847, "grad_norm": 2.9548287391662598, "learning_rate": 4.902464343744358e-05, "loss": 5.0287, "step": 2162 }, { "epoch": 0.01952518505145333, "grad_norm": 4.7140936851501465, "learning_rate": 4.902419209243546e-05, "loss": 4.1038, "step": 2163 }, { "epoch": 0.019534211951615817, "grad_norm": 3.5858802795410156, "learning_rate": 4.902374074742734e-05, "loss": 4.8356, "step": 2164 }, { "epoch": 0.019543238851778298, "grad_norm": 5.545526504516602, "learning_rate": 4.9023289402419206e-05, "loss": 4.8959, "step": 2165 }, { "epoch": 0.019552265751940783, "grad_norm": 3.2658565044403076, "learning_rate": 4.9022838057411085e-05, "loss": 5.0301, "step": 2166 }, { "epoch": 0.019561292652103268, "grad_norm": 3.935749053955078, "learning_rate": 4.9022386712402964e-05, "loss": 5.0349, "step": 2167 }, { "epoch": 0.019570319552265753, "grad_norm": 3.610900640487671, "learning_rate": 4.902193536739484e-05, "loss": 4.8226, "step": 2168 }, { "epoch": 0.019579346452428235, "grad_norm": 3.485023260116577, "learning_rate": 4.9021484022386716e-05, "loss": 4.5511, "step": 2169 }, { "epoch": 0.01958837335259072, "grad_norm": 3.2670419216156006, "learning_rate": 4.9021032677378595e-05, "loss": 5.0123, "step": 2170 }, { "epoch": 0.019597400252753205, "grad_norm": 3.247119903564453, "learning_rate": 4.902058133237047e-05, "loss": 5.0075, "step": 2171 }, { "epoch": 0.01960642715291569, "grad_norm": 3.933055877685547, "learning_rate": 4.902012998736234e-05, "loss": 4.6949, "step": 2172 }, { "epoch": 0.01961545405307817, "grad_norm": 4.24975061416626, "learning_rate": 4.901967864235422e-05, "loss": 4.9652, "step": 2173 }, { "epoch": 0.019624480953240656, "grad_norm": 2.957040309906006, "learning_rate": 4.901922729734609e-05, "loss": 4.8284, "step": 2174 }, { "epoch": 0.01963350785340314, "grad_norm": 4.147572994232178, "learning_rate": 4.901877595233797e-05, "loss": 4.9759, "step": 2175 }, { "epoch": 0.019642534753565626, "grad_norm": 3.3250739574432373, "learning_rate": 4.9018324607329844e-05, "loss": 5.1231, "step": 2176 }, { "epoch": 0.01965156165372811, "grad_norm": 4.566120147705078, "learning_rate": 4.901787326232172e-05, "loss": 4.502, "step": 2177 }, { "epoch": 0.019660588553890593, "grad_norm": 5.256094455718994, "learning_rate": 4.90174219173136e-05, "loss": 3.9352, "step": 2178 }, { "epoch": 0.019669615454053078, "grad_norm": 9.277670860290527, "learning_rate": 4.901697057230547e-05, "loss": 4.2964, "step": 2179 }, { "epoch": 0.019678642354215563, "grad_norm": 3.7941269874572754, "learning_rate": 4.901651922729735e-05, "loss": 4.7112, "step": 2180 }, { "epoch": 0.019687669254378048, "grad_norm": 3.91740345954895, "learning_rate": 4.9016067882289226e-05, "loss": 4.5958, "step": 2181 }, { "epoch": 0.01969669615454053, "grad_norm": 3.9976847171783447, "learning_rate": 4.90156165372811e-05, "loss": 5.1025, "step": 2182 }, { "epoch": 0.019705723054703014, "grad_norm": 3.229430675506592, "learning_rate": 4.901516519227298e-05, "loss": 4.7907, "step": 2183 }, { "epoch": 0.0197147499548655, "grad_norm": 3.8619847297668457, "learning_rate": 4.901471384726485e-05, "loss": 4.6887, "step": 2184 }, { "epoch": 0.019723776855027984, "grad_norm": 4.025402069091797, "learning_rate": 4.901426250225672e-05, "loss": 4.5767, "step": 2185 }, { "epoch": 0.019732803755190466, "grad_norm": 3.8231358528137207, "learning_rate": 4.90138111572486e-05, "loss": 4.6941, "step": 2186 }, { "epoch": 0.01974183065535295, "grad_norm": 3.063735246658325, "learning_rate": 4.901335981224048e-05, "loss": 3.9714, "step": 2187 }, { "epoch": 0.019750857555515436, "grad_norm": 4.678351402282715, "learning_rate": 4.9012908467232354e-05, "loss": 3.7781, "step": 2188 }, { "epoch": 0.01975988445567792, "grad_norm": 7.140661239624023, "learning_rate": 4.901245712222423e-05, "loss": 4.6429, "step": 2189 }, { "epoch": 0.019768911355840406, "grad_norm": 3.51670241355896, "learning_rate": 4.9012005777216105e-05, "loss": 4.5855, "step": 2190 }, { "epoch": 0.019777938256002887, "grad_norm": 3.626699447631836, "learning_rate": 4.9011554432207985e-05, "loss": 4.74, "step": 2191 }, { "epoch": 0.019786965156165372, "grad_norm": 7.269057273864746, "learning_rate": 4.901110308719986e-05, "loss": 4.7664, "step": 2192 }, { "epoch": 0.019795992056327857, "grad_norm": 3.836357355117798, "learning_rate": 4.901065174219173e-05, "loss": 4.842, "step": 2193 }, { "epoch": 0.019805018956490342, "grad_norm": 3.9234819412231445, "learning_rate": 4.901020039718361e-05, "loss": 5.2505, "step": 2194 }, { "epoch": 0.019814045856652824, "grad_norm": 3.4445767402648926, "learning_rate": 4.900974905217549e-05, "loss": 5.2928, "step": 2195 }, { "epoch": 0.01982307275681531, "grad_norm": 3.1856279373168945, "learning_rate": 4.900929770716736e-05, "loss": 4.8878, "step": 2196 }, { "epoch": 0.019832099656977794, "grad_norm": 3.4881350994110107, "learning_rate": 4.900884636215924e-05, "loss": 5.0016, "step": 2197 }, { "epoch": 0.01984112655714028, "grad_norm": 3.995633840560913, "learning_rate": 4.900839501715111e-05, "loss": 5.0238, "step": 2198 }, { "epoch": 0.019850153457302764, "grad_norm": 4.105778217315674, "learning_rate": 4.9007943672142985e-05, "loss": 5.1778, "step": 2199 }, { "epoch": 0.019859180357465245, "grad_norm": 3.555814266204834, "learning_rate": 4.9007492327134864e-05, "loss": 4.6355, "step": 2200 }, { "epoch": 0.01986820725762773, "grad_norm": 3.71405291557312, "learning_rate": 4.9007040982126736e-05, "loss": 4.6359, "step": 2201 }, { "epoch": 0.019877234157790215, "grad_norm": 3.446345090866089, "learning_rate": 4.9006589637118616e-05, "loss": 5.3662, "step": 2202 }, { "epoch": 0.0198862610579527, "grad_norm": 4.899185657501221, "learning_rate": 4.9006138292110495e-05, "loss": 4.7255, "step": 2203 }, { "epoch": 0.019895287958115182, "grad_norm": 3.4806723594665527, "learning_rate": 4.900568694710237e-05, "loss": 4.6855, "step": 2204 }, { "epoch": 0.019904314858277667, "grad_norm": 3.839052677154541, "learning_rate": 4.900523560209424e-05, "loss": 4.6484, "step": 2205 }, { "epoch": 0.019913341758440152, "grad_norm": 3.9612693786621094, "learning_rate": 4.900478425708612e-05, "loss": 5.35, "step": 2206 }, { "epoch": 0.019922368658602637, "grad_norm": 5.589084148406982, "learning_rate": 4.900433291207799e-05, "loss": 5.3694, "step": 2207 }, { "epoch": 0.01993139555876512, "grad_norm": 4.4648823738098145, "learning_rate": 4.900388156706987e-05, "loss": 5.1296, "step": 2208 }, { "epoch": 0.019940422458927604, "grad_norm": 3.9645586013793945, "learning_rate": 4.900343022206175e-05, "loss": 5.0545, "step": 2209 }, { "epoch": 0.01994944935909009, "grad_norm": 4.080862998962402, "learning_rate": 4.900297887705362e-05, "loss": 4.7835, "step": 2210 }, { "epoch": 0.019958476259252574, "grad_norm": 5.137067794799805, "learning_rate": 4.90025275320455e-05, "loss": 4.6892, "step": 2211 }, { "epoch": 0.01996750315941506, "grad_norm": 3.8021907806396484, "learning_rate": 4.9002076187037374e-05, "loss": 4.9839, "step": 2212 }, { "epoch": 0.01997653005957754, "grad_norm": 2.898737668991089, "learning_rate": 4.900162484202925e-05, "loss": 4.924, "step": 2213 }, { "epoch": 0.019985556959740025, "grad_norm": 6.471345901489258, "learning_rate": 4.9001173497021126e-05, "loss": 4.7766, "step": 2214 }, { "epoch": 0.01999458385990251, "grad_norm": 4.507112979888916, "learning_rate": 4.9000722152013e-05, "loss": 4.7373, "step": 2215 }, { "epoch": 0.020003610760064995, "grad_norm": 4.146306037902832, "learning_rate": 4.900027080700488e-05, "loss": 4.8709, "step": 2216 }, { "epoch": 0.020012637660227477, "grad_norm": 3.821702003479004, "learning_rate": 4.899981946199676e-05, "loss": 5.1452, "step": 2217 }, { "epoch": 0.02002166456038996, "grad_norm": 3.0335283279418945, "learning_rate": 4.899936811698863e-05, "loss": 5.1388, "step": 2218 }, { "epoch": 0.020030691460552447, "grad_norm": 5.508620738983154, "learning_rate": 4.89989167719805e-05, "loss": 4.237, "step": 2219 }, { "epoch": 0.02003971836071493, "grad_norm": 2.8284881114959717, "learning_rate": 4.899846542697238e-05, "loss": 5.2399, "step": 2220 }, { "epoch": 0.020048745260877413, "grad_norm": 3.2184221744537354, "learning_rate": 4.8998014081964254e-05, "loss": 4.2429, "step": 2221 }, { "epoch": 0.020057772161039898, "grad_norm": 4.171184062957764, "learning_rate": 4.899756273695613e-05, "loss": 5.0622, "step": 2222 }, { "epoch": 0.020066799061202383, "grad_norm": 4.816786766052246, "learning_rate": 4.8997111391948005e-05, "loss": 5.3074, "step": 2223 }, { "epoch": 0.020075825961364868, "grad_norm": 3.9916255474090576, "learning_rate": 4.8996660046939884e-05, "loss": 5.3658, "step": 2224 }, { "epoch": 0.020084852861527353, "grad_norm": 4.451657295227051, "learning_rate": 4.8996208701931764e-05, "loss": 5.5382, "step": 2225 }, { "epoch": 0.020093879761689835, "grad_norm": 3.717599868774414, "learning_rate": 4.899575735692363e-05, "loss": 5.4274, "step": 2226 }, { "epoch": 0.02010290666185232, "grad_norm": 18.395784378051758, "learning_rate": 4.899530601191551e-05, "loss": 5.4677, "step": 2227 }, { "epoch": 0.020111933562014805, "grad_norm": 5.087182521820068, "learning_rate": 4.899485466690739e-05, "loss": 5.0069, "step": 2228 }, { "epoch": 0.02012096046217729, "grad_norm": 3.1679539680480957, "learning_rate": 4.899440332189926e-05, "loss": 4.6954, "step": 2229 }, { "epoch": 0.02012998736233977, "grad_norm": 7.861802101135254, "learning_rate": 4.899395197689114e-05, "loss": 4.8414, "step": 2230 }, { "epoch": 0.020139014262502256, "grad_norm": 4.255713939666748, "learning_rate": 4.899350063188302e-05, "loss": 4.8322, "step": 2231 }, { "epoch": 0.02014804116266474, "grad_norm": 3.5866479873657227, "learning_rate": 4.8993049286874885e-05, "loss": 4.606, "step": 2232 }, { "epoch": 0.020157068062827226, "grad_norm": 4.168229579925537, "learning_rate": 4.8992597941866764e-05, "loss": 4.5223, "step": 2233 }, { "epoch": 0.020166094962989708, "grad_norm": 4.955402374267578, "learning_rate": 4.899214659685864e-05, "loss": 4.7799, "step": 2234 }, { "epoch": 0.020175121863152193, "grad_norm": 4.7436203956604, "learning_rate": 4.8991695251850515e-05, "loss": 4.4919, "step": 2235 }, { "epoch": 0.020184148763314678, "grad_norm": 3.4240097999572754, "learning_rate": 4.8991243906842395e-05, "loss": 5.3487, "step": 2236 }, { "epoch": 0.020193175663477163, "grad_norm": 4.801209926605225, "learning_rate": 4.899079256183427e-05, "loss": 3.9394, "step": 2237 }, { "epoch": 0.020202202563639648, "grad_norm": 2.894502878189087, "learning_rate": 4.8990341216826146e-05, "loss": 5.0876, "step": 2238 }, { "epoch": 0.02021122946380213, "grad_norm": 3.4079418182373047, "learning_rate": 4.898988987181802e-05, "loss": 5.06, "step": 2239 }, { "epoch": 0.020220256363964614, "grad_norm": 3.9236114025115967, "learning_rate": 4.898943852680989e-05, "loss": 4.5498, "step": 2240 }, { "epoch": 0.0202292832641271, "grad_norm": 3.7197587490081787, "learning_rate": 4.898898718180177e-05, "loss": 4.7156, "step": 2241 }, { "epoch": 0.020238310164289584, "grad_norm": 2.8777782917022705, "learning_rate": 4.898853583679365e-05, "loss": 4.5882, "step": 2242 }, { "epoch": 0.020247337064452066, "grad_norm": 4.145806789398193, "learning_rate": 4.898808449178552e-05, "loss": 5.0886, "step": 2243 }, { "epoch": 0.02025636396461455, "grad_norm": 4.777920246124268, "learning_rate": 4.89876331467774e-05, "loss": 5.1458, "step": 2244 }, { "epoch": 0.020265390864777036, "grad_norm": 3.6017820835113525, "learning_rate": 4.898718180176928e-05, "loss": 5.0734, "step": 2245 }, { "epoch": 0.02027441776493952, "grad_norm": 4.032859802246094, "learning_rate": 4.8986730456761146e-05, "loss": 4.7774, "step": 2246 }, { "epoch": 0.020283444665102002, "grad_norm": 2.8428430557250977, "learning_rate": 4.8986279111753026e-05, "loss": 4.9242, "step": 2247 }, { "epoch": 0.020292471565264487, "grad_norm": 4.275248050689697, "learning_rate": 4.8985827766744905e-05, "loss": 4.1701, "step": 2248 }, { "epoch": 0.020301498465426972, "grad_norm": 3.9870996475219727, "learning_rate": 4.898537642173678e-05, "loss": 4.4487, "step": 2249 }, { "epoch": 0.020310525365589457, "grad_norm": 5.170536041259766, "learning_rate": 4.898492507672866e-05, "loss": 5.3323, "step": 2250 }, { "epoch": 0.020319552265751942, "grad_norm": 3.3743841648101807, "learning_rate": 4.898447373172053e-05, "loss": 4.621, "step": 2251 }, { "epoch": 0.020328579165914424, "grad_norm": 3.9962258338928223, "learning_rate": 4.89840223867124e-05, "loss": 4.6512, "step": 2252 }, { "epoch": 0.02033760606607691, "grad_norm": 5.509644508361816, "learning_rate": 4.898357104170428e-05, "loss": 3.6041, "step": 2253 }, { "epoch": 0.020346632966239394, "grad_norm": 4.376550197601318, "learning_rate": 4.898311969669615e-05, "loss": 3.7563, "step": 2254 }, { "epoch": 0.02035565986640188, "grad_norm": 3.5747666358947754, "learning_rate": 4.898266835168803e-05, "loss": 5.2235, "step": 2255 }, { "epoch": 0.02036468676656436, "grad_norm": 3.507148504257202, "learning_rate": 4.898221700667991e-05, "loss": 4.8415, "step": 2256 }, { "epoch": 0.020373713666726846, "grad_norm": 2.7094309329986572, "learning_rate": 4.8981765661671784e-05, "loss": 5.0107, "step": 2257 }, { "epoch": 0.02038274056688933, "grad_norm": 3.9087345600128174, "learning_rate": 4.8981314316663663e-05, "loss": 4.6192, "step": 2258 }, { "epoch": 0.020391767467051816, "grad_norm": 4.164226055145264, "learning_rate": 4.8980862971655536e-05, "loss": 3.9205, "step": 2259 }, { "epoch": 0.020400794367214297, "grad_norm": 3.4427671432495117, "learning_rate": 4.898041162664741e-05, "loss": 4.5035, "step": 2260 }, { "epoch": 0.020409821267376782, "grad_norm": 5.815817832946777, "learning_rate": 4.897996028163929e-05, "loss": 4.3426, "step": 2261 }, { "epoch": 0.020418848167539267, "grad_norm": 3.7097086906433105, "learning_rate": 4.897950893663116e-05, "loss": 4.7012, "step": 2262 }, { "epoch": 0.020427875067701752, "grad_norm": 3.639451265335083, "learning_rate": 4.897905759162304e-05, "loss": 4.9547, "step": 2263 }, { "epoch": 0.020436901967864237, "grad_norm": 3.162842035293579, "learning_rate": 4.897860624661492e-05, "loss": 4.8606, "step": 2264 }, { "epoch": 0.02044592886802672, "grad_norm": 4.394318580627441, "learning_rate": 4.897815490160679e-05, "loss": 4.8507, "step": 2265 }, { "epoch": 0.020454955768189204, "grad_norm": 3.297410726547241, "learning_rate": 4.8977703556598664e-05, "loss": 4.6754, "step": 2266 }, { "epoch": 0.02046398266835169, "grad_norm": 2.527470827102661, "learning_rate": 4.897725221159054e-05, "loss": 4.655, "step": 2267 }, { "epoch": 0.020473009568514174, "grad_norm": 4.127908229827881, "learning_rate": 4.8976800866582415e-05, "loss": 4.7862, "step": 2268 }, { "epoch": 0.020482036468676655, "grad_norm": 4.059902667999268, "learning_rate": 4.8976349521574294e-05, "loss": 4.7026, "step": 2269 }, { "epoch": 0.02049106336883914, "grad_norm": 3.7126901149749756, "learning_rate": 4.8975898176566174e-05, "loss": 5.0606, "step": 2270 }, { "epoch": 0.020500090269001625, "grad_norm": 4.493852615356445, "learning_rate": 4.8975446831558046e-05, "loss": 5.0701, "step": 2271 }, { "epoch": 0.02050911716916411, "grad_norm": 4.190615177154541, "learning_rate": 4.8974995486549925e-05, "loss": 5.1939, "step": 2272 }, { "epoch": 0.02051814406932659, "grad_norm": 5.385256290435791, "learning_rate": 4.89745441415418e-05, "loss": 4.4391, "step": 2273 }, { "epoch": 0.020527170969489077, "grad_norm": 3.3228137493133545, "learning_rate": 4.897409279653367e-05, "loss": 5.5739, "step": 2274 }, { "epoch": 0.02053619786965156, "grad_norm": 3.54113507270813, "learning_rate": 4.897364145152555e-05, "loss": 4.5764, "step": 2275 }, { "epoch": 0.020545224769814047, "grad_norm": 3.7365427017211914, "learning_rate": 4.897319010651742e-05, "loss": 4.2616, "step": 2276 }, { "epoch": 0.02055425166997653, "grad_norm": 6.104538440704346, "learning_rate": 4.89727387615093e-05, "loss": 4.4608, "step": 2277 }, { "epoch": 0.020563278570139013, "grad_norm": 7.042216777801514, "learning_rate": 4.897228741650118e-05, "loss": 4.1969, "step": 2278 }, { "epoch": 0.020572305470301498, "grad_norm": 4.584319591522217, "learning_rate": 4.8971836071493046e-05, "loss": 4.9151, "step": 2279 }, { "epoch": 0.020581332370463983, "grad_norm": 3.9642326831817627, "learning_rate": 4.8971384726484925e-05, "loss": 4.8621, "step": 2280 }, { "epoch": 0.020590359270626468, "grad_norm": 5.239227771759033, "learning_rate": 4.8970933381476805e-05, "loss": 4.5743, "step": 2281 }, { "epoch": 0.02059938617078895, "grad_norm": 4.057293891906738, "learning_rate": 4.897048203646868e-05, "loss": 5.0576, "step": 2282 }, { "epoch": 0.020608413070951435, "grad_norm": 4.934315204620361, "learning_rate": 4.8970030691460556e-05, "loss": 4.7293, "step": 2283 }, { "epoch": 0.02061743997111392, "grad_norm": 4.660378932952881, "learning_rate": 4.896957934645243e-05, "loss": 4.9787, "step": 2284 }, { "epoch": 0.020626466871276405, "grad_norm": 3.659912109375, "learning_rate": 4.896912800144431e-05, "loss": 4.6857, "step": 2285 }, { "epoch": 0.020635493771438886, "grad_norm": 3.202472686767578, "learning_rate": 4.896867665643618e-05, "loss": 4.5384, "step": 2286 }, { "epoch": 0.02064452067160137, "grad_norm": 3.587529182434082, "learning_rate": 4.896822531142806e-05, "loss": 4.8368, "step": 2287 }, { "epoch": 0.020653547571763856, "grad_norm": 4.772915840148926, "learning_rate": 4.896777396641993e-05, "loss": 3.8257, "step": 2288 }, { "epoch": 0.02066257447192634, "grad_norm": 4.049070835113525, "learning_rate": 4.896732262141181e-05, "loss": 5.4557, "step": 2289 }, { "epoch": 0.020671601372088826, "grad_norm": 3.8500213623046875, "learning_rate": 4.8966871276403684e-05, "loss": 4.8758, "step": 2290 }, { "epoch": 0.020680628272251308, "grad_norm": 4.447655200958252, "learning_rate": 4.896641993139556e-05, "loss": 4.8991, "step": 2291 }, { "epoch": 0.020689655172413793, "grad_norm": 3.7658133506774902, "learning_rate": 4.896596858638744e-05, "loss": 5.0227, "step": 2292 }, { "epoch": 0.020698682072576278, "grad_norm": 4.494457244873047, "learning_rate": 4.896551724137931e-05, "loss": 4.6617, "step": 2293 }, { "epoch": 0.020707708972738763, "grad_norm": 3.7933428287506104, "learning_rate": 4.896506589637119e-05, "loss": 5.4903, "step": 2294 }, { "epoch": 0.020716735872901244, "grad_norm": 4.29585599899292, "learning_rate": 4.896461455136307e-05, "loss": 5.2767, "step": 2295 }, { "epoch": 0.02072576277306373, "grad_norm": 3.85876202583313, "learning_rate": 4.896416320635494e-05, "loss": 4.4114, "step": 2296 }, { "epoch": 0.020734789673226214, "grad_norm": 4.9228315353393555, "learning_rate": 4.896371186134682e-05, "loss": 5.3083, "step": 2297 }, { "epoch": 0.0207438165733887, "grad_norm": 4.452271938323975, "learning_rate": 4.896326051633869e-05, "loss": 5.2521, "step": 2298 }, { "epoch": 0.02075284347355118, "grad_norm": 4.659542560577393, "learning_rate": 4.896280917133056e-05, "loss": 4.815, "step": 2299 }, { "epoch": 0.020761870373713666, "grad_norm": 4.902070999145508, "learning_rate": 4.896235782632244e-05, "loss": 5.2881, "step": 2300 }, { "epoch": 0.02077089727387615, "grad_norm": 4.316460132598877, "learning_rate": 4.8961906481314315e-05, "loss": 5.1953, "step": 2301 }, { "epoch": 0.020779924174038636, "grad_norm": 3.537468433380127, "learning_rate": 4.8961455136306194e-05, "loss": 5.0149, "step": 2302 }, { "epoch": 0.02078895107420112, "grad_norm": 4.438207149505615, "learning_rate": 4.8961003791298073e-05, "loss": 4.8077, "step": 2303 }, { "epoch": 0.020797977974363602, "grad_norm": 6.136965751647949, "learning_rate": 4.8960552446289946e-05, "loss": 5.1996, "step": 2304 }, { "epoch": 0.020807004874526087, "grad_norm": 4.828881740570068, "learning_rate": 4.8960101101281825e-05, "loss": 3.8165, "step": 2305 }, { "epoch": 0.020816031774688572, "grad_norm": 3.332846164703369, "learning_rate": 4.89596497562737e-05, "loss": 5.2864, "step": 2306 }, { "epoch": 0.020825058674851057, "grad_norm": 3.6147894859313965, "learning_rate": 4.895919841126557e-05, "loss": 5.0069, "step": 2307 }, { "epoch": 0.02083408557501354, "grad_norm": 4.589515209197998, "learning_rate": 4.895874706625745e-05, "loss": 4.6689, "step": 2308 }, { "epoch": 0.020843112475176024, "grad_norm": 7.490836143493652, "learning_rate": 4.895829572124933e-05, "loss": 5.0677, "step": 2309 }, { "epoch": 0.02085213937533851, "grad_norm": 3.953176736831665, "learning_rate": 4.89578443762412e-05, "loss": 5.0721, "step": 2310 }, { "epoch": 0.020861166275500994, "grad_norm": 10.887181282043457, "learning_rate": 4.895739303123308e-05, "loss": 4.0238, "step": 2311 }, { "epoch": 0.020870193175663476, "grad_norm": 3.829178810119629, "learning_rate": 4.895694168622495e-05, "loss": 5.4207, "step": 2312 }, { "epoch": 0.02087922007582596, "grad_norm": 7.4344258308410645, "learning_rate": 4.8956490341216825e-05, "loss": 5.192, "step": 2313 }, { "epoch": 0.020888246975988446, "grad_norm": 3.42988920211792, "learning_rate": 4.8956038996208704e-05, "loss": 5.0859, "step": 2314 }, { "epoch": 0.02089727387615093, "grad_norm": 5.837100028991699, "learning_rate": 4.895558765120058e-05, "loss": 4.6322, "step": 2315 }, { "epoch": 0.020906300776313416, "grad_norm": 3.6857283115386963, "learning_rate": 4.8955136306192456e-05, "loss": 4.633, "step": 2316 }, { "epoch": 0.020915327676475897, "grad_norm": 3.3874950408935547, "learning_rate": 4.8954684961184335e-05, "loss": 4.5543, "step": 2317 }, { "epoch": 0.020924354576638382, "grad_norm": 4.288236141204834, "learning_rate": 4.895423361617621e-05, "loss": 5.1166, "step": 2318 }, { "epoch": 0.020933381476800867, "grad_norm": 3.8316240310668945, "learning_rate": 4.895378227116808e-05, "loss": 4.704, "step": 2319 }, { "epoch": 0.020942408376963352, "grad_norm": 4.595935344696045, "learning_rate": 4.895333092615996e-05, "loss": 4.8012, "step": 2320 }, { "epoch": 0.020951435277125834, "grad_norm": 4.184163570404053, "learning_rate": 4.895287958115183e-05, "loss": 5.2339, "step": 2321 }, { "epoch": 0.02096046217728832, "grad_norm": 4.442155838012695, "learning_rate": 4.895242823614371e-05, "loss": 4.8942, "step": 2322 }, { "epoch": 0.020969489077450804, "grad_norm": 3.4774885177612305, "learning_rate": 4.8951976891135584e-05, "loss": 4.6236, "step": 2323 }, { "epoch": 0.02097851597761329, "grad_norm": 4.1495537757873535, "learning_rate": 4.895152554612746e-05, "loss": 3.7141, "step": 2324 }, { "epoch": 0.02098754287777577, "grad_norm": 5.403359889984131, "learning_rate": 4.895107420111934e-05, "loss": 3.9752, "step": 2325 }, { "epoch": 0.020996569777938255, "grad_norm": 4.15499210357666, "learning_rate": 4.895062285611121e-05, "loss": 4.2555, "step": 2326 }, { "epoch": 0.02100559667810074, "grad_norm": 3.515854835510254, "learning_rate": 4.895017151110309e-05, "loss": 4.5884, "step": 2327 }, { "epoch": 0.021014623578263225, "grad_norm": 4.851097583770752, "learning_rate": 4.8949720166094966e-05, "loss": 3.7885, "step": 2328 }, { "epoch": 0.02102365047842571, "grad_norm": 4.662252902984619, "learning_rate": 4.894926882108684e-05, "loss": 4.0343, "step": 2329 }, { "epoch": 0.02103267737858819, "grad_norm": 5.097886085510254, "learning_rate": 4.894881747607872e-05, "loss": 5.3278, "step": 2330 }, { "epoch": 0.021041704278750677, "grad_norm": 3.1396994590759277, "learning_rate": 4.89483661310706e-05, "loss": 4.4554, "step": 2331 }, { "epoch": 0.02105073117891316, "grad_norm": 4.0733208656311035, "learning_rate": 4.894791478606247e-05, "loss": 4.9778, "step": 2332 }, { "epoch": 0.021059758079075647, "grad_norm": 4.071434497833252, "learning_rate": 4.894746344105434e-05, "loss": 4.6095, "step": 2333 }, { "epoch": 0.021068784979238128, "grad_norm": 3.3036465644836426, "learning_rate": 4.894701209604622e-05, "loss": 5.5016, "step": 2334 }, { "epoch": 0.021077811879400613, "grad_norm": 4.808426380157471, "learning_rate": 4.8946560751038094e-05, "loss": 4.7353, "step": 2335 }, { "epoch": 0.021086838779563098, "grad_norm": 4.547458171844482, "learning_rate": 4.894610940602997e-05, "loss": 4.4068, "step": 2336 }, { "epoch": 0.021095865679725583, "grad_norm": 4.602736473083496, "learning_rate": 4.8945658061021846e-05, "loss": 5.0616, "step": 2337 }, { "epoch": 0.021104892579888065, "grad_norm": 3.9004359245300293, "learning_rate": 4.8945206716013725e-05, "loss": 5.0237, "step": 2338 }, { "epoch": 0.02111391948005055, "grad_norm": 3.4868288040161133, "learning_rate": 4.8944755371005604e-05, "loss": 5.5264, "step": 2339 }, { "epoch": 0.021122946380213035, "grad_norm": 4.337189197540283, "learning_rate": 4.894430402599747e-05, "loss": 4.9703, "step": 2340 }, { "epoch": 0.02113197328037552, "grad_norm": 3.656524181365967, "learning_rate": 4.894385268098935e-05, "loss": 5.2824, "step": 2341 }, { "epoch": 0.021141000180538005, "grad_norm": 4.452551364898682, "learning_rate": 4.894340133598123e-05, "loss": 4.937, "step": 2342 }, { "epoch": 0.021150027080700486, "grad_norm": 4.429129123687744, "learning_rate": 4.89429499909731e-05, "loss": 5.2321, "step": 2343 }, { "epoch": 0.02115905398086297, "grad_norm": 4.202645778656006, "learning_rate": 4.894249864596498e-05, "loss": 4.8392, "step": 2344 }, { "epoch": 0.021168080881025456, "grad_norm": 4.402524948120117, "learning_rate": 4.894204730095685e-05, "loss": 4.1011, "step": 2345 }, { "epoch": 0.02117710778118794, "grad_norm": 4.549295902252197, "learning_rate": 4.8941595955948725e-05, "loss": 4.1355, "step": 2346 }, { "epoch": 0.021186134681350423, "grad_norm": 3.330125331878662, "learning_rate": 4.8941144610940604e-05, "loss": 4.4796, "step": 2347 }, { "epoch": 0.021195161581512908, "grad_norm": 4.1525774002075195, "learning_rate": 4.8940693265932483e-05, "loss": 4.8059, "step": 2348 }, { "epoch": 0.021204188481675393, "grad_norm": 3.5510058403015137, "learning_rate": 4.8940241920924356e-05, "loss": 4.5419, "step": 2349 }, { "epoch": 0.021213215381837878, "grad_norm": 3.2684552669525146, "learning_rate": 4.8939790575916235e-05, "loss": 4.5388, "step": 2350 }, { "epoch": 0.02122224228200036, "grad_norm": 4.259711742401123, "learning_rate": 4.893933923090811e-05, "loss": 4.1465, "step": 2351 }, { "epoch": 0.021231269182162844, "grad_norm": 3.3491384983062744, "learning_rate": 4.893888788589999e-05, "loss": 4.33, "step": 2352 }, { "epoch": 0.02124029608232533, "grad_norm": 3.7165324687957764, "learning_rate": 4.893843654089186e-05, "loss": 4.9203, "step": 2353 }, { "epoch": 0.021249322982487814, "grad_norm": 3.8257460594177246, "learning_rate": 4.893798519588373e-05, "loss": 4.2208, "step": 2354 }, { "epoch": 0.0212583498826503, "grad_norm": 2.93959903717041, "learning_rate": 4.893753385087561e-05, "loss": 5.0501, "step": 2355 }, { "epoch": 0.02126737678281278, "grad_norm": 4.193984031677246, "learning_rate": 4.893708250586749e-05, "loss": 4.2475, "step": 2356 }, { "epoch": 0.021276403682975266, "grad_norm": 3.053098678588867, "learning_rate": 4.893663116085936e-05, "loss": 4.9044, "step": 2357 }, { "epoch": 0.02128543058313775, "grad_norm": 3.5306577682495117, "learning_rate": 4.893617981585124e-05, "loss": 5.0543, "step": 2358 }, { "epoch": 0.021294457483300236, "grad_norm": 4.12740421295166, "learning_rate": 4.8935728470843114e-05, "loss": 4.7149, "step": 2359 }, { "epoch": 0.021303484383462717, "grad_norm": 2.5092146396636963, "learning_rate": 4.893527712583499e-05, "loss": 5.0817, "step": 2360 }, { "epoch": 0.021312511283625202, "grad_norm": 4.670712947845459, "learning_rate": 4.8934825780826866e-05, "loss": 5.4274, "step": 2361 }, { "epoch": 0.021321538183787687, "grad_norm": 3.0326619148254395, "learning_rate": 4.893437443581874e-05, "loss": 4.9602, "step": 2362 }, { "epoch": 0.021330565083950172, "grad_norm": 4.448103904724121, "learning_rate": 4.893392309081062e-05, "loss": 4.9886, "step": 2363 }, { "epoch": 0.021339591984112654, "grad_norm": 4.681454181671143, "learning_rate": 4.89334717458025e-05, "loss": 4.9581, "step": 2364 }, { "epoch": 0.02134861888427514, "grad_norm": 2.97066593170166, "learning_rate": 4.893302040079437e-05, "loss": 4.7964, "step": 2365 }, { "epoch": 0.021357645784437624, "grad_norm": 4.457521915435791, "learning_rate": 4.893256905578624e-05, "loss": 4.4075, "step": 2366 }, { "epoch": 0.02136667268460011, "grad_norm": 3.1871681213378906, "learning_rate": 4.893211771077812e-05, "loss": 4.7642, "step": 2367 }, { "epoch": 0.021375699584762594, "grad_norm": 3.9886999130249023, "learning_rate": 4.8931666365769994e-05, "loss": 4.4215, "step": 2368 }, { "epoch": 0.021384726484925076, "grad_norm": 4.156190395355225, "learning_rate": 4.893121502076187e-05, "loss": 4.8589, "step": 2369 }, { "epoch": 0.02139375338508756, "grad_norm": 4.651899337768555, "learning_rate": 4.893076367575375e-05, "loss": 3.913, "step": 2370 }, { "epoch": 0.021402780285250046, "grad_norm": 3.66949725151062, "learning_rate": 4.8930312330745625e-05, "loss": 4.879, "step": 2371 }, { "epoch": 0.02141180718541253, "grad_norm": 3.9278831481933594, "learning_rate": 4.8929860985737504e-05, "loss": 4.8123, "step": 2372 }, { "epoch": 0.021420834085575012, "grad_norm": 3.710794687271118, "learning_rate": 4.8929409640729376e-05, "loss": 5.4876, "step": 2373 }, { "epoch": 0.021429860985737497, "grad_norm": 3.831993579864502, "learning_rate": 4.892895829572125e-05, "loss": 4.6314, "step": 2374 }, { "epoch": 0.021438887885899982, "grad_norm": 3.3488855361938477, "learning_rate": 4.892850695071313e-05, "loss": 4.6485, "step": 2375 }, { "epoch": 0.021447914786062467, "grad_norm": 4.030371189117432, "learning_rate": 4.8928055605705e-05, "loss": 4.7039, "step": 2376 }, { "epoch": 0.021456941686224952, "grad_norm": 3.7886855602264404, "learning_rate": 4.892760426069688e-05, "loss": 4.7825, "step": 2377 }, { "epoch": 0.021465968586387434, "grad_norm": 4.077943801879883, "learning_rate": 4.892715291568876e-05, "loss": 4.4289, "step": 2378 }, { "epoch": 0.02147499548654992, "grad_norm": 3.608121395111084, "learning_rate": 4.892670157068063e-05, "loss": 5.348, "step": 2379 }, { "epoch": 0.021484022386712404, "grad_norm": 6.168067455291748, "learning_rate": 4.8926250225672504e-05, "loss": 4.626, "step": 2380 }, { "epoch": 0.02149304928687489, "grad_norm": 5.030329704284668, "learning_rate": 4.892579888066438e-05, "loss": 5.5501, "step": 2381 }, { "epoch": 0.02150207618703737, "grad_norm": 3.2637977600097656, "learning_rate": 4.8925347535656256e-05, "loss": 4.6501, "step": 2382 }, { "epoch": 0.021511103087199855, "grad_norm": 4.056182861328125, "learning_rate": 4.8924896190648135e-05, "loss": 5.0421, "step": 2383 }, { "epoch": 0.02152012998736234, "grad_norm": 4.097090244293213, "learning_rate": 4.892444484564001e-05, "loss": 4.7693, "step": 2384 }, { "epoch": 0.021529156887524825, "grad_norm": 3.5993173122406006, "learning_rate": 4.892399350063189e-05, "loss": 4.8876, "step": 2385 }, { "epoch": 0.021538183787687307, "grad_norm": 4.431774139404297, "learning_rate": 4.8923542155623766e-05, "loss": 4.9631, "step": 2386 }, { "epoch": 0.02154721068784979, "grad_norm": 3.7706823348999023, "learning_rate": 4.892309081061563e-05, "loss": 5.005, "step": 2387 }, { "epoch": 0.021556237588012277, "grad_norm": 3.7709298133850098, "learning_rate": 4.892263946560751e-05, "loss": 4.4503, "step": 2388 }, { "epoch": 0.02156526448817476, "grad_norm": 5.0454206466674805, "learning_rate": 4.892218812059939e-05, "loss": 4.186, "step": 2389 }, { "epoch": 0.021574291388337247, "grad_norm": 3.16567063331604, "learning_rate": 4.892173677559126e-05, "loss": 4.6533, "step": 2390 }, { "epoch": 0.021583318288499728, "grad_norm": 3.983018636703491, "learning_rate": 4.892128543058314e-05, "loss": 5.0823, "step": 2391 }, { "epoch": 0.021592345188662213, "grad_norm": 6.034430027008057, "learning_rate": 4.892083408557502e-05, "loss": 4.8728, "step": 2392 }, { "epoch": 0.021601372088824698, "grad_norm": 4.7604570388793945, "learning_rate": 4.892038274056689e-05, "loss": 4.7146, "step": 2393 }, { "epoch": 0.021610398988987183, "grad_norm": 3.540813446044922, "learning_rate": 4.8919931395558766e-05, "loss": 4.4189, "step": 2394 }, { "epoch": 0.021619425889149665, "grad_norm": 6.3559417724609375, "learning_rate": 4.8919480050550645e-05, "loss": 4.5113, "step": 2395 }, { "epoch": 0.02162845278931215, "grad_norm": 5.110837459564209, "learning_rate": 4.891902870554252e-05, "loss": 5.1396, "step": 2396 }, { "epoch": 0.021637479689474635, "grad_norm": 3.512455701828003, "learning_rate": 4.89185773605344e-05, "loss": 4.8803, "step": 2397 }, { "epoch": 0.02164650658963712, "grad_norm": 4.050800800323486, "learning_rate": 4.891812601552627e-05, "loss": 4.6137, "step": 2398 }, { "epoch": 0.0216555334897996, "grad_norm": 5.293701171875, "learning_rate": 4.891767467051815e-05, "loss": 4.1325, "step": 2399 }, { "epoch": 0.021664560389962086, "grad_norm": 4.429774761199951, "learning_rate": 4.891722332551002e-05, "loss": 4.8866, "step": 2400 }, { "epoch": 0.02167358729012457, "grad_norm": 8.100939750671387, "learning_rate": 4.8916771980501893e-05, "loss": 5.5059, "step": 2401 }, { "epoch": 0.021682614190287056, "grad_norm": 3.3652679920196533, "learning_rate": 4.891632063549377e-05, "loss": 4.6038, "step": 2402 }, { "epoch": 0.02169164109044954, "grad_norm": 3.4692437648773193, "learning_rate": 4.891586929048565e-05, "loss": 4.4108, "step": 2403 }, { "epoch": 0.021700667990612023, "grad_norm": 5.352600574493408, "learning_rate": 4.8915417945477524e-05, "loss": 4.6354, "step": 2404 }, { "epoch": 0.021709694890774508, "grad_norm": 4.559988021850586, "learning_rate": 4.8914966600469404e-05, "loss": 5.0846, "step": 2405 }, { "epoch": 0.021718721790936993, "grad_norm": 4.4177656173706055, "learning_rate": 4.891451525546128e-05, "loss": 5.0661, "step": 2406 }, { "epoch": 0.021727748691099478, "grad_norm": 4.362973213195801, "learning_rate": 4.891406391045315e-05, "loss": 5.112, "step": 2407 }, { "epoch": 0.02173677559126196, "grad_norm": 3.786949872970581, "learning_rate": 4.891361256544503e-05, "loss": 4.1428, "step": 2408 }, { "epoch": 0.021745802491424444, "grad_norm": 3.4599153995513916, "learning_rate": 4.891316122043691e-05, "loss": 4.8983, "step": 2409 }, { "epoch": 0.02175482939158693, "grad_norm": 4.294157981872559, "learning_rate": 4.891270987542878e-05, "loss": 4.7504, "step": 2410 }, { "epoch": 0.021763856291749414, "grad_norm": 3.9164462089538574, "learning_rate": 4.891225853042066e-05, "loss": 4.4201, "step": 2411 }, { "epoch": 0.021772883191911896, "grad_norm": 3.7598671913146973, "learning_rate": 4.891180718541253e-05, "loss": 4.9258, "step": 2412 }, { "epoch": 0.02178191009207438, "grad_norm": 3.2685513496398926, "learning_rate": 4.8911355840404404e-05, "loss": 5.0129, "step": 2413 }, { "epoch": 0.021790936992236866, "grad_norm": 3.082566261291504, "learning_rate": 4.891090449539628e-05, "loss": 4.6971, "step": 2414 }, { "epoch": 0.02179996389239935, "grad_norm": 3.5816969871520996, "learning_rate": 4.8910453150388155e-05, "loss": 4.6561, "step": 2415 }, { "epoch": 0.021808990792561836, "grad_norm": 3.181408643722534, "learning_rate": 4.8910001805380035e-05, "loss": 4.344, "step": 2416 }, { "epoch": 0.021818017692724317, "grad_norm": 3.9394021034240723, "learning_rate": 4.8909550460371914e-05, "loss": 4.7114, "step": 2417 }, { "epoch": 0.021827044592886802, "grad_norm": 5.1140522956848145, "learning_rate": 4.8909099115363786e-05, "loss": 4.689, "step": 2418 }, { "epoch": 0.021836071493049287, "grad_norm": 3.945066213607788, "learning_rate": 4.8908647770355666e-05, "loss": 4.7885, "step": 2419 }, { "epoch": 0.021845098393211772, "grad_norm": 3.3662190437316895, "learning_rate": 4.890819642534754e-05, "loss": 5.4166, "step": 2420 }, { "epoch": 0.021854125293374254, "grad_norm": 4.573988437652588, "learning_rate": 4.890774508033941e-05, "loss": 5.2008, "step": 2421 }, { "epoch": 0.02186315219353674, "grad_norm": 2.85129451751709, "learning_rate": 4.890729373533129e-05, "loss": 5.2279, "step": 2422 }, { "epoch": 0.021872179093699224, "grad_norm": 3.9711506366729736, "learning_rate": 4.890684239032316e-05, "loss": 4.9569, "step": 2423 }, { "epoch": 0.02188120599386171, "grad_norm": 4.740024089813232, "learning_rate": 4.890639104531504e-05, "loss": 4.3137, "step": 2424 }, { "epoch": 0.02189023289402419, "grad_norm": 5.3427934646606445, "learning_rate": 4.890593970030692e-05, "loss": 4.9375, "step": 2425 }, { "epoch": 0.021899259794186676, "grad_norm": 5.210928916931152, "learning_rate": 4.890548835529879e-05, "loss": 4.5933, "step": 2426 }, { "epoch": 0.02190828669434916, "grad_norm": 3.205674409866333, "learning_rate": 4.8905037010290666e-05, "loss": 4.4853, "step": 2427 }, { "epoch": 0.021917313594511646, "grad_norm": 3.3120858669281006, "learning_rate": 4.8904585665282545e-05, "loss": 4.7864, "step": 2428 }, { "epoch": 0.02192634049467413, "grad_norm": 3.801454544067383, "learning_rate": 4.890413432027442e-05, "loss": 4.3731, "step": 2429 }, { "epoch": 0.021935367394836612, "grad_norm": 3.6407039165496826, "learning_rate": 4.89036829752663e-05, "loss": 3.8949, "step": 2430 }, { "epoch": 0.021944394294999097, "grad_norm": 2.8622772693634033, "learning_rate": 4.8903231630258176e-05, "loss": 4.6776, "step": 2431 }, { "epoch": 0.021953421195161582, "grad_norm": 4.0390095710754395, "learning_rate": 4.890278028525005e-05, "loss": 4.4962, "step": 2432 }, { "epoch": 0.021962448095324067, "grad_norm": 3.0943033695220947, "learning_rate": 4.890232894024193e-05, "loss": 5.1056, "step": 2433 }, { "epoch": 0.02197147499548655, "grad_norm": 2.9589388370513916, "learning_rate": 4.89018775952338e-05, "loss": 5.6241, "step": 2434 }, { "epoch": 0.021980501895649034, "grad_norm": 3.3833725452423096, "learning_rate": 4.890142625022567e-05, "loss": 4.7255, "step": 2435 }, { "epoch": 0.02198952879581152, "grad_norm": 3.346743583679199, "learning_rate": 4.890097490521755e-05, "loss": 4.5393, "step": 2436 }, { "epoch": 0.021998555695974004, "grad_norm": 3.8800606727600098, "learning_rate": 4.8900523560209424e-05, "loss": 4.0168, "step": 2437 }, { "epoch": 0.022007582596136485, "grad_norm": 3.943397045135498, "learning_rate": 4.8900072215201303e-05, "loss": 4.3555, "step": 2438 }, { "epoch": 0.02201660949629897, "grad_norm": 3.2874529361724854, "learning_rate": 4.889962087019318e-05, "loss": 4.5876, "step": 2439 }, { "epoch": 0.022025636396461455, "grad_norm": 3.7696712017059326, "learning_rate": 4.889916952518505e-05, "loss": 4.3884, "step": 2440 }, { "epoch": 0.02203466329662394, "grad_norm": 4.79374885559082, "learning_rate": 4.889871818017693e-05, "loss": 3.5924, "step": 2441 }, { "epoch": 0.022043690196786425, "grad_norm": 5.534491539001465, "learning_rate": 4.889826683516881e-05, "loss": 4.6168, "step": 2442 }, { "epoch": 0.022052717096948907, "grad_norm": 5.061854839324951, "learning_rate": 4.889781549016068e-05, "loss": 4.3242, "step": 2443 }, { "epoch": 0.02206174399711139, "grad_norm": 4.131048202514648, "learning_rate": 4.889736414515256e-05, "loss": 4.5842, "step": 2444 }, { "epoch": 0.022070770897273877, "grad_norm": 4.185222625732422, "learning_rate": 4.889691280014443e-05, "loss": 4.7728, "step": 2445 }, { "epoch": 0.02207979779743636, "grad_norm": 4.987896919250488, "learning_rate": 4.889646145513631e-05, "loss": 4.0614, "step": 2446 }, { "epoch": 0.022088824697598843, "grad_norm": 4.115236282348633, "learning_rate": 4.889601011012818e-05, "loss": 5.2544, "step": 2447 }, { "epoch": 0.022097851597761328, "grad_norm": 3.2218339443206787, "learning_rate": 4.889555876512006e-05, "loss": 5.2029, "step": 2448 }, { "epoch": 0.022106878497923813, "grad_norm": 3.889946222305298, "learning_rate": 4.8895107420111934e-05, "loss": 4.6418, "step": 2449 }, { "epoch": 0.022115905398086298, "grad_norm": 4.412356376647949, "learning_rate": 4.8894656075103814e-05, "loss": 3.7127, "step": 2450 }, { "epoch": 0.02212493229824878, "grad_norm": 4.682716369628906, "learning_rate": 4.8894204730095686e-05, "loss": 5.49, "step": 2451 }, { "epoch": 0.022133959198411265, "grad_norm": 4.628910541534424, "learning_rate": 4.8893753385087565e-05, "loss": 4.7063, "step": 2452 }, { "epoch": 0.02214298609857375, "grad_norm": 3.7342519760131836, "learning_rate": 4.8893302040079445e-05, "loss": 4.5901, "step": 2453 }, { "epoch": 0.022152012998736235, "grad_norm": 4.04961633682251, "learning_rate": 4.889285069507131e-05, "loss": 4.3954, "step": 2454 }, { "epoch": 0.02216103989889872, "grad_norm": 3.2779347896575928, "learning_rate": 4.889239935006319e-05, "loss": 4.0338, "step": 2455 }, { "epoch": 0.0221700667990612, "grad_norm": 4.310755252838135, "learning_rate": 4.889194800505507e-05, "loss": 4.8824, "step": 2456 }, { "epoch": 0.022179093699223686, "grad_norm": 3.516616106033325, "learning_rate": 4.889149666004694e-05, "loss": 4.9224, "step": 2457 }, { "epoch": 0.02218812059938617, "grad_norm": 3.193073034286499, "learning_rate": 4.889104531503882e-05, "loss": 4.7154, "step": 2458 }, { "epoch": 0.022197147499548656, "grad_norm": 4.22499942779541, "learning_rate": 4.889059397003069e-05, "loss": 4.9942, "step": 2459 }, { "epoch": 0.022206174399711138, "grad_norm": 3.793757438659668, "learning_rate": 4.8890142625022565e-05, "loss": 4.6177, "step": 2460 }, { "epoch": 0.022215201299873623, "grad_norm": 3.800037384033203, "learning_rate": 4.8889691280014445e-05, "loss": 4.3244, "step": 2461 }, { "epoch": 0.022224228200036108, "grad_norm": 3.8118910789489746, "learning_rate": 4.888923993500632e-05, "loss": 5.4287, "step": 2462 }, { "epoch": 0.022233255100198593, "grad_norm": 3.769009590148926, "learning_rate": 4.8888788589998196e-05, "loss": 4.9817, "step": 2463 }, { "epoch": 0.022242282000361074, "grad_norm": 3.9854650497436523, "learning_rate": 4.8888337244990076e-05, "loss": 5.1987, "step": 2464 }, { "epoch": 0.02225130890052356, "grad_norm": 3.3311710357666016, "learning_rate": 4.888788589998195e-05, "loss": 5.1035, "step": 2465 }, { "epoch": 0.022260335800686044, "grad_norm": 7.801204681396484, "learning_rate": 4.888743455497383e-05, "loss": 4.7611, "step": 2466 }, { "epoch": 0.02226936270084853, "grad_norm": 5.013595104217529, "learning_rate": 4.88869832099657e-05, "loss": 4.8936, "step": 2467 }, { "epoch": 0.022278389601011014, "grad_norm": 4.262542724609375, "learning_rate": 4.888653186495757e-05, "loss": 4.9257, "step": 2468 }, { "epoch": 0.022287416501173496, "grad_norm": 2.8740451335906982, "learning_rate": 4.888608051994945e-05, "loss": 4.9364, "step": 2469 }, { "epoch": 0.02229644340133598, "grad_norm": 3.173470973968506, "learning_rate": 4.888562917494133e-05, "loss": 5.1479, "step": 2470 }, { "epoch": 0.022305470301498466, "grad_norm": 3.0808050632476807, "learning_rate": 4.88851778299332e-05, "loss": 4.748, "step": 2471 }, { "epoch": 0.02231449720166095, "grad_norm": 3.474363327026367, "learning_rate": 4.888472648492508e-05, "loss": 4.7034, "step": 2472 }, { "epoch": 0.022323524101823432, "grad_norm": 3.9639739990234375, "learning_rate": 4.8884275139916955e-05, "loss": 4.8232, "step": 2473 }, { "epoch": 0.022332551001985917, "grad_norm": 4.566375255584717, "learning_rate": 4.888382379490883e-05, "loss": 4.7823, "step": 2474 }, { "epoch": 0.022341577902148402, "grad_norm": 3.736955404281616, "learning_rate": 4.8883372449900707e-05, "loss": 5.0271, "step": 2475 }, { "epoch": 0.022350604802310887, "grad_norm": 3.065598249435425, "learning_rate": 4.888292110489258e-05, "loss": 5.1624, "step": 2476 }, { "epoch": 0.02235963170247337, "grad_norm": 4.566951274871826, "learning_rate": 4.888246975988446e-05, "loss": 4.9576, "step": 2477 }, { "epoch": 0.022368658602635854, "grad_norm": 5.653655529022217, "learning_rate": 4.888201841487634e-05, "loss": 4.9971, "step": 2478 }, { "epoch": 0.02237768550279834, "grad_norm": 4.916644096374512, "learning_rate": 4.888156706986821e-05, "loss": 4.5494, "step": 2479 }, { "epoch": 0.022386712402960824, "grad_norm": 5.032508850097656, "learning_rate": 4.888111572486008e-05, "loss": 4.366, "step": 2480 }, { "epoch": 0.02239573930312331, "grad_norm": 2.8424015045166016, "learning_rate": 4.888066437985196e-05, "loss": 5.0724, "step": 2481 }, { "epoch": 0.02240476620328579, "grad_norm": 4.503474712371826, "learning_rate": 4.8880213034843834e-05, "loss": 4.2786, "step": 2482 }, { "epoch": 0.022413793103448276, "grad_norm": 3.9832019805908203, "learning_rate": 4.8879761689835713e-05, "loss": 4.8848, "step": 2483 }, { "epoch": 0.02242282000361076, "grad_norm": 3.846277952194214, "learning_rate": 4.8879310344827586e-05, "loss": 4.6358, "step": 2484 }, { "epoch": 0.022431846903773246, "grad_norm": 3.611438035964966, "learning_rate": 4.8878858999819465e-05, "loss": 4.8683, "step": 2485 }, { "epoch": 0.022440873803935727, "grad_norm": 5.298641204833984, "learning_rate": 4.8878407654811344e-05, "loss": 3.7451, "step": 2486 }, { "epoch": 0.022449900704098212, "grad_norm": 3.039792537689209, "learning_rate": 4.887795630980321e-05, "loss": 4.7293, "step": 2487 }, { "epoch": 0.022458927604260697, "grad_norm": 4.472402095794678, "learning_rate": 4.887750496479509e-05, "loss": 5.458, "step": 2488 }, { "epoch": 0.022467954504423182, "grad_norm": 4.104997158050537, "learning_rate": 4.887705361978697e-05, "loss": 4.7818, "step": 2489 }, { "epoch": 0.022476981404585664, "grad_norm": 4.048990726470947, "learning_rate": 4.887660227477884e-05, "loss": 4.4347, "step": 2490 }, { "epoch": 0.02248600830474815, "grad_norm": 4.369925498962402, "learning_rate": 4.887615092977072e-05, "loss": 5.1655, "step": 2491 }, { "epoch": 0.022495035204910634, "grad_norm": 3.298154354095459, "learning_rate": 4.88756995847626e-05, "loss": 4.4996, "step": 2492 }, { "epoch": 0.02250406210507312, "grad_norm": 3.8274495601654053, "learning_rate": 4.887524823975447e-05, "loss": 4.6251, "step": 2493 }, { "epoch": 0.022513089005235604, "grad_norm": 5.651111125946045, "learning_rate": 4.8874796894746344e-05, "loss": 5.2025, "step": 2494 }, { "epoch": 0.022522115905398085, "grad_norm": 3.486948013305664, "learning_rate": 4.8874345549738224e-05, "loss": 4.6788, "step": 2495 }, { "epoch": 0.02253114280556057, "grad_norm": 3.9176411628723145, "learning_rate": 4.8873894204730096e-05, "loss": 4.6098, "step": 2496 }, { "epoch": 0.022540169705723055, "grad_norm": 2.7619333267211914, "learning_rate": 4.8873442859721975e-05, "loss": 4.4558, "step": 2497 }, { "epoch": 0.02254919660588554, "grad_norm": 3.8267412185668945, "learning_rate": 4.887299151471385e-05, "loss": 4.5157, "step": 2498 }, { "epoch": 0.02255822350604802, "grad_norm": 3.650839328765869, "learning_rate": 4.887254016970573e-05, "loss": 5.0932, "step": 2499 }, { "epoch": 0.022567250406210507, "grad_norm": 6.052186489105225, "learning_rate": 4.8872088824697606e-05, "loss": 5.0242, "step": 2500 }, { "epoch": 0.02257627730637299, "grad_norm": 5.868923664093018, "learning_rate": 4.887163747968947e-05, "loss": 4.9538, "step": 2501 }, { "epoch": 0.022585304206535477, "grad_norm": 4.681241035461426, "learning_rate": 4.887118613468135e-05, "loss": 3.6687, "step": 2502 }, { "epoch": 0.022594331106697958, "grad_norm": 3.427708387374878, "learning_rate": 4.887073478967323e-05, "loss": 5.1568, "step": 2503 }, { "epoch": 0.022603358006860443, "grad_norm": 4.182109355926514, "learning_rate": 4.88702834446651e-05, "loss": 4.4104, "step": 2504 }, { "epoch": 0.022612384907022928, "grad_norm": 4.388371467590332, "learning_rate": 4.886983209965698e-05, "loss": 5.0789, "step": 2505 }, { "epoch": 0.022621411807185413, "grad_norm": 3.9871582984924316, "learning_rate": 4.8869380754648855e-05, "loss": 5.0572, "step": 2506 }, { "epoch": 0.022630438707347898, "grad_norm": 3.5635714530944824, "learning_rate": 4.886892940964073e-05, "loss": 4.4251, "step": 2507 }, { "epoch": 0.02263946560751038, "grad_norm": 4.4753522872924805, "learning_rate": 4.8868478064632606e-05, "loss": 4.8338, "step": 2508 }, { "epoch": 0.022648492507672865, "grad_norm": 4.281540393829346, "learning_rate": 4.8868026719624486e-05, "loss": 4.991, "step": 2509 }, { "epoch": 0.02265751940783535, "grad_norm": 3.6427001953125, "learning_rate": 4.886757537461636e-05, "loss": 5.3111, "step": 2510 }, { "epoch": 0.022666546307997835, "grad_norm": 3.4804422855377197, "learning_rate": 4.886712402960824e-05, "loss": 5.0629, "step": 2511 }, { "epoch": 0.022675573208160316, "grad_norm": 5.032567501068115, "learning_rate": 4.886667268460011e-05, "loss": 5.105, "step": 2512 }, { "epoch": 0.0226846001083228, "grad_norm": 3.0663881301879883, "learning_rate": 4.886622133959199e-05, "loss": 4.9734, "step": 2513 }, { "epoch": 0.022693627008485286, "grad_norm": 2.261155843734741, "learning_rate": 4.886576999458386e-05, "loss": 5.3324, "step": 2514 }, { "epoch": 0.02270265390864777, "grad_norm": 4.846475601196289, "learning_rate": 4.8865318649575734e-05, "loss": 5.4299, "step": 2515 }, { "epoch": 0.022711680808810253, "grad_norm": 3.8683106899261475, "learning_rate": 4.886486730456761e-05, "loss": 4.5884, "step": 2516 }, { "epoch": 0.022720707708972738, "grad_norm": 5.027445316314697, "learning_rate": 4.886441595955949e-05, "loss": 4.5176, "step": 2517 }, { "epoch": 0.022729734609135223, "grad_norm": 3.9510862827301025, "learning_rate": 4.8863964614551365e-05, "loss": 4.9211, "step": 2518 }, { "epoch": 0.022738761509297708, "grad_norm": 4.514431953430176, "learning_rate": 4.8863513269543244e-05, "loss": 5.0486, "step": 2519 }, { "epoch": 0.022747788409460193, "grad_norm": 4.01195764541626, "learning_rate": 4.8863061924535117e-05, "loss": 4.9809, "step": 2520 }, { "epoch": 0.022756815309622674, "grad_norm": 3.3760323524475098, "learning_rate": 4.886261057952699e-05, "loss": 5.108, "step": 2521 }, { "epoch": 0.02276584220978516, "grad_norm": 3.267336845397949, "learning_rate": 4.886215923451887e-05, "loss": 5.2745, "step": 2522 }, { "epoch": 0.022774869109947644, "grad_norm": 4.1731061935424805, "learning_rate": 4.886170788951074e-05, "loss": 4.9564, "step": 2523 }, { "epoch": 0.02278389601011013, "grad_norm": 3.895155191421509, "learning_rate": 4.886125654450262e-05, "loss": 4.6731, "step": 2524 }, { "epoch": 0.02279292291027261, "grad_norm": 4.134814262390137, "learning_rate": 4.88608051994945e-05, "loss": 5.0214, "step": 2525 }, { "epoch": 0.022801949810435096, "grad_norm": 3.4610462188720703, "learning_rate": 4.886035385448637e-05, "loss": 4.9934, "step": 2526 }, { "epoch": 0.02281097671059758, "grad_norm": 2.859302043914795, "learning_rate": 4.8859902509478244e-05, "loss": 4.8328, "step": 2527 }, { "epoch": 0.022820003610760066, "grad_norm": 3.19594144821167, "learning_rate": 4.8859451164470123e-05, "loss": 4.7685, "step": 2528 }, { "epoch": 0.022829030510922547, "grad_norm": 3.7315192222595215, "learning_rate": 4.8858999819461996e-05, "loss": 4.7789, "step": 2529 }, { "epoch": 0.022838057411085032, "grad_norm": 4.2291669845581055, "learning_rate": 4.8858548474453875e-05, "loss": 5.28, "step": 2530 }, { "epoch": 0.022847084311247517, "grad_norm": 5.710243225097656, "learning_rate": 4.8858097129445754e-05, "loss": 3.6937, "step": 2531 }, { "epoch": 0.022856111211410002, "grad_norm": 3.9313111305236816, "learning_rate": 4.885764578443763e-05, "loss": 5.2384, "step": 2532 }, { "epoch": 0.022865138111572487, "grad_norm": 3.409585952758789, "learning_rate": 4.8857194439429506e-05, "loss": 4.5562, "step": 2533 }, { "epoch": 0.02287416501173497, "grad_norm": 4.167289733886719, "learning_rate": 4.885674309442138e-05, "loss": 5.2512, "step": 2534 }, { "epoch": 0.022883191911897454, "grad_norm": 3.6107711791992188, "learning_rate": 4.885629174941325e-05, "loss": 4.7409, "step": 2535 }, { "epoch": 0.02289221881205994, "grad_norm": 4.2831549644470215, "learning_rate": 4.885584040440513e-05, "loss": 4.5314, "step": 2536 }, { "epoch": 0.022901245712222424, "grad_norm": 2.967843770980835, "learning_rate": 4.8855389059397e-05, "loss": 4.9676, "step": 2537 }, { "epoch": 0.022910272612384906, "grad_norm": 4.237166404724121, "learning_rate": 4.885493771438888e-05, "loss": 4.6957, "step": 2538 }, { "epoch": 0.02291929951254739, "grad_norm": 3.459202289581299, "learning_rate": 4.885448636938076e-05, "loss": 4.8435, "step": 2539 }, { "epoch": 0.022928326412709876, "grad_norm": 4.139927387237549, "learning_rate": 4.8854035024372634e-05, "loss": 5.426, "step": 2540 }, { "epoch": 0.02293735331287236, "grad_norm": 4.133938789367676, "learning_rate": 4.8853583679364506e-05, "loss": 4.5086, "step": 2541 }, { "epoch": 0.022946380213034846, "grad_norm": 3.169327735900879, "learning_rate": 4.8853132334356385e-05, "loss": 3.0688, "step": 2542 }, { "epoch": 0.022955407113197327, "grad_norm": 3.3999600410461426, "learning_rate": 4.885268098934826e-05, "loss": 4.7602, "step": 2543 }, { "epoch": 0.022964434013359812, "grad_norm": 4.489657402038574, "learning_rate": 4.885222964434014e-05, "loss": 5.4855, "step": 2544 }, { "epoch": 0.022973460913522297, "grad_norm": 3.447618007659912, "learning_rate": 4.885177829933201e-05, "loss": 4.9229, "step": 2545 }, { "epoch": 0.022982487813684782, "grad_norm": 3.9164838790893555, "learning_rate": 4.885132695432389e-05, "loss": 4.9638, "step": 2546 }, { "epoch": 0.022991514713847264, "grad_norm": 4.8694024085998535, "learning_rate": 4.885087560931577e-05, "loss": 4.4742, "step": 2547 }, { "epoch": 0.02300054161400975, "grad_norm": 3.783936023712158, "learning_rate": 4.8850424264307634e-05, "loss": 4.8836, "step": 2548 }, { "epoch": 0.023009568514172234, "grad_norm": 4.624107360839844, "learning_rate": 4.884997291929951e-05, "loss": 5.0765, "step": 2549 }, { "epoch": 0.02301859541433472, "grad_norm": 3.564811944961548, "learning_rate": 4.884952157429139e-05, "loss": 5.0767, "step": 2550 }, { "epoch": 0.0230276223144972, "grad_norm": 4.589054107666016, "learning_rate": 4.8849070229283265e-05, "loss": 4.5018, "step": 2551 }, { "epoch": 0.023036649214659685, "grad_norm": 2.239837169647217, "learning_rate": 4.8848618884275144e-05, "loss": 4.8985, "step": 2552 }, { "epoch": 0.02304567611482217, "grad_norm": 3.218024969100952, "learning_rate": 4.884816753926702e-05, "loss": 4.414, "step": 2553 }, { "epoch": 0.023054703014984655, "grad_norm": 3.3773696422576904, "learning_rate": 4.884771619425889e-05, "loss": 5.1359, "step": 2554 }, { "epoch": 0.02306372991514714, "grad_norm": 4.020918846130371, "learning_rate": 4.884726484925077e-05, "loss": 4.3958, "step": 2555 }, { "epoch": 0.02307275681530962, "grad_norm": 3.0429635047912598, "learning_rate": 4.884681350424265e-05, "loss": 4.1033, "step": 2556 }, { "epoch": 0.023081783715472107, "grad_norm": 4.113393783569336, "learning_rate": 4.884636215923452e-05, "loss": 4.209, "step": 2557 }, { "epoch": 0.02309081061563459, "grad_norm": 5.470997333526611, "learning_rate": 4.88459108142264e-05, "loss": 4.5835, "step": 2558 }, { "epoch": 0.023099837515797077, "grad_norm": 3.55336332321167, "learning_rate": 4.884545946921827e-05, "loss": 4.9415, "step": 2559 }, { "epoch": 0.023108864415959558, "grad_norm": 3.9679648876190186, "learning_rate": 4.884500812421015e-05, "loss": 4.708, "step": 2560 }, { "epoch": 0.023117891316122043, "grad_norm": 4.228524208068848, "learning_rate": 4.884455677920202e-05, "loss": 5.2805, "step": 2561 }, { "epoch": 0.023126918216284528, "grad_norm": 3.5432560443878174, "learning_rate": 4.8844105434193896e-05, "loss": 4.8021, "step": 2562 }, { "epoch": 0.023135945116447013, "grad_norm": 4.594267845153809, "learning_rate": 4.8843654089185775e-05, "loss": 3.9311, "step": 2563 }, { "epoch": 0.023144972016609495, "grad_norm": 3.8872807025909424, "learning_rate": 4.8843202744177654e-05, "loss": 4.9472, "step": 2564 }, { "epoch": 0.02315399891677198, "grad_norm": 2.668703556060791, "learning_rate": 4.8842751399169527e-05, "loss": 5.1901, "step": 2565 }, { "epoch": 0.023163025816934465, "grad_norm": 3.854614734649658, "learning_rate": 4.8842300054161406e-05, "loss": 4.291, "step": 2566 }, { "epoch": 0.02317205271709695, "grad_norm": 4.633874893188477, "learning_rate": 4.8841848709153285e-05, "loss": 4.7986, "step": 2567 }, { "epoch": 0.023181079617259435, "grad_norm": 3.672232151031494, "learning_rate": 4.884139736414515e-05, "loss": 4.9254, "step": 2568 }, { "epoch": 0.023190106517421916, "grad_norm": 4.235375881195068, "learning_rate": 4.884094601913703e-05, "loss": 4.7704, "step": 2569 }, { "epoch": 0.0231991334175844, "grad_norm": 4.544433116912842, "learning_rate": 4.884049467412891e-05, "loss": 4.68, "step": 2570 }, { "epoch": 0.023208160317746886, "grad_norm": 4.21722936630249, "learning_rate": 4.884004332912078e-05, "loss": 4.295, "step": 2571 }, { "epoch": 0.02321718721790937, "grad_norm": 3.064173460006714, "learning_rate": 4.883959198411266e-05, "loss": 4.3564, "step": 2572 }, { "epoch": 0.023226214118071853, "grad_norm": 5.921563148498535, "learning_rate": 4.8839140639104533e-05, "loss": 4.6552, "step": 2573 }, { "epoch": 0.023235241018234338, "grad_norm": 4.290645122528076, "learning_rate": 4.8838689294096406e-05, "loss": 4.3762, "step": 2574 }, { "epoch": 0.023244267918396823, "grad_norm": 4.050984859466553, "learning_rate": 4.8838237949088285e-05, "loss": 4.2502, "step": 2575 }, { "epoch": 0.023253294818559308, "grad_norm": 3.7503316402435303, "learning_rate": 4.883778660408016e-05, "loss": 5.5726, "step": 2576 }, { "epoch": 0.02326232171872179, "grad_norm": 3.938753604888916, "learning_rate": 4.883733525907204e-05, "loss": 4.4322, "step": 2577 }, { "epoch": 0.023271348618884274, "grad_norm": 5.808314800262451, "learning_rate": 4.8836883914063916e-05, "loss": 5.3401, "step": 2578 }, { "epoch": 0.02328037551904676, "grad_norm": 2.994663953781128, "learning_rate": 4.883643256905579e-05, "loss": 5.1477, "step": 2579 }, { "epoch": 0.023289402419209244, "grad_norm": 4.968782901763916, "learning_rate": 4.883598122404767e-05, "loss": 4.8293, "step": 2580 }, { "epoch": 0.02329842931937173, "grad_norm": 3.5077102184295654, "learning_rate": 4.883552987903954e-05, "loss": 4.9393, "step": 2581 }, { "epoch": 0.02330745621953421, "grad_norm": 6.431130409240723, "learning_rate": 4.883507853403141e-05, "loss": 4.9696, "step": 2582 }, { "epoch": 0.023316483119696696, "grad_norm": 3.9356908798217773, "learning_rate": 4.883462718902329e-05, "loss": 4.495, "step": 2583 }, { "epoch": 0.02332551001985918, "grad_norm": 3.130466938018799, "learning_rate": 4.8834175844015164e-05, "loss": 4.1023, "step": 2584 }, { "epoch": 0.023334536920021666, "grad_norm": 5.415238857269287, "learning_rate": 4.8833724499007044e-05, "loss": 5.8168, "step": 2585 }, { "epoch": 0.023343563820184147, "grad_norm": 3.9140477180480957, "learning_rate": 4.883327315399892e-05, "loss": 5.4665, "step": 2586 }, { "epoch": 0.023352590720346632, "grad_norm": 3.4380276203155518, "learning_rate": 4.8832821808990795e-05, "loss": 4.7342, "step": 2587 }, { "epoch": 0.023361617620509117, "grad_norm": 4.170627593994141, "learning_rate": 4.883237046398267e-05, "loss": 4.5469, "step": 2588 }, { "epoch": 0.023370644520671603, "grad_norm": 3.930490255355835, "learning_rate": 4.883191911897455e-05, "loss": 4.9323, "step": 2589 }, { "epoch": 0.023379671420834084, "grad_norm": 4.195336818695068, "learning_rate": 4.883146777396642e-05, "loss": 4.6995, "step": 2590 }, { "epoch": 0.02338869832099657, "grad_norm": 3.9498212337493896, "learning_rate": 4.88310164289583e-05, "loss": 5.3167, "step": 2591 }, { "epoch": 0.023397725221159054, "grad_norm": 3.699143409729004, "learning_rate": 4.883056508395018e-05, "loss": 5.2415, "step": 2592 }, { "epoch": 0.02340675212132154, "grad_norm": 4.086746692657471, "learning_rate": 4.883011373894205e-05, "loss": 4.677, "step": 2593 }, { "epoch": 0.023415779021484024, "grad_norm": 4.054769515991211, "learning_rate": 4.882966239393393e-05, "loss": 4.7031, "step": 2594 }, { "epoch": 0.023424805921646506, "grad_norm": 4.542843818664551, "learning_rate": 4.88292110489258e-05, "loss": 5.1913, "step": 2595 }, { "epoch": 0.02343383282180899, "grad_norm": 3.2555761337280273, "learning_rate": 4.8828759703917675e-05, "loss": 4.4243, "step": 2596 }, { "epoch": 0.023442859721971476, "grad_norm": 3.443850040435791, "learning_rate": 4.8828308358909554e-05, "loss": 5.0649, "step": 2597 }, { "epoch": 0.02345188662213396, "grad_norm": 3.432905912399292, "learning_rate": 4.8827857013901426e-05, "loss": 5.104, "step": 2598 }, { "epoch": 0.023460913522296442, "grad_norm": 2.9491355419158936, "learning_rate": 4.8827405668893306e-05, "loss": 4.6441, "step": 2599 }, { "epoch": 0.023469940422458927, "grad_norm": 4.453527927398682, "learning_rate": 4.8826954323885185e-05, "loss": 4.6579, "step": 2600 }, { "epoch": 0.023478967322621412, "grad_norm": 4.1935811042785645, "learning_rate": 4.882650297887705e-05, "loss": 5.2947, "step": 2601 }, { "epoch": 0.023487994222783897, "grad_norm": 3.2034194469451904, "learning_rate": 4.882605163386893e-05, "loss": 4.5213, "step": 2602 }, { "epoch": 0.02349702112294638, "grad_norm": 4.59647798538208, "learning_rate": 4.882560028886081e-05, "loss": 4.8742, "step": 2603 }, { "epoch": 0.023506048023108864, "grad_norm": 3.8435704708099365, "learning_rate": 4.882514894385268e-05, "loss": 4.3804, "step": 2604 }, { "epoch": 0.02351507492327135, "grad_norm": 3.2333738803863525, "learning_rate": 4.882469759884456e-05, "loss": 5.0342, "step": 2605 }, { "epoch": 0.023524101823433834, "grad_norm": 3.489391565322876, "learning_rate": 4.882424625383643e-05, "loss": 4.7499, "step": 2606 }, { "epoch": 0.02353312872359632, "grad_norm": 3.303689956665039, "learning_rate": 4.882379490882831e-05, "loss": 4.9412, "step": 2607 }, { "epoch": 0.0235421556237588, "grad_norm": 3.7524852752685547, "learning_rate": 4.8823343563820185e-05, "loss": 5.2521, "step": 2608 }, { "epoch": 0.023551182523921285, "grad_norm": 4.562972068786621, "learning_rate": 4.8822892218812064e-05, "loss": 4.8673, "step": 2609 }, { "epoch": 0.02356020942408377, "grad_norm": 4.45949125289917, "learning_rate": 4.8822440873803937e-05, "loss": 4.5165, "step": 2610 }, { "epoch": 0.023569236324246255, "grad_norm": 3.7003819942474365, "learning_rate": 4.8821989528795816e-05, "loss": 4.1453, "step": 2611 }, { "epoch": 0.023578263224408737, "grad_norm": 3.8373329639434814, "learning_rate": 4.882153818378769e-05, "loss": 4.9075, "step": 2612 }, { "epoch": 0.02358729012457122, "grad_norm": 3.6967599391937256, "learning_rate": 4.882108683877957e-05, "loss": 4.451, "step": 2613 }, { "epoch": 0.023596317024733707, "grad_norm": 3.0235207080841064, "learning_rate": 4.882063549377145e-05, "loss": 4.9709, "step": 2614 }, { "epoch": 0.023605343924896192, "grad_norm": 3.6188840866088867, "learning_rate": 4.882018414876331e-05, "loss": 4.5968, "step": 2615 }, { "epoch": 0.023614370825058673, "grad_norm": 3.7883870601654053, "learning_rate": 4.881973280375519e-05, "loss": 5.1174, "step": 2616 }, { "epoch": 0.02362339772522116, "grad_norm": 3.8391239643096924, "learning_rate": 4.881928145874707e-05, "loss": 4.8938, "step": 2617 }, { "epoch": 0.023632424625383643, "grad_norm": 3.626437187194824, "learning_rate": 4.8818830113738943e-05, "loss": 4.4464, "step": 2618 }, { "epoch": 0.02364145152554613, "grad_norm": 4.738600254058838, "learning_rate": 4.881837876873082e-05, "loss": 4.2569, "step": 2619 }, { "epoch": 0.023650478425708613, "grad_norm": 2.7332212924957275, "learning_rate": 4.8817927423722695e-05, "loss": 5.1095, "step": 2620 }, { "epoch": 0.023659505325871095, "grad_norm": 3.9827215671539307, "learning_rate": 4.881747607871457e-05, "loss": 4.8102, "step": 2621 }, { "epoch": 0.02366853222603358, "grad_norm": 2.848378896713257, "learning_rate": 4.881702473370645e-05, "loss": 4.0918, "step": 2622 }, { "epoch": 0.023677559126196065, "grad_norm": 5.275362491607666, "learning_rate": 4.881657338869832e-05, "loss": 5.4022, "step": 2623 }, { "epoch": 0.02368658602635855, "grad_norm": 3.2399771213531494, "learning_rate": 4.88161220436902e-05, "loss": 5.0866, "step": 2624 }, { "epoch": 0.02369561292652103, "grad_norm": 4.760037422180176, "learning_rate": 4.881567069868208e-05, "loss": 4.5393, "step": 2625 }, { "epoch": 0.023704639826683516, "grad_norm": 3.0283000469207764, "learning_rate": 4.881521935367395e-05, "loss": 4.7026, "step": 2626 }, { "epoch": 0.023713666726846, "grad_norm": 4.098232269287109, "learning_rate": 4.881476800866583e-05, "loss": 5.1041, "step": 2627 }, { "epoch": 0.023722693627008486, "grad_norm": 5.149988174438477, "learning_rate": 4.88143166636577e-05, "loss": 4.031, "step": 2628 }, { "epoch": 0.023731720527170968, "grad_norm": 4.30027437210083, "learning_rate": 4.8813865318649574e-05, "loss": 5.2127, "step": 2629 }, { "epoch": 0.023740747427333453, "grad_norm": 4.711564540863037, "learning_rate": 4.8813413973641454e-05, "loss": 5.2281, "step": 2630 }, { "epoch": 0.023749774327495938, "grad_norm": 4.768317222595215, "learning_rate": 4.881296262863333e-05, "loss": 4.6135, "step": 2631 }, { "epoch": 0.023758801227658423, "grad_norm": 3.1340696811676025, "learning_rate": 4.8812511283625205e-05, "loss": 4.5638, "step": 2632 }, { "epoch": 0.023767828127820908, "grad_norm": 3.622084140777588, "learning_rate": 4.8812059938617085e-05, "loss": 4.9276, "step": 2633 }, { "epoch": 0.02377685502798339, "grad_norm": 3.60921311378479, "learning_rate": 4.881160859360896e-05, "loss": 4.8537, "step": 2634 }, { "epoch": 0.023785881928145874, "grad_norm": 3.405851125717163, "learning_rate": 4.881115724860083e-05, "loss": 4.0571, "step": 2635 }, { "epoch": 0.02379490882830836, "grad_norm": 3.292637348175049, "learning_rate": 4.881070590359271e-05, "loss": 4.7042, "step": 2636 }, { "epoch": 0.023803935728470844, "grad_norm": 3.835688829421997, "learning_rate": 4.881025455858458e-05, "loss": 4.6662, "step": 2637 }, { "epoch": 0.023812962628633326, "grad_norm": 3.2330312728881836, "learning_rate": 4.880980321357646e-05, "loss": 4.7244, "step": 2638 }, { "epoch": 0.02382198952879581, "grad_norm": 3.685270071029663, "learning_rate": 4.880935186856834e-05, "loss": 5.2148, "step": 2639 }, { "epoch": 0.023831016428958296, "grad_norm": 4.376675128936768, "learning_rate": 4.880890052356021e-05, "loss": 4.2182, "step": 2640 }, { "epoch": 0.02384004332912078, "grad_norm": 4.66953182220459, "learning_rate": 4.880844917855209e-05, "loss": 5.0997, "step": 2641 }, { "epoch": 0.023849070229283263, "grad_norm": 4.21575927734375, "learning_rate": 4.8807997833543964e-05, "loss": 4.8129, "step": 2642 }, { "epoch": 0.023858097129445748, "grad_norm": 6.236120223999023, "learning_rate": 4.8807546488535836e-05, "loss": 4.5452, "step": 2643 }, { "epoch": 0.023867124029608233, "grad_norm": 4.085434436798096, "learning_rate": 4.8807095143527716e-05, "loss": 4.7141, "step": 2644 }, { "epoch": 0.023876150929770718, "grad_norm": 3.0471513271331787, "learning_rate": 4.880664379851959e-05, "loss": 4.8479, "step": 2645 }, { "epoch": 0.023885177829933203, "grad_norm": 3.9657180309295654, "learning_rate": 4.880619245351147e-05, "loss": 4.8698, "step": 2646 }, { "epoch": 0.023894204730095684, "grad_norm": 4.310367107391357, "learning_rate": 4.8805741108503347e-05, "loss": 4.9593, "step": 2647 }, { "epoch": 0.02390323163025817, "grad_norm": 4.460762023925781, "learning_rate": 4.880528976349521e-05, "loss": 5.3139, "step": 2648 }, { "epoch": 0.023912258530420654, "grad_norm": 2.8321805000305176, "learning_rate": 4.880483841848709e-05, "loss": 4.7106, "step": 2649 }, { "epoch": 0.02392128543058314, "grad_norm": 3.262131452560425, "learning_rate": 4.880438707347897e-05, "loss": 4.8413, "step": 2650 }, { "epoch": 0.02393031233074562, "grad_norm": 4.064083576202393, "learning_rate": 4.880393572847084e-05, "loss": 4.568, "step": 2651 }, { "epoch": 0.023939339230908106, "grad_norm": 3.6241633892059326, "learning_rate": 4.880348438346272e-05, "loss": 4.5537, "step": 2652 }, { "epoch": 0.02394836613107059, "grad_norm": 4.495134353637695, "learning_rate": 4.88030330384546e-05, "loss": 4.1514, "step": 2653 }, { "epoch": 0.023957393031233076, "grad_norm": 4.3893280029296875, "learning_rate": 4.8802581693446474e-05, "loss": 3.8128, "step": 2654 }, { "epoch": 0.023966419931395557, "grad_norm": 3.488978147506714, "learning_rate": 4.8802130348438347e-05, "loss": 5.0643, "step": 2655 }, { "epoch": 0.023975446831558042, "grad_norm": 4.0124053955078125, "learning_rate": 4.8801679003430226e-05, "loss": 4.5857, "step": 2656 }, { "epoch": 0.023984473731720527, "grad_norm": 3.400162696838379, "learning_rate": 4.88012276584221e-05, "loss": 5.2637, "step": 2657 }, { "epoch": 0.023993500631883012, "grad_norm": 3.8102617263793945, "learning_rate": 4.880077631341398e-05, "loss": 4.3506, "step": 2658 }, { "epoch": 0.024002527532045497, "grad_norm": 3.5363945960998535, "learning_rate": 4.880032496840585e-05, "loss": 5.1319, "step": 2659 }, { "epoch": 0.02401155443220798, "grad_norm": 3.4010188579559326, "learning_rate": 4.879987362339773e-05, "loss": 4.821, "step": 2660 }, { "epoch": 0.024020581332370464, "grad_norm": 3.0875651836395264, "learning_rate": 4.879942227838961e-05, "loss": 4.4028, "step": 2661 }, { "epoch": 0.02402960823253295, "grad_norm": 2.917893648147583, "learning_rate": 4.8798970933381474e-05, "loss": 5.1236, "step": 2662 }, { "epoch": 0.024038635132695434, "grad_norm": 4.209592342376709, "learning_rate": 4.8798519588373353e-05, "loss": 4.9487, "step": 2663 }, { "epoch": 0.024047662032857915, "grad_norm": 7.5813469886779785, "learning_rate": 4.879806824336523e-05, "loss": 5.1521, "step": 2664 }, { "epoch": 0.0240566889330204, "grad_norm": 3.7502784729003906, "learning_rate": 4.8797616898357105e-05, "loss": 4.6542, "step": 2665 }, { "epoch": 0.024065715833182885, "grad_norm": 5.727113246917725, "learning_rate": 4.8797165553348984e-05, "loss": 3.8132, "step": 2666 }, { "epoch": 0.02407474273334537, "grad_norm": 3.6422321796417236, "learning_rate": 4.879671420834086e-05, "loss": 5.3526, "step": 2667 }, { "epoch": 0.024083769633507852, "grad_norm": 5.827818870544434, "learning_rate": 4.879626286333273e-05, "loss": 4.363, "step": 2668 }, { "epoch": 0.024092796533670337, "grad_norm": 3.0263946056365967, "learning_rate": 4.879581151832461e-05, "loss": 5.1711, "step": 2669 }, { "epoch": 0.024101823433832822, "grad_norm": 3.5697827339172363, "learning_rate": 4.879536017331649e-05, "loss": 4.3493, "step": 2670 }, { "epoch": 0.024110850333995307, "grad_norm": 3.9017040729522705, "learning_rate": 4.879490882830836e-05, "loss": 5.336, "step": 2671 }, { "epoch": 0.024119877234157792, "grad_norm": 4.115163803100586, "learning_rate": 4.879445748330024e-05, "loss": 5.1259, "step": 2672 }, { "epoch": 0.024128904134320273, "grad_norm": 4.4237380027771, "learning_rate": 4.879400613829211e-05, "loss": 4.2909, "step": 2673 }, { "epoch": 0.02413793103448276, "grad_norm": 2.701739549636841, "learning_rate": 4.879355479328399e-05, "loss": 4.4608, "step": 2674 }, { "epoch": 0.024146957934645243, "grad_norm": 4.145193576812744, "learning_rate": 4.8793103448275864e-05, "loss": 4.7233, "step": 2675 }, { "epoch": 0.02415598483480773, "grad_norm": 3.2616655826568604, "learning_rate": 4.8792652103267736e-05, "loss": 5.0113, "step": 2676 }, { "epoch": 0.02416501173497021, "grad_norm": 3.5643694400787354, "learning_rate": 4.8792200758259615e-05, "loss": 4.7871, "step": 2677 }, { "epoch": 0.024174038635132695, "grad_norm": 6.3014116287231445, "learning_rate": 4.8791749413251495e-05, "loss": 4.8259, "step": 2678 }, { "epoch": 0.02418306553529518, "grad_norm": 2.7380056381225586, "learning_rate": 4.879129806824337e-05, "loss": 5.1079, "step": 2679 }, { "epoch": 0.024192092435457665, "grad_norm": 4.006518363952637, "learning_rate": 4.8790846723235246e-05, "loss": 4.0396, "step": 2680 }, { "epoch": 0.024201119335620146, "grad_norm": 4.003290176391602, "learning_rate": 4.879039537822712e-05, "loss": 4.7184, "step": 2681 }, { "epoch": 0.02421014623578263, "grad_norm": 2.5680418014526367, "learning_rate": 4.878994403321899e-05, "loss": 5.4781, "step": 2682 }, { "epoch": 0.024219173135945116, "grad_norm": 4.492859840393066, "learning_rate": 4.878949268821087e-05, "loss": 4.6201, "step": 2683 }, { "epoch": 0.0242282000361076, "grad_norm": 3.68095326423645, "learning_rate": 4.878904134320274e-05, "loss": 4.6751, "step": 2684 }, { "epoch": 0.024237226936270086, "grad_norm": 3.4634017944335938, "learning_rate": 4.878858999819462e-05, "loss": 4.9268, "step": 2685 }, { "epoch": 0.024246253836432568, "grad_norm": 5.404234886169434, "learning_rate": 4.87881386531865e-05, "loss": 4.4801, "step": 2686 }, { "epoch": 0.024255280736595053, "grad_norm": 3.7384870052337646, "learning_rate": 4.8787687308178374e-05, "loss": 4.6899, "step": 2687 }, { "epoch": 0.024264307636757538, "grad_norm": 4.331826686859131, "learning_rate": 4.8787235963170246e-05, "loss": 4.6297, "step": 2688 }, { "epoch": 0.024273334536920023, "grad_norm": 5.037558555603027, "learning_rate": 4.8786784618162126e-05, "loss": 5.1836, "step": 2689 }, { "epoch": 0.024282361437082504, "grad_norm": 3.3435580730438232, "learning_rate": 4.8786333273154e-05, "loss": 5.0254, "step": 2690 }, { "epoch": 0.02429138833724499, "grad_norm": 2.662531852722168, "learning_rate": 4.878588192814588e-05, "loss": 4.7858, "step": 2691 }, { "epoch": 0.024300415237407474, "grad_norm": 3.934635639190674, "learning_rate": 4.8785430583137757e-05, "loss": 5.1062, "step": 2692 }, { "epoch": 0.02430944213756996, "grad_norm": 3.927600622177124, "learning_rate": 4.878497923812963e-05, "loss": 4.9705, "step": 2693 }, { "epoch": 0.02431846903773244, "grad_norm": 3.616149425506592, "learning_rate": 4.878452789312151e-05, "loss": 5.2544, "step": 2694 }, { "epoch": 0.024327495937894926, "grad_norm": 4.675205230712891, "learning_rate": 4.878407654811338e-05, "loss": 4.9681, "step": 2695 }, { "epoch": 0.02433652283805741, "grad_norm": 4.556786060333252, "learning_rate": 4.878362520310525e-05, "loss": 5.3446, "step": 2696 }, { "epoch": 0.024345549738219896, "grad_norm": 5.100498676300049, "learning_rate": 4.878317385809713e-05, "loss": 5.3973, "step": 2697 }, { "epoch": 0.02435457663838238, "grad_norm": 4.282364845275879, "learning_rate": 4.8782722513089005e-05, "loss": 5.1965, "step": 2698 }, { "epoch": 0.024363603538544863, "grad_norm": 3.2669074535369873, "learning_rate": 4.8782271168080884e-05, "loss": 4.4641, "step": 2699 }, { "epoch": 0.024372630438707348, "grad_norm": 3.009031057357788, "learning_rate": 4.878181982307276e-05, "loss": 4.907, "step": 2700 }, { "epoch": 0.024381657338869833, "grad_norm": 10.980446815490723, "learning_rate": 4.8781368478064636e-05, "loss": 4.9116, "step": 2701 }, { "epoch": 0.024390684239032318, "grad_norm": 3.2279067039489746, "learning_rate": 4.878091713305651e-05, "loss": 5.2285, "step": 2702 }, { "epoch": 0.0243997111391948, "grad_norm": 3.460554838180542, "learning_rate": 4.878046578804839e-05, "loss": 4.8441, "step": 2703 }, { "epoch": 0.024408738039357284, "grad_norm": 3.784773349761963, "learning_rate": 4.878001444304026e-05, "loss": 4.3089, "step": 2704 }, { "epoch": 0.02441776493951977, "grad_norm": 2.686802387237549, "learning_rate": 4.877956309803214e-05, "loss": 4.1098, "step": 2705 }, { "epoch": 0.024426791839682254, "grad_norm": 3.574000835418701, "learning_rate": 4.877911175302401e-05, "loss": 4.5647, "step": 2706 }, { "epoch": 0.024435818739844736, "grad_norm": 4.135365009307861, "learning_rate": 4.877866040801589e-05, "loss": 4.5797, "step": 2707 }, { "epoch": 0.02444484564000722, "grad_norm": 4.605866432189941, "learning_rate": 4.877820906300777e-05, "loss": 4.3373, "step": 2708 }, { "epoch": 0.024453872540169706, "grad_norm": 3.742689371109009, "learning_rate": 4.8777757717999636e-05, "loss": 5.0376, "step": 2709 }, { "epoch": 0.02446289944033219, "grad_norm": 3.9423396587371826, "learning_rate": 4.8777306372991515e-05, "loss": 3.5689, "step": 2710 }, { "epoch": 0.024471926340494676, "grad_norm": 3.333768367767334, "learning_rate": 4.8776855027983394e-05, "loss": 5.4698, "step": 2711 }, { "epoch": 0.024480953240657157, "grad_norm": 4.955144882202148, "learning_rate": 4.877640368297527e-05, "loss": 3.0707, "step": 2712 }, { "epoch": 0.024489980140819642, "grad_norm": 4.076716899871826, "learning_rate": 4.8775952337967146e-05, "loss": 4.6027, "step": 2713 }, { "epoch": 0.024499007040982127, "grad_norm": 2.874211549758911, "learning_rate": 4.8775500992959025e-05, "loss": 5.0146, "step": 2714 }, { "epoch": 0.024508033941144612, "grad_norm": 4.093982696533203, "learning_rate": 4.877504964795089e-05, "loss": 4.6771, "step": 2715 }, { "epoch": 0.024517060841307094, "grad_norm": 3.8242650032043457, "learning_rate": 4.877459830294277e-05, "loss": 4.7402, "step": 2716 }, { "epoch": 0.02452608774146958, "grad_norm": 4.034634590148926, "learning_rate": 4.877414695793465e-05, "loss": 4.2431, "step": 2717 }, { "epoch": 0.024535114641632064, "grad_norm": 4.14249324798584, "learning_rate": 4.877369561292652e-05, "loss": 4.3701, "step": 2718 }, { "epoch": 0.02454414154179455, "grad_norm": 3.390752077102661, "learning_rate": 4.87732442679184e-05, "loss": 5.3448, "step": 2719 }, { "epoch": 0.024553168441957034, "grad_norm": 4.760272979736328, "learning_rate": 4.8772792922910274e-05, "loss": 4.2559, "step": 2720 }, { "epoch": 0.024562195342119515, "grad_norm": 3.119943618774414, "learning_rate": 4.877234157790215e-05, "loss": 4.9517, "step": 2721 }, { "epoch": 0.024571222242282, "grad_norm": 3.3441038131713867, "learning_rate": 4.8771890232894025e-05, "loss": 5.1865, "step": 2722 }, { "epoch": 0.024580249142444485, "grad_norm": 3.576397657394409, "learning_rate": 4.87714388878859e-05, "loss": 5.1083, "step": 2723 }, { "epoch": 0.02458927604260697, "grad_norm": 3.822031259536743, "learning_rate": 4.877098754287778e-05, "loss": 4.3732, "step": 2724 }, { "epoch": 0.024598302942769452, "grad_norm": 4.290449619293213, "learning_rate": 4.8770536197869656e-05, "loss": 3.779, "step": 2725 }, { "epoch": 0.024607329842931937, "grad_norm": 3.263704776763916, "learning_rate": 4.877008485286153e-05, "loss": 5.1427, "step": 2726 }, { "epoch": 0.024616356743094422, "grad_norm": 5.789054870605469, "learning_rate": 4.876963350785341e-05, "loss": 5.2076, "step": 2727 }, { "epoch": 0.024625383643256907, "grad_norm": 4.498304843902588, "learning_rate": 4.876918216284529e-05, "loss": 4.628, "step": 2728 }, { "epoch": 0.02463441054341939, "grad_norm": 3.554743766784668, "learning_rate": 4.876873081783715e-05, "loss": 4.8776, "step": 2729 }, { "epoch": 0.024643437443581873, "grad_norm": 4.68992805480957, "learning_rate": 4.876827947282903e-05, "loss": 4.825, "step": 2730 }, { "epoch": 0.02465246434374436, "grad_norm": 3.534183979034424, "learning_rate": 4.876782812782091e-05, "loss": 4.7708, "step": 2731 }, { "epoch": 0.024661491243906843, "grad_norm": 4.223320007324219, "learning_rate": 4.8767376782812784e-05, "loss": 4.5779, "step": 2732 }, { "epoch": 0.02467051814406933, "grad_norm": 4.034117698669434, "learning_rate": 4.876692543780466e-05, "loss": 5.1033, "step": 2733 }, { "epoch": 0.02467954504423181, "grad_norm": 3.4592134952545166, "learning_rate": 4.8766474092796536e-05, "loss": 4.6887, "step": 2734 }, { "epoch": 0.024688571944394295, "grad_norm": 3.2277209758758545, "learning_rate": 4.876602274778841e-05, "loss": 4.4299, "step": 2735 }, { "epoch": 0.02469759884455678, "grad_norm": 4.967749118804932, "learning_rate": 4.876557140278029e-05, "loss": 4.5687, "step": 2736 }, { "epoch": 0.024706625744719265, "grad_norm": 3.649949550628662, "learning_rate": 4.876512005777216e-05, "loss": 4.7497, "step": 2737 }, { "epoch": 0.024715652644881746, "grad_norm": 3.3901360034942627, "learning_rate": 4.876466871276404e-05, "loss": 5.1038, "step": 2738 }, { "epoch": 0.02472467954504423, "grad_norm": 3.1384551525115967, "learning_rate": 4.876421736775592e-05, "loss": 4.7252, "step": 2739 }, { "epoch": 0.024733706445206716, "grad_norm": 3.30941104888916, "learning_rate": 4.876376602274779e-05, "loss": 4.9749, "step": 2740 }, { "epoch": 0.0247427333453692, "grad_norm": 3.1989777088165283, "learning_rate": 4.876331467773967e-05, "loss": 5.0549, "step": 2741 }, { "epoch": 0.024751760245531683, "grad_norm": 6.009932518005371, "learning_rate": 4.876286333273154e-05, "loss": 3.7655, "step": 2742 }, { "epoch": 0.024760787145694168, "grad_norm": 3.257307529449463, "learning_rate": 4.8762411987723415e-05, "loss": 4.5537, "step": 2743 }, { "epoch": 0.024769814045856653, "grad_norm": 3.7387077808380127, "learning_rate": 4.8761960642715294e-05, "loss": 5.1828, "step": 2744 }, { "epoch": 0.024778840946019138, "grad_norm": 3.675539493560791, "learning_rate": 4.8761509297707167e-05, "loss": 4.6037, "step": 2745 }, { "epoch": 0.024787867846181623, "grad_norm": 3.9149348735809326, "learning_rate": 4.8761057952699046e-05, "loss": 3.5502, "step": 2746 }, { "epoch": 0.024796894746344104, "grad_norm": 4.044651508331299, "learning_rate": 4.8760606607690925e-05, "loss": 4.7308, "step": 2747 }, { "epoch": 0.02480592164650659, "grad_norm": 3.7244997024536133, "learning_rate": 4.87601552626828e-05, "loss": 4.9004, "step": 2748 }, { "epoch": 0.024814948546669074, "grad_norm": 4.04166841506958, "learning_rate": 4.875970391767467e-05, "loss": 4.8686, "step": 2749 }, { "epoch": 0.02482397544683156, "grad_norm": 5.043482303619385, "learning_rate": 4.875925257266655e-05, "loss": 4.7835, "step": 2750 }, { "epoch": 0.02483300234699404, "grad_norm": 3.228835105895996, "learning_rate": 4.875880122765842e-05, "loss": 4.7231, "step": 2751 }, { "epoch": 0.024842029247156526, "grad_norm": 3.5961122512817383, "learning_rate": 4.87583498826503e-05, "loss": 4.9135, "step": 2752 }, { "epoch": 0.02485105614731901, "grad_norm": 3.2476577758789062, "learning_rate": 4.875789853764218e-05, "loss": 5.1183, "step": 2753 }, { "epoch": 0.024860083047481496, "grad_norm": 3.1624515056610107, "learning_rate": 4.875744719263405e-05, "loss": 5.0159, "step": 2754 }, { "epoch": 0.024869109947643978, "grad_norm": 9.02939510345459, "learning_rate": 4.875699584762593e-05, "loss": 4.8892, "step": 2755 }, { "epoch": 0.024878136847806463, "grad_norm": 2.8629114627838135, "learning_rate": 4.8756544502617804e-05, "loss": 4.7411, "step": 2756 }, { "epoch": 0.024887163747968948, "grad_norm": 4.287327766418457, "learning_rate": 4.875609315760968e-05, "loss": 4.9737, "step": 2757 }, { "epoch": 0.024896190648131433, "grad_norm": 3.334132194519043, "learning_rate": 4.8755641812601556e-05, "loss": 4.9453, "step": 2758 }, { "epoch": 0.024905217548293918, "grad_norm": 4.902735710144043, "learning_rate": 4.875519046759343e-05, "loss": 4.4994, "step": 2759 }, { "epoch": 0.0249142444484564, "grad_norm": 3.0542757511138916, "learning_rate": 4.875473912258531e-05, "loss": 4.8307, "step": 2760 }, { "epoch": 0.024923271348618884, "grad_norm": 5.216578006744385, "learning_rate": 4.875428777757719e-05, "loss": 5.0731, "step": 2761 }, { "epoch": 0.02493229824878137, "grad_norm": 3.798877477645874, "learning_rate": 4.875383643256905e-05, "loss": 4.5134, "step": 2762 }, { "epoch": 0.024941325148943854, "grad_norm": 3.6948184967041016, "learning_rate": 4.875338508756093e-05, "loss": 5.4531, "step": 2763 }, { "epoch": 0.024950352049106336, "grad_norm": 3.005117416381836, "learning_rate": 4.875293374255281e-05, "loss": 4.3171, "step": 2764 }, { "epoch": 0.02495937894926882, "grad_norm": 3.834566116333008, "learning_rate": 4.8752482397544684e-05, "loss": 5.6559, "step": 2765 }, { "epoch": 0.024968405849431306, "grad_norm": 5.709918975830078, "learning_rate": 4.875203105253656e-05, "loss": 4.9848, "step": 2766 }, { "epoch": 0.02497743274959379, "grad_norm": 3.070237874984741, "learning_rate": 4.8751579707528435e-05, "loss": 4.735, "step": 2767 }, { "epoch": 0.024986459649756272, "grad_norm": 4.118338108062744, "learning_rate": 4.8751128362520315e-05, "loss": 4.5985, "step": 2768 }, { "epoch": 0.024995486549918757, "grad_norm": 3.628772258758545, "learning_rate": 4.875067701751219e-05, "loss": 4.7673, "step": 2769 }, { "epoch": 0.025004513450081242, "grad_norm": 3.258568286895752, "learning_rate": 4.8750225672504066e-05, "loss": 4.9643, "step": 2770 }, { "epoch": 0.025013540350243727, "grad_norm": 4.46489953994751, "learning_rate": 4.874977432749594e-05, "loss": 4.3821, "step": 2771 }, { "epoch": 0.025022567250406212, "grad_norm": 3.425124406814575, "learning_rate": 4.874932298248782e-05, "loss": 4.9269, "step": 2772 }, { "epoch": 0.025031594150568694, "grad_norm": 3.2277228832244873, "learning_rate": 4.874887163747969e-05, "loss": 4.9284, "step": 2773 }, { "epoch": 0.02504062105073118, "grad_norm": 2.926863431930542, "learning_rate": 4.874842029247157e-05, "loss": 5.0212, "step": 2774 }, { "epoch": 0.025049647950893664, "grad_norm": 3.7291436195373535, "learning_rate": 4.874796894746345e-05, "loss": 4.8444, "step": 2775 }, { "epoch": 0.02505867485105615, "grad_norm": 4.4859161376953125, "learning_rate": 4.8747517602455315e-05, "loss": 4.965, "step": 2776 }, { "epoch": 0.02506770175121863, "grad_norm": 3.3671886920928955, "learning_rate": 4.8747066257447194e-05, "loss": 4.4698, "step": 2777 }, { "epoch": 0.025076728651381115, "grad_norm": 5.116353988647461, "learning_rate": 4.874661491243907e-05, "loss": 4.7053, "step": 2778 }, { "epoch": 0.0250857555515436, "grad_norm": 2.623157501220703, "learning_rate": 4.8746163567430946e-05, "loss": 3.5449, "step": 2779 }, { "epoch": 0.025094782451706085, "grad_norm": 6.334542274475098, "learning_rate": 4.8745712222422825e-05, "loss": 3.4691, "step": 2780 }, { "epoch": 0.025103809351868567, "grad_norm": 5.0561628341674805, "learning_rate": 4.87452608774147e-05, "loss": 5.1146, "step": 2781 }, { "epoch": 0.025112836252031052, "grad_norm": 6.553138256072998, "learning_rate": 4.874480953240657e-05, "loss": 4.9826, "step": 2782 }, { "epoch": 0.025121863152193537, "grad_norm": 3.557804822921753, "learning_rate": 4.874435818739845e-05, "loss": 4.8372, "step": 2783 }, { "epoch": 0.025130890052356022, "grad_norm": 3.5614492893218994, "learning_rate": 4.874390684239032e-05, "loss": 4.9153, "step": 2784 }, { "epoch": 0.025139916952518507, "grad_norm": 3.6130998134613037, "learning_rate": 4.87434554973822e-05, "loss": 4.7252, "step": 2785 }, { "epoch": 0.02514894385268099, "grad_norm": 6.617212295532227, "learning_rate": 4.874300415237408e-05, "loss": 4.0992, "step": 2786 }, { "epoch": 0.025157970752843473, "grad_norm": 3.546860694885254, "learning_rate": 4.874255280736595e-05, "loss": 4.7791, "step": 2787 }, { "epoch": 0.02516699765300596, "grad_norm": 4.142953395843506, "learning_rate": 4.874210146235783e-05, "loss": 4.7367, "step": 2788 }, { "epoch": 0.025176024553168443, "grad_norm": 3.9138247966766357, "learning_rate": 4.8741650117349704e-05, "loss": 5.3751, "step": 2789 }, { "epoch": 0.025185051453330925, "grad_norm": 3.4542882442474365, "learning_rate": 4.8741198772341577e-05, "loss": 5.0056, "step": 2790 }, { "epoch": 0.02519407835349341, "grad_norm": 3.5246660709381104, "learning_rate": 4.8740747427333456e-05, "loss": 4.9899, "step": 2791 }, { "epoch": 0.025203105253655895, "grad_norm": 4.984328746795654, "learning_rate": 4.8740296082325335e-05, "loss": 5.0125, "step": 2792 }, { "epoch": 0.02521213215381838, "grad_norm": 2.3484392166137695, "learning_rate": 4.873984473731721e-05, "loss": 5.4632, "step": 2793 }, { "epoch": 0.02522115905398086, "grad_norm": 4.239717483520508, "learning_rate": 4.873939339230909e-05, "loss": 4.3817, "step": 2794 }, { "epoch": 0.025230185954143346, "grad_norm": 3.9730224609375, "learning_rate": 4.873894204730096e-05, "loss": 4.5062, "step": 2795 }, { "epoch": 0.02523921285430583, "grad_norm": 3.245349884033203, "learning_rate": 4.873849070229283e-05, "loss": 5.2581, "step": 2796 }, { "epoch": 0.025248239754468316, "grad_norm": 4.303569793701172, "learning_rate": 4.873803935728471e-05, "loss": 4.9313, "step": 2797 }, { "epoch": 0.0252572666546308, "grad_norm": 3.334035634994507, "learning_rate": 4.873758801227658e-05, "loss": 4.8135, "step": 2798 }, { "epoch": 0.025266293554793283, "grad_norm": 3.785619020462036, "learning_rate": 4.873713666726846e-05, "loss": 5.1, "step": 2799 }, { "epoch": 0.025275320454955768, "grad_norm": 2.9330503940582275, "learning_rate": 4.873668532226034e-05, "loss": 5.0245, "step": 2800 }, { "epoch": 0.025284347355118253, "grad_norm": 4.0583648681640625, "learning_rate": 4.8736233977252214e-05, "loss": 5.0982, "step": 2801 }, { "epoch": 0.025293374255280738, "grad_norm": 3.4990506172180176, "learning_rate": 4.8735782632244094e-05, "loss": 5.4218, "step": 2802 }, { "epoch": 0.02530240115544322, "grad_norm": 4.26165246963501, "learning_rate": 4.8735331287235966e-05, "loss": 4.9204, "step": 2803 }, { "epoch": 0.025311428055605704, "grad_norm": 3.7602601051330566, "learning_rate": 4.873487994222784e-05, "loss": 4.4296, "step": 2804 }, { "epoch": 0.02532045495576819, "grad_norm": 4.095961093902588, "learning_rate": 4.873442859721972e-05, "loss": 4.7095, "step": 2805 }, { "epoch": 0.025329481855930674, "grad_norm": 7.789320945739746, "learning_rate": 4.873397725221159e-05, "loss": 5.2101, "step": 2806 }, { "epoch": 0.025338508756093156, "grad_norm": 3.2901322841644287, "learning_rate": 4.873352590720347e-05, "loss": 4.5882, "step": 2807 }, { "epoch": 0.02534753565625564, "grad_norm": 3.215641975402832, "learning_rate": 4.873307456219535e-05, "loss": 5.0845, "step": 2808 }, { "epoch": 0.025356562556418126, "grad_norm": 3.826274871826172, "learning_rate": 4.8732623217187214e-05, "loss": 5.0841, "step": 2809 }, { "epoch": 0.02536558945658061, "grad_norm": 5.096938133239746, "learning_rate": 4.8732171872179094e-05, "loss": 4.6151, "step": 2810 }, { "epoch": 0.025374616356743096, "grad_norm": 3.7044198513031006, "learning_rate": 4.873172052717097e-05, "loss": 5.2742, "step": 2811 }, { "epoch": 0.025383643256905578, "grad_norm": 5.355084419250488, "learning_rate": 4.8731269182162845e-05, "loss": 4.8438, "step": 2812 }, { "epoch": 0.025392670157068063, "grad_norm": 2.8698208332061768, "learning_rate": 4.8730817837154725e-05, "loss": 4.655, "step": 2813 }, { "epoch": 0.025401697057230548, "grad_norm": 5.369996070861816, "learning_rate": 4.8730366492146604e-05, "loss": 4.4233, "step": 2814 }, { "epoch": 0.025410723957393033, "grad_norm": 4.079714775085449, "learning_rate": 4.8729915147138476e-05, "loss": 5.0052, "step": 2815 }, { "epoch": 0.025419750857555514, "grad_norm": 3.4048736095428467, "learning_rate": 4.872946380213035e-05, "loss": 4.4845, "step": 2816 }, { "epoch": 0.025428777757718, "grad_norm": 3.7393901348114014, "learning_rate": 4.872901245712223e-05, "loss": 4.0242, "step": 2817 }, { "epoch": 0.025437804657880484, "grad_norm": 3.899024248123169, "learning_rate": 4.87285611121141e-05, "loss": 4.8664, "step": 2818 }, { "epoch": 0.02544683155804297, "grad_norm": 3.292801856994629, "learning_rate": 4.872810976710598e-05, "loss": 5.4154, "step": 2819 }, { "epoch": 0.02545585845820545, "grad_norm": 3.5477752685546875, "learning_rate": 4.872765842209785e-05, "loss": 4.4921, "step": 2820 }, { "epoch": 0.025464885358367936, "grad_norm": 3.154323101043701, "learning_rate": 4.872720707708973e-05, "loss": 5.2352, "step": 2821 }, { "epoch": 0.02547391225853042, "grad_norm": 3.4175407886505127, "learning_rate": 4.872675573208161e-05, "loss": 4.6496, "step": 2822 }, { "epoch": 0.025482939158692906, "grad_norm": 3.5746166706085205, "learning_rate": 4.8726304387073476e-05, "loss": 4.8868, "step": 2823 }, { "epoch": 0.02549196605885539, "grad_norm": 3.382338762283325, "learning_rate": 4.8725853042065356e-05, "loss": 4.196, "step": 2824 }, { "epoch": 0.025500992959017872, "grad_norm": 3.881223201751709, "learning_rate": 4.8725401697057235e-05, "loss": 4.4103, "step": 2825 }, { "epoch": 0.025510019859180357, "grad_norm": 3.288895606994629, "learning_rate": 4.872495035204911e-05, "loss": 4.5865, "step": 2826 }, { "epoch": 0.025519046759342842, "grad_norm": 2.6988749504089355, "learning_rate": 4.8724499007040987e-05, "loss": 4.0385, "step": 2827 }, { "epoch": 0.025528073659505327, "grad_norm": 3.5938642024993896, "learning_rate": 4.872404766203286e-05, "loss": 4.991, "step": 2828 }, { "epoch": 0.02553710055966781, "grad_norm": 3.2208926677703857, "learning_rate": 4.872359631702473e-05, "loss": 5.0486, "step": 2829 }, { "epoch": 0.025546127459830294, "grad_norm": 3.3705296516418457, "learning_rate": 4.872314497201661e-05, "loss": 4.8716, "step": 2830 }, { "epoch": 0.02555515435999278, "grad_norm": 4.520303726196289, "learning_rate": 4.872269362700849e-05, "loss": 4.8469, "step": 2831 }, { "epoch": 0.025564181260155264, "grad_norm": 3.517604351043701, "learning_rate": 4.872224228200036e-05, "loss": 4.4402, "step": 2832 }, { "epoch": 0.025573208160317745, "grad_norm": 5.570470809936523, "learning_rate": 4.872179093699224e-05, "loss": 4.4548, "step": 2833 }, { "epoch": 0.02558223506048023, "grad_norm": 3.280771017074585, "learning_rate": 4.8721339591984114e-05, "loss": 4.7109, "step": 2834 }, { "epoch": 0.025591261960642715, "grad_norm": 4.2771992683410645, "learning_rate": 4.872088824697599e-05, "loss": 4.3262, "step": 2835 }, { "epoch": 0.0256002888608052, "grad_norm": 4.050005912780762, "learning_rate": 4.8720436901967866e-05, "loss": 4.1611, "step": 2836 }, { "epoch": 0.025609315760967685, "grad_norm": 3.73769211769104, "learning_rate": 4.871998555695974e-05, "loss": 4.1624, "step": 2837 }, { "epoch": 0.025618342661130167, "grad_norm": 5.26196813583374, "learning_rate": 4.871953421195162e-05, "loss": 3.9087, "step": 2838 }, { "epoch": 0.025627369561292652, "grad_norm": 5.648410320281982, "learning_rate": 4.87190828669435e-05, "loss": 3.6615, "step": 2839 }, { "epoch": 0.025636396461455137, "grad_norm": 3.7297170162200928, "learning_rate": 4.871863152193537e-05, "loss": 4.1897, "step": 2840 }, { "epoch": 0.025645423361617622, "grad_norm": 3.7471516132354736, "learning_rate": 4.871818017692725e-05, "loss": 4.4419, "step": 2841 }, { "epoch": 0.025654450261780103, "grad_norm": 3.299349069595337, "learning_rate": 4.871772883191912e-05, "loss": 4.994, "step": 2842 }, { "epoch": 0.02566347716194259, "grad_norm": 4.292388916015625, "learning_rate": 4.871727748691099e-05, "loss": 5.3774, "step": 2843 }, { "epoch": 0.025672504062105073, "grad_norm": 3.614367723464966, "learning_rate": 4.871682614190287e-05, "loss": 4.7488, "step": 2844 }, { "epoch": 0.02568153096226756, "grad_norm": 3.056459903717041, "learning_rate": 4.8716374796894745e-05, "loss": 5.1754, "step": 2845 }, { "epoch": 0.02569055786243004, "grad_norm": 5.133590221405029, "learning_rate": 4.8715923451886624e-05, "loss": 5.3818, "step": 2846 }, { "epoch": 0.025699584762592525, "grad_norm": 3.870086908340454, "learning_rate": 4.8715472106878504e-05, "loss": 4.6818, "step": 2847 }, { "epoch": 0.02570861166275501, "grad_norm": 4.143945217132568, "learning_rate": 4.8715020761870376e-05, "loss": 4.1389, "step": 2848 }, { "epoch": 0.025717638562917495, "grad_norm": 3.7002313137054443, "learning_rate": 4.871456941686225e-05, "loss": 5.1393, "step": 2849 }, { "epoch": 0.02572666546307998, "grad_norm": 3.7289071083068848, "learning_rate": 4.871411807185413e-05, "loss": 4.2242, "step": 2850 }, { "epoch": 0.02573569236324246, "grad_norm": 3.893921375274658, "learning_rate": 4.8713666726846e-05, "loss": 4.6797, "step": 2851 }, { "epoch": 0.025744719263404946, "grad_norm": 2.7182068824768066, "learning_rate": 4.871321538183788e-05, "loss": 4.5356, "step": 2852 }, { "epoch": 0.02575374616356743, "grad_norm": 5.315722942352295, "learning_rate": 4.871276403682976e-05, "loss": 5.1082, "step": 2853 }, { "epoch": 0.025762773063729916, "grad_norm": 3.4196319580078125, "learning_rate": 4.871231269182163e-05, "loss": 4.534, "step": 2854 }, { "epoch": 0.025771799963892398, "grad_norm": 3.3112950325012207, "learning_rate": 4.871186134681351e-05, "loss": 5.2151, "step": 2855 }, { "epoch": 0.025780826864054883, "grad_norm": 2.932799816131592, "learning_rate": 4.871141000180538e-05, "loss": 4.4503, "step": 2856 }, { "epoch": 0.025789853764217368, "grad_norm": 3.567491292953491, "learning_rate": 4.8710958656797255e-05, "loss": 4.4298, "step": 2857 }, { "epoch": 0.025798880664379853, "grad_norm": 3.7595911026000977, "learning_rate": 4.8710507311789135e-05, "loss": 4.9858, "step": 2858 }, { "epoch": 0.025807907564542334, "grad_norm": 3.4540607929229736, "learning_rate": 4.871005596678101e-05, "loss": 4.4702, "step": 2859 }, { "epoch": 0.02581693446470482, "grad_norm": 5.03973388671875, "learning_rate": 4.8709604621772886e-05, "loss": 4.1408, "step": 2860 }, { "epoch": 0.025825961364867304, "grad_norm": 3.2349770069122314, "learning_rate": 4.8709153276764765e-05, "loss": 4.6316, "step": 2861 }, { "epoch": 0.02583498826502979, "grad_norm": 4.374045372009277, "learning_rate": 4.870870193175664e-05, "loss": 4.8556, "step": 2862 }, { "epoch": 0.025844015165192274, "grad_norm": 6.8986101150512695, "learning_rate": 4.870825058674851e-05, "loss": 4.8683, "step": 2863 }, { "epoch": 0.025853042065354756, "grad_norm": 4.470893383026123, "learning_rate": 4.870779924174039e-05, "loss": 4.7828, "step": 2864 }, { "epoch": 0.02586206896551724, "grad_norm": 3.1098079681396484, "learning_rate": 4.870734789673226e-05, "loss": 4.4928, "step": 2865 }, { "epoch": 0.025871095865679726, "grad_norm": 3.912426710128784, "learning_rate": 4.870689655172414e-05, "loss": 3.8226, "step": 2866 }, { "epoch": 0.02588012276584221, "grad_norm": 3.390272855758667, "learning_rate": 4.8706445206716014e-05, "loss": 4.8357, "step": 2867 }, { "epoch": 0.025889149666004693, "grad_norm": 10.156221389770508, "learning_rate": 4.870599386170789e-05, "loss": 3.9699, "step": 2868 }, { "epoch": 0.025898176566167178, "grad_norm": 3.568025588989258, "learning_rate": 4.870554251669977e-05, "loss": 4.3577, "step": 2869 }, { "epoch": 0.025907203466329663, "grad_norm": 3.1478400230407715, "learning_rate": 4.870509117169164e-05, "loss": 4.3633, "step": 2870 }, { "epoch": 0.025916230366492148, "grad_norm": 3.7637529373168945, "learning_rate": 4.870463982668352e-05, "loss": 5.0444, "step": 2871 }, { "epoch": 0.02592525726665463, "grad_norm": 3.427062511444092, "learning_rate": 4.8704188481675396e-05, "loss": 5.4617, "step": 2872 }, { "epoch": 0.025934284166817114, "grad_norm": 4.569692134857178, "learning_rate": 4.870373713666727e-05, "loss": 4.8666, "step": 2873 }, { "epoch": 0.0259433110669796, "grad_norm": 4.267848491668701, "learning_rate": 4.870328579165915e-05, "loss": 5.2821, "step": 2874 }, { "epoch": 0.025952337967142084, "grad_norm": 4.9628801345825195, "learning_rate": 4.870283444665103e-05, "loss": 4.6808, "step": 2875 }, { "epoch": 0.02596136486730457, "grad_norm": 3.2811925411224365, "learning_rate": 4.870238310164289e-05, "loss": 4.246, "step": 2876 }, { "epoch": 0.02597039176746705, "grad_norm": 6.932538032531738, "learning_rate": 4.870193175663477e-05, "loss": 4.2798, "step": 2877 }, { "epoch": 0.025979418667629536, "grad_norm": 5.552825927734375, "learning_rate": 4.870148041162665e-05, "loss": 4.1902, "step": 2878 }, { "epoch": 0.02598844556779202, "grad_norm": 5.721077919006348, "learning_rate": 4.8701029066618524e-05, "loss": 5.3877, "step": 2879 }, { "epoch": 0.025997472467954506, "grad_norm": 3.7524983882904053, "learning_rate": 4.87005777216104e-05, "loss": 4.9523, "step": 2880 }, { "epoch": 0.026006499368116987, "grad_norm": 3.1564576625823975, "learning_rate": 4.8700126376602276e-05, "loss": 4.6667, "step": 2881 }, { "epoch": 0.026015526268279472, "grad_norm": 2.8473589420318604, "learning_rate": 4.8699675031594155e-05, "loss": 4.3212, "step": 2882 }, { "epoch": 0.026024553168441957, "grad_norm": 4.040017604827881, "learning_rate": 4.869922368658603e-05, "loss": 4.8072, "step": 2883 }, { "epoch": 0.026033580068604442, "grad_norm": 4.139339447021484, "learning_rate": 4.86987723415779e-05, "loss": 5.3354, "step": 2884 }, { "epoch": 0.026042606968766924, "grad_norm": 3.4946672916412354, "learning_rate": 4.869832099656978e-05, "loss": 4.6076, "step": 2885 }, { "epoch": 0.02605163386892941, "grad_norm": 3.6022861003875732, "learning_rate": 4.869786965156166e-05, "loss": 4.5629, "step": 2886 }, { "epoch": 0.026060660769091894, "grad_norm": 3.4958138465881348, "learning_rate": 4.869741830655353e-05, "loss": 4.7143, "step": 2887 }, { "epoch": 0.02606968766925438, "grad_norm": 4.3633131980896, "learning_rate": 4.869696696154541e-05, "loss": 4.4762, "step": 2888 }, { "epoch": 0.026078714569416864, "grad_norm": 2.8609635829925537, "learning_rate": 4.869651561653729e-05, "loss": 5.5866, "step": 2889 }, { "epoch": 0.026087741469579345, "grad_norm": 3.9322147369384766, "learning_rate": 4.8696064271529155e-05, "loss": 4.2984, "step": 2890 }, { "epoch": 0.02609676836974183, "grad_norm": 3.3129358291625977, "learning_rate": 4.8695612926521034e-05, "loss": 5.0594, "step": 2891 }, { "epoch": 0.026105795269904315, "grad_norm": 5.471735954284668, "learning_rate": 4.8695161581512914e-05, "loss": 5.3735, "step": 2892 }, { "epoch": 0.0261148221700668, "grad_norm": 3.975573778152466, "learning_rate": 4.8694710236504786e-05, "loss": 4.7673, "step": 2893 }, { "epoch": 0.026123849070229282, "grad_norm": 4.4524054527282715, "learning_rate": 4.8694258891496665e-05, "loss": 3.9503, "step": 2894 }, { "epoch": 0.026132875970391767, "grad_norm": 3.505549192428589, "learning_rate": 4.869380754648854e-05, "loss": 4.545, "step": 2895 }, { "epoch": 0.026141902870554252, "grad_norm": 3.556391477584839, "learning_rate": 4.869335620148041e-05, "loss": 4.8088, "step": 2896 }, { "epoch": 0.026150929770716737, "grad_norm": 3.6070687770843506, "learning_rate": 4.869290485647229e-05, "loss": 4.1798, "step": 2897 }, { "epoch": 0.026159956670879222, "grad_norm": 3.895012140274048, "learning_rate": 4.869245351146416e-05, "loss": 5.2115, "step": 2898 }, { "epoch": 0.026168983571041703, "grad_norm": 3.4919590950012207, "learning_rate": 4.869200216645604e-05, "loss": 5.3118, "step": 2899 }, { "epoch": 0.02617801047120419, "grad_norm": 2.753095865249634, "learning_rate": 4.869155082144792e-05, "loss": 4.7604, "step": 2900 }, { "epoch": 0.026187037371366673, "grad_norm": 2.914966344833374, "learning_rate": 4.869109947643979e-05, "loss": 5.2728, "step": 2901 }, { "epoch": 0.02619606427152916, "grad_norm": 3.4643733501434326, "learning_rate": 4.869064813143167e-05, "loss": 4.5109, "step": 2902 }, { "epoch": 0.02620509117169164, "grad_norm": 3.2434234619140625, "learning_rate": 4.8690196786423545e-05, "loss": 5.1122, "step": 2903 }, { "epoch": 0.026214118071854125, "grad_norm": 3.625154733657837, "learning_rate": 4.868974544141542e-05, "loss": 4.5527, "step": 2904 }, { "epoch": 0.02622314497201661, "grad_norm": 4.38958215713501, "learning_rate": 4.8689294096407296e-05, "loss": 4.1361, "step": 2905 }, { "epoch": 0.026232171872179095, "grad_norm": 4.001317501068115, "learning_rate": 4.868884275139917e-05, "loss": 5.1606, "step": 2906 }, { "epoch": 0.026241198772341576, "grad_norm": 2.8096158504486084, "learning_rate": 4.868839140639105e-05, "loss": 4.7255, "step": 2907 }, { "epoch": 0.02625022567250406, "grad_norm": 3.519559860229492, "learning_rate": 4.868794006138293e-05, "loss": 5.2991, "step": 2908 }, { "epoch": 0.026259252572666546, "grad_norm": 4.435334205627441, "learning_rate": 4.86874887163748e-05, "loss": 4.6937, "step": 2909 }, { "epoch": 0.02626827947282903, "grad_norm": 3.719085931777954, "learning_rate": 4.868703737136667e-05, "loss": 5.1185, "step": 2910 }, { "epoch": 0.026277306372991516, "grad_norm": 4.085842609405518, "learning_rate": 4.868658602635855e-05, "loss": 4.3289, "step": 2911 }, { "epoch": 0.026286333273153998, "grad_norm": 3.451251745223999, "learning_rate": 4.8686134681350424e-05, "loss": 4.6221, "step": 2912 }, { "epoch": 0.026295360173316483, "grad_norm": 3.384294033050537, "learning_rate": 4.86856833363423e-05, "loss": 4.933, "step": 2913 }, { "epoch": 0.026304387073478968, "grad_norm": 3.223203420639038, "learning_rate": 4.868523199133418e-05, "loss": 4.6273, "step": 2914 }, { "epoch": 0.026313413973641453, "grad_norm": 3.5560200214385986, "learning_rate": 4.8684780646326055e-05, "loss": 4.6516, "step": 2915 }, { "epoch": 0.026322440873803934, "grad_norm": 3.98654842376709, "learning_rate": 4.8684329301317934e-05, "loss": 4.3417, "step": 2916 }, { "epoch": 0.02633146777396642, "grad_norm": 3.2516283988952637, "learning_rate": 4.8683877956309806e-05, "loss": 4.0603, "step": 2917 }, { "epoch": 0.026340494674128904, "grad_norm": 4.04024076461792, "learning_rate": 4.868342661130168e-05, "loss": 5.1804, "step": 2918 }, { "epoch": 0.02634952157429139, "grad_norm": 4.076931953430176, "learning_rate": 4.868297526629356e-05, "loss": 4.8971, "step": 2919 }, { "epoch": 0.02635854847445387, "grad_norm": 4.291697978973389, "learning_rate": 4.868252392128543e-05, "loss": 4.6464, "step": 2920 }, { "epoch": 0.026367575374616356, "grad_norm": 3.4711782932281494, "learning_rate": 4.868207257627731e-05, "loss": 4.4724, "step": 2921 }, { "epoch": 0.02637660227477884, "grad_norm": 4.635206699371338, "learning_rate": 4.868162123126919e-05, "loss": 4.8462, "step": 2922 }, { "epoch": 0.026385629174941326, "grad_norm": 4.31581449508667, "learning_rate": 4.8681169886261055e-05, "loss": 5.3989, "step": 2923 }, { "epoch": 0.02639465607510381, "grad_norm": 2.876969814300537, "learning_rate": 4.8680718541252934e-05, "loss": 4.7906, "step": 2924 }, { "epoch": 0.026403682975266293, "grad_norm": 3.184758424758911, "learning_rate": 4.868026719624481e-05, "loss": 5.0711, "step": 2925 }, { "epoch": 0.026412709875428778, "grad_norm": 4.00273323059082, "learning_rate": 4.8679815851236686e-05, "loss": 5.1038, "step": 2926 }, { "epoch": 0.026421736775591263, "grad_norm": 3.1349411010742188, "learning_rate": 4.8679364506228565e-05, "loss": 5.0592, "step": 2927 }, { "epoch": 0.026430763675753748, "grad_norm": 4.844033718109131, "learning_rate": 4.867891316122044e-05, "loss": 4.6823, "step": 2928 }, { "epoch": 0.02643979057591623, "grad_norm": 3.9885776042938232, "learning_rate": 4.867846181621232e-05, "loss": 4.4552, "step": 2929 }, { "epoch": 0.026448817476078714, "grad_norm": 3.034698247909546, "learning_rate": 4.867801047120419e-05, "loss": 4.5764, "step": 2930 }, { "epoch": 0.0264578443762412, "grad_norm": 2.995758056640625, "learning_rate": 4.867755912619607e-05, "loss": 4.9824, "step": 2931 }, { "epoch": 0.026466871276403684, "grad_norm": 4.726057529449463, "learning_rate": 4.867710778118794e-05, "loss": 4.4848, "step": 2932 }, { "epoch": 0.026475898176566166, "grad_norm": 4.101163387298584, "learning_rate": 4.867665643617982e-05, "loss": 4.7725, "step": 2933 }, { "epoch": 0.02648492507672865, "grad_norm": 3.3773441314697266, "learning_rate": 4.867620509117169e-05, "loss": 5.0125, "step": 2934 }, { "epoch": 0.026493951976891136, "grad_norm": 5.8425397872924805, "learning_rate": 4.867575374616357e-05, "loss": 4.6524, "step": 2935 }, { "epoch": 0.02650297887705362, "grad_norm": 3.4071805477142334, "learning_rate": 4.867530240115545e-05, "loss": 4.5547, "step": 2936 }, { "epoch": 0.026512005777216106, "grad_norm": 3.8986120223999023, "learning_rate": 4.867485105614732e-05, "loss": 4.3628, "step": 2937 }, { "epoch": 0.026521032677378587, "grad_norm": 3.927957773208618, "learning_rate": 4.8674399711139196e-05, "loss": 4.6495, "step": 2938 }, { "epoch": 0.026530059577541072, "grad_norm": 2.9058780670166016, "learning_rate": 4.8673948366131075e-05, "loss": 4.9172, "step": 2939 }, { "epoch": 0.026539086477703557, "grad_norm": 3.222194194793701, "learning_rate": 4.867349702112295e-05, "loss": 4.5053, "step": 2940 }, { "epoch": 0.026548113377866042, "grad_norm": 4.172971725463867, "learning_rate": 4.867304567611483e-05, "loss": 4.8239, "step": 2941 }, { "epoch": 0.026557140278028524, "grad_norm": 3.471001148223877, "learning_rate": 4.86725943311067e-05, "loss": 4.2564, "step": 2942 }, { "epoch": 0.02656616717819101, "grad_norm": 3.790534019470215, "learning_rate": 4.867214298609857e-05, "loss": 4.4747, "step": 2943 }, { "epoch": 0.026575194078353494, "grad_norm": 3.884469509124756, "learning_rate": 4.867169164109045e-05, "loss": 5.089, "step": 2944 }, { "epoch": 0.02658422097851598, "grad_norm": 3.633124828338623, "learning_rate": 4.8671240296082324e-05, "loss": 5.23, "step": 2945 }, { "epoch": 0.02659324787867846, "grad_norm": 4.006176948547363, "learning_rate": 4.86707889510742e-05, "loss": 4.9354, "step": 2946 }, { "epoch": 0.026602274778840945, "grad_norm": 4.18398904800415, "learning_rate": 4.867033760606608e-05, "loss": 4.743, "step": 2947 }, { "epoch": 0.02661130167900343, "grad_norm": 4.331464767456055, "learning_rate": 4.8669886261057955e-05, "loss": 4.6231, "step": 2948 }, { "epoch": 0.026620328579165915, "grad_norm": 3.6201138496398926, "learning_rate": 4.8669434916049834e-05, "loss": 4.91, "step": 2949 }, { "epoch": 0.0266293554793284, "grad_norm": 3.2837517261505127, "learning_rate": 4.8668983571041706e-05, "loss": 4.6051, "step": 2950 }, { "epoch": 0.026638382379490882, "grad_norm": 4.568305969238281, "learning_rate": 4.866853222603358e-05, "loss": 5.0641, "step": 2951 }, { "epoch": 0.026647409279653367, "grad_norm": 5.065301418304443, "learning_rate": 4.866808088102546e-05, "loss": 4.6552, "step": 2952 }, { "epoch": 0.026656436179815852, "grad_norm": 3.4954586029052734, "learning_rate": 4.866762953601734e-05, "loss": 4.4613, "step": 2953 }, { "epoch": 0.026665463079978337, "grad_norm": 3.6814897060394287, "learning_rate": 4.866717819100921e-05, "loss": 4.3699, "step": 2954 }, { "epoch": 0.02667448998014082, "grad_norm": 3.202329158782959, "learning_rate": 4.866672684600109e-05, "loss": 4.7856, "step": 2955 }, { "epoch": 0.026683516880303303, "grad_norm": 2.7152252197265625, "learning_rate": 4.866627550099296e-05, "loss": 4.318, "step": 2956 }, { "epoch": 0.02669254378046579, "grad_norm": 6.881696701049805, "learning_rate": 4.8665824155984834e-05, "loss": 5.6133, "step": 2957 }, { "epoch": 0.026701570680628273, "grad_norm": 3.4210708141326904, "learning_rate": 4.866537281097671e-05, "loss": 4.2972, "step": 2958 }, { "epoch": 0.026710597580790755, "grad_norm": 4.3448052406311035, "learning_rate": 4.8664921465968586e-05, "loss": 4.5609, "step": 2959 }, { "epoch": 0.02671962448095324, "grad_norm": 4.191941738128662, "learning_rate": 4.8664470120960465e-05, "loss": 4.7823, "step": 2960 }, { "epoch": 0.026728651381115725, "grad_norm": 4.325887680053711, "learning_rate": 4.8664018775952344e-05, "loss": 4.284, "step": 2961 }, { "epoch": 0.02673767828127821, "grad_norm": 4.016884803771973, "learning_rate": 4.8663567430944216e-05, "loss": 4.6609, "step": 2962 }, { "epoch": 0.026746705181440695, "grad_norm": 4.669596195220947, "learning_rate": 4.8663116085936096e-05, "loss": 4.1627, "step": 2963 }, { "epoch": 0.026755732081603176, "grad_norm": 4.663469314575195, "learning_rate": 4.866266474092797e-05, "loss": 5.1808, "step": 2964 }, { "epoch": 0.02676475898176566, "grad_norm": 3.988818883895874, "learning_rate": 4.866221339591984e-05, "loss": 4.3918, "step": 2965 }, { "epoch": 0.026773785881928146, "grad_norm": 3.44157338142395, "learning_rate": 4.866176205091172e-05, "loss": 4.4889, "step": 2966 }, { "epoch": 0.02678281278209063, "grad_norm": 4.129915237426758, "learning_rate": 4.866131070590359e-05, "loss": 5.5609, "step": 2967 }, { "epoch": 0.026791839682253113, "grad_norm": 3.048875093460083, "learning_rate": 4.866085936089547e-05, "loss": 4.7071, "step": 2968 }, { "epoch": 0.026800866582415598, "grad_norm": 5.389829158782959, "learning_rate": 4.866040801588735e-05, "loss": 4.6977, "step": 2969 }, { "epoch": 0.026809893482578083, "grad_norm": 4.660400390625, "learning_rate": 4.8659956670879217e-05, "loss": 5.0908, "step": 2970 }, { "epoch": 0.026818920382740568, "grad_norm": 5.1364641189575195, "learning_rate": 4.8659505325871096e-05, "loss": 5.2356, "step": 2971 }, { "epoch": 0.02682794728290305, "grad_norm": 4.444894790649414, "learning_rate": 4.8659053980862975e-05, "loss": 5.0219, "step": 2972 }, { "epoch": 0.026836974183065535, "grad_norm": 3.301750659942627, "learning_rate": 4.865860263585485e-05, "loss": 4.5964, "step": 2973 }, { "epoch": 0.02684600108322802, "grad_norm": 3.4406659603118896, "learning_rate": 4.865815129084673e-05, "loss": 5.0346, "step": 2974 }, { "epoch": 0.026855027983390505, "grad_norm": 2.896117925643921, "learning_rate": 4.8657699945838606e-05, "loss": 4.7089, "step": 2975 }, { "epoch": 0.02686405488355299, "grad_norm": 3.2598323822021484, "learning_rate": 4.865724860083048e-05, "loss": 5.2225, "step": 2976 }, { "epoch": 0.02687308178371547, "grad_norm": 3.1488871574401855, "learning_rate": 4.865679725582235e-05, "loss": 5.0893, "step": 2977 }, { "epoch": 0.026882108683877956, "grad_norm": 3.595120429992676, "learning_rate": 4.865634591081423e-05, "loss": 4.6252, "step": 2978 }, { "epoch": 0.02689113558404044, "grad_norm": 3.9446022510528564, "learning_rate": 4.86558945658061e-05, "loss": 3.861, "step": 2979 }, { "epoch": 0.026900162484202926, "grad_norm": 3.9365475177764893, "learning_rate": 4.865544322079798e-05, "loss": 4.2193, "step": 2980 }, { "epoch": 0.026909189384365408, "grad_norm": 3.819478988647461, "learning_rate": 4.8654991875789854e-05, "loss": 4.8326, "step": 2981 }, { "epoch": 0.026918216284527893, "grad_norm": 3.9440369606018066, "learning_rate": 4.8654540530781734e-05, "loss": 4.941, "step": 2982 }, { "epoch": 0.026927243184690378, "grad_norm": 4.249216079711914, "learning_rate": 4.865408918577361e-05, "loss": 4.5277, "step": 2983 }, { "epoch": 0.026936270084852863, "grad_norm": 5.599897384643555, "learning_rate": 4.865363784076548e-05, "loss": 4.9884, "step": 2984 }, { "epoch": 0.026945296985015344, "grad_norm": 4.27876615524292, "learning_rate": 4.865318649575736e-05, "loss": 5.2817, "step": 2985 }, { "epoch": 0.02695432388517783, "grad_norm": 6.18069314956665, "learning_rate": 4.865273515074924e-05, "loss": 4.8192, "step": 2986 }, { "epoch": 0.026963350785340314, "grad_norm": 4.5977783203125, "learning_rate": 4.865228380574111e-05, "loss": 3.2904, "step": 2987 }, { "epoch": 0.0269723776855028, "grad_norm": 3.76609206199646, "learning_rate": 4.865183246073299e-05, "loss": 4.2565, "step": 2988 }, { "epoch": 0.026981404585665284, "grad_norm": 2.554642677307129, "learning_rate": 4.865138111572486e-05, "loss": 4.5891, "step": 2989 }, { "epoch": 0.026990431485827766, "grad_norm": 4.073953151702881, "learning_rate": 4.8650929770716734e-05, "loss": 4.4986, "step": 2990 }, { "epoch": 0.02699945838599025, "grad_norm": 3.0616331100463867, "learning_rate": 4.865047842570861e-05, "loss": 4.8218, "step": 2991 }, { "epoch": 0.027008485286152736, "grad_norm": 3.778743028640747, "learning_rate": 4.865002708070049e-05, "loss": 4.7583, "step": 2992 }, { "epoch": 0.02701751218631522, "grad_norm": 3.4877729415893555, "learning_rate": 4.8649575735692365e-05, "loss": 4.7391, "step": 2993 }, { "epoch": 0.027026539086477702, "grad_norm": 4.1688642501831055, "learning_rate": 4.8649124390684244e-05, "loss": 4.9633, "step": 2994 }, { "epoch": 0.027035565986640187, "grad_norm": 4.668467998504639, "learning_rate": 4.8648673045676116e-05, "loss": 5.1105, "step": 2995 }, { "epoch": 0.027044592886802672, "grad_norm": 4.309675693511963, "learning_rate": 4.8648221700667995e-05, "loss": 3.6456, "step": 2996 }, { "epoch": 0.027053619786965157, "grad_norm": 4.246055603027344, "learning_rate": 4.864777035565987e-05, "loss": 5.641, "step": 2997 }, { "epoch": 0.02706264668712764, "grad_norm": 3.3765408992767334, "learning_rate": 4.864731901065174e-05, "loss": 4.4483, "step": 2998 }, { "epoch": 0.027071673587290124, "grad_norm": 3.2233006954193115, "learning_rate": 4.864686766564362e-05, "loss": 4.957, "step": 2999 }, { "epoch": 0.02708070048745261, "grad_norm": 3.493006467819214, "learning_rate": 4.86464163206355e-05, "loss": 4.3528, "step": 3000 }, { "epoch": 0.027089727387615094, "grad_norm": 3.9568819999694824, "learning_rate": 4.864596497562737e-05, "loss": 4.82, "step": 3001 }, { "epoch": 0.02709875428777758, "grad_norm": 4.663990497589111, "learning_rate": 4.864551363061925e-05, "loss": 4.1712, "step": 3002 }, { "epoch": 0.02710778118794006, "grad_norm": 2.7399351596832275, "learning_rate": 4.864506228561112e-05, "loss": 5.2687, "step": 3003 }, { "epoch": 0.027116808088102545, "grad_norm": 3.4296939373016357, "learning_rate": 4.8644610940602996e-05, "loss": 4.887, "step": 3004 }, { "epoch": 0.02712583498826503, "grad_norm": 4.854471206665039, "learning_rate": 4.8644159595594875e-05, "loss": 4.5237, "step": 3005 }, { "epoch": 0.027134861888427515, "grad_norm": 4.167217254638672, "learning_rate": 4.864370825058675e-05, "loss": 4.5472, "step": 3006 }, { "epoch": 0.027143888788589997, "grad_norm": 2.834540605545044, "learning_rate": 4.8643256905578626e-05, "loss": 4.8145, "step": 3007 }, { "epoch": 0.027152915688752482, "grad_norm": 2.668982982635498, "learning_rate": 4.8642805560570506e-05, "loss": 4.9118, "step": 3008 }, { "epoch": 0.027161942588914967, "grad_norm": 4.436773777008057, "learning_rate": 4.864235421556238e-05, "loss": 4.6937, "step": 3009 }, { "epoch": 0.027170969489077452, "grad_norm": 4.508744239807129, "learning_rate": 4.864190287055426e-05, "loss": 5.0095, "step": 3010 }, { "epoch": 0.027179996389239933, "grad_norm": 6.163975715637207, "learning_rate": 4.864145152554613e-05, "loss": 4.3421, "step": 3011 }, { "epoch": 0.02718902328940242, "grad_norm": 4.311012268066406, "learning_rate": 4.8641000180538e-05, "loss": 4.7883, "step": 3012 }, { "epoch": 0.027198050189564903, "grad_norm": 6.027900218963623, "learning_rate": 4.864054883552988e-05, "loss": 5.3309, "step": 3013 }, { "epoch": 0.02720707708972739, "grad_norm": 2.8110101222991943, "learning_rate": 4.864009749052176e-05, "loss": 5.1747, "step": 3014 }, { "epoch": 0.027216103989889873, "grad_norm": 4.419084548950195, "learning_rate": 4.863964614551363e-05, "loss": 3.5212, "step": 3015 }, { "epoch": 0.027225130890052355, "grad_norm": 4.322904586791992, "learning_rate": 4.863919480050551e-05, "loss": 4.3902, "step": 3016 }, { "epoch": 0.02723415779021484, "grad_norm": 4.034881114959717, "learning_rate": 4.8638743455497385e-05, "loss": 4.4608, "step": 3017 }, { "epoch": 0.027243184690377325, "grad_norm": 2.965008020401001, "learning_rate": 4.863829211048926e-05, "loss": 5.1148, "step": 3018 }, { "epoch": 0.02725221159053981, "grad_norm": 3.574204206466675, "learning_rate": 4.863784076548114e-05, "loss": 4.9727, "step": 3019 }, { "epoch": 0.02726123849070229, "grad_norm": 3.8689186573028564, "learning_rate": 4.863738942047301e-05, "loss": 4.7049, "step": 3020 }, { "epoch": 0.027270265390864776, "grad_norm": 3.1406874656677246, "learning_rate": 4.863693807546489e-05, "loss": 5.1806, "step": 3021 }, { "epoch": 0.02727929229102726, "grad_norm": 4.10140323638916, "learning_rate": 4.863648673045677e-05, "loss": 4.8733, "step": 3022 }, { "epoch": 0.027288319191189746, "grad_norm": 6.317760944366455, "learning_rate": 4.863603538544864e-05, "loss": 4.3455, "step": 3023 }, { "epoch": 0.027297346091352228, "grad_norm": 3.7376577854156494, "learning_rate": 4.863558404044051e-05, "loss": 4.6659, "step": 3024 }, { "epoch": 0.027306372991514713, "grad_norm": 4.470543384552002, "learning_rate": 4.863513269543239e-05, "loss": 4.2378, "step": 3025 }, { "epoch": 0.027315399891677198, "grad_norm": 3.2487006187438965, "learning_rate": 4.8634681350424264e-05, "loss": 5.2829, "step": 3026 }, { "epoch": 0.027324426791839683, "grad_norm": 3.6753318309783936, "learning_rate": 4.8634230005416144e-05, "loss": 4.8897, "step": 3027 }, { "epoch": 0.027333453692002168, "grad_norm": 3.7040202617645264, "learning_rate": 4.8633778660408016e-05, "loss": 4.7596, "step": 3028 }, { "epoch": 0.02734248059216465, "grad_norm": 3.05545711517334, "learning_rate": 4.8633327315399895e-05, "loss": 4.5816, "step": 3029 }, { "epoch": 0.027351507492327135, "grad_norm": 3.606288194656372, "learning_rate": 4.8632875970391774e-05, "loss": 4.9598, "step": 3030 }, { "epoch": 0.02736053439248962, "grad_norm": 3.6272804737091064, "learning_rate": 4.863242462538364e-05, "loss": 5.2794, "step": 3031 }, { "epoch": 0.027369561292652105, "grad_norm": 3.18796706199646, "learning_rate": 4.863197328037552e-05, "loss": 4.7027, "step": 3032 }, { "epoch": 0.027378588192814586, "grad_norm": 4.405904293060303, "learning_rate": 4.86315219353674e-05, "loss": 4.8976, "step": 3033 }, { "epoch": 0.02738761509297707, "grad_norm": 3.590923309326172, "learning_rate": 4.863107059035927e-05, "loss": 4.5193, "step": 3034 }, { "epoch": 0.027396641993139556, "grad_norm": 3.673614025115967, "learning_rate": 4.863061924535115e-05, "loss": 4.6985, "step": 3035 }, { "epoch": 0.02740566889330204, "grad_norm": 3.7150917053222656, "learning_rate": 4.863016790034303e-05, "loss": 4.7856, "step": 3036 }, { "epoch": 0.027414695793464523, "grad_norm": 3.6256232261657715, "learning_rate": 4.8629716555334895e-05, "loss": 4.8178, "step": 3037 }, { "epoch": 0.027423722693627008, "grad_norm": 3.749350070953369, "learning_rate": 4.8629265210326775e-05, "loss": 4.5395, "step": 3038 }, { "epoch": 0.027432749593789493, "grad_norm": 4.222991943359375, "learning_rate": 4.8628813865318654e-05, "loss": 4.3451, "step": 3039 }, { "epoch": 0.027441776493951978, "grad_norm": 6.946835517883301, "learning_rate": 4.8628362520310526e-05, "loss": 4.5433, "step": 3040 }, { "epoch": 0.027450803394114463, "grad_norm": 3.8252882957458496, "learning_rate": 4.8627911175302405e-05, "loss": 4.5322, "step": 3041 }, { "epoch": 0.027459830294276944, "grad_norm": 4.374979019165039, "learning_rate": 4.862745983029428e-05, "loss": 5.6191, "step": 3042 }, { "epoch": 0.02746885719443943, "grad_norm": 3.998344659805298, "learning_rate": 4.862700848528616e-05, "loss": 5.7882, "step": 3043 }, { "epoch": 0.027477884094601914, "grad_norm": 3.9839766025543213, "learning_rate": 4.862655714027803e-05, "loss": 4.4377, "step": 3044 }, { "epoch": 0.0274869109947644, "grad_norm": 3.886669635772705, "learning_rate": 4.86261057952699e-05, "loss": 4.9337, "step": 3045 }, { "epoch": 0.02749593789492688, "grad_norm": 6.600761413574219, "learning_rate": 4.862565445026178e-05, "loss": 4.7642, "step": 3046 }, { "epoch": 0.027504964795089366, "grad_norm": 3.2590272426605225, "learning_rate": 4.862520310525366e-05, "loss": 4.2764, "step": 3047 }, { "epoch": 0.02751399169525185, "grad_norm": 4.651957035064697, "learning_rate": 4.862475176024553e-05, "loss": 4.7963, "step": 3048 }, { "epoch": 0.027523018595414336, "grad_norm": 3.695996046066284, "learning_rate": 4.862430041523741e-05, "loss": 5.0888, "step": 3049 }, { "epoch": 0.027532045495576817, "grad_norm": 5.686885356903076, "learning_rate": 4.862384907022929e-05, "loss": 4.8654, "step": 3050 }, { "epoch": 0.027541072395739302, "grad_norm": 3.888847589492798, "learning_rate": 4.862339772522116e-05, "loss": 4.6809, "step": 3051 }, { "epoch": 0.027550099295901787, "grad_norm": 4.074769973754883, "learning_rate": 4.8622946380213036e-05, "loss": 4.3557, "step": 3052 }, { "epoch": 0.027559126196064272, "grad_norm": 4.071746826171875, "learning_rate": 4.8622495035204916e-05, "loss": 4.4842, "step": 3053 }, { "epoch": 0.027568153096226757, "grad_norm": 7.0709452629089355, "learning_rate": 4.862204369019679e-05, "loss": 4.9515, "step": 3054 }, { "epoch": 0.02757717999638924, "grad_norm": 3.2315216064453125, "learning_rate": 4.862159234518867e-05, "loss": 5.1777, "step": 3055 }, { "epoch": 0.027586206896551724, "grad_norm": 4.08125114440918, "learning_rate": 4.862114100018054e-05, "loss": 4.8952, "step": 3056 }, { "epoch": 0.02759523379671421, "grad_norm": 7.139258861541748, "learning_rate": 4.862068965517241e-05, "loss": 5.2515, "step": 3057 }, { "epoch": 0.027604260696876694, "grad_norm": 4.076920032501221, "learning_rate": 4.862023831016429e-05, "loss": 5.5101, "step": 3058 }, { "epoch": 0.027613287597039175, "grad_norm": 2.5614054203033447, "learning_rate": 4.8619786965156164e-05, "loss": 4.444, "step": 3059 }, { "epoch": 0.02762231449720166, "grad_norm": 4.103952884674072, "learning_rate": 4.861933562014804e-05, "loss": 5.1383, "step": 3060 }, { "epoch": 0.027631341397364145, "grad_norm": 3.4607479572296143, "learning_rate": 4.861888427513992e-05, "loss": 4.9435, "step": 3061 }, { "epoch": 0.02764036829752663, "grad_norm": 3.9601025581359863, "learning_rate": 4.8618432930131795e-05, "loss": 4.7658, "step": 3062 }, { "epoch": 0.027649395197689112, "grad_norm": 4.022968292236328, "learning_rate": 4.8617981585123674e-05, "loss": 3.8047, "step": 3063 }, { "epoch": 0.027658422097851597, "grad_norm": 3.4508395195007324, "learning_rate": 4.861753024011555e-05, "loss": 4.6074, "step": 3064 }, { "epoch": 0.027667448998014082, "grad_norm": 3.225405216217041, "learning_rate": 4.861707889510742e-05, "loss": 4.3657, "step": 3065 }, { "epoch": 0.027676475898176567, "grad_norm": 3.799318552017212, "learning_rate": 4.86166275500993e-05, "loss": 4.1393, "step": 3066 }, { "epoch": 0.027685502798339052, "grad_norm": 3.7321949005126953, "learning_rate": 4.861617620509117e-05, "loss": 4.5635, "step": 3067 }, { "epoch": 0.027694529698501533, "grad_norm": 3.8585715293884277, "learning_rate": 4.861572486008305e-05, "loss": 5.1979, "step": 3068 }, { "epoch": 0.02770355659866402, "grad_norm": 3.1059420108795166, "learning_rate": 4.861527351507493e-05, "loss": 5.5563, "step": 3069 }, { "epoch": 0.027712583498826503, "grad_norm": 3.194164276123047, "learning_rate": 4.86148221700668e-05, "loss": 5.1501, "step": 3070 }, { "epoch": 0.02772161039898899, "grad_norm": 3.9895520210266113, "learning_rate": 4.8614370825058674e-05, "loss": 4.1905, "step": 3071 }, { "epoch": 0.02773063729915147, "grad_norm": 4.334673881530762, "learning_rate": 4.8613919480050554e-05, "loss": 4.9581, "step": 3072 }, { "epoch": 0.027739664199313955, "grad_norm": 4.428213119506836, "learning_rate": 4.8613468135042426e-05, "loss": 3.3644, "step": 3073 }, { "epoch": 0.02774869109947644, "grad_norm": 3.254685401916504, "learning_rate": 4.8613016790034305e-05, "loss": 4.8955, "step": 3074 }, { "epoch": 0.027757717999638925, "grad_norm": 3.548696279525757, "learning_rate": 4.8612565445026184e-05, "loss": 4.6767, "step": 3075 }, { "epoch": 0.02776674489980141, "grad_norm": 3.6561498641967773, "learning_rate": 4.861211410001806e-05, "loss": 4.4207, "step": 3076 }, { "epoch": 0.02777577179996389, "grad_norm": 4.687497138977051, "learning_rate": 4.8611662755009936e-05, "loss": 4.9552, "step": 3077 }, { "epoch": 0.027784798700126376, "grad_norm": 3.449702739715576, "learning_rate": 4.861121141000181e-05, "loss": 4.5047, "step": 3078 }, { "epoch": 0.02779382560028886, "grad_norm": 5.239184856414795, "learning_rate": 4.861076006499368e-05, "loss": 4.7063, "step": 3079 }, { "epoch": 0.027802852500451346, "grad_norm": 4.00656795501709, "learning_rate": 4.861030871998556e-05, "loss": 4.6121, "step": 3080 }, { "epoch": 0.027811879400613828, "grad_norm": 3.3127286434173584, "learning_rate": 4.860985737497743e-05, "loss": 4.0733, "step": 3081 }, { "epoch": 0.027820906300776313, "grad_norm": 4.218668460845947, "learning_rate": 4.860940602996931e-05, "loss": 4.5038, "step": 3082 }, { "epoch": 0.027829933200938798, "grad_norm": 5.285073280334473, "learning_rate": 4.860895468496119e-05, "loss": 4.9939, "step": 3083 }, { "epoch": 0.027838960101101283, "grad_norm": 3.947221040725708, "learning_rate": 4.860850333995306e-05, "loss": 4.818, "step": 3084 }, { "epoch": 0.027847987001263765, "grad_norm": 5.08430814743042, "learning_rate": 4.8608051994944936e-05, "loss": 5.4881, "step": 3085 }, { "epoch": 0.02785701390142625, "grad_norm": 4.677674293518066, "learning_rate": 4.8607600649936815e-05, "loss": 5.2226, "step": 3086 }, { "epoch": 0.027866040801588735, "grad_norm": 4.535323619842529, "learning_rate": 4.860714930492869e-05, "loss": 3.7971, "step": 3087 }, { "epoch": 0.02787506770175122, "grad_norm": 3.8188652992248535, "learning_rate": 4.860669795992057e-05, "loss": 4.8547, "step": 3088 }, { "epoch": 0.027884094601913705, "grad_norm": 4.470146179199219, "learning_rate": 4.860624661491244e-05, "loss": 4.5859, "step": 3089 }, { "epoch": 0.027893121502076186, "grad_norm": 3.8319599628448486, "learning_rate": 4.860579526990432e-05, "loss": 5.0941, "step": 3090 }, { "epoch": 0.02790214840223867, "grad_norm": 5.720096588134766, "learning_rate": 4.860534392489619e-05, "loss": 5.1719, "step": 3091 }, { "epoch": 0.027911175302401156, "grad_norm": 3.3065249919891357, "learning_rate": 4.860489257988807e-05, "loss": 3.4755, "step": 3092 }, { "epoch": 0.02792020220256364, "grad_norm": 3.8307220935821533, "learning_rate": 4.860444123487994e-05, "loss": 4.4125, "step": 3093 }, { "epoch": 0.027929229102726123, "grad_norm": 4.1806817054748535, "learning_rate": 4.860398988987182e-05, "loss": 5.0026, "step": 3094 }, { "epoch": 0.027938256002888608, "grad_norm": 3.2753350734710693, "learning_rate": 4.8603538544863695e-05, "loss": 4.5758, "step": 3095 }, { "epoch": 0.027947282903051093, "grad_norm": 2.843348264694214, "learning_rate": 4.8603087199855574e-05, "loss": 5.1469, "step": 3096 }, { "epoch": 0.027956309803213578, "grad_norm": 4.325685977935791, "learning_rate": 4.860263585484745e-05, "loss": 4.1096, "step": 3097 }, { "epoch": 0.02796533670337606, "grad_norm": 3.8770365715026855, "learning_rate": 4.860218450983932e-05, "loss": 4.7341, "step": 3098 }, { "epoch": 0.027974363603538544, "grad_norm": 3.8648109436035156, "learning_rate": 4.86017331648312e-05, "loss": 4.3568, "step": 3099 }, { "epoch": 0.02798339050370103, "grad_norm": 3.879591941833496, "learning_rate": 4.860128181982308e-05, "loss": 3.6814, "step": 3100 }, { "epoch": 0.027992417403863514, "grad_norm": 3.452491283416748, "learning_rate": 4.860083047481495e-05, "loss": 4.1424, "step": 3101 }, { "epoch": 0.028001444304026, "grad_norm": 9.396004676818848, "learning_rate": 4.860037912980683e-05, "loss": 4.7196, "step": 3102 }, { "epoch": 0.02801047120418848, "grad_norm": 4.13032865524292, "learning_rate": 4.85999277847987e-05, "loss": 4.8085, "step": 3103 }, { "epoch": 0.028019498104350966, "grad_norm": 3.0247600078582764, "learning_rate": 4.8599476439790574e-05, "loss": 4.722, "step": 3104 }, { "epoch": 0.02802852500451345, "grad_norm": 2.9665982723236084, "learning_rate": 4.859902509478245e-05, "loss": 3.8298, "step": 3105 }, { "epoch": 0.028037551904675936, "grad_norm": 4.900688171386719, "learning_rate": 4.8598573749774326e-05, "loss": 3.5314, "step": 3106 }, { "epoch": 0.028046578804838417, "grad_norm": 4.2499680519104, "learning_rate": 4.8598122404766205e-05, "loss": 4.5282, "step": 3107 }, { "epoch": 0.028055605705000902, "grad_norm": 10.667872428894043, "learning_rate": 4.8597671059758084e-05, "loss": 4.7198, "step": 3108 }, { "epoch": 0.028064632605163387, "grad_norm": 3.645470380783081, "learning_rate": 4.859721971474996e-05, "loss": 4.5453, "step": 3109 }, { "epoch": 0.028073659505325872, "grad_norm": 3.750183343887329, "learning_rate": 4.8596768369741836e-05, "loss": 5.1958, "step": 3110 }, { "epoch": 0.028082686405488354, "grad_norm": 4.088392734527588, "learning_rate": 4.859631702473371e-05, "loss": 4.3085, "step": 3111 }, { "epoch": 0.02809171330565084, "grad_norm": 4.090463638305664, "learning_rate": 4.859586567972558e-05, "loss": 4.3384, "step": 3112 }, { "epoch": 0.028100740205813324, "grad_norm": 3.8684661388397217, "learning_rate": 4.859541433471746e-05, "loss": 5.088, "step": 3113 }, { "epoch": 0.02810976710597581, "grad_norm": 3.8946642875671387, "learning_rate": 4.859496298970934e-05, "loss": 4.6815, "step": 3114 }, { "epoch": 0.028118794006138294, "grad_norm": 3.8557119369506836, "learning_rate": 4.859451164470121e-05, "loss": 4.756, "step": 3115 }, { "epoch": 0.028127820906300775, "grad_norm": 4.9173502922058105, "learning_rate": 4.859406029969309e-05, "loss": 5.4537, "step": 3116 }, { "epoch": 0.02813684780646326, "grad_norm": 3.233271360397339, "learning_rate": 4.8593608954684964e-05, "loss": 5.1558, "step": 3117 }, { "epoch": 0.028145874706625745, "grad_norm": 6.017108917236328, "learning_rate": 4.8593157609676836e-05, "loss": 4.6695, "step": 3118 }, { "epoch": 0.02815490160678823, "grad_norm": 3.655550956726074, "learning_rate": 4.8592706264668715e-05, "loss": 5.2291, "step": 3119 }, { "epoch": 0.028163928506950712, "grad_norm": 3.9032230377197266, "learning_rate": 4.859225491966059e-05, "loss": 5.1069, "step": 3120 }, { "epoch": 0.028172955407113197, "grad_norm": 4.574434757232666, "learning_rate": 4.859180357465247e-05, "loss": 5.3955, "step": 3121 }, { "epoch": 0.028181982307275682, "grad_norm": 9.436904907226562, "learning_rate": 4.8591352229644346e-05, "loss": 4.8449, "step": 3122 }, { "epoch": 0.028191009207438167, "grad_norm": 3.996143341064453, "learning_rate": 4.859090088463622e-05, "loss": 4.491, "step": 3123 }, { "epoch": 0.02820003610760065, "grad_norm": 4.1731390953063965, "learning_rate": 4.85904495396281e-05, "loss": 4.6705, "step": 3124 }, { "epoch": 0.028209063007763133, "grad_norm": 4.049007415771484, "learning_rate": 4.858999819461997e-05, "loss": 5.4925, "step": 3125 }, { "epoch": 0.02821808990792562, "grad_norm": 3.5157792568206787, "learning_rate": 4.858954684961184e-05, "loss": 4.5947, "step": 3126 }, { "epoch": 0.028227116808088103, "grad_norm": 3.512770652770996, "learning_rate": 4.858909550460372e-05, "loss": 4.8205, "step": 3127 }, { "epoch": 0.02823614370825059, "grad_norm": 3.78352427482605, "learning_rate": 4.8588644159595595e-05, "loss": 4.3645, "step": 3128 }, { "epoch": 0.02824517060841307, "grad_norm": 3.6252410411834717, "learning_rate": 4.8588192814587474e-05, "loss": 4.6468, "step": 3129 }, { "epoch": 0.028254197508575555, "grad_norm": 4.134542465209961, "learning_rate": 4.858774146957935e-05, "loss": 4.1447, "step": 3130 }, { "epoch": 0.02826322440873804, "grad_norm": 3.8481061458587646, "learning_rate": 4.858729012457122e-05, "loss": 5.3482, "step": 3131 }, { "epoch": 0.028272251308900525, "grad_norm": 5.7010178565979, "learning_rate": 4.85868387795631e-05, "loss": 4.5175, "step": 3132 }, { "epoch": 0.028281278209063006, "grad_norm": 3.2597713470458984, "learning_rate": 4.858638743455498e-05, "loss": 4.8352, "step": 3133 }, { "epoch": 0.02829030510922549, "grad_norm": 3.9154090881347656, "learning_rate": 4.858593608954685e-05, "loss": 5.2234, "step": 3134 }, { "epoch": 0.028299332009387976, "grad_norm": 3.5372376441955566, "learning_rate": 4.858548474453873e-05, "loss": 4.8856, "step": 3135 }, { "epoch": 0.02830835890955046, "grad_norm": 3.8698360919952393, "learning_rate": 4.858503339953061e-05, "loss": 4.3108, "step": 3136 }, { "epoch": 0.028317385809712943, "grad_norm": 3.3576419353485107, "learning_rate": 4.858458205452248e-05, "loss": 4.5082, "step": 3137 }, { "epoch": 0.028326412709875428, "grad_norm": 3.0630111694335938, "learning_rate": 4.858413070951435e-05, "loss": 5.5861, "step": 3138 }, { "epoch": 0.028335439610037913, "grad_norm": 3.619741916656494, "learning_rate": 4.858367936450623e-05, "loss": 4.6761, "step": 3139 }, { "epoch": 0.028344466510200398, "grad_norm": 3.3455045223236084, "learning_rate": 4.8583228019498105e-05, "loss": 5.167, "step": 3140 }, { "epoch": 0.028353493410362883, "grad_norm": 2.619802236557007, "learning_rate": 4.8582776674489984e-05, "loss": 4.4395, "step": 3141 }, { "epoch": 0.028362520310525365, "grad_norm": 3.7154922485351562, "learning_rate": 4.8582325329481856e-05, "loss": 4.6134, "step": 3142 }, { "epoch": 0.02837154721068785, "grad_norm": 3.873932123184204, "learning_rate": 4.8581873984473736e-05, "loss": 4.5905, "step": 3143 }, { "epoch": 0.028380574110850335, "grad_norm": 3.8936984539031982, "learning_rate": 4.8581422639465615e-05, "loss": 4.8443, "step": 3144 }, { "epoch": 0.02838960101101282, "grad_norm": 4.186642646789551, "learning_rate": 4.858097129445748e-05, "loss": 5.1441, "step": 3145 }, { "epoch": 0.0283986279111753, "grad_norm": 3.6890347003936768, "learning_rate": 4.858051994944936e-05, "loss": 5.0098, "step": 3146 }, { "epoch": 0.028407654811337786, "grad_norm": 2.3706753253936768, "learning_rate": 4.858006860444124e-05, "loss": 5.2003, "step": 3147 }, { "epoch": 0.02841668171150027, "grad_norm": 3.254042625427246, "learning_rate": 4.857961725943311e-05, "loss": 5.0291, "step": 3148 }, { "epoch": 0.028425708611662756, "grad_norm": 2.666705846786499, "learning_rate": 4.857916591442499e-05, "loss": 4.9124, "step": 3149 }, { "epoch": 0.028434735511825238, "grad_norm": 3.7120556831359863, "learning_rate": 4.857871456941686e-05, "loss": 4.3976, "step": 3150 }, { "epoch": 0.028443762411987723, "grad_norm": 3.4232914447784424, "learning_rate": 4.8578263224408736e-05, "loss": 4.4253, "step": 3151 }, { "epoch": 0.028452789312150208, "grad_norm": 4.8421854972839355, "learning_rate": 4.8577811879400615e-05, "loss": 4.5197, "step": 3152 }, { "epoch": 0.028461816212312693, "grad_norm": 2.925854444503784, "learning_rate": 4.8577360534392494e-05, "loss": 4.5415, "step": 3153 }, { "epoch": 0.028470843112475178, "grad_norm": 4.46476936340332, "learning_rate": 4.857690918938437e-05, "loss": 5.1884, "step": 3154 }, { "epoch": 0.02847987001263766, "grad_norm": 4.241002559661865, "learning_rate": 4.8576457844376246e-05, "loss": 4.5286, "step": 3155 }, { "epoch": 0.028488896912800144, "grad_norm": 2.7257509231567383, "learning_rate": 4.857600649936812e-05, "loss": 4.6729, "step": 3156 }, { "epoch": 0.02849792381296263, "grad_norm": 3.4347009658813477, "learning_rate": 4.857555515436e-05, "loss": 3.5054, "step": 3157 }, { "epoch": 0.028506950713125114, "grad_norm": 3.7914369106292725, "learning_rate": 4.857510380935187e-05, "loss": 2.9916, "step": 3158 }, { "epoch": 0.028515977613287596, "grad_norm": 3.5112056732177734, "learning_rate": 4.857465246434374e-05, "loss": 3.9126, "step": 3159 }, { "epoch": 0.02852500451345008, "grad_norm": 5.12378454208374, "learning_rate": 4.857420111933562e-05, "loss": 5.32, "step": 3160 }, { "epoch": 0.028534031413612566, "grad_norm": 3.1824986934661865, "learning_rate": 4.85737497743275e-05, "loss": 4.4717, "step": 3161 }, { "epoch": 0.02854305831377505, "grad_norm": 4.140571594238281, "learning_rate": 4.8573298429319374e-05, "loss": 4.8098, "step": 3162 }, { "epoch": 0.028552085213937532, "grad_norm": 3.1423799991607666, "learning_rate": 4.857284708431125e-05, "loss": 5.0921, "step": 3163 }, { "epoch": 0.028561112114100017, "grad_norm": 4.287172794342041, "learning_rate": 4.8572395739303125e-05, "loss": 4.859, "step": 3164 }, { "epoch": 0.028570139014262502, "grad_norm": 3.388424873352051, "learning_rate": 4.8571944394295e-05, "loss": 4.6593, "step": 3165 }, { "epoch": 0.028579165914424987, "grad_norm": 3.986053943634033, "learning_rate": 4.857149304928688e-05, "loss": 4.7338, "step": 3166 }, { "epoch": 0.028588192814587472, "grad_norm": 7.209136009216309, "learning_rate": 4.857104170427875e-05, "loss": 3.5601, "step": 3167 }, { "epoch": 0.028597219714749954, "grad_norm": 3.038358449935913, "learning_rate": 4.857059035927063e-05, "loss": 4.7343, "step": 3168 }, { "epoch": 0.02860624661491244, "grad_norm": 3.5803775787353516, "learning_rate": 4.857013901426251e-05, "loss": 4.1049, "step": 3169 }, { "epoch": 0.028615273515074924, "grad_norm": 3.517868757247925, "learning_rate": 4.856968766925438e-05, "loss": 5.0899, "step": 3170 }, { "epoch": 0.02862430041523741, "grad_norm": 3.7466683387756348, "learning_rate": 4.856923632424626e-05, "loss": 4.5059, "step": 3171 }, { "epoch": 0.02863332731539989, "grad_norm": 3.8458330631256104, "learning_rate": 4.856878497923813e-05, "loss": 4.7373, "step": 3172 }, { "epoch": 0.028642354215562375, "grad_norm": 3.620401382446289, "learning_rate": 4.8568333634230004e-05, "loss": 4.2467, "step": 3173 }, { "epoch": 0.02865138111572486, "grad_norm": 3.2609503269195557, "learning_rate": 4.8567882289221884e-05, "loss": 4.502, "step": 3174 }, { "epoch": 0.028660408015887345, "grad_norm": 3.7694175243377686, "learning_rate": 4.856743094421376e-05, "loss": 4.5101, "step": 3175 }, { "epoch": 0.028669434916049827, "grad_norm": 3.369333028793335, "learning_rate": 4.8566979599205635e-05, "loss": 4.8915, "step": 3176 }, { "epoch": 0.028678461816212312, "grad_norm": 3.2635185718536377, "learning_rate": 4.8566528254197515e-05, "loss": 4.613, "step": 3177 }, { "epoch": 0.028687488716374797, "grad_norm": 3.8089025020599365, "learning_rate": 4.856607690918939e-05, "loss": 4.7998, "step": 3178 }, { "epoch": 0.028696515616537282, "grad_norm": 4.131897926330566, "learning_rate": 4.856562556418126e-05, "loss": 4.7463, "step": 3179 }, { "epoch": 0.028705542516699767, "grad_norm": 4.2127485275268555, "learning_rate": 4.856517421917314e-05, "loss": 4.9042, "step": 3180 }, { "epoch": 0.02871456941686225, "grad_norm": 5.3731255531311035, "learning_rate": 4.856472287416501e-05, "loss": 4.5854, "step": 3181 }, { "epoch": 0.028723596317024733, "grad_norm": 4.505436420440674, "learning_rate": 4.856427152915689e-05, "loss": 4.1119, "step": 3182 }, { "epoch": 0.02873262321718722, "grad_norm": 3.772588014602661, "learning_rate": 4.856382018414877e-05, "loss": 4.6624, "step": 3183 }, { "epoch": 0.028741650117349703, "grad_norm": 3.9937901496887207, "learning_rate": 4.856336883914064e-05, "loss": 4.7948, "step": 3184 }, { "epoch": 0.028750677017512185, "grad_norm": 3.6137640476226807, "learning_rate": 4.8562917494132515e-05, "loss": 5.4135, "step": 3185 }, { "epoch": 0.02875970391767467, "grad_norm": 3.9802191257476807, "learning_rate": 4.8562466149124394e-05, "loss": 5.579, "step": 3186 }, { "epoch": 0.028768730817837155, "grad_norm": 3.4897348880767822, "learning_rate": 4.8562014804116266e-05, "loss": 4.2852, "step": 3187 }, { "epoch": 0.02877775771799964, "grad_norm": 3.796509265899658, "learning_rate": 4.8561563459108146e-05, "loss": 4.4128, "step": 3188 }, { "epoch": 0.02878678461816212, "grad_norm": 5.157465934753418, "learning_rate": 4.856111211410002e-05, "loss": 5.2967, "step": 3189 }, { "epoch": 0.028795811518324606, "grad_norm": 3.607098340988159, "learning_rate": 4.85606607690919e-05, "loss": 4.6315, "step": 3190 }, { "epoch": 0.02880483841848709, "grad_norm": 3.7854692935943604, "learning_rate": 4.856020942408378e-05, "loss": 4.0892, "step": 3191 }, { "epoch": 0.028813865318649576, "grad_norm": 3.428436040878296, "learning_rate": 4.855975807907564e-05, "loss": 4.2474, "step": 3192 }, { "epoch": 0.02882289221881206, "grad_norm": 3.7840981483459473, "learning_rate": 4.855930673406752e-05, "loss": 4.6857, "step": 3193 }, { "epoch": 0.028831919118974543, "grad_norm": 3.0508313179016113, "learning_rate": 4.85588553890594e-05, "loss": 5.0272, "step": 3194 }, { "epoch": 0.028840946019137028, "grad_norm": 3.435270309448242, "learning_rate": 4.855840404405127e-05, "loss": 4.9046, "step": 3195 }, { "epoch": 0.028849972919299513, "grad_norm": 2.9024834632873535, "learning_rate": 4.855795269904315e-05, "loss": 4.9319, "step": 3196 }, { "epoch": 0.028858999819461998, "grad_norm": 3.363957405090332, "learning_rate": 4.855750135403503e-05, "loss": 3.9558, "step": 3197 }, { "epoch": 0.02886802671962448, "grad_norm": 4.3866868019104, "learning_rate": 4.85570500090269e-05, "loss": 3.9213, "step": 3198 }, { "epoch": 0.028877053619786965, "grad_norm": 3.437765598297119, "learning_rate": 4.855659866401878e-05, "loss": 4.352, "step": 3199 }, { "epoch": 0.02888608051994945, "grad_norm": 4.321303844451904, "learning_rate": 4.8556147319010656e-05, "loss": 4.9003, "step": 3200 }, { "epoch": 0.028895107420111935, "grad_norm": 3.4338979721069336, "learning_rate": 4.855569597400253e-05, "loss": 5.0172, "step": 3201 }, { "epoch": 0.028904134320274416, "grad_norm": 3.7487573623657227, "learning_rate": 4.855524462899441e-05, "loss": 4.3198, "step": 3202 }, { "epoch": 0.0289131612204369, "grad_norm": 3.7399213314056396, "learning_rate": 4.855479328398628e-05, "loss": 5.1318, "step": 3203 }, { "epoch": 0.028922188120599386, "grad_norm": 3.813289165496826, "learning_rate": 4.855434193897816e-05, "loss": 4.8063, "step": 3204 }, { "epoch": 0.02893121502076187, "grad_norm": 3.7774863243103027, "learning_rate": 4.855389059397003e-05, "loss": 4.5453, "step": 3205 }, { "epoch": 0.028940241920924356, "grad_norm": 3.3746838569641113, "learning_rate": 4.8553439248961904e-05, "loss": 4.7323, "step": 3206 }, { "epoch": 0.028949268821086838, "grad_norm": 3.3977999687194824, "learning_rate": 4.8552987903953783e-05, "loss": 4.6214, "step": 3207 }, { "epoch": 0.028958295721249323, "grad_norm": 4.20737886428833, "learning_rate": 4.855253655894566e-05, "loss": 4.1534, "step": 3208 }, { "epoch": 0.028967322621411808, "grad_norm": 2.747857093811035, "learning_rate": 4.8552085213937535e-05, "loss": 4.8629, "step": 3209 }, { "epoch": 0.028976349521574293, "grad_norm": 3.4753901958465576, "learning_rate": 4.8551633868929414e-05, "loss": 5.0226, "step": 3210 }, { "epoch": 0.028985376421736774, "grad_norm": 4.122272491455078, "learning_rate": 4.8551182523921294e-05, "loss": 4.6885, "step": 3211 }, { "epoch": 0.02899440332189926, "grad_norm": 3.748608350753784, "learning_rate": 4.855073117891316e-05, "loss": 4.0718, "step": 3212 }, { "epoch": 0.029003430222061744, "grad_norm": 3.380566358566284, "learning_rate": 4.855027983390504e-05, "loss": 4.3582, "step": 3213 }, { "epoch": 0.02901245712222423, "grad_norm": 3.4004299640655518, "learning_rate": 4.854982848889692e-05, "loss": 3.9241, "step": 3214 }, { "epoch": 0.02902148402238671, "grad_norm": 3.0311784744262695, "learning_rate": 4.854937714388879e-05, "loss": 5.6184, "step": 3215 }, { "epoch": 0.029030510922549196, "grad_norm": 3.7231593132019043, "learning_rate": 4.854892579888067e-05, "loss": 4.6108, "step": 3216 }, { "epoch": 0.02903953782271168, "grad_norm": 3.953400135040283, "learning_rate": 4.854847445387254e-05, "loss": 4.983, "step": 3217 }, { "epoch": 0.029048564722874166, "grad_norm": 3.7770252227783203, "learning_rate": 4.8548023108864414e-05, "loss": 4.7712, "step": 3218 }, { "epoch": 0.02905759162303665, "grad_norm": 3.0492191314697266, "learning_rate": 4.8547571763856294e-05, "loss": 4.7134, "step": 3219 }, { "epoch": 0.029066618523199132, "grad_norm": 4.282784938812256, "learning_rate": 4.8547120418848166e-05, "loss": 3.7946, "step": 3220 }, { "epoch": 0.029075645423361617, "grad_norm": 3.5892481803894043, "learning_rate": 4.8546669073840045e-05, "loss": 4.8424, "step": 3221 }, { "epoch": 0.029084672323524102, "grad_norm": 4.826938152313232, "learning_rate": 4.8546217728831925e-05, "loss": 4.982, "step": 3222 }, { "epoch": 0.029093699223686587, "grad_norm": 3.0396082401275635, "learning_rate": 4.85457663838238e-05, "loss": 4.7527, "step": 3223 }, { "epoch": 0.02910272612384907, "grad_norm": 3.124699831008911, "learning_rate": 4.8545315038815676e-05, "loss": 4.4134, "step": 3224 }, { "epoch": 0.029111753024011554, "grad_norm": 4.2153215408325195, "learning_rate": 4.854486369380755e-05, "loss": 4.8513, "step": 3225 }, { "epoch": 0.02912077992417404, "grad_norm": 3.262352466583252, "learning_rate": 4.854441234879942e-05, "loss": 4.2272, "step": 3226 }, { "epoch": 0.029129806824336524, "grad_norm": 3.7712457180023193, "learning_rate": 4.85439610037913e-05, "loss": 4.2185, "step": 3227 }, { "epoch": 0.029138833724499005, "grad_norm": 3.242792844772339, "learning_rate": 4.854350965878317e-05, "loss": 4.7375, "step": 3228 }, { "epoch": 0.02914786062466149, "grad_norm": 4.064382553100586, "learning_rate": 4.854305831377505e-05, "loss": 4.8938, "step": 3229 }, { "epoch": 0.029156887524823975, "grad_norm": 3.5136170387268066, "learning_rate": 4.854260696876693e-05, "loss": 4.6387, "step": 3230 }, { "epoch": 0.02916591442498646, "grad_norm": 3.2688281536102295, "learning_rate": 4.8542155623758804e-05, "loss": 4.9231, "step": 3231 }, { "epoch": 0.029174941325148945, "grad_norm": 4.0424017906188965, "learning_rate": 4.8541704278750676e-05, "loss": 4.5831, "step": 3232 }, { "epoch": 0.029183968225311427, "grad_norm": 4.039957523345947, "learning_rate": 4.8541252933742556e-05, "loss": 4.9995, "step": 3233 }, { "epoch": 0.029192995125473912, "grad_norm": 3.8661890029907227, "learning_rate": 4.854080158873443e-05, "loss": 5.0708, "step": 3234 }, { "epoch": 0.029202022025636397, "grad_norm": 4.555014610290527, "learning_rate": 4.854035024372631e-05, "loss": 5.3825, "step": 3235 }, { "epoch": 0.029211048925798882, "grad_norm": 3.424449920654297, "learning_rate": 4.853989889871819e-05, "loss": 4.3335, "step": 3236 }, { "epoch": 0.029220075825961363, "grad_norm": 3.105963945388794, "learning_rate": 4.853944755371006e-05, "loss": 4.6114, "step": 3237 }, { "epoch": 0.02922910272612385, "grad_norm": 2.7691810131073, "learning_rate": 4.853899620870194e-05, "loss": 5.1493, "step": 3238 }, { "epoch": 0.029238129626286333, "grad_norm": 3.938718557357788, "learning_rate": 4.853854486369381e-05, "loss": 3.6032, "step": 3239 }, { "epoch": 0.02924715652644882, "grad_norm": 5.188103199005127, "learning_rate": 4.853809351868568e-05, "loss": 4.6435, "step": 3240 }, { "epoch": 0.0292561834266113, "grad_norm": 3.4975481033325195, "learning_rate": 4.853764217367756e-05, "loss": 5.3251, "step": 3241 }, { "epoch": 0.029265210326773785, "grad_norm": 4.043394565582275, "learning_rate": 4.8537190828669435e-05, "loss": 4.2581, "step": 3242 }, { "epoch": 0.02927423722693627, "grad_norm": 3.218714714050293, "learning_rate": 4.8536739483661314e-05, "loss": 5.1975, "step": 3243 }, { "epoch": 0.029283264127098755, "grad_norm": 3.368457794189453, "learning_rate": 4.8536288138653193e-05, "loss": 4.7104, "step": 3244 }, { "epoch": 0.02929229102726124, "grad_norm": 3.9224019050598145, "learning_rate": 4.853583679364506e-05, "loss": 4.8494, "step": 3245 }, { "epoch": 0.02930131792742372, "grad_norm": 3.2392165660858154, "learning_rate": 4.853538544863694e-05, "loss": 4.9859, "step": 3246 }, { "epoch": 0.029310344827586206, "grad_norm": 3.5190374851226807, "learning_rate": 4.853493410362882e-05, "loss": 4.5512, "step": 3247 }, { "epoch": 0.02931937172774869, "grad_norm": 3.4365313053131104, "learning_rate": 4.853448275862069e-05, "loss": 4.6854, "step": 3248 }, { "epoch": 0.029328398627911176, "grad_norm": 2.3701767921447754, "learning_rate": 4.853403141361257e-05, "loss": 4.8456, "step": 3249 }, { "epoch": 0.029337425528073658, "grad_norm": 3.9331257343292236, "learning_rate": 4.853358006860444e-05, "loss": 4.3242, "step": 3250 }, { "epoch": 0.029346452428236143, "grad_norm": 6.613626003265381, "learning_rate": 4.853312872359632e-05, "loss": 4.1642, "step": 3251 }, { "epoch": 0.029355479328398628, "grad_norm": 2.996788263320923, "learning_rate": 4.8532677378588193e-05, "loss": 5.2257, "step": 3252 }, { "epoch": 0.029364506228561113, "grad_norm": 2.873953342437744, "learning_rate": 4.853222603358007e-05, "loss": 4.8349, "step": 3253 }, { "epoch": 0.029373533128723598, "grad_norm": 2.940882682800293, "learning_rate": 4.8531774688571945e-05, "loss": 4.5072, "step": 3254 }, { "epoch": 0.02938256002888608, "grad_norm": 3.6502773761749268, "learning_rate": 4.8531323343563824e-05, "loss": 4.305, "step": 3255 }, { "epoch": 0.029391586929048565, "grad_norm": 4.3017258644104, "learning_rate": 4.85308719985557e-05, "loss": 5.0545, "step": 3256 }, { "epoch": 0.02940061382921105, "grad_norm": 3.4763684272766113, "learning_rate": 4.8530420653547576e-05, "loss": 4.4585, "step": 3257 }, { "epoch": 0.029409640729373535, "grad_norm": 4.760830879211426, "learning_rate": 4.8529969308539455e-05, "loss": 5.1953, "step": 3258 }, { "epoch": 0.029418667629536016, "grad_norm": 3.9324183464050293, "learning_rate": 4.852951796353132e-05, "loss": 5.2526, "step": 3259 }, { "epoch": 0.0294276945296985, "grad_norm": 3.3861920833587646, "learning_rate": 4.85290666185232e-05, "loss": 4.3167, "step": 3260 }, { "epoch": 0.029436721429860986, "grad_norm": 4.248815059661865, "learning_rate": 4.852861527351508e-05, "loss": 4.3906, "step": 3261 }, { "epoch": 0.02944574833002347, "grad_norm": 3.617396116256714, "learning_rate": 4.852816392850695e-05, "loss": 4.9129, "step": 3262 }, { "epoch": 0.029454775230185953, "grad_norm": 3.7370872497558594, "learning_rate": 4.852771258349883e-05, "loss": 4.7581, "step": 3263 }, { "epoch": 0.029463802130348438, "grad_norm": 3.381821870803833, "learning_rate": 4.8527261238490704e-05, "loss": 4.8552, "step": 3264 }, { "epoch": 0.029472829030510923, "grad_norm": 3.1184256076812744, "learning_rate": 4.8526809893482576e-05, "loss": 4.9872, "step": 3265 }, { "epoch": 0.029481855930673408, "grad_norm": 3.631035566329956, "learning_rate": 4.8526358548474455e-05, "loss": 3.7021, "step": 3266 }, { "epoch": 0.029490882830835893, "grad_norm": 3.1006832122802734, "learning_rate": 4.852590720346633e-05, "loss": 5.2655, "step": 3267 }, { "epoch": 0.029499909730998374, "grad_norm": 3.987962484359741, "learning_rate": 4.852545585845821e-05, "loss": 4.6057, "step": 3268 }, { "epoch": 0.02950893663116086, "grad_norm": 4.046266555786133, "learning_rate": 4.8525004513450086e-05, "loss": 5.1628, "step": 3269 }, { "epoch": 0.029517963531323344, "grad_norm": 3.614309072494507, "learning_rate": 4.852455316844196e-05, "loss": 5.1697, "step": 3270 }, { "epoch": 0.02952699043148583, "grad_norm": 2.849760055541992, "learning_rate": 4.852410182343384e-05, "loss": 4.6169, "step": 3271 }, { "epoch": 0.02953601733164831, "grad_norm": 4.1575093269348145, "learning_rate": 4.852365047842571e-05, "loss": 4.7284, "step": 3272 }, { "epoch": 0.029545044231810796, "grad_norm": 3.501034736633301, "learning_rate": 4.852319913341758e-05, "loss": 4.7884, "step": 3273 }, { "epoch": 0.02955407113197328, "grad_norm": 2.9260575771331787, "learning_rate": 4.852274778840946e-05, "loss": 4.3882, "step": 3274 }, { "epoch": 0.029563098032135766, "grad_norm": 3.499221086502075, "learning_rate": 4.852229644340134e-05, "loss": 4.5432, "step": 3275 }, { "epoch": 0.029572124932298247, "grad_norm": 4.057992935180664, "learning_rate": 4.8521845098393214e-05, "loss": 4.8199, "step": 3276 }, { "epoch": 0.029581151832460732, "grad_norm": 3.519599199295044, "learning_rate": 4.852139375338509e-05, "loss": 4.5922, "step": 3277 }, { "epoch": 0.029590178732623217, "grad_norm": 4.178920269012451, "learning_rate": 4.8520942408376966e-05, "loss": 4.449, "step": 3278 }, { "epoch": 0.029599205632785702, "grad_norm": 3.354627847671509, "learning_rate": 4.852049106336884e-05, "loss": 4.4547, "step": 3279 }, { "epoch": 0.029608232532948187, "grad_norm": 4.7290167808532715, "learning_rate": 4.852003971836072e-05, "loss": 5.135, "step": 3280 }, { "epoch": 0.02961725943311067, "grad_norm": 3.5589661598205566, "learning_rate": 4.851958837335259e-05, "loss": 5.5765, "step": 3281 }, { "epoch": 0.029626286333273154, "grad_norm": 4.91150426864624, "learning_rate": 4.851913702834447e-05, "loss": 5.1208, "step": 3282 }, { "epoch": 0.02963531323343564, "grad_norm": 3.761775255203247, "learning_rate": 4.851868568333635e-05, "loss": 5.0666, "step": 3283 }, { "epoch": 0.029644340133598124, "grad_norm": 3.757783889770508, "learning_rate": 4.851823433832822e-05, "loss": 4.6033, "step": 3284 }, { "epoch": 0.029653367033760605, "grad_norm": 3.882687568664551, "learning_rate": 4.85177829933201e-05, "loss": 4.7181, "step": 3285 }, { "epoch": 0.02966239393392309, "grad_norm": 3.06083607673645, "learning_rate": 4.851733164831197e-05, "loss": 5.0183, "step": 3286 }, { "epoch": 0.029671420834085575, "grad_norm": 2.639705181121826, "learning_rate": 4.8516880303303845e-05, "loss": 5.2646, "step": 3287 }, { "epoch": 0.02968044773424806, "grad_norm": 4.498006820678711, "learning_rate": 4.8516428958295724e-05, "loss": 5.0354, "step": 3288 }, { "epoch": 0.029689474634410542, "grad_norm": 3.7185654640197754, "learning_rate": 4.85159776132876e-05, "loss": 4.4061, "step": 3289 }, { "epoch": 0.029698501534573027, "grad_norm": 3.826876640319824, "learning_rate": 4.8515526268279476e-05, "loss": 4.8399, "step": 3290 }, { "epoch": 0.029707528434735512, "grad_norm": 3.508195638656616, "learning_rate": 4.8515074923271355e-05, "loss": 4.9428, "step": 3291 }, { "epoch": 0.029716555334897997, "grad_norm": 3.8829593658447266, "learning_rate": 4.851462357826322e-05, "loss": 4.436, "step": 3292 }, { "epoch": 0.029725582235060482, "grad_norm": 3.144883871078491, "learning_rate": 4.85141722332551e-05, "loss": 4.0316, "step": 3293 }, { "epoch": 0.029734609135222963, "grad_norm": 4.856651782989502, "learning_rate": 4.851372088824698e-05, "loss": 4.737, "step": 3294 }, { "epoch": 0.02974363603538545, "grad_norm": 3.3448431491851807, "learning_rate": 4.851326954323885e-05, "loss": 4.8298, "step": 3295 }, { "epoch": 0.029752662935547933, "grad_norm": 2.923678398132324, "learning_rate": 4.851281819823073e-05, "loss": 4.679, "step": 3296 }, { "epoch": 0.02976168983571042, "grad_norm": 4.412843704223633, "learning_rate": 4.851236685322261e-05, "loss": 4.7004, "step": 3297 }, { "epoch": 0.0297707167358729, "grad_norm": 4.139839172363281, "learning_rate": 4.851191550821448e-05, "loss": 5.0443, "step": 3298 }, { "epoch": 0.029779743636035385, "grad_norm": 4.753286838531494, "learning_rate": 4.8511464163206355e-05, "loss": 4.8385, "step": 3299 }, { "epoch": 0.02978877053619787, "grad_norm": 3.868368148803711, "learning_rate": 4.8511012818198234e-05, "loss": 4.5961, "step": 3300 }, { "epoch": 0.029797797436360355, "grad_norm": 3.3410167694091797, "learning_rate": 4.851056147319011e-05, "loss": 4.9327, "step": 3301 }, { "epoch": 0.029806824336522836, "grad_norm": 3.2341649532318115, "learning_rate": 4.8510110128181986e-05, "loss": 4.0574, "step": 3302 }, { "epoch": 0.02981585123668532, "grad_norm": 5.53917121887207, "learning_rate": 4.850965878317386e-05, "loss": 4.5809, "step": 3303 }, { "epoch": 0.029824878136847807, "grad_norm": 3.9396159648895264, "learning_rate": 4.850920743816574e-05, "loss": 5.4587, "step": 3304 }, { "epoch": 0.02983390503701029, "grad_norm": 6.173807144165039, "learning_rate": 4.850875609315762e-05, "loss": 4.968, "step": 3305 }, { "epoch": 0.029842931937172777, "grad_norm": 4.498316287994385, "learning_rate": 4.850830474814948e-05, "loss": 4.8443, "step": 3306 }, { "epoch": 0.029851958837335258, "grad_norm": 2.8552088737487793, "learning_rate": 4.850785340314136e-05, "loss": 4.6542, "step": 3307 }, { "epoch": 0.029860985737497743, "grad_norm": 3.822702407836914, "learning_rate": 4.850740205813324e-05, "loss": 3.8178, "step": 3308 }, { "epoch": 0.029870012637660228, "grad_norm": 3.8505847454071045, "learning_rate": 4.8506950713125114e-05, "loss": 4.3949, "step": 3309 }, { "epoch": 0.029879039537822713, "grad_norm": 5.024055480957031, "learning_rate": 4.850649936811699e-05, "loss": 4.804, "step": 3310 }, { "epoch": 0.029888066437985195, "grad_norm": 3.442960023880005, "learning_rate": 4.8506048023108865e-05, "loss": 4.9243, "step": 3311 }, { "epoch": 0.02989709333814768, "grad_norm": 4.167220592498779, "learning_rate": 4.850559667810074e-05, "loss": 4.4184, "step": 3312 }, { "epoch": 0.029906120238310165, "grad_norm": 4.1509881019592285, "learning_rate": 4.850514533309262e-05, "loss": 4.8477, "step": 3313 }, { "epoch": 0.02991514713847265, "grad_norm": 3.2893946170806885, "learning_rate": 4.8504693988084496e-05, "loss": 4.8234, "step": 3314 }, { "epoch": 0.02992417403863513, "grad_norm": 4.772158145904541, "learning_rate": 4.850424264307637e-05, "loss": 5.2189, "step": 3315 }, { "epoch": 0.029933200938797616, "grad_norm": 3.952890634536743, "learning_rate": 4.850379129806825e-05, "loss": 5.1295, "step": 3316 }, { "epoch": 0.0299422278389601, "grad_norm": 4.1492109298706055, "learning_rate": 4.850333995306012e-05, "loss": 5.0783, "step": 3317 }, { "epoch": 0.029951254739122586, "grad_norm": 3.7512950897216797, "learning_rate": 4.8502888608052e-05, "loss": 5.0982, "step": 3318 }, { "epoch": 0.02996028163928507, "grad_norm": 5.215361595153809, "learning_rate": 4.850243726304387e-05, "loss": 3.4606, "step": 3319 }, { "epoch": 0.029969308539447553, "grad_norm": 3.586840867996216, "learning_rate": 4.8501985918035745e-05, "loss": 5.1215, "step": 3320 }, { "epoch": 0.029978335439610038, "grad_norm": 3.0326387882232666, "learning_rate": 4.8501534573027624e-05, "loss": 5.2308, "step": 3321 }, { "epoch": 0.029987362339772523, "grad_norm": 4.185133457183838, "learning_rate": 4.85010832280195e-05, "loss": 5.4439, "step": 3322 }, { "epoch": 0.029996389239935008, "grad_norm": 3.4749345779418945, "learning_rate": 4.8500631883011376e-05, "loss": 5.0174, "step": 3323 }, { "epoch": 0.03000541614009749, "grad_norm": 3.5073516368865967, "learning_rate": 4.8500180538003255e-05, "loss": 4.875, "step": 3324 }, { "epoch": 0.030014443040259974, "grad_norm": 3.221311092376709, "learning_rate": 4.849972919299513e-05, "loss": 4.7208, "step": 3325 }, { "epoch": 0.03002346994042246, "grad_norm": 5.00808572769165, "learning_rate": 4.8499277847987e-05, "loss": 5.0976, "step": 3326 }, { "epoch": 0.030032496840584944, "grad_norm": 3.086181879043579, "learning_rate": 4.849882650297888e-05, "loss": 4.897, "step": 3327 }, { "epoch": 0.030041523740747426, "grad_norm": 3.2220778465270996, "learning_rate": 4.849837515797075e-05, "loss": 4.1307, "step": 3328 }, { "epoch": 0.03005055064090991, "grad_norm": 3.872194528579712, "learning_rate": 4.849792381296263e-05, "loss": 4.8795, "step": 3329 }, { "epoch": 0.030059577541072396, "grad_norm": 3.979012966156006, "learning_rate": 4.849747246795451e-05, "loss": 5.1876, "step": 3330 }, { "epoch": 0.03006860444123488, "grad_norm": 5.83949613571167, "learning_rate": 4.849702112294638e-05, "loss": 4.7519, "step": 3331 }, { "epoch": 0.030077631341397366, "grad_norm": 4.554261684417725, "learning_rate": 4.849656977793826e-05, "loss": 4.5048, "step": 3332 }, { "epoch": 0.030086658241559847, "grad_norm": 3.84932279586792, "learning_rate": 4.8496118432930134e-05, "loss": 4.9512, "step": 3333 }, { "epoch": 0.030095685141722332, "grad_norm": 4.339049339294434, "learning_rate": 4.849566708792201e-05, "loss": 5.0028, "step": 3334 }, { "epoch": 0.030104712041884817, "grad_norm": 4.584746360778809, "learning_rate": 4.8495215742913886e-05, "loss": 4.8265, "step": 3335 }, { "epoch": 0.030113738942047302, "grad_norm": 4.192340850830078, "learning_rate": 4.8494764397905765e-05, "loss": 4.7606, "step": 3336 }, { "epoch": 0.030122765842209784, "grad_norm": 2.810896635055542, "learning_rate": 4.849431305289764e-05, "loss": 4.9908, "step": 3337 }, { "epoch": 0.03013179274237227, "grad_norm": 3.1467671394348145, "learning_rate": 4.849386170788952e-05, "loss": 4.8001, "step": 3338 }, { "epoch": 0.030140819642534754, "grad_norm": 3.3721776008605957, "learning_rate": 4.849341036288139e-05, "loss": 4.853, "step": 3339 }, { "epoch": 0.03014984654269724, "grad_norm": 3.482556104660034, "learning_rate": 4.849295901787326e-05, "loss": 3.9868, "step": 3340 }, { "epoch": 0.03015887344285972, "grad_norm": 5.022948741912842, "learning_rate": 4.849250767286514e-05, "loss": 4.5244, "step": 3341 }, { "epoch": 0.030167900343022205, "grad_norm": 4.199187278747559, "learning_rate": 4.8492056327857013e-05, "loss": 4.842, "step": 3342 }, { "epoch": 0.03017692724318469, "grad_norm": 4.332314968109131, "learning_rate": 4.849160498284889e-05, "loss": 4.9142, "step": 3343 }, { "epoch": 0.030185954143347175, "grad_norm": 2.8702049255371094, "learning_rate": 4.849115363784077e-05, "loss": 4.9386, "step": 3344 }, { "epoch": 0.03019498104350966, "grad_norm": 3.2811057567596436, "learning_rate": 4.8490702292832644e-05, "loss": 4.8577, "step": 3345 }, { "epoch": 0.030204007943672142, "grad_norm": 3.2569217681884766, "learning_rate": 4.849025094782452e-05, "loss": 4.9346, "step": 3346 }, { "epoch": 0.030213034843834627, "grad_norm": 3.268425464630127, "learning_rate": 4.8489799602816396e-05, "loss": 4.7873, "step": 3347 }, { "epoch": 0.030222061743997112, "grad_norm": 3.548861026763916, "learning_rate": 4.848934825780827e-05, "loss": 4.7111, "step": 3348 }, { "epoch": 0.030231088644159597, "grad_norm": 4.196589946746826, "learning_rate": 4.848889691280015e-05, "loss": 5.0666, "step": 3349 }, { "epoch": 0.03024011554432208, "grad_norm": 3.3405189514160156, "learning_rate": 4.848844556779202e-05, "loss": 4.9673, "step": 3350 }, { "epoch": 0.030249142444484563, "grad_norm": 3.9628353118896484, "learning_rate": 4.84879942227839e-05, "loss": 4.872, "step": 3351 }, { "epoch": 0.03025816934464705, "grad_norm": 3.4124231338500977, "learning_rate": 4.848754287777578e-05, "loss": 4.3262, "step": 3352 }, { "epoch": 0.030267196244809533, "grad_norm": 4.118551731109619, "learning_rate": 4.8487091532767644e-05, "loss": 4.8823, "step": 3353 }, { "epoch": 0.030276223144972015, "grad_norm": 2.898083209991455, "learning_rate": 4.8486640187759524e-05, "loss": 5.0047, "step": 3354 }, { "epoch": 0.0302852500451345, "grad_norm": 3.7627737522125244, "learning_rate": 4.84861888427514e-05, "loss": 4.6711, "step": 3355 }, { "epoch": 0.030294276945296985, "grad_norm": 4.318266868591309, "learning_rate": 4.8485737497743275e-05, "loss": 5.1461, "step": 3356 }, { "epoch": 0.03030330384545947, "grad_norm": 3.562786817550659, "learning_rate": 4.8485286152735155e-05, "loss": 4.8115, "step": 3357 }, { "epoch": 0.030312330745621955, "grad_norm": 3.2364695072174072, "learning_rate": 4.8484834807727034e-05, "loss": 4.182, "step": 3358 }, { "epoch": 0.030321357645784437, "grad_norm": 2.7051358222961426, "learning_rate": 4.84843834627189e-05, "loss": 4.6192, "step": 3359 }, { "epoch": 0.03033038454594692, "grad_norm": 3.872195243835449, "learning_rate": 4.848393211771078e-05, "loss": 4.6247, "step": 3360 }, { "epoch": 0.030339411446109407, "grad_norm": 3.5870933532714844, "learning_rate": 4.848348077270266e-05, "loss": 5.1706, "step": 3361 }, { "epoch": 0.03034843834627189, "grad_norm": 5.290770530700684, "learning_rate": 4.848302942769453e-05, "loss": 4.3962, "step": 3362 }, { "epoch": 0.030357465246434373, "grad_norm": 6.373859882354736, "learning_rate": 4.848257808268641e-05, "loss": 3.8928, "step": 3363 }, { "epoch": 0.030366492146596858, "grad_norm": 5.42241096496582, "learning_rate": 4.848212673767828e-05, "loss": 4.5308, "step": 3364 }, { "epoch": 0.030375519046759343, "grad_norm": 3.477766513824463, "learning_rate": 4.848167539267016e-05, "loss": 5.0243, "step": 3365 }, { "epoch": 0.030384545946921828, "grad_norm": 3.6216001510620117, "learning_rate": 4.8481224047662034e-05, "loss": 4.8798, "step": 3366 }, { "epoch": 0.03039357284708431, "grad_norm": 3.163393259048462, "learning_rate": 4.8480772702653906e-05, "loss": 5.3277, "step": 3367 }, { "epoch": 0.030402599747246795, "grad_norm": 4.440666675567627, "learning_rate": 4.8480321357645786e-05, "loss": 4.9344, "step": 3368 }, { "epoch": 0.03041162664740928, "grad_norm": 2.72446870803833, "learning_rate": 4.8479870012637665e-05, "loss": 5.1141, "step": 3369 }, { "epoch": 0.030420653547571765, "grad_norm": 4.274067401885986, "learning_rate": 4.847941866762954e-05, "loss": 4.7116, "step": 3370 }, { "epoch": 0.03042968044773425, "grad_norm": 3.1258089542388916, "learning_rate": 4.8478967322621417e-05, "loss": 5.1308, "step": 3371 }, { "epoch": 0.03043870734789673, "grad_norm": 3.781973123550415, "learning_rate": 4.847851597761329e-05, "loss": 3.5517, "step": 3372 }, { "epoch": 0.030447734248059216, "grad_norm": 3.7603871822357178, "learning_rate": 4.847806463260516e-05, "loss": 4.4405, "step": 3373 }, { "epoch": 0.0304567611482217, "grad_norm": 5.192245960235596, "learning_rate": 4.847761328759704e-05, "loss": 3.6455, "step": 3374 }, { "epoch": 0.030465788048384186, "grad_norm": 3.1610097885131836, "learning_rate": 4.847716194258892e-05, "loss": 4.2498, "step": 3375 }, { "epoch": 0.030474814948546668, "grad_norm": 2.7141454219818115, "learning_rate": 4.847671059758079e-05, "loss": 4.6866, "step": 3376 }, { "epoch": 0.030483841848709153, "grad_norm": 3.18090558052063, "learning_rate": 4.847625925257267e-05, "loss": 5.0723, "step": 3377 }, { "epoch": 0.030492868748871638, "grad_norm": 3.3589043617248535, "learning_rate": 4.8475807907564544e-05, "loss": 5.0974, "step": 3378 }, { "epoch": 0.030501895649034123, "grad_norm": 3.3957273960113525, "learning_rate": 4.8475356562556423e-05, "loss": 5.4179, "step": 3379 }, { "epoch": 0.030510922549196604, "grad_norm": 3.4728775024414062, "learning_rate": 4.8474905217548296e-05, "loss": 4.5546, "step": 3380 }, { "epoch": 0.03051994944935909, "grad_norm": 3.582341194152832, "learning_rate": 4.847445387254017e-05, "loss": 4.1271, "step": 3381 }, { "epoch": 0.030528976349521574, "grad_norm": 3.418152093887329, "learning_rate": 4.847400252753205e-05, "loss": 4.9123, "step": 3382 }, { "epoch": 0.03053800324968406, "grad_norm": 3.229334592819214, "learning_rate": 4.847355118252393e-05, "loss": 4.9769, "step": 3383 }, { "epoch": 0.030547030149846544, "grad_norm": 4.6371378898620605, "learning_rate": 4.84730998375158e-05, "loss": 4.9415, "step": 3384 }, { "epoch": 0.030556057050009026, "grad_norm": 3.909541606903076, "learning_rate": 4.847264849250768e-05, "loss": 4.0375, "step": 3385 }, { "epoch": 0.03056508395017151, "grad_norm": 3.6174662113189697, "learning_rate": 4.847219714749955e-05, "loss": 4.6574, "step": 3386 }, { "epoch": 0.030574110850333996, "grad_norm": 4.117972373962402, "learning_rate": 4.8471745802491423e-05, "loss": 4.9977, "step": 3387 }, { "epoch": 0.03058313775049648, "grad_norm": 4.093498706817627, "learning_rate": 4.84712944574833e-05, "loss": 4.2254, "step": 3388 }, { "epoch": 0.030592164650658962, "grad_norm": 3.682971477508545, "learning_rate": 4.8470843112475175e-05, "loss": 4.5948, "step": 3389 }, { "epoch": 0.030601191550821447, "grad_norm": 4.208061695098877, "learning_rate": 4.8470391767467054e-05, "loss": 4.4592, "step": 3390 }, { "epoch": 0.030610218450983932, "grad_norm": 3.5603387355804443, "learning_rate": 4.8469940422458934e-05, "loss": 4.399, "step": 3391 }, { "epoch": 0.030619245351146417, "grad_norm": 5.035390853881836, "learning_rate": 4.8469489077450806e-05, "loss": 4.71, "step": 3392 }, { "epoch": 0.0306282722513089, "grad_norm": 3.4169929027557373, "learning_rate": 4.846903773244268e-05, "loss": 4.9488, "step": 3393 }, { "epoch": 0.030637299151471384, "grad_norm": 3.4276387691497803, "learning_rate": 4.846858638743456e-05, "loss": 4.3696, "step": 3394 }, { "epoch": 0.03064632605163387, "grad_norm": 5.661010265350342, "learning_rate": 4.846813504242643e-05, "loss": 4.8444, "step": 3395 }, { "epoch": 0.030655352951796354, "grad_norm": 3.9336819648742676, "learning_rate": 4.846768369741831e-05, "loss": 5.1186, "step": 3396 }, { "epoch": 0.03066437985195884, "grad_norm": 4.150436878204346, "learning_rate": 4.846723235241019e-05, "loss": 5.6673, "step": 3397 }, { "epoch": 0.03067340675212132, "grad_norm": 3.9557976722717285, "learning_rate": 4.846678100740206e-05, "loss": 4.6693, "step": 3398 }, { "epoch": 0.030682433652283805, "grad_norm": 3.039964199066162, "learning_rate": 4.846632966239394e-05, "loss": 4.87, "step": 3399 }, { "epoch": 0.03069146055244629, "grad_norm": 3.454141616821289, "learning_rate": 4.846587831738581e-05, "loss": 5.1364, "step": 3400 }, { "epoch": 0.030700487452608775, "grad_norm": 3.6670475006103516, "learning_rate": 4.8465426972377685e-05, "loss": 4.5222, "step": 3401 }, { "epoch": 0.030709514352771257, "grad_norm": 4.656257629394531, "learning_rate": 4.8464975627369565e-05, "loss": 4.7772, "step": 3402 }, { "epoch": 0.030718541252933742, "grad_norm": 4.750531196594238, "learning_rate": 4.846452428236144e-05, "loss": 4.2138, "step": 3403 }, { "epoch": 0.030727568153096227, "grad_norm": 4.060482978820801, "learning_rate": 4.8464072937353316e-05, "loss": 5.5021, "step": 3404 }, { "epoch": 0.030736595053258712, "grad_norm": 3.886070489883423, "learning_rate": 4.8463621592345196e-05, "loss": 4.0148, "step": 3405 }, { "epoch": 0.030745621953421193, "grad_norm": 3.6102325916290283, "learning_rate": 4.846317024733706e-05, "loss": 4.9687, "step": 3406 }, { "epoch": 0.03075464885358368, "grad_norm": 3.0034375190734863, "learning_rate": 4.846271890232894e-05, "loss": 4.893, "step": 3407 }, { "epoch": 0.030763675753746163, "grad_norm": 2.801868438720703, "learning_rate": 4.846226755732082e-05, "loss": 5.5203, "step": 3408 }, { "epoch": 0.03077270265390865, "grad_norm": 3.2301480770111084, "learning_rate": 4.846181621231269e-05, "loss": 4.6103, "step": 3409 }, { "epoch": 0.030781729554071133, "grad_norm": 5.724645137786865, "learning_rate": 4.846136486730457e-05, "loss": 3.0882, "step": 3410 }, { "epoch": 0.030790756454233615, "grad_norm": 4.029398441314697, "learning_rate": 4.8460913522296444e-05, "loss": 5.5183, "step": 3411 }, { "epoch": 0.0307997833543961, "grad_norm": 4.093636512756348, "learning_rate": 4.846046217728832e-05, "loss": 5.0804, "step": 3412 }, { "epoch": 0.030808810254558585, "grad_norm": 4.343344211578369, "learning_rate": 4.8460010832280196e-05, "loss": 5.2487, "step": 3413 }, { "epoch": 0.03081783715472107, "grad_norm": 3.1026203632354736, "learning_rate": 4.8459559487272075e-05, "loss": 4.8332, "step": 3414 }, { "epoch": 0.03082686405488355, "grad_norm": 2.9617109298706055, "learning_rate": 4.845910814226395e-05, "loss": 4.1793, "step": 3415 }, { "epoch": 0.030835890955046037, "grad_norm": 3.7267544269561768, "learning_rate": 4.8458656797255827e-05, "loss": 4.7342, "step": 3416 }, { "epoch": 0.03084491785520852, "grad_norm": 3.2489356994628906, "learning_rate": 4.84582054522477e-05, "loss": 4.9189, "step": 3417 }, { "epoch": 0.030853944755371007, "grad_norm": 2.7897942066192627, "learning_rate": 4.845775410723958e-05, "loss": 5.176, "step": 3418 }, { "epoch": 0.03086297165553349, "grad_norm": 4.365467071533203, "learning_rate": 4.845730276223146e-05, "loss": 4.8546, "step": 3419 }, { "epoch": 0.030871998555695973, "grad_norm": 4.00057315826416, "learning_rate": 4.845685141722332e-05, "loss": 4.9314, "step": 3420 }, { "epoch": 0.030881025455858458, "grad_norm": 4.4135565757751465, "learning_rate": 4.84564000722152e-05, "loss": 4.8008, "step": 3421 }, { "epoch": 0.030890052356020943, "grad_norm": 3.1609933376312256, "learning_rate": 4.845594872720708e-05, "loss": 5.3212, "step": 3422 }, { "epoch": 0.030899079256183428, "grad_norm": 2.8549370765686035, "learning_rate": 4.8455497382198954e-05, "loss": 5.1297, "step": 3423 }, { "epoch": 0.03090810615634591, "grad_norm": 3.634878396987915, "learning_rate": 4.8455046037190833e-05, "loss": 4.8581, "step": 3424 }, { "epoch": 0.030917133056508395, "grad_norm": 3.859683036804199, "learning_rate": 4.8454594692182706e-05, "loss": 4.7168, "step": 3425 }, { "epoch": 0.03092615995667088, "grad_norm": 5.046659469604492, "learning_rate": 4.845414334717458e-05, "loss": 5.2872, "step": 3426 }, { "epoch": 0.030935186856833365, "grad_norm": 4.938817977905273, "learning_rate": 4.845369200216646e-05, "loss": 4.7544, "step": 3427 }, { "epoch": 0.030944213756995846, "grad_norm": 4.3775105476379395, "learning_rate": 4.845324065715833e-05, "loss": 4.5619, "step": 3428 }, { "epoch": 0.03095324065715833, "grad_norm": 3.3972103595733643, "learning_rate": 4.845278931215021e-05, "loss": 4.9859, "step": 3429 }, { "epoch": 0.030962267557320816, "grad_norm": 4.02109432220459, "learning_rate": 4.845233796714209e-05, "loss": 4.4495, "step": 3430 }, { "epoch": 0.0309712944574833, "grad_norm": 2.8470921516418457, "learning_rate": 4.845188662213396e-05, "loss": 4.7195, "step": 3431 }, { "epoch": 0.030980321357645786, "grad_norm": 9.997748374938965, "learning_rate": 4.845143527712584e-05, "loss": 4.858, "step": 3432 }, { "epoch": 0.030989348257808268, "grad_norm": 3.349097490310669, "learning_rate": 4.845098393211771e-05, "loss": 4.9197, "step": 3433 }, { "epoch": 0.030998375157970753, "grad_norm": 3.921010732650757, "learning_rate": 4.8450532587109585e-05, "loss": 5.204, "step": 3434 }, { "epoch": 0.031007402058133238, "grad_norm": 3.415839195251465, "learning_rate": 4.8450081242101464e-05, "loss": 5.1587, "step": 3435 }, { "epoch": 0.031016428958295723, "grad_norm": 3.0282416343688965, "learning_rate": 4.8449629897093344e-05, "loss": 4.9542, "step": 3436 }, { "epoch": 0.031025455858458204, "grad_norm": 4.789288520812988, "learning_rate": 4.8449178552085216e-05, "loss": 5.1537, "step": 3437 }, { "epoch": 0.03103448275862069, "grad_norm": 4.611804008483887, "learning_rate": 4.8448727207077095e-05, "loss": 3.6604, "step": 3438 }, { "epoch": 0.031043509658783174, "grad_norm": 5.759646892547607, "learning_rate": 4.844827586206897e-05, "loss": 4.9183, "step": 3439 }, { "epoch": 0.03105253655894566, "grad_norm": 3.683340072631836, "learning_rate": 4.844782451706084e-05, "loss": 4.2625, "step": 3440 }, { "epoch": 0.03106156345910814, "grad_norm": 4.17112398147583, "learning_rate": 4.844737317205272e-05, "loss": 4.8739, "step": 3441 }, { "epoch": 0.031070590359270626, "grad_norm": 3.726879358291626, "learning_rate": 4.844692182704459e-05, "loss": 4.669, "step": 3442 }, { "epoch": 0.03107961725943311, "grad_norm": 2.9476027488708496, "learning_rate": 4.844647048203647e-05, "loss": 5.1847, "step": 3443 }, { "epoch": 0.031088644159595596, "grad_norm": 3.581543207168579, "learning_rate": 4.844601913702835e-05, "loss": 4.2708, "step": 3444 }, { "epoch": 0.03109767105975808, "grad_norm": 4.080275535583496, "learning_rate": 4.844556779202022e-05, "loss": 4.9385, "step": 3445 }, { "epoch": 0.031106697959920562, "grad_norm": 3.2600133419036865, "learning_rate": 4.84451164470121e-05, "loss": 4.9575, "step": 3446 }, { "epoch": 0.031115724860083047, "grad_norm": 3.141758441925049, "learning_rate": 4.8444665102003975e-05, "loss": 5.0245, "step": 3447 }, { "epoch": 0.031124751760245532, "grad_norm": 3.385939598083496, "learning_rate": 4.844421375699585e-05, "loss": 5.165, "step": 3448 }, { "epoch": 0.031133778660408017, "grad_norm": 4.139033794403076, "learning_rate": 4.8443762411987726e-05, "loss": 4.932, "step": 3449 }, { "epoch": 0.0311428055605705, "grad_norm": 4.372469425201416, "learning_rate": 4.84433110669796e-05, "loss": 4.7375, "step": 3450 }, { "epoch": 0.031151832460732984, "grad_norm": 4.626279354095459, "learning_rate": 4.844285972197148e-05, "loss": 4.6145, "step": 3451 }, { "epoch": 0.03116085936089547, "grad_norm": 4.165383338928223, "learning_rate": 4.844240837696336e-05, "loss": 4.8907, "step": 3452 }, { "epoch": 0.031169886261057954, "grad_norm": 3.9324281215667725, "learning_rate": 4.844195703195522e-05, "loss": 4.8278, "step": 3453 }, { "epoch": 0.031178913161220435, "grad_norm": 6.560367584228516, "learning_rate": 4.84415056869471e-05, "loss": 4.7675, "step": 3454 }, { "epoch": 0.03118794006138292, "grad_norm": 4.093046188354492, "learning_rate": 4.844105434193898e-05, "loss": 4.8304, "step": 3455 }, { "epoch": 0.031196966961545405, "grad_norm": 2.9703259468078613, "learning_rate": 4.8440602996930854e-05, "loss": 5.0916, "step": 3456 }, { "epoch": 0.03120599386170789, "grad_norm": 3.212397336959839, "learning_rate": 4.844015165192273e-05, "loss": 4.8048, "step": 3457 }, { "epoch": 0.031215020761870375, "grad_norm": 3.638359785079956, "learning_rate": 4.843970030691461e-05, "loss": 4.7702, "step": 3458 }, { "epoch": 0.031224047662032857, "grad_norm": 4.119572639465332, "learning_rate": 4.8439248961906485e-05, "loss": 4.6047, "step": 3459 }, { "epoch": 0.031233074562195342, "grad_norm": 3.8485286235809326, "learning_rate": 4.843879761689836e-05, "loss": 4.3201, "step": 3460 }, { "epoch": 0.031242101462357827, "grad_norm": 3.05790114402771, "learning_rate": 4.8438346271890237e-05, "loss": 4.8552, "step": 3461 }, { "epoch": 0.03125112836252031, "grad_norm": 3.5941710472106934, "learning_rate": 4.843789492688211e-05, "loss": 4.9299, "step": 3462 }, { "epoch": 0.0312601552626828, "grad_norm": 3.672642707824707, "learning_rate": 4.843744358187399e-05, "loss": 4.9525, "step": 3463 }, { "epoch": 0.03126918216284528, "grad_norm": 4.042890548706055, "learning_rate": 4.843699223686586e-05, "loss": 4.6494, "step": 3464 }, { "epoch": 0.03127820906300776, "grad_norm": 3.869009256362915, "learning_rate": 4.843654089185774e-05, "loss": 5.0772, "step": 3465 }, { "epoch": 0.03128723596317025, "grad_norm": 3.6827783584594727, "learning_rate": 4.843608954684962e-05, "loss": 4.9267, "step": 3466 }, { "epoch": 0.03129626286333273, "grad_norm": 3.5017776489257812, "learning_rate": 4.8435638201841485e-05, "loss": 5.0787, "step": 3467 }, { "epoch": 0.03130528976349522, "grad_norm": 4.339308738708496, "learning_rate": 4.8435186856833364e-05, "loss": 4.3932, "step": 3468 }, { "epoch": 0.0313143166636577, "grad_norm": 4.313255786895752, "learning_rate": 4.8434735511825243e-05, "loss": 4.6952, "step": 3469 }, { "epoch": 0.03132334356382018, "grad_norm": 4.056318283081055, "learning_rate": 4.8434284166817116e-05, "loss": 4.92, "step": 3470 }, { "epoch": 0.03133237046398267, "grad_norm": 2.8643908500671387, "learning_rate": 4.8433832821808995e-05, "loss": 4.413, "step": 3471 }, { "epoch": 0.03134139736414515, "grad_norm": 3.242748498916626, "learning_rate": 4.843338147680087e-05, "loss": 4.8825, "step": 3472 }, { "epoch": 0.03135042426430764, "grad_norm": 3.594971179962158, "learning_rate": 4.843293013179274e-05, "loss": 4.6555, "step": 3473 }, { "epoch": 0.03135945116447012, "grad_norm": 4.116180419921875, "learning_rate": 4.843247878678462e-05, "loss": 3.843, "step": 3474 }, { "epoch": 0.0313684780646326, "grad_norm": 3.611402750015259, "learning_rate": 4.84320274417765e-05, "loss": 4.4345, "step": 3475 }, { "epoch": 0.03137750496479509, "grad_norm": 3.4819321632385254, "learning_rate": 4.843157609676837e-05, "loss": 5.3196, "step": 3476 }, { "epoch": 0.03138653186495757, "grad_norm": 3.7700788974761963, "learning_rate": 4.843112475176025e-05, "loss": 5.0877, "step": 3477 }, { "epoch": 0.031395558765120055, "grad_norm": 3.6203033924102783, "learning_rate": 4.843067340675212e-05, "loss": 5.0381, "step": 3478 }, { "epoch": 0.03140458566528254, "grad_norm": 2.9733262062072754, "learning_rate": 4.8430222061744e-05, "loss": 4.6786, "step": 3479 }, { "epoch": 0.031413612565445025, "grad_norm": 3.9393768310546875, "learning_rate": 4.8429770716735874e-05, "loss": 4.7756, "step": 3480 }, { "epoch": 0.03142263946560751, "grad_norm": 3.2669548988342285, "learning_rate": 4.842931937172775e-05, "loss": 4.6791, "step": 3481 }, { "epoch": 0.031431666365769995, "grad_norm": 3.093104124069214, "learning_rate": 4.8428868026719626e-05, "loss": 4.6618, "step": 3482 }, { "epoch": 0.031440693265932476, "grad_norm": 3.6329615116119385, "learning_rate": 4.8428416681711505e-05, "loss": 4.7846, "step": 3483 }, { "epoch": 0.031449720166094965, "grad_norm": 7.691922664642334, "learning_rate": 4.842796533670338e-05, "loss": 5.3456, "step": 3484 }, { "epoch": 0.031458747066257446, "grad_norm": 3.224261522293091, "learning_rate": 4.842751399169526e-05, "loss": 5.2031, "step": 3485 }, { "epoch": 0.031467773966419935, "grad_norm": 4.4034528732299805, "learning_rate": 4.842706264668713e-05, "loss": 4.6424, "step": 3486 }, { "epoch": 0.031476800866582416, "grad_norm": 4.051950454711914, "learning_rate": 4.8426611301679e-05, "loss": 4.8442, "step": 3487 }, { "epoch": 0.0314858277667449, "grad_norm": 3.7909765243530273, "learning_rate": 4.842615995667088e-05, "loss": 4.9556, "step": 3488 }, { "epoch": 0.031494854666907386, "grad_norm": 3.6491427421569824, "learning_rate": 4.8425708611662754e-05, "loss": 4.599, "step": 3489 }, { "epoch": 0.03150388156706987, "grad_norm": 2.942965269088745, "learning_rate": 4.842525726665463e-05, "loss": 4.6802, "step": 3490 }, { "epoch": 0.03151290846723235, "grad_norm": 4.515268802642822, "learning_rate": 4.842480592164651e-05, "loss": 4.5978, "step": 3491 }, { "epoch": 0.03152193536739484, "grad_norm": 4.103769302368164, "learning_rate": 4.8424354576638385e-05, "loss": 4.6561, "step": 3492 }, { "epoch": 0.03153096226755732, "grad_norm": 3.1834819316864014, "learning_rate": 4.8423903231630264e-05, "loss": 4.8185, "step": 3493 }, { "epoch": 0.03153998916771981, "grad_norm": 4.224174499511719, "learning_rate": 4.8423451886622136e-05, "loss": 3.8951, "step": 3494 }, { "epoch": 0.03154901606788229, "grad_norm": 3.435270309448242, "learning_rate": 4.842300054161401e-05, "loss": 4.4907, "step": 3495 }, { "epoch": 0.03155804296804477, "grad_norm": 3.5137557983398438, "learning_rate": 4.842254919660589e-05, "loss": 4.9375, "step": 3496 }, { "epoch": 0.03156706986820726, "grad_norm": 4.50969123840332, "learning_rate": 4.842209785159777e-05, "loss": 5.2965, "step": 3497 }, { "epoch": 0.03157609676836974, "grad_norm": 3.132523536682129, "learning_rate": 4.842164650658964e-05, "loss": 5.1741, "step": 3498 }, { "epoch": 0.03158512366853223, "grad_norm": 4.129655361175537, "learning_rate": 4.842119516158152e-05, "loss": 4.6535, "step": 3499 }, { "epoch": 0.03159415056869471, "grad_norm": 4.910271644592285, "learning_rate": 4.842074381657339e-05, "loss": 4.794, "step": 3500 }, { "epoch": 0.03160317746885719, "grad_norm": 2.9595775604248047, "learning_rate": 4.8420292471565264e-05, "loss": 4.6017, "step": 3501 }, { "epoch": 0.03161220436901968, "grad_norm": 2.91837739944458, "learning_rate": 4.841984112655714e-05, "loss": 4.9364, "step": 3502 }, { "epoch": 0.03162123126918216, "grad_norm": 3.629044771194458, "learning_rate": 4.8419389781549016e-05, "loss": 4.8895, "step": 3503 }, { "epoch": 0.031630258169344644, "grad_norm": 6.001712799072266, "learning_rate": 4.8418938436540895e-05, "loss": 4.8543, "step": 3504 }, { "epoch": 0.03163928506950713, "grad_norm": 4.300166606903076, "learning_rate": 4.8418487091532774e-05, "loss": 5.0662, "step": 3505 }, { "epoch": 0.031648311969669614, "grad_norm": 3.4698233604431152, "learning_rate": 4.8418035746524647e-05, "loss": 5.3423, "step": 3506 }, { "epoch": 0.0316573388698321, "grad_norm": 3.103037118911743, "learning_rate": 4.841758440151652e-05, "loss": 4.7228, "step": 3507 }, { "epoch": 0.031666365769994584, "grad_norm": 3.456434488296509, "learning_rate": 4.84171330565084e-05, "loss": 5.0102, "step": 3508 }, { "epoch": 0.031675392670157065, "grad_norm": 3.0159411430358887, "learning_rate": 4.841668171150027e-05, "loss": 5.3902, "step": 3509 }, { "epoch": 0.031684419570319554, "grad_norm": 3.7207677364349365, "learning_rate": 4.841623036649215e-05, "loss": 4.7925, "step": 3510 }, { "epoch": 0.031693446470482035, "grad_norm": 3.122673273086548, "learning_rate": 4.841577902148402e-05, "loss": 4.48, "step": 3511 }, { "epoch": 0.031702473370644524, "grad_norm": 3.805525541305542, "learning_rate": 4.84153276764759e-05, "loss": 4.9482, "step": 3512 }, { "epoch": 0.031711500270807005, "grad_norm": 6.533178329467773, "learning_rate": 4.841487633146778e-05, "loss": 4.5065, "step": 3513 }, { "epoch": 0.03172052717096949, "grad_norm": 3.2794930934906006, "learning_rate": 4.841442498645965e-05, "loss": 4.5445, "step": 3514 }, { "epoch": 0.031729554071131975, "grad_norm": 2.8568687438964844, "learning_rate": 4.8413973641451526e-05, "loss": 5.0424, "step": 3515 }, { "epoch": 0.03173858097129446, "grad_norm": 3.9663290977478027, "learning_rate": 4.8413522296443405e-05, "loss": 5.095, "step": 3516 }, { "epoch": 0.03174760787145694, "grad_norm": 3.2267532348632812, "learning_rate": 4.841307095143528e-05, "loss": 4.5248, "step": 3517 }, { "epoch": 0.03175663477161943, "grad_norm": 4.086264133453369, "learning_rate": 4.841261960642716e-05, "loss": 4.5126, "step": 3518 }, { "epoch": 0.03176566167178191, "grad_norm": 3.894740104675293, "learning_rate": 4.8412168261419036e-05, "loss": 4.8963, "step": 3519 }, { "epoch": 0.0317746885719444, "grad_norm": 3.818692207336426, "learning_rate": 4.84117169164109e-05, "loss": 4.5613, "step": 3520 }, { "epoch": 0.03178371547210688, "grad_norm": 5.052786827087402, "learning_rate": 4.841126557140278e-05, "loss": 4.5148, "step": 3521 }, { "epoch": 0.03179274237226936, "grad_norm": 3.1831068992614746, "learning_rate": 4.841081422639466e-05, "loss": 5.1324, "step": 3522 }, { "epoch": 0.03180176927243185, "grad_norm": 3.96500563621521, "learning_rate": 4.841036288138653e-05, "loss": 4.785, "step": 3523 }, { "epoch": 0.03181079617259433, "grad_norm": 3.5634734630584717, "learning_rate": 4.840991153637841e-05, "loss": 4.1756, "step": 3524 }, { "epoch": 0.03181982307275682, "grad_norm": 3.586442470550537, "learning_rate": 4.8409460191370284e-05, "loss": 5.8388, "step": 3525 }, { "epoch": 0.0318288499729193, "grad_norm": 3.032468557357788, "learning_rate": 4.8409008846362164e-05, "loss": 5.4049, "step": 3526 }, { "epoch": 0.03183787687308178, "grad_norm": 5.313611030578613, "learning_rate": 4.8408557501354036e-05, "loss": 5.1402, "step": 3527 }, { "epoch": 0.03184690377324427, "grad_norm": 3.312870979309082, "learning_rate": 4.840810615634591e-05, "loss": 5.264, "step": 3528 }, { "epoch": 0.03185593067340675, "grad_norm": 3.265488862991333, "learning_rate": 4.840765481133779e-05, "loss": 4.725, "step": 3529 }, { "epoch": 0.03186495757356923, "grad_norm": 4.9342265129089355, "learning_rate": 4.840720346632967e-05, "loss": 4.8774, "step": 3530 }, { "epoch": 0.03187398447373172, "grad_norm": 3.901932954788208, "learning_rate": 4.840675212132154e-05, "loss": 5.1766, "step": 3531 }, { "epoch": 0.0318830113738942, "grad_norm": 4.68517541885376, "learning_rate": 4.840630077631342e-05, "loss": 4.0216, "step": 3532 }, { "epoch": 0.03189203827405669, "grad_norm": 2.8329145908355713, "learning_rate": 4.840584943130529e-05, "loss": 4.2138, "step": 3533 }, { "epoch": 0.03190106517421917, "grad_norm": 3.8161094188690186, "learning_rate": 4.8405398086297164e-05, "loss": 4.9565, "step": 3534 }, { "epoch": 0.031910092074381655, "grad_norm": 3.1757404804229736, "learning_rate": 4.840494674128904e-05, "loss": 4.575, "step": 3535 }, { "epoch": 0.03191911897454414, "grad_norm": 4.6295881271362305, "learning_rate": 4.840449539628092e-05, "loss": 4.1351, "step": 3536 }, { "epoch": 0.031928145874706625, "grad_norm": 4.562087059020996, "learning_rate": 4.8404044051272795e-05, "loss": 4.5927, "step": 3537 }, { "epoch": 0.03193717277486911, "grad_norm": 3.5834903717041016, "learning_rate": 4.8403592706264674e-05, "loss": 4.8646, "step": 3538 }, { "epoch": 0.031946199675031595, "grad_norm": 4.1958537101745605, "learning_rate": 4.8403141361256546e-05, "loss": 4.7276, "step": 3539 }, { "epoch": 0.031955226575194076, "grad_norm": 3.2038862705230713, "learning_rate": 4.8402690016248426e-05, "loss": 4.6733, "step": 3540 }, { "epoch": 0.031964253475356565, "grad_norm": 3.3743484020233154, "learning_rate": 4.84022386712403e-05, "loss": 4.5721, "step": 3541 }, { "epoch": 0.031973280375519046, "grad_norm": 5.1261372566223145, "learning_rate": 4.840178732623217e-05, "loss": 5.1987, "step": 3542 }, { "epoch": 0.03198230727568153, "grad_norm": 3.9338080883026123, "learning_rate": 4.840133598122405e-05, "loss": 4.8545, "step": 3543 }, { "epoch": 0.031991334175844016, "grad_norm": 5.569693565368652, "learning_rate": 4.840088463621593e-05, "loss": 4.5533, "step": 3544 }, { "epoch": 0.0320003610760065, "grad_norm": 5.409559726715088, "learning_rate": 4.84004332912078e-05, "loss": 4.1344, "step": 3545 }, { "epoch": 0.032009387976168986, "grad_norm": 4.885926246643066, "learning_rate": 4.839998194619968e-05, "loss": 4.513, "step": 3546 }, { "epoch": 0.03201841487633147, "grad_norm": 4.043652057647705, "learning_rate": 4.839953060119155e-05, "loss": 4.4565, "step": 3547 }, { "epoch": 0.03202744177649395, "grad_norm": 3.883072853088379, "learning_rate": 4.8399079256183426e-05, "loss": 4.4602, "step": 3548 }, { "epoch": 0.03203646867665644, "grad_norm": 3.7516067028045654, "learning_rate": 4.8398627911175305e-05, "loss": 4.7372, "step": 3549 }, { "epoch": 0.03204549557681892, "grad_norm": 4.018538475036621, "learning_rate": 4.839817656616718e-05, "loss": 4.5138, "step": 3550 }, { "epoch": 0.03205452247698141, "grad_norm": 3.301360845565796, "learning_rate": 4.8397725221159057e-05, "loss": 5.0134, "step": 3551 }, { "epoch": 0.03206354937714389, "grad_norm": 3.102703809738159, "learning_rate": 4.8397273876150936e-05, "loss": 4.3744, "step": 3552 }, { "epoch": 0.03207257627730637, "grad_norm": 4.303585052490234, "learning_rate": 4.839682253114281e-05, "loss": 4.5632, "step": 3553 }, { "epoch": 0.03208160317746886, "grad_norm": 4.763572692871094, "learning_rate": 4.839637118613468e-05, "loss": 4.8142, "step": 3554 }, { "epoch": 0.03209063007763134, "grad_norm": 3.506671667098999, "learning_rate": 4.839591984112656e-05, "loss": 4.39, "step": 3555 }, { "epoch": 0.03209965697779382, "grad_norm": 3.8989224433898926, "learning_rate": 4.839546849611843e-05, "loss": 4.6093, "step": 3556 }, { "epoch": 0.03210868387795631, "grad_norm": 3.17997670173645, "learning_rate": 4.839501715111031e-05, "loss": 5.0256, "step": 3557 }, { "epoch": 0.03211771077811879, "grad_norm": 4.203868389129639, "learning_rate": 4.839456580610219e-05, "loss": 5.3177, "step": 3558 }, { "epoch": 0.03212673767828128, "grad_norm": 5.465549945831299, "learning_rate": 4.839411446109406e-05, "loss": 4.3622, "step": 3559 }, { "epoch": 0.03213576457844376, "grad_norm": 2.9996626377105713, "learning_rate": 4.839366311608594e-05, "loss": 4.8187, "step": 3560 }, { "epoch": 0.032144791478606244, "grad_norm": 5.006460189819336, "learning_rate": 4.8393211771077815e-05, "loss": 4.77, "step": 3561 }, { "epoch": 0.03215381837876873, "grad_norm": 3.9044792652130127, "learning_rate": 4.839276042606969e-05, "loss": 5.2749, "step": 3562 }, { "epoch": 0.032162845278931214, "grad_norm": 7.14399528503418, "learning_rate": 4.839230908106157e-05, "loss": 5.0357, "step": 3563 }, { "epoch": 0.0321718721790937, "grad_norm": 6.102655410766602, "learning_rate": 4.839185773605344e-05, "loss": 4.4964, "step": 3564 }, { "epoch": 0.032180899079256184, "grad_norm": 3.768951416015625, "learning_rate": 4.839140639104532e-05, "loss": 4.256, "step": 3565 }, { "epoch": 0.032189925979418665, "grad_norm": 2.861323833465576, "learning_rate": 4.83909550460372e-05, "loss": 5.2292, "step": 3566 }, { "epoch": 0.032198952879581154, "grad_norm": 4.538455486297607, "learning_rate": 4.8390503701029063e-05, "loss": 4.9915, "step": 3567 }, { "epoch": 0.032207979779743635, "grad_norm": 3.582021951675415, "learning_rate": 4.839005235602094e-05, "loss": 4.8246, "step": 3568 }, { "epoch": 0.03221700667990612, "grad_norm": 3.5852503776550293, "learning_rate": 4.838960101101282e-05, "loss": 3.8304, "step": 3569 }, { "epoch": 0.032226033580068605, "grad_norm": 3.1892635822296143, "learning_rate": 4.8389149666004694e-05, "loss": 4.3727, "step": 3570 }, { "epoch": 0.03223506048023109, "grad_norm": 3.6219370365142822, "learning_rate": 4.8388698320996574e-05, "loss": 5.0763, "step": 3571 }, { "epoch": 0.032244087380393575, "grad_norm": 3.9415078163146973, "learning_rate": 4.8388246975988446e-05, "loss": 4.7206, "step": 3572 }, { "epoch": 0.03225311428055606, "grad_norm": 2.997028350830078, "learning_rate": 4.8387795630980325e-05, "loss": 4.0564, "step": 3573 }, { "epoch": 0.03226214118071854, "grad_norm": 3.952941417694092, "learning_rate": 4.83873442859722e-05, "loss": 4.955, "step": 3574 }, { "epoch": 0.03227116808088103, "grad_norm": 3.4903652667999268, "learning_rate": 4.838689294096408e-05, "loss": 4.7651, "step": 3575 }, { "epoch": 0.03228019498104351, "grad_norm": 3.444216728210449, "learning_rate": 4.838644159595595e-05, "loss": 4.3138, "step": 3576 }, { "epoch": 0.032289221881206, "grad_norm": 4.406518936157227, "learning_rate": 4.838599025094783e-05, "loss": 4.4885, "step": 3577 }, { "epoch": 0.03229824878136848, "grad_norm": 3.7229397296905518, "learning_rate": 4.83855389059397e-05, "loss": 5.0988, "step": 3578 }, { "epoch": 0.03230727568153096, "grad_norm": 4.208820343017578, "learning_rate": 4.838508756093158e-05, "loss": 5.3116, "step": 3579 }, { "epoch": 0.03231630258169345, "grad_norm": 4.805359840393066, "learning_rate": 4.838463621592346e-05, "loss": 4.7376, "step": 3580 }, { "epoch": 0.03232532948185593, "grad_norm": 4.334683418273926, "learning_rate": 4.8384184870915325e-05, "loss": 3.7727, "step": 3581 }, { "epoch": 0.03233435638201841, "grad_norm": 4.311980247497559, "learning_rate": 4.8383733525907205e-05, "loss": 4.8453, "step": 3582 }, { "epoch": 0.0323433832821809, "grad_norm": 3.4514858722686768, "learning_rate": 4.8383282180899084e-05, "loss": 4.2319, "step": 3583 }, { "epoch": 0.03235241018234338, "grad_norm": 3.8952529430389404, "learning_rate": 4.8382830835890956e-05, "loss": 5.0027, "step": 3584 }, { "epoch": 0.03236143708250587, "grad_norm": 2.840019941329956, "learning_rate": 4.8382379490882836e-05, "loss": 5.2398, "step": 3585 }, { "epoch": 0.03237046398266835, "grad_norm": 4.03184175491333, "learning_rate": 4.838192814587471e-05, "loss": 4.9925, "step": 3586 }, { "epoch": 0.03237949088283083, "grad_norm": 3.4772417545318604, "learning_rate": 4.838147680086658e-05, "loss": 4.6412, "step": 3587 }, { "epoch": 0.03238851778299332, "grad_norm": 3.6410717964172363, "learning_rate": 4.838102545585846e-05, "loss": 5.53, "step": 3588 }, { "epoch": 0.0323975446831558, "grad_norm": 3.6260623931884766, "learning_rate": 4.838057411085033e-05, "loss": 4.8705, "step": 3589 }, { "epoch": 0.03240657158331829, "grad_norm": 2.649137496948242, "learning_rate": 4.838012276584221e-05, "loss": 4.8736, "step": 3590 }, { "epoch": 0.03241559848348077, "grad_norm": 4.0383477210998535, "learning_rate": 4.837967142083409e-05, "loss": 3.9499, "step": 3591 }, { "epoch": 0.032424625383643255, "grad_norm": 3.4892077445983887, "learning_rate": 4.837922007582596e-05, "loss": 4.4685, "step": 3592 }, { "epoch": 0.03243365228380574, "grad_norm": 3.0648677349090576, "learning_rate": 4.837876873081784e-05, "loss": 4.5705, "step": 3593 }, { "epoch": 0.032442679183968225, "grad_norm": 3.6552014350891113, "learning_rate": 4.8378317385809715e-05, "loss": 5.1735, "step": 3594 }, { "epoch": 0.032451706084130706, "grad_norm": 3.2931697368621826, "learning_rate": 4.837786604080159e-05, "loss": 4.6777, "step": 3595 }, { "epoch": 0.032460732984293195, "grad_norm": 3.216130018234253, "learning_rate": 4.8377414695793467e-05, "loss": 4.9464, "step": 3596 }, { "epoch": 0.032469759884455676, "grad_norm": 3.2931642532348633, "learning_rate": 4.8376963350785346e-05, "loss": 4.1999, "step": 3597 }, { "epoch": 0.032478786784618165, "grad_norm": 4.040933132171631, "learning_rate": 4.837651200577722e-05, "loss": 5.0725, "step": 3598 }, { "epoch": 0.032487813684780646, "grad_norm": 4.18287467956543, "learning_rate": 4.83760606607691e-05, "loss": 5.2632, "step": 3599 }, { "epoch": 0.03249684058494313, "grad_norm": 3.461662769317627, "learning_rate": 4.837560931576097e-05, "loss": 4.5784, "step": 3600 }, { "epoch": 0.032505867485105616, "grad_norm": 3.321751832962036, "learning_rate": 4.837515797075284e-05, "loss": 4.4291, "step": 3601 }, { "epoch": 0.0325148943852681, "grad_norm": 3.7726824283599854, "learning_rate": 4.837470662574472e-05, "loss": 5.1186, "step": 3602 }, { "epoch": 0.032523921285430586, "grad_norm": 3.6348378658294678, "learning_rate": 4.8374255280736594e-05, "loss": 4.5255, "step": 3603 }, { "epoch": 0.03253294818559307, "grad_norm": 4.207329750061035, "learning_rate": 4.837380393572847e-05, "loss": 4.6704, "step": 3604 }, { "epoch": 0.03254197508575555, "grad_norm": 3.7623536586761475, "learning_rate": 4.837335259072035e-05, "loss": 4.4874, "step": 3605 }, { "epoch": 0.03255100198591804, "grad_norm": 3.9521076679229736, "learning_rate": 4.8372901245712225e-05, "loss": 4.9918, "step": 3606 }, { "epoch": 0.03256002888608052, "grad_norm": 3.773902654647827, "learning_rate": 4.8372449900704104e-05, "loss": 5.1097, "step": 3607 }, { "epoch": 0.032569055786243, "grad_norm": 3.5820059776306152, "learning_rate": 4.837199855569598e-05, "loss": 4.937, "step": 3608 }, { "epoch": 0.03257808268640549, "grad_norm": 6.698265552520752, "learning_rate": 4.837154721068785e-05, "loss": 3.7057, "step": 3609 }, { "epoch": 0.03258710958656797, "grad_norm": 4.430299758911133, "learning_rate": 4.837109586567973e-05, "loss": 5.452, "step": 3610 }, { "epoch": 0.03259613648673046, "grad_norm": 2.6563470363616943, "learning_rate": 4.83706445206716e-05, "loss": 5.0351, "step": 3611 }, { "epoch": 0.03260516338689294, "grad_norm": 4.4049248695373535, "learning_rate": 4.837019317566348e-05, "loss": 4.9462, "step": 3612 }, { "epoch": 0.03261419028705542, "grad_norm": 4.017927646636963, "learning_rate": 4.836974183065536e-05, "loss": 3.8852, "step": 3613 }, { "epoch": 0.03262321718721791, "grad_norm": 3.0205774307250977, "learning_rate": 4.8369290485647225e-05, "loss": 3.8537, "step": 3614 }, { "epoch": 0.03263224408738039, "grad_norm": 4.382291793823242, "learning_rate": 4.8368839140639104e-05, "loss": 4.913, "step": 3615 }, { "epoch": 0.03264127098754288, "grad_norm": 3.6289191246032715, "learning_rate": 4.8368387795630984e-05, "loss": 4.7647, "step": 3616 }, { "epoch": 0.03265029788770536, "grad_norm": 3.66219162940979, "learning_rate": 4.8367936450622856e-05, "loss": 4.9713, "step": 3617 }, { "epoch": 0.032659324787867844, "grad_norm": 6.095943927764893, "learning_rate": 4.8367485105614735e-05, "loss": 4.8246, "step": 3618 }, { "epoch": 0.03266835168803033, "grad_norm": 3.089181661605835, "learning_rate": 4.8367033760606615e-05, "loss": 4.9128, "step": 3619 }, { "epoch": 0.032677378588192814, "grad_norm": 3.1967685222625732, "learning_rate": 4.836658241559849e-05, "loss": 4.0489, "step": 3620 }, { "epoch": 0.032686405488355295, "grad_norm": 5.3659987449646, "learning_rate": 4.836613107059036e-05, "loss": 4.5695, "step": 3621 }, { "epoch": 0.032695432388517784, "grad_norm": 3.6606173515319824, "learning_rate": 4.836567972558224e-05, "loss": 5.0556, "step": 3622 }, { "epoch": 0.032704459288680265, "grad_norm": 3.0363850593566895, "learning_rate": 4.836522838057411e-05, "loss": 4.9533, "step": 3623 }, { "epoch": 0.032713486188842754, "grad_norm": 4.32514762878418, "learning_rate": 4.836477703556599e-05, "loss": 4.5951, "step": 3624 }, { "epoch": 0.032722513089005235, "grad_norm": 4.961574077606201, "learning_rate": 4.836432569055786e-05, "loss": 4.648, "step": 3625 }, { "epoch": 0.03273153998916772, "grad_norm": 3.4815468788146973, "learning_rate": 4.836387434554974e-05, "loss": 5.3169, "step": 3626 }, { "epoch": 0.032740566889330205, "grad_norm": 4.250156879425049, "learning_rate": 4.836342300054162e-05, "loss": 4.0507, "step": 3627 }, { "epoch": 0.03274959378949269, "grad_norm": 3.222417116165161, "learning_rate": 4.836297165553349e-05, "loss": 5.1725, "step": 3628 }, { "epoch": 0.032758620689655175, "grad_norm": 4.5642828941345215, "learning_rate": 4.8362520310525366e-05, "loss": 4.788, "step": 3629 }, { "epoch": 0.03276764758981766, "grad_norm": 3.860959768295288, "learning_rate": 4.8362068965517246e-05, "loss": 4.9539, "step": 3630 }, { "epoch": 0.03277667448998014, "grad_norm": 3.351715326309204, "learning_rate": 4.836161762050912e-05, "loss": 4.7767, "step": 3631 }, { "epoch": 0.03278570139014263, "grad_norm": 3.1930043697357178, "learning_rate": 4.8361166275501e-05, "loss": 5.0618, "step": 3632 }, { "epoch": 0.03279472829030511, "grad_norm": 3.5903844833374023, "learning_rate": 4.836071493049287e-05, "loss": 4.9225, "step": 3633 }, { "epoch": 0.03280375519046759, "grad_norm": 3.5053999423980713, "learning_rate": 4.836026358548474e-05, "loss": 4.3903, "step": 3634 }, { "epoch": 0.03281278209063008, "grad_norm": 4.335473537445068, "learning_rate": 4.835981224047662e-05, "loss": 4.9915, "step": 3635 }, { "epoch": 0.03282180899079256, "grad_norm": 4.460337162017822, "learning_rate": 4.83593608954685e-05, "loss": 5.2195, "step": 3636 }, { "epoch": 0.03283083589095505, "grad_norm": 4.430839538574219, "learning_rate": 4.835890955046037e-05, "loss": 4.0293, "step": 3637 }, { "epoch": 0.03283986279111753, "grad_norm": 3.369394540786743, "learning_rate": 4.835845820545225e-05, "loss": 4.5568, "step": 3638 }, { "epoch": 0.03284888969128001, "grad_norm": 4.428982257843018, "learning_rate": 4.8358006860444125e-05, "loss": 5.1996, "step": 3639 }, { "epoch": 0.0328579165914425, "grad_norm": 3.3465914726257324, "learning_rate": 4.8357555515436004e-05, "loss": 4.9635, "step": 3640 }, { "epoch": 0.03286694349160498, "grad_norm": 4.447437286376953, "learning_rate": 4.8357104170427877e-05, "loss": 4.4929, "step": 3641 }, { "epoch": 0.03287597039176747, "grad_norm": 3.530672550201416, "learning_rate": 4.835665282541975e-05, "loss": 3.9962, "step": 3642 }, { "epoch": 0.03288499729192995, "grad_norm": 3.7719433307647705, "learning_rate": 4.835620148041163e-05, "loss": 4.8222, "step": 3643 }, { "epoch": 0.03289402419209243, "grad_norm": 3.7868893146514893, "learning_rate": 4.835575013540351e-05, "loss": 4.9164, "step": 3644 }, { "epoch": 0.03290305109225492, "grad_norm": 5.383219242095947, "learning_rate": 4.835529879039538e-05, "loss": 4.7021, "step": 3645 }, { "epoch": 0.0329120779924174, "grad_norm": 2.935929775238037, "learning_rate": 4.835484744538726e-05, "loss": 5.0, "step": 3646 }, { "epoch": 0.032921104892579885, "grad_norm": 3.328254461288452, "learning_rate": 4.835439610037913e-05, "loss": 4.5632, "step": 3647 }, { "epoch": 0.03293013179274237, "grad_norm": 4.688518524169922, "learning_rate": 4.8353944755371004e-05, "loss": 4.6948, "step": 3648 }, { "epoch": 0.032939158692904855, "grad_norm": 3.8006818294525146, "learning_rate": 4.835349341036288e-05, "loss": 3.9925, "step": 3649 }, { "epoch": 0.03294818559306734, "grad_norm": 3.986586570739746, "learning_rate": 4.8353042065354756e-05, "loss": 4.6748, "step": 3650 }, { "epoch": 0.032957212493229825, "grad_norm": 5.245904922485352, "learning_rate": 4.8352590720346635e-05, "loss": 4.7378, "step": 3651 }, { "epoch": 0.032966239393392306, "grad_norm": 5.128523349761963, "learning_rate": 4.8352139375338514e-05, "loss": 5.09, "step": 3652 }, { "epoch": 0.032975266293554795, "grad_norm": 3.2230026721954346, "learning_rate": 4.835168803033039e-05, "loss": 5.3103, "step": 3653 }, { "epoch": 0.032984293193717276, "grad_norm": 3.6233022212982178, "learning_rate": 4.8351236685322266e-05, "loss": 4.5081, "step": 3654 }, { "epoch": 0.032993320093879765, "grad_norm": 3.5365238189697266, "learning_rate": 4.835078534031414e-05, "loss": 4.6664, "step": 3655 }, { "epoch": 0.033002346994042246, "grad_norm": 3.158656120300293, "learning_rate": 4.835033399530601e-05, "loss": 4.5124, "step": 3656 }, { "epoch": 0.03301137389420473, "grad_norm": 3.148014783859253, "learning_rate": 4.834988265029789e-05, "loss": 4.9374, "step": 3657 }, { "epoch": 0.033020400794367216, "grad_norm": 4.1053147315979, "learning_rate": 4.834943130528977e-05, "loss": 5.0935, "step": 3658 }, { "epoch": 0.0330294276945297, "grad_norm": 2.5627613067626953, "learning_rate": 4.834897996028164e-05, "loss": 5.0965, "step": 3659 }, { "epoch": 0.03303845459469218, "grad_norm": 4.176539897918701, "learning_rate": 4.834852861527352e-05, "loss": 5.1788, "step": 3660 }, { "epoch": 0.03304748149485467, "grad_norm": 5.208218097686768, "learning_rate": 4.8348077270265394e-05, "loss": 4.3451, "step": 3661 }, { "epoch": 0.03305650839501715, "grad_norm": 3.3718221187591553, "learning_rate": 4.8347625925257266e-05, "loss": 5.2026, "step": 3662 }, { "epoch": 0.03306553529517964, "grad_norm": 3.5695223808288574, "learning_rate": 4.8347174580249145e-05, "loss": 4.8057, "step": 3663 }, { "epoch": 0.03307456219534212, "grad_norm": 2.987309217453003, "learning_rate": 4.834672323524102e-05, "loss": 4.7858, "step": 3664 }, { "epoch": 0.0330835890955046, "grad_norm": 2.9906718730926514, "learning_rate": 4.83462718902329e-05, "loss": 4.9353, "step": 3665 }, { "epoch": 0.03309261599566709, "grad_norm": 4.328796863555908, "learning_rate": 4.8345820545224776e-05, "loss": 4.2798, "step": 3666 }, { "epoch": 0.03310164289582957, "grad_norm": 3.995879650115967, "learning_rate": 4.834536920021665e-05, "loss": 4.9499, "step": 3667 }, { "epoch": 0.03311066979599206, "grad_norm": 4.7177228927612305, "learning_rate": 4.834491785520852e-05, "loss": 3.7938, "step": 3668 }, { "epoch": 0.03311969669615454, "grad_norm": 4.423151969909668, "learning_rate": 4.83444665102004e-05, "loss": 4.9108, "step": 3669 }, { "epoch": 0.03312872359631702, "grad_norm": 4.109916687011719, "learning_rate": 4.834401516519227e-05, "loss": 4.6052, "step": 3670 }, { "epoch": 0.03313775049647951, "grad_norm": 3.576702833175659, "learning_rate": 4.834356382018415e-05, "loss": 4.6357, "step": 3671 }, { "epoch": 0.03314677739664199, "grad_norm": 3.7068541049957275, "learning_rate": 4.8343112475176025e-05, "loss": 4.9592, "step": 3672 }, { "epoch": 0.033155804296804474, "grad_norm": 3.554081439971924, "learning_rate": 4.8342661130167904e-05, "loss": 4.5427, "step": 3673 }, { "epoch": 0.03316483119696696, "grad_norm": 3.0325279235839844, "learning_rate": 4.834220978515978e-05, "loss": 5.0171, "step": 3674 }, { "epoch": 0.033173858097129444, "grad_norm": 4.177008628845215, "learning_rate": 4.834175844015165e-05, "loss": 4.7781, "step": 3675 }, { "epoch": 0.03318288499729193, "grad_norm": 3.5797128677368164, "learning_rate": 4.834130709514353e-05, "loss": 5.3975, "step": 3676 }, { "epoch": 0.033191911897454414, "grad_norm": 4.139972686767578, "learning_rate": 4.834085575013541e-05, "loss": 5.0149, "step": 3677 }, { "epoch": 0.033200938797616895, "grad_norm": 4.549894332885742, "learning_rate": 4.834040440512728e-05, "loss": 4.6315, "step": 3678 }, { "epoch": 0.033209965697779384, "grad_norm": 3.11281156539917, "learning_rate": 4.833995306011916e-05, "loss": 5.4177, "step": 3679 }, { "epoch": 0.033218992597941865, "grad_norm": 2.785627603530884, "learning_rate": 4.833950171511104e-05, "loss": 4.74, "step": 3680 }, { "epoch": 0.033228019498104354, "grad_norm": 6.709392547607422, "learning_rate": 4.8339050370102904e-05, "loss": 4.786, "step": 3681 }, { "epoch": 0.033237046398266835, "grad_norm": 3.559232711791992, "learning_rate": 4.833859902509478e-05, "loss": 4.7278, "step": 3682 }, { "epoch": 0.03324607329842932, "grad_norm": 3.5936734676361084, "learning_rate": 4.833814768008666e-05, "loss": 3.9924, "step": 3683 }, { "epoch": 0.033255100198591805, "grad_norm": 3.7294833660125732, "learning_rate": 4.8337696335078535e-05, "loss": 5.2344, "step": 3684 }, { "epoch": 0.03326412709875429, "grad_norm": 3.531684398651123, "learning_rate": 4.8337244990070414e-05, "loss": 4.3064, "step": 3685 }, { "epoch": 0.033273153998916775, "grad_norm": 3.0712990760803223, "learning_rate": 4.8336793645062287e-05, "loss": 4.6147, "step": 3686 }, { "epoch": 0.03328218089907926, "grad_norm": 3.647547960281372, "learning_rate": 4.8336342300054166e-05, "loss": 4.5263, "step": 3687 }, { "epoch": 0.03329120779924174, "grad_norm": 2.9595863819122314, "learning_rate": 4.833589095504604e-05, "loss": 4.9832, "step": 3688 }, { "epoch": 0.03330023469940423, "grad_norm": 4.055741786956787, "learning_rate": 4.833543961003791e-05, "loss": 3.8925, "step": 3689 }, { "epoch": 0.03330926159956671, "grad_norm": 3.1858913898468018, "learning_rate": 4.833498826502979e-05, "loss": 4.2339, "step": 3690 }, { "epoch": 0.03331828849972919, "grad_norm": 4.447527885437012, "learning_rate": 4.833453692002167e-05, "loss": 4.458, "step": 3691 }, { "epoch": 0.03332731539989168, "grad_norm": 2.670304298400879, "learning_rate": 4.833408557501354e-05, "loss": 4.9729, "step": 3692 }, { "epoch": 0.03333634230005416, "grad_norm": 3.293139696121216, "learning_rate": 4.833363423000542e-05, "loss": 3.9889, "step": 3693 }, { "epoch": 0.03334536920021665, "grad_norm": 3.2726633548736572, "learning_rate": 4.833318288499729e-05, "loss": 5.0151, "step": 3694 }, { "epoch": 0.03335439610037913, "grad_norm": 3.9983198642730713, "learning_rate": 4.8332731539989166e-05, "loss": 4.4063, "step": 3695 }, { "epoch": 0.03336342300054161, "grad_norm": 11.4457368850708, "learning_rate": 4.8332280194981045e-05, "loss": 4.1606, "step": 3696 }, { "epoch": 0.0333724499007041, "grad_norm": 3.454402208328247, "learning_rate": 4.8331828849972924e-05, "loss": 4.4899, "step": 3697 }, { "epoch": 0.03338147680086658, "grad_norm": 2.657290458679199, "learning_rate": 4.83313775049648e-05, "loss": 4.8572, "step": 3698 }, { "epoch": 0.03339050370102907, "grad_norm": 4.207672119140625, "learning_rate": 4.8330926159956676e-05, "loss": 4.0306, "step": 3699 }, { "epoch": 0.03339953060119155, "grad_norm": 3.852318286895752, "learning_rate": 4.833047481494855e-05, "loss": 4.6968, "step": 3700 }, { "epoch": 0.03340855750135403, "grad_norm": 4.523581504821777, "learning_rate": 4.833002346994043e-05, "loss": 4.6195, "step": 3701 }, { "epoch": 0.03341758440151652, "grad_norm": 3.5536227226257324, "learning_rate": 4.83295721249323e-05, "loss": 4.8863, "step": 3702 }, { "epoch": 0.033426611301679, "grad_norm": 3.9750611782073975, "learning_rate": 4.832912077992417e-05, "loss": 5.0563, "step": 3703 }, { "epoch": 0.033435638201841485, "grad_norm": 3.744800329208374, "learning_rate": 4.832866943491605e-05, "loss": 4.65, "step": 3704 }, { "epoch": 0.03344466510200397, "grad_norm": 4.132030010223389, "learning_rate": 4.832821808990793e-05, "loss": 4.7742, "step": 3705 }, { "epoch": 0.033453692002166455, "grad_norm": 2.986196756362915, "learning_rate": 4.8327766744899804e-05, "loss": 5.2194, "step": 3706 }, { "epoch": 0.03346271890232894, "grad_norm": 3.514080286026001, "learning_rate": 4.832731539989168e-05, "loss": 4.061, "step": 3707 }, { "epoch": 0.033471745802491425, "grad_norm": 4.647726535797119, "learning_rate": 4.8326864054883555e-05, "loss": 4.4939, "step": 3708 }, { "epoch": 0.033480772702653906, "grad_norm": 2.864807367324829, "learning_rate": 4.832641270987543e-05, "loss": 5.2541, "step": 3709 }, { "epoch": 0.033489799602816395, "grad_norm": 3.45284104347229, "learning_rate": 4.832596136486731e-05, "loss": 4.8561, "step": 3710 }, { "epoch": 0.033498826502978876, "grad_norm": 4.062947750091553, "learning_rate": 4.832551001985918e-05, "loss": 4.9334, "step": 3711 }, { "epoch": 0.033507853403141365, "grad_norm": 4.467643737792969, "learning_rate": 4.832505867485106e-05, "loss": 4.7164, "step": 3712 }, { "epoch": 0.033516880303303846, "grad_norm": 5.463357925415039, "learning_rate": 4.832460732984294e-05, "loss": 5.3808, "step": 3713 }, { "epoch": 0.03352590720346633, "grad_norm": 3.597999334335327, "learning_rate": 4.832415598483481e-05, "loss": 5.0985, "step": 3714 }, { "epoch": 0.033534934103628816, "grad_norm": 3.1648919582366943, "learning_rate": 4.832370463982668e-05, "loss": 4.8718, "step": 3715 }, { "epoch": 0.0335439610037913, "grad_norm": 3.3868908882141113, "learning_rate": 4.832325329481856e-05, "loss": 5.144, "step": 3716 }, { "epoch": 0.03355298790395378, "grad_norm": 4.607978343963623, "learning_rate": 4.8322801949810435e-05, "loss": 3.9504, "step": 3717 }, { "epoch": 0.03356201480411627, "grad_norm": 3.0833535194396973, "learning_rate": 4.8322350604802314e-05, "loss": 4.7063, "step": 3718 }, { "epoch": 0.03357104170427875, "grad_norm": 3.312537908554077, "learning_rate": 4.832189925979419e-05, "loss": 4.557, "step": 3719 }, { "epoch": 0.03358006860444124, "grad_norm": 3.174110174179077, "learning_rate": 4.8321447914786066e-05, "loss": 5.0172, "step": 3720 }, { "epoch": 0.03358909550460372, "grad_norm": 5.529688358306885, "learning_rate": 4.8320996569777945e-05, "loss": 4.9706, "step": 3721 }, { "epoch": 0.0335981224047662, "grad_norm": 2.7871575355529785, "learning_rate": 4.832054522476982e-05, "loss": 4.9679, "step": 3722 }, { "epoch": 0.03360714930492869, "grad_norm": 2.489248514175415, "learning_rate": 4.832009387976169e-05, "loss": 5.0716, "step": 3723 }, { "epoch": 0.03361617620509117, "grad_norm": 2.258244752883911, "learning_rate": 4.831964253475357e-05, "loss": 5.0341, "step": 3724 }, { "epoch": 0.03362520310525366, "grad_norm": 2.446521520614624, "learning_rate": 4.831919118974544e-05, "loss": 4.5694, "step": 3725 }, { "epoch": 0.03363423000541614, "grad_norm": 4.256535053253174, "learning_rate": 4.831873984473732e-05, "loss": 4.6167, "step": 3726 }, { "epoch": 0.03364325690557862, "grad_norm": 2.970505475997925, "learning_rate": 4.83182884997292e-05, "loss": 5.0334, "step": 3727 }, { "epoch": 0.03365228380574111, "grad_norm": 4.30146598815918, "learning_rate": 4.8317837154721066e-05, "loss": 4.345, "step": 3728 }, { "epoch": 0.03366131070590359, "grad_norm": 4.321875095367432, "learning_rate": 4.8317385809712945e-05, "loss": 5.0982, "step": 3729 }, { "epoch": 0.033670337606066074, "grad_norm": 3.4774248600006104, "learning_rate": 4.8316934464704824e-05, "loss": 4.873, "step": 3730 }, { "epoch": 0.03367936450622856, "grad_norm": 3.8298728466033936, "learning_rate": 4.8316483119696697e-05, "loss": 4.263, "step": 3731 }, { "epoch": 0.033688391406391044, "grad_norm": 4.1353759765625, "learning_rate": 4.8316031774688576e-05, "loss": 5.0902, "step": 3732 }, { "epoch": 0.03369741830655353, "grad_norm": 3.6870086193084717, "learning_rate": 4.831558042968045e-05, "loss": 5.3041, "step": 3733 }, { "epoch": 0.033706445206716014, "grad_norm": 3.9405171871185303, "learning_rate": 4.831512908467233e-05, "loss": 4.8768, "step": 3734 }, { "epoch": 0.033715472106878495, "grad_norm": 2.850506544113159, "learning_rate": 4.83146777396642e-05, "loss": 4.9444, "step": 3735 }, { "epoch": 0.033724499007040984, "grad_norm": 2.354868173599243, "learning_rate": 4.831422639465608e-05, "loss": 5.0206, "step": 3736 }, { "epoch": 0.033733525907203465, "grad_norm": 3.1451523303985596, "learning_rate": 4.831377504964795e-05, "loss": 4.8079, "step": 3737 }, { "epoch": 0.033742552807365954, "grad_norm": 3.4424614906311035, "learning_rate": 4.831332370463983e-05, "loss": 4.5225, "step": 3738 }, { "epoch": 0.033751579707528435, "grad_norm": 3.5737972259521484, "learning_rate": 4.83128723596317e-05, "loss": 4.4778, "step": 3739 }, { "epoch": 0.03376060660769092, "grad_norm": 3.820354461669922, "learning_rate": 4.831242101462358e-05, "loss": 5.2005, "step": 3740 }, { "epoch": 0.033769633507853405, "grad_norm": 3.4460880756378174, "learning_rate": 4.831196966961546e-05, "loss": 4.7301, "step": 3741 }, { "epoch": 0.03377866040801589, "grad_norm": 3.6931073665618896, "learning_rate": 4.831151832460733e-05, "loss": 4.7755, "step": 3742 }, { "epoch": 0.03378768730817837, "grad_norm": 5.762036323547363, "learning_rate": 4.831106697959921e-05, "loss": 4.5975, "step": 3743 }, { "epoch": 0.03379671420834086, "grad_norm": 2.877563714981079, "learning_rate": 4.8310615634591086e-05, "loss": 4.5445, "step": 3744 }, { "epoch": 0.03380574110850334, "grad_norm": 4.863344192504883, "learning_rate": 4.831016428958296e-05, "loss": 5.7507, "step": 3745 }, { "epoch": 0.03381476800866583, "grad_norm": 3.4981629848480225, "learning_rate": 4.830971294457484e-05, "loss": 4.9092, "step": 3746 }, { "epoch": 0.03382379490882831, "grad_norm": 2.9642536640167236, "learning_rate": 4.830926159956671e-05, "loss": 5.3195, "step": 3747 }, { "epoch": 0.03383282180899079, "grad_norm": 4.509363174438477, "learning_rate": 4.830881025455859e-05, "loss": 4.3633, "step": 3748 }, { "epoch": 0.03384184870915328, "grad_norm": 3.4637398719787598, "learning_rate": 4.830835890955046e-05, "loss": 4.5503, "step": 3749 }, { "epoch": 0.03385087560931576, "grad_norm": 3.221369981765747, "learning_rate": 4.8307907564542334e-05, "loss": 5.2556, "step": 3750 }, { "epoch": 0.03385990250947825, "grad_norm": 3.3075995445251465, "learning_rate": 4.8307456219534214e-05, "loss": 5.1248, "step": 3751 }, { "epoch": 0.03386892940964073, "grad_norm": 3.446286678314209, "learning_rate": 4.830700487452609e-05, "loss": 4.7026, "step": 3752 }, { "epoch": 0.03387795630980321, "grad_norm": 4.795200347900391, "learning_rate": 4.8306553529517965e-05, "loss": 5.0889, "step": 3753 }, { "epoch": 0.0338869832099657, "grad_norm": 3.257246971130371, "learning_rate": 4.8306102184509845e-05, "loss": 4.7096, "step": 3754 }, { "epoch": 0.03389601011012818, "grad_norm": 3.486264705657959, "learning_rate": 4.830565083950172e-05, "loss": 4.411, "step": 3755 }, { "epoch": 0.03390503701029066, "grad_norm": 10.710612297058105, "learning_rate": 4.830519949449359e-05, "loss": 3.4445, "step": 3756 }, { "epoch": 0.03391406391045315, "grad_norm": 3.998920440673828, "learning_rate": 4.830474814948547e-05, "loss": 5.0358, "step": 3757 }, { "epoch": 0.03392309081061563, "grad_norm": 3.762742042541504, "learning_rate": 4.830429680447735e-05, "loss": 4.3503, "step": 3758 }, { "epoch": 0.03393211771077812, "grad_norm": 3.4893288612365723, "learning_rate": 4.830384545946922e-05, "loss": 3.807, "step": 3759 }, { "epoch": 0.0339411446109406, "grad_norm": 3.721986770629883, "learning_rate": 4.83033941144611e-05, "loss": 4.4479, "step": 3760 }, { "epoch": 0.033950171511103085, "grad_norm": 2.515918731689453, "learning_rate": 4.830294276945297e-05, "loss": 4.8984, "step": 3761 }, { "epoch": 0.03395919841126557, "grad_norm": 2.5821003913879395, "learning_rate": 4.8302491424444845e-05, "loss": 4.542, "step": 3762 }, { "epoch": 0.033968225311428055, "grad_norm": 3.835158586502075, "learning_rate": 4.8302040079436724e-05, "loss": 4.9372, "step": 3763 }, { "epoch": 0.03397725221159054, "grad_norm": 3.8251683712005615, "learning_rate": 4.8301588734428596e-05, "loss": 4.5024, "step": 3764 }, { "epoch": 0.033986279111753025, "grad_norm": 3.124645233154297, "learning_rate": 4.8301137389420476e-05, "loss": 5.0337, "step": 3765 }, { "epoch": 0.033995306011915506, "grad_norm": 4.2890167236328125, "learning_rate": 4.8300686044412355e-05, "loss": 4.5211, "step": 3766 }, { "epoch": 0.034004332912077995, "grad_norm": 4.850132465362549, "learning_rate": 4.830023469940423e-05, "loss": 4.8454, "step": 3767 }, { "epoch": 0.034013359812240476, "grad_norm": 4.744113922119141, "learning_rate": 4.8299783354396106e-05, "loss": 3.353, "step": 3768 }, { "epoch": 0.03402238671240296, "grad_norm": 3.1196274757385254, "learning_rate": 4.829933200938798e-05, "loss": 5.0065, "step": 3769 }, { "epoch": 0.034031413612565446, "grad_norm": 3.1976778507232666, "learning_rate": 4.829888066437985e-05, "loss": 4.9521, "step": 3770 }, { "epoch": 0.03404044051272793, "grad_norm": 3.7436511516571045, "learning_rate": 4.829842931937173e-05, "loss": 5.1062, "step": 3771 }, { "epoch": 0.034049467412890416, "grad_norm": 3.728922128677368, "learning_rate": 4.82979779743636e-05, "loss": 4.6462, "step": 3772 }, { "epoch": 0.0340584943130529, "grad_norm": 3.401128053665161, "learning_rate": 4.829752662935548e-05, "loss": 3.9939, "step": 3773 }, { "epoch": 0.03406752121321538, "grad_norm": 5.878567695617676, "learning_rate": 4.829707528434736e-05, "loss": 4.0279, "step": 3774 }, { "epoch": 0.03407654811337787, "grad_norm": 3.12731671333313, "learning_rate": 4.829662393933923e-05, "loss": 4.4302, "step": 3775 }, { "epoch": 0.03408557501354035, "grad_norm": 3.7235777378082275, "learning_rate": 4.8296172594331107e-05, "loss": 4.7285, "step": 3776 }, { "epoch": 0.03409460191370284, "grad_norm": 2.855882167816162, "learning_rate": 4.8295721249322986e-05, "loss": 5.3501, "step": 3777 }, { "epoch": 0.03410362881386532, "grad_norm": 4.516730308532715, "learning_rate": 4.829526990431486e-05, "loss": 5.2248, "step": 3778 }, { "epoch": 0.0341126557140278, "grad_norm": 3.313906192779541, "learning_rate": 4.829481855930674e-05, "loss": 3.7578, "step": 3779 }, { "epoch": 0.03412168261419029, "grad_norm": 3.2287070751190186, "learning_rate": 4.829436721429862e-05, "loss": 5.0748, "step": 3780 }, { "epoch": 0.03413070951435277, "grad_norm": 3.3914549350738525, "learning_rate": 4.829391586929049e-05, "loss": 5.1304, "step": 3781 }, { "epoch": 0.03413973641451525, "grad_norm": 3.2812771797180176, "learning_rate": 4.829346452428236e-05, "loss": 4.7061, "step": 3782 }, { "epoch": 0.03414876331467774, "grad_norm": 3.799837112426758, "learning_rate": 4.829301317927424e-05, "loss": 4.9882, "step": 3783 }, { "epoch": 0.03415779021484022, "grad_norm": 5.387024879455566, "learning_rate": 4.829256183426611e-05, "loss": 4.6641, "step": 3784 }, { "epoch": 0.03416681711500271, "grad_norm": 3.72953462600708, "learning_rate": 4.829211048925799e-05, "loss": 4.5982, "step": 3785 }, { "epoch": 0.03417584401516519, "grad_norm": 3.13053822517395, "learning_rate": 4.8291659144249865e-05, "loss": 4.858, "step": 3786 }, { "epoch": 0.034184870915327674, "grad_norm": 10.576807022094727, "learning_rate": 4.8291207799241744e-05, "loss": 4.0915, "step": 3787 }, { "epoch": 0.03419389781549016, "grad_norm": 3.3496856689453125, "learning_rate": 4.8290756454233624e-05, "loss": 5.1559, "step": 3788 }, { "epoch": 0.034202924715652644, "grad_norm": 2.8413243293762207, "learning_rate": 4.829030510922549e-05, "loss": 5.0388, "step": 3789 }, { "epoch": 0.03421195161581513, "grad_norm": 3.219440460205078, "learning_rate": 4.828985376421737e-05, "loss": 5.2052, "step": 3790 }, { "epoch": 0.034220978515977614, "grad_norm": 3.2613790035247803, "learning_rate": 4.828940241920925e-05, "loss": 4.2633, "step": 3791 }, { "epoch": 0.034230005416140095, "grad_norm": 3.832995891571045, "learning_rate": 4.828895107420112e-05, "loss": 4.7974, "step": 3792 }, { "epoch": 0.034239032316302584, "grad_norm": 5.0693793296813965, "learning_rate": 4.8288499729193e-05, "loss": 5.5007, "step": 3793 }, { "epoch": 0.034248059216465065, "grad_norm": 3.5538268089294434, "learning_rate": 4.828804838418487e-05, "loss": 4.406, "step": 3794 }, { "epoch": 0.03425708611662755, "grad_norm": 4.339923858642578, "learning_rate": 4.8287597039176744e-05, "loss": 4.6524, "step": 3795 }, { "epoch": 0.034266113016790035, "grad_norm": 3.1430089473724365, "learning_rate": 4.8287145694168624e-05, "loss": 5.1419, "step": 3796 }, { "epoch": 0.03427513991695252, "grad_norm": 3.9147696495056152, "learning_rate": 4.82866943491605e-05, "loss": 5.3023, "step": 3797 }, { "epoch": 0.034284166817115005, "grad_norm": 3.4922025203704834, "learning_rate": 4.8286243004152375e-05, "loss": 4.7213, "step": 3798 }, { "epoch": 0.03429319371727749, "grad_norm": 5.022582530975342, "learning_rate": 4.8285791659144255e-05, "loss": 4.5949, "step": 3799 }, { "epoch": 0.03430222061743997, "grad_norm": 2.930610179901123, "learning_rate": 4.828534031413613e-05, "loss": 5.0424, "step": 3800 }, { "epoch": 0.03431124751760246, "grad_norm": 3.8329572677612305, "learning_rate": 4.8284888969128006e-05, "loss": 4.765, "step": 3801 }, { "epoch": 0.03432027441776494, "grad_norm": 3.656865119934082, "learning_rate": 4.828443762411988e-05, "loss": 5.0947, "step": 3802 }, { "epoch": 0.03432930131792743, "grad_norm": 3.5000405311584473, "learning_rate": 4.828398627911175e-05, "loss": 4.7416, "step": 3803 }, { "epoch": 0.03433832821808991, "grad_norm": 4.570240497589111, "learning_rate": 4.828353493410363e-05, "loss": 4.8109, "step": 3804 }, { "epoch": 0.03434735511825239, "grad_norm": 3.4377241134643555, "learning_rate": 4.828308358909551e-05, "loss": 4.8763, "step": 3805 }, { "epoch": 0.03435638201841488, "grad_norm": 4.166487216949463, "learning_rate": 4.828263224408738e-05, "loss": 4.7658, "step": 3806 }, { "epoch": 0.03436540891857736, "grad_norm": 3.9194629192352295, "learning_rate": 4.828218089907926e-05, "loss": 4.514, "step": 3807 }, { "epoch": 0.03437443581873984, "grad_norm": 2.7217981815338135, "learning_rate": 4.8281729554071134e-05, "loss": 4.2971, "step": 3808 }, { "epoch": 0.03438346271890233, "grad_norm": 4.665152549743652, "learning_rate": 4.8281278209063006e-05, "loss": 3.9293, "step": 3809 }, { "epoch": 0.03439248961906481, "grad_norm": 4.607171058654785, "learning_rate": 4.8280826864054886e-05, "loss": 4.6516, "step": 3810 }, { "epoch": 0.0344015165192273, "grad_norm": 3.166826009750366, "learning_rate": 4.828037551904676e-05, "loss": 4.7297, "step": 3811 }, { "epoch": 0.03441054341938978, "grad_norm": 2.7791759967803955, "learning_rate": 4.827992417403864e-05, "loss": 4.9335, "step": 3812 }, { "epoch": 0.03441957031955226, "grad_norm": 3.7610127925872803, "learning_rate": 4.8279472829030516e-05, "loss": 5.565, "step": 3813 }, { "epoch": 0.03442859721971475, "grad_norm": 5.048781871795654, "learning_rate": 4.827902148402239e-05, "loss": 4.6751, "step": 3814 }, { "epoch": 0.03443762411987723, "grad_norm": 2.9886696338653564, "learning_rate": 4.827857013901427e-05, "loss": 4.5876, "step": 3815 }, { "epoch": 0.03444665102003972, "grad_norm": 3.3026881217956543, "learning_rate": 4.827811879400614e-05, "loss": 5.0899, "step": 3816 }, { "epoch": 0.0344556779202022, "grad_norm": 3.074068069458008, "learning_rate": 4.827766744899801e-05, "loss": 4.2968, "step": 3817 }, { "epoch": 0.034464704820364685, "grad_norm": 3.3121559619903564, "learning_rate": 4.827721610398989e-05, "loss": 4.9079, "step": 3818 }, { "epoch": 0.03447373172052717, "grad_norm": 3.8149168491363525, "learning_rate": 4.827676475898177e-05, "loss": 4.3889, "step": 3819 }, { "epoch": 0.034482758620689655, "grad_norm": 5.750323295593262, "learning_rate": 4.8276313413973644e-05, "loss": 4.6541, "step": 3820 }, { "epoch": 0.034491785520852136, "grad_norm": 3.7433955669403076, "learning_rate": 4.827586206896552e-05, "loss": 4.2456, "step": 3821 }, { "epoch": 0.034500812421014625, "grad_norm": 3.8123619556427, "learning_rate": 4.8275410723957396e-05, "loss": 4.6895, "step": 3822 }, { "epoch": 0.034509839321177106, "grad_norm": 4.182777404785156, "learning_rate": 4.827495937894927e-05, "loss": 4.6696, "step": 3823 }, { "epoch": 0.034518866221339595, "grad_norm": 3.798529863357544, "learning_rate": 4.827450803394115e-05, "loss": 4.9989, "step": 3824 }, { "epoch": 0.034527893121502076, "grad_norm": 6.557467460632324, "learning_rate": 4.827405668893302e-05, "loss": 4.3726, "step": 3825 }, { "epoch": 0.03453692002166456, "grad_norm": 3.619450807571411, "learning_rate": 4.82736053439249e-05, "loss": 4.1688, "step": 3826 }, { "epoch": 0.034545946921827046, "grad_norm": 4.207652568817139, "learning_rate": 4.827315399891678e-05, "loss": 4.4564, "step": 3827 }, { "epoch": 0.03455497382198953, "grad_norm": 4.112683296203613, "learning_rate": 4.827270265390865e-05, "loss": 4.6798, "step": 3828 }, { "epoch": 0.034564000722152016, "grad_norm": 3.815298557281494, "learning_rate": 4.827225130890052e-05, "loss": 4.8727, "step": 3829 }, { "epoch": 0.0345730276223145, "grad_norm": 3.051193952560425, "learning_rate": 4.82717999638924e-05, "loss": 4.7313, "step": 3830 }, { "epoch": 0.03458205452247698, "grad_norm": 3.4975924491882324, "learning_rate": 4.8271348618884275e-05, "loss": 4.2213, "step": 3831 }, { "epoch": 0.03459108142263947, "grad_norm": 3.6644227504730225, "learning_rate": 4.8270897273876154e-05, "loss": 4.5054, "step": 3832 }, { "epoch": 0.03460010832280195, "grad_norm": 4.608572483062744, "learning_rate": 4.827044592886803e-05, "loss": 5.0611, "step": 3833 }, { "epoch": 0.03460913522296443, "grad_norm": 3.2349154949188232, "learning_rate": 4.8269994583859906e-05, "loss": 4.7594, "step": 3834 }, { "epoch": 0.03461816212312692, "grad_norm": 3.336632013320923, "learning_rate": 4.8269543238851785e-05, "loss": 4.5348, "step": 3835 }, { "epoch": 0.0346271890232894, "grad_norm": 2.8159303665161133, "learning_rate": 4.826909189384365e-05, "loss": 4.5371, "step": 3836 }, { "epoch": 0.03463621592345189, "grad_norm": 3.2721736431121826, "learning_rate": 4.826864054883553e-05, "loss": 5.0515, "step": 3837 }, { "epoch": 0.03464524282361437, "grad_norm": 2.6124887466430664, "learning_rate": 4.826818920382741e-05, "loss": 5.1692, "step": 3838 }, { "epoch": 0.03465426972377685, "grad_norm": 3.136704206466675, "learning_rate": 4.826773785881928e-05, "loss": 4.4983, "step": 3839 }, { "epoch": 0.03466329662393934, "grad_norm": 4.855647087097168, "learning_rate": 4.826728651381116e-05, "loss": 3.9873, "step": 3840 }, { "epoch": 0.03467232352410182, "grad_norm": 3.901838779449463, "learning_rate": 4.826683516880304e-05, "loss": 4.6632, "step": 3841 }, { "epoch": 0.03468135042426431, "grad_norm": 4.297654628753662, "learning_rate": 4.8266383823794906e-05, "loss": 4.9013, "step": 3842 }, { "epoch": 0.03469037732442679, "grad_norm": 3.8995442390441895, "learning_rate": 4.8265932478786785e-05, "loss": 4.6967, "step": 3843 }, { "epoch": 0.034699404224589274, "grad_norm": 3.491460084915161, "learning_rate": 4.8265481133778665e-05, "loss": 4.6619, "step": 3844 }, { "epoch": 0.03470843112475176, "grad_norm": 4.617280960083008, "learning_rate": 4.826502978877054e-05, "loss": 5.4284, "step": 3845 }, { "epoch": 0.034717458024914244, "grad_norm": 2.9275283813476562, "learning_rate": 4.8264578443762416e-05, "loss": 4.0321, "step": 3846 }, { "epoch": 0.034726484925076725, "grad_norm": 3.1531124114990234, "learning_rate": 4.826412709875429e-05, "loss": 4.4844, "step": 3847 }, { "epoch": 0.034735511825239214, "grad_norm": 3.7945287227630615, "learning_rate": 4.826367575374617e-05, "loss": 4.9398, "step": 3848 }, { "epoch": 0.034744538725401695, "grad_norm": 3.5869827270507812, "learning_rate": 4.826322440873804e-05, "loss": 4.6153, "step": 3849 }, { "epoch": 0.034753565625564184, "grad_norm": 3.5215067863464355, "learning_rate": 4.826277306372991e-05, "loss": 4.8123, "step": 3850 }, { "epoch": 0.034762592525726665, "grad_norm": 3.8329708576202393, "learning_rate": 4.826232171872179e-05, "loss": 5.5405, "step": 3851 }, { "epoch": 0.03477161942588915, "grad_norm": 3.6129934787750244, "learning_rate": 4.826187037371367e-05, "loss": 4.9623, "step": 3852 }, { "epoch": 0.034780646326051635, "grad_norm": 3.7357828617095947, "learning_rate": 4.8261419028705544e-05, "loss": 4.5359, "step": 3853 }, { "epoch": 0.03478967322621412, "grad_norm": 4.032566547393799, "learning_rate": 4.826096768369742e-05, "loss": 5.0292, "step": 3854 }, { "epoch": 0.034798700126376605, "grad_norm": 3.6873462200164795, "learning_rate": 4.8260516338689296e-05, "loss": 4.4588, "step": 3855 }, { "epoch": 0.03480772702653909, "grad_norm": 4.580665588378906, "learning_rate": 4.826006499368117e-05, "loss": 4.3413, "step": 3856 }, { "epoch": 0.03481675392670157, "grad_norm": 4.4482421875, "learning_rate": 4.825961364867305e-05, "loss": 3.7779, "step": 3857 }, { "epoch": 0.03482578082686406, "grad_norm": 3.613659143447876, "learning_rate": 4.8259162303664926e-05, "loss": 4.9628, "step": 3858 }, { "epoch": 0.03483480772702654, "grad_norm": 5.097867488861084, "learning_rate": 4.82587109586568e-05, "loss": 5.2325, "step": 3859 }, { "epoch": 0.03484383462718902, "grad_norm": 3.13553524017334, "learning_rate": 4.825825961364868e-05, "loss": 5.1454, "step": 3860 }, { "epoch": 0.03485286152735151, "grad_norm": 3.108873128890991, "learning_rate": 4.825780826864055e-05, "loss": 4.6408, "step": 3861 }, { "epoch": 0.03486188842751399, "grad_norm": 2.8233561515808105, "learning_rate": 4.825735692363243e-05, "loss": 4.6595, "step": 3862 }, { "epoch": 0.03487091532767648, "grad_norm": 2.9327425956726074, "learning_rate": 4.82569055786243e-05, "loss": 4.9145, "step": 3863 }, { "epoch": 0.03487994222783896, "grad_norm": 2.9076998233795166, "learning_rate": 4.8256454233616175e-05, "loss": 4.6904, "step": 3864 }, { "epoch": 0.03488896912800144, "grad_norm": 5.1652021408081055, "learning_rate": 4.8256002888608054e-05, "loss": 4.1605, "step": 3865 }, { "epoch": 0.03489799602816393, "grad_norm": 5.899582862854004, "learning_rate": 4.825555154359993e-05, "loss": 4.5815, "step": 3866 }, { "epoch": 0.03490702292832641, "grad_norm": 3.9578030109405518, "learning_rate": 4.8255100198591806e-05, "loss": 5.1725, "step": 3867 }, { "epoch": 0.0349160498284889, "grad_norm": 4.192164421081543, "learning_rate": 4.8254648853583685e-05, "loss": 4.4083, "step": 3868 }, { "epoch": 0.03492507672865138, "grad_norm": 2.535956859588623, "learning_rate": 4.825419750857556e-05, "loss": 4.9226, "step": 3869 }, { "epoch": 0.03493410362881386, "grad_norm": 2.8172028064727783, "learning_rate": 4.825374616356743e-05, "loss": 4.3985, "step": 3870 }, { "epoch": 0.03494313052897635, "grad_norm": 4.128332614898682, "learning_rate": 4.825329481855931e-05, "loss": 4.3371, "step": 3871 }, { "epoch": 0.03495215742913883, "grad_norm": 2.7720470428466797, "learning_rate": 4.825284347355118e-05, "loss": 4.9598, "step": 3872 }, { "epoch": 0.034961184329301315, "grad_norm": 3.7112040519714355, "learning_rate": 4.825239212854306e-05, "loss": 4.991, "step": 3873 }, { "epoch": 0.0349702112294638, "grad_norm": 4.149596691131592, "learning_rate": 4.825194078353494e-05, "loss": 4.0971, "step": 3874 }, { "epoch": 0.034979238129626285, "grad_norm": 3.9269025325775146, "learning_rate": 4.825148943852681e-05, "loss": 4.3763, "step": 3875 }, { "epoch": 0.03498826502978877, "grad_norm": 3.2945849895477295, "learning_rate": 4.8251038093518685e-05, "loss": 4.0871, "step": 3876 }, { "epoch": 0.034997291929951255, "grad_norm": 3.6264429092407227, "learning_rate": 4.8250586748510564e-05, "loss": 4.8894, "step": 3877 }, { "epoch": 0.035006318830113736, "grad_norm": 2.7627196311950684, "learning_rate": 4.825013540350244e-05, "loss": 5.0401, "step": 3878 }, { "epoch": 0.035015345730276225, "grad_norm": 5.932417392730713, "learning_rate": 4.8249684058494316e-05, "loss": 4.9481, "step": 3879 }, { "epoch": 0.035024372630438706, "grad_norm": 3.5307297706604004, "learning_rate": 4.8249232713486195e-05, "loss": 4.5842, "step": 3880 }, { "epoch": 0.035033399530601195, "grad_norm": 5.105662822723389, "learning_rate": 4.824878136847807e-05, "loss": 3.3222, "step": 3881 }, { "epoch": 0.035042426430763676, "grad_norm": 5.871695518493652, "learning_rate": 4.824833002346995e-05, "loss": 4.9359, "step": 3882 }, { "epoch": 0.03505145333092616, "grad_norm": 4.209343910217285, "learning_rate": 4.824787867846182e-05, "loss": 3.7759, "step": 3883 }, { "epoch": 0.035060480231088646, "grad_norm": 2.8493216037750244, "learning_rate": 4.824742733345369e-05, "loss": 5.2435, "step": 3884 }, { "epoch": 0.03506950713125113, "grad_norm": 2.6676549911499023, "learning_rate": 4.824697598844557e-05, "loss": 4.9142, "step": 3885 }, { "epoch": 0.03507853403141361, "grad_norm": 5.4439215660095215, "learning_rate": 4.8246524643437444e-05, "loss": 5.2212, "step": 3886 }, { "epoch": 0.0350875609315761, "grad_norm": 3.781567335128784, "learning_rate": 4.824607329842932e-05, "loss": 4.5984, "step": 3887 }, { "epoch": 0.03509658783173858, "grad_norm": 5.570319175720215, "learning_rate": 4.82456219534212e-05, "loss": 5.4314, "step": 3888 }, { "epoch": 0.03510561473190107, "grad_norm": 3.256971597671509, "learning_rate": 4.824517060841307e-05, "loss": 4.8277, "step": 3889 }, { "epoch": 0.03511464163206355, "grad_norm": 4.7141523361206055, "learning_rate": 4.824471926340495e-05, "loss": 3.721, "step": 3890 }, { "epoch": 0.03512366853222603, "grad_norm": 3.559037685394287, "learning_rate": 4.8244267918396826e-05, "loss": 5.0375, "step": 3891 }, { "epoch": 0.03513269543238852, "grad_norm": 4.0323100090026855, "learning_rate": 4.82438165733887e-05, "loss": 3.8748, "step": 3892 }, { "epoch": 0.035141722332551, "grad_norm": 3.2221617698669434, "learning_rate": 4.824336522838058e-05, "loss": 4.3618, "step": 3893 }, { "epoch": 0.03515074923271349, "grad_norm": 3.9748969078063965, "learning_rate": 4.824291388337245e-05, "loss": 4.6474, "step": 3894 }, { "epoch": 0.03515977613287597, "grad_norm": 3.8896615505218506, "learning_rate": 4.824246253836433e-05, "loss": 4.3098, "step": 3895 }, { "epoch": 0.03516880303303845, "grad_norm": 4.372281551361084, "learning_rate": 4.82420111933562e-05, "loss": 4.2627, "step": 3896 }, { "epoch": 0.03517782993320094, "grad_norm": 3.9331135749816895, "learning_rate": 4.824155984834808e-05, "loss": 4.5838, "step": 3897 }, { "epoch": 0.03518685683336342, "grad_norm": 4.167125701904297, "learning_rate": 4.8241108503339954e-05, "loss": 4.42, "step": 3898 }, { "epoch": 0.035195883733525904, "grad_norm": 4.297623634338379, "learning_rate": 4.824065715833183e-05, "loss": 4.6513, "step": 3899 }, { "epoch": 0.03520491063368839, "grad_norm": 2.2789533138275146, "learning_rate": 4.8240205813323706e-05, "loss": 4.7853, "step": 3900 }, { "epoch": 0.035213937533850874, "grad_norm": 4.691030502319336, "learning_rate": 4.8239754468315585e-05, "loss": 5.5706, "step": 3901 }, { "epoch": 0.03522296443401336, "grad_norm": 4.299862861633301, "learning_rate": 4.8239303123307464e-05, "loss": 4.9801, "step": 3902 }, { "epoch": 0.035231991334175844, "grad_norm": 3.827704906463623, "learning_rate": 4.823885177829933e-05, "loss": 3.9546, "step": 3903 }, { "epoch": 0.035241018234338325, "grad_norm": 3.988767385482788, "learning_rate": 4.823840043329121e-05, "loss": 4.5167, "step": 3904 }, { "epoch": 0.035250045134500814, "grad_norm": 3.1437489986419678, "learning_rate": 4.823794908828309e-05, "loss": 5.2734, "step": 3905 }, { "epoch": 0.035259072034663295, "grad_norm": 3.0311193466186523, "learning_rate": 4.823749774327496e-05, "loss": 4.4019, "step": 3906 }, { "epoch": 0.035268098934825784, "grad_norm": 4.370110511779785, "learning_rate": 4.823704639826684e-05, "loss": 4.7087, "step": 3907 }, { "epoch": 0.035277125834988265, "grad_norm": 3.6542937755584717, "learning_rate": 4.823659505325871e-05, "loss": 4.9469, "step": 3908 }, { "epoch": 0.03528615273515075, "grad_norm": 3.0721230506896973, "learning_rate": 4.823614370825059e-05, "loss": 4.4361, "step": 3909 }, { "epoch": 0.035295179635313235, "grad_norm": 2.383978843688965, "learning_rate": 4.8235692363242464e-05, "loss": 4.9106, "step": 3910 }, { "epoch": 0.03530420653547572, "grad_norm": 2.5384366512298584, "learning_rate": 4.8235241018234337e-05, "loss": 4.649, "step": 3911 }, { "epoch": 0.0353132334356382, "grad_norm": 3.4951648712158203, "learning_rate": 4.8234789673226216e-05, "loss": 3.7933, "step": 3912 }, { "epoch": 0.03532226033580069, "grad_norm": 4.325876235961914, "learning_rate": 4.8234338328218095e-05, "loss": 4.8833, "step": 3913 }, { "epoch": 0.03533128723596317, "grad_norm": 3.941798448562622, "learning_rate": 4.823388698320997e-05, "loss": 5.1476, "step": 3914 }, { "epoch": 0.03534031413612566, "grad_norm": 4.072367191314697, "learning_rate": 4.823343563820185e-05, "loss": 4.6245, "step": 3915 }, { "epoch": 0.03534934103628814, "grad_norm": 3.4150500297546387, "learning_rate": 4.823298429319372e-05, "loss": 4.5532, "step": 3916 }, { "epoch": 0.03535836793645062, "grad_norm": 3.2143633365631104, "learning_rate": 4.823253294818559e-05, "loss": 4.5546, "step": 3917 }, { "epoch": 0.03536739483661311, "grad_norm": 4.175321102142334, "learning_rate": 4.823208160317747e-05, "loss": 4.7093, "step": 3918 }, { "epoch": 0.03537642173677559, "grad_norm": 3.523869276046753, "learning_rate": 4.823163025816935e-05, "loss": 4.9096, "step": 3919 }, { "epoch": 0.03538544863693808, "grad_norm": 3.6539084911346436, "learning_rate": 4.823117891316122e-05, "loss": 4.5926, "step": 3920 }, { "epoch": 0.03539447553710056, "grad_norm": 3.2046501636505127, "learning_rate": 4.82307275681531e-05, "loss": 4.8661, "step": 3921 }, { "epoch": 0.03540350243726304, "grad_norm": 3.7269086837768555, "learning_rate": 4.8230276223144974e-05, "loss": 4.886, "step": 3922 }, { "epoch": 0.03541252933742553, "grad_norm": 3.7231974601745605, "learning_rate": 4.822982487813685e-05, "loss": 5.0455, "step": 3923 }, { "epoch": 0.03542155623758801, "grad_norm": 3.8077332973480225, "learning_rate": 4.8229373533128726e-05, "loss": 5.1436, "step": 3924 }, { "epoch": 0.03543058313775049, "grad_norm": 2.843940019607544, "learning_rate": 4.82289221881206e-05, "loss": 4.4131, "step": 3925 }, { "epoch": 0.03543961003791298, "grad_norm": 3.249410390853882, "learning_rate": 4.822847084311248e-05, "loss": 4.8603, "step": 3926 }, { "epoch": 0.03544863693807546, "grad_norm": 3.8091378211975098, "learning_rate": 4.822801949810436e-05, "loss": 5.3002, "step": 3927 }, { "epoch": 0.03545766383823795, "grad_norm": 4.0242390632629395, "learning_rate": 4.822756815309623e-05, "loss": 4.4208, "step": 3928 }, { "epoch": 0.03546669073840043, "grad_norm": 3.5335569381713867, "learning_rate": 4.822711680808811e-05, "loss": 5.4134, "step": 3929 }, { "epoch": 0.035475717638562915, "grad_norm": 6.724981307983398, "learning_rate": 4.822666546307998e-05, "loss": 4.435, "step": 3930 }, { "epoch": 0.0354847445387254, "grad_norm": 3.9163200855255127, "learning_rate": 4.8226214118071854e-05, "loss": 4.6401, "step": 3931 }, { "epoch": 0.035493771438887885, "grad_norm": 3.567357063293457, "learning_rate": 4.822576277306373e-05, "loss": 5.1013, "step": 3932 }, { "epoch": 0.03550279833905037, "grad_norm": 3.814525842666626, "learning_rate": 4.8225311428055605e-05, "loss": 3.8505, "step": 3933 }, { "epoch": 0.035511825239212855, "grad_norm": 3.111436128616333, "learning_rate": 4.8224860083047485e-05, "loss": 4.6665, "step": 3934 }, { "epoch": 0.035520852139375336, "grad_norm": 4.6287970542907715, "learning_rate": 4.8224408738039364e-05, "loss": 4.9029, "step": 3935 }, { "epoch": 0.035529879039537825, "grad_norm": 3.99828839302063, "learning_rate": 4.822395739303123e-05, "loss": 4.8274, "step": 3936 }, { "epoch": 0.035538905939700306, "grad_norm": 3.606262683868408, "learning_rate": 4.822350604802311e-05, "loss": 4.3361, "step": 3937 }, { "epoch": 0.03554793283986279, "grad_norm": 3.3661675453186035, "learning_rate": 4.822305470301499e-05, "loss": 5.0147, "step": 3938 }, { "epoch": 0.035556959740025276, "grad_norm": 2.866846799850464, "learning_rate": 4.822260335800686e-05, "loss": 4.8391, "step": 3939 }, { "epoch": 0.03556598664018776, "grad_norm": 3.828765630722046, "learning_rate": 4.822215201299874e-05, "loss": 4.4584, "step": 3940 }, { "epoch": 0.035575013540350246, "grad_norm": 3.7834177017211914, "learning_rate": 4.822170066799062e-05, "loss": 4.4205, "step": 3941 }, { "epoch": 0.03558404044051273, "grad_norm": 3.221306324005127, "learning_rate": 4.822124932298249e-05, "loss": 5.0051, "step": 3942 }, { "epoch": 0.03559306734067521, "grad_norm": 4.056507587432861, "learning_rate": 4.8220797977974364e-05, "loss": 4.2552, "step": 3943 }, { "epoch": 0.0356020942408377, "grad_norm": 4.058862686157227, "learning_rate": 4.822034663296624e-05, "loss": 5.3337, "step": 3944 }, { "epoch": 0.03561112114100018, "grad_norm": 3.0893795490264893, "learning_rate": 4.8219895287958116e-05, "loss": 5.0016, "step": 3945 }, { "epoch": 0.03562014804116267, "grad_norm": 3.9594292640686035, "learning_rate": 4.8219443942949995e-05, "loss": 5.829, "step": 3946 }, { "epoch": 0.03562917494132515, "grad_norm": 2.7720842361450195, "learning_rate": 4.821899259794187e-05, "loss": 5.0255, "step": 3947 }, { "epoch": 0.03563820184148763, "grad_norm": 1.8980810642242432, "learning_rate": 4.8218541252933746e-05, "loss": 4.8765, "step": 3948 }, { "epoch": 0.03564722874165012, "grad_norm": 2.9781038761138916, "learning_rate": 4.8218089907925626e-05, "loss": 4.8487, "step": 3949 }, { "epoch": 0.0356562556418126, "grad_norm": 6.105467796325684, "learning_rate": 4.821763856291749e-05, "loss": 4.6589, "step": 3950 }, { "epoch": 0.03566528254197508, "grad_norm": 4.264915466308594, "learning_rate": 4.821718721790937e-05, "loss": 4.535, "step": 3951 }, { "epoch": 0.03567430944213757, "grad_norm": 2.999363899230957, "learning_rate": 4.821673587290125e-05, "loss": 4.5302, "step": 3952 }, { "epoch": 0.03568333634230005, "grad_norm": 3.5197975635528564, "learning_rate": 4.821628452789312e-05, "loss": 4.7699, "step": 3953 }, { "epoch": 0.03569236324246254, "grad_norm": 3.2132599353790283, "learning_rate": 4.8215833182885e-05, "loss": 4.9215, "step": 3954 }, { "epoch": 0.03570139014262502, "grad_norm": 4.245574951171875, "learning_rate": 4.8215381837876874e-05, "loss": 5.4675, "step": 3955 }, { "epoch": 0.035710417042787504, "grad_norm": 4.408749580383301, "learning_rate": 4.8214930492868746e-05, "loss": 3.5949, "step": 3956 }, { "epoch": 0.03571944394294999, "grad_norm": 2.851726770401001, "learning_rate": 4.8214479147860626e-05, "loss": 4.5972, "step": 3957 }, { "epoch": 0.035728470843112474, "grad_norm": 3.2058467864990234, "learning_rate": 4.8214027802852505e-05, "loss": 4.7891, "step": 3958 }, { "epoch": 0.03573749774327496, "grad_norm": 4.6362409591674805, "learning_rate": 4.821357645784438e-05, "loss": 3.5307, "step": 3959 }, { "epoch": 0.035746524643437444, "grad_norm": 3.3901896476745605, "learning_rate": 4.821312511283626e-05, "loss": 5.111, "step": 3960 }, { "epoch": 0.035755551543599925, "grad_norm": 3.2503674030303955, "learning_rate": 4.821267376782813e-05, "loss": 4.7252, "step": 3961 }, { "epoch": 0.035764578443762414, "grad_norm": 3.062080144882202, "learning_rate": 4.821222242282001e-05, "loss": 4.788, "step": 3962 }, { "epoch": 0.035773605343924895, "grad_norm": 3.102651357650757, "learning_rate": 4.821177107781188e-05, "loss": 4.9856, "step": 3963 }, { "epoch": 0.03578263224408738, "grad_norm": 3.327611207962036, "learning_rate": 4.821131973280375e-05, "loss": 4.9268, "step": 3964 }, { "epoch": 0.035791659144249865, "grad_norm": 3.8680217266082764, "learning_rate": 4.821086838779563e-05, "loss": 4.5283, "step": 3965 }, { "epoch": 0.03580068604441235, "grad_norm": 3.2215518951416016, "learning_rate": 4.821041704278751e-05, "loss": 3.7069, "step": 3966 }, { "epoch": 0.035809712944574836, "grad_norm": 3.214787721633911, "learning_rate": 4.8209965697779384e-05, "loss": 4.8347, "step": 3967 }, { "epoch": 0.03581873984473732, "grad_norm": 3.7877197265625, "learning_rate": 4.8209514352771264e-05, "loss": 4.9564, "step": 3968 }, { "epoch": 0.0358277667448998, "grad_norm": 4.469936370849609, "learning_rate": 4.8209063007763136e-05, "loss": 4.6401, "step": 3969 }, { "epoch": 0.03583679364506229, "grad_norm": 4.299330711364746, "learning_rate": 4.820861166275501e-05, "loss": 4.3258, "step": 3970 }, { "epoch": 0.03584582054522477, "grad_norm": 3.432515859603882, "learning_rate": 4.820816031774689e-05, "loss": 4.966, "step": 3971 }, { "epoch": 0.03585484744538726, "grad_norm": 3.883976697921753, "learning_rate": 4.820770897273876e-05, "loss": 4.4294, "step": 3972 }, { "epoch": 0.03586387434554974, "grad_norm": 4.293912887573242, "learning_rate": 4.820725762773064e-05, "loss": 4.1269, "step": 3973 }, { "epoch": 0.03587290124571222, "grad_norm": 3.94132399559021, "learning_rate": 4.820680628272252e-05, "loss": 5.0834, "step": 3974 }, { "epoch": 0.03588192814587471, "grad_norm": 4.60013484954834, "learning_rate": 4.820635493771439e-05, "loss": 4.9641, "step": 3975 }, { "epoch": 0.03589095504603719, "grad_norm": 4.13368034362793, "learning_rate": 4.820590359270627e-05, "loss": 5.4041, "step": 3976 }, { "epoch": 0.03589998194619967, "grad_norm": 4.760021686553955, "learning_rate": 4.820545224769814e-05, "loss": 5.0623, "step": 3977 }, { "epoch": 0.03590900884636216, "grad_norm": 5.943287372589111, "learning_rate": 4.8205000902690015e-05, "loss": 4.314, "step": 3978 }, { "epoch": 0.03591803574652464, "grad_norm": 5.020456314086914, "learning_rate": 4.8204549557681894e-05, "loss": 4.2022, "step": 3979 }, { "epoch": 0.03592706264668713, "grad_norm": 3.3631515502929688, "learning_rate": 4.8204098212673774e-05, "loss": 4.5118, "step": 3980 }, { "epoch": 0.03593608954684961, "grad_norm": 3.4402964115142822, "learning_rate": 4.8203646867665646e-05, "loss": 5.0867, "step": 3981 }, { "epoch": 0.03594511644701209, "grad_norm": 4.090478897094727, "learning_rate": 4.8203195522657525e-05, "loss": 4.9325, "step": 3982 }, { "epoch": 0.03595414334717458, "grad_norm": 2.9008712768554688, "learning_rate": 4.82027441776494e-05, "loss": 4.6545, "step": 3983 }, { "epoch": 0.03596317024733706, "grad_norm": 2.81693959236145, "learning_rate": 4.820229283264127e-05, "loss": 4.4265, "step": 3984 }, { "epoch": 0.03597219714749955, "grad_norm": 4.229250431060791, "learning_rate": 4.820184148763315e-05, "loss": 4.6351, "step": 3985 }, { "epoch": 0.03598122404766203, "grad_norm": 4.995671272277832, "learning_rate": 4.820139014262502e-05, "loss": 4.1888, "step": 3986 }, { "epoch": 0.035990250947824515, "grad_norm": 3.31851863861084, "learning_rate": 4.82009387976169e-05, "loss": 4.6293, "step": 3987 }, { "epoch": 0.035999277847987, "grad_norm": 3.868403196334839, "learning_rate": 4.820048745260878e-05, "loss": 5.0535, "step": 3988 }, { "epoch": 0.036008304748149485, "grad_norm": 5.051089763641357, "learning_rate": 4.820003610760065e-05, "loss": 4.45, "step": 3989 }, { "epoch": 0.036017331648311966, "grad_norm": 3.7665491104125977, "learning_rate": 4.8199584762592525e-05, "loss": 5.4084, "step": 3990 }, { "epoch": 0.036026358548474455, "grad_norm": 3.726625919342041, "learning_rate": 4.8199133417584405e-05, "loss": 4.5849, "step": 3991 }, { "epoch": 0.036035385448636936, "grad_norm": 4.915926933288574, "learning_rate": 4.819868207257628e-05, "loss": 4.5729, "step": 3992 }, { "epoch": 0.036044412348799425, "grad_norm": 3.0999021530151367, "learning_rate": 4.8198230727568156e-05, "loss": 4.8738, "step": 3993 }, { "epoch": 0.036053439248961906, "grad_norm": 4.170443058013916, "learning_rate": 4.819777938256003e-05, "loss": 3.847, "step": 3994 }, { "epoch": 0.03606246614912439, "grad_norm": 3.108262062072754, "learning_rate": 4.819732803755191e-05, "loss": 4.6091, "step": 3995 }, { "epoch": 0.036071493049286876, "grad_norm": 3.1702942848205566, "learning_rate": 4.819687669254379e-05, "loss": 4.5529, "step": 3996 }, { "epoch": 0.03608051994944936, "grad_norm": 3.2884180545806885, "learning_rate": 4.819642534753565e-05, "loss": 4.9263, "step": 3997 }, { "epoch": 0.036089546849611846, "grad_norm": 2.6884164810180664, "learning_rate": 4.819597400252753e-05, "loss": 4.1763, "step": 3998 }, { "epoch": 0.03609857374977433, "grad_norm": 3.4351806640625, "learning_rate": 4.819552265751941e-05, "loss": 4.5005, "step": 3999 }, { "epoch": 0.03610760064993681, "grad_norm": 2.8470468521118164, "learning_rate": 4.8195071312511284e-05, "loss": 4.7395, "step": 4000 }, { "epoch": 0.0361166275500993, "grad_norm": 4.139713287353516, "learning_rate": 4.819461996750316e-05, "loss": 3.6535, "step": 4001 }, { "epoch": 0.03612565445026178, "grad_norm": 3.644576072692871, "learning_rate": 4.819416862249504e-05, "loss": 5.3859, "step": 4002 }, { "epoch": 0.03613468135042426, "grad_norm": 3.7776780128479004, "learning_rate": 4.819371727748691e-05, "loss": 4.33, "step": 4003 }, { "epoch": 0.03614370825058675, "grad_norm": 3.9026241302490234, "learning_rate": 4.819326593247879e-05, "loss": 5.3068, "step": 4004 }, { "epoch": 0.03615273515074923, "grad_norm": 3.520277738571167, "learning_rate": 4.819281458747067e-05, "loss": 4.9271, "step": 4005 }, { "epoch": 0.03616176205091172, "grad_norm": 3.6003153324127197, "learning_rate": 4.819236324246254e-05, "loss": 5.2132, "step": 4006 }, { "epoch": 0.0361707889510742, "grad_norm": 3.685558319091797, "learning_rate": 4.819191189745442e-05, "loss": 4.543, "step": 4007 }, { "epoch": 0.03617981585123668, "grad_norm": 3.583677053451538, "learning_rate": 4.819146055244629e-05, "loss": 4.6069, "step": 4008 }, { "epoch": 0.03618884275139917, "grad_norm": 2.739734411239624, "learning_rate": 4.819100920743817e-05, "loss": 4.5467, "step": 4009 }, { "epoch": 0.03619786965156165, "grad_norm": 3.079679250717163, "learning_rate": 4.819055786243004e-05, "loss": 5.1923, "step": 4010 }, { "epoch": 0.03620689655172414, "grad_norm": 2.8889319896698, "learning_rate": 4.8190106517421915e-05, "loss": 5.0002, "step": 4011 }, { "epoch": 0.03621592345188662, "grad_norm": 3.9941728115081787, "learning_rate": 4.8189655172413794e-05, "loss": 5.1809, "step": 4012 }, { "epoch": 0.036224950352049104, "grad_norm": 3.467205762863159, "learning_rate": 4.8189203827405673e-05, "loss": 4.3409, "step": 4013 }, { "epoch": 0.03623397725221159, "grad_norm": 4.037084102630615, "learning_rate": 4.8188752482397546e-05, "loss": 4.6433, "step": 4014 }, { "epoch": 0.036243004152374074, "grad_norm": 4.735999584197998, "learning_rate": 4.8188301137389425e-05, "loss": 5.8013, "step": 4015 }, { "epoch": 0.036252031052536555, "grad_norm": 2.6446244716644287, "learning_rate": 4.81878497923813e-05, "loss": 5.206, "step": 4016 }, { "epoch": 0.036261057952699044, "grad_norm": 3.2645561695098877, "learning_rate": 4.818739844737317e-05, "loss": 4.8602, "step": 4017 }, { "epoch": 0.036270084852861526, "grad_norm": 3.95912766456604, "learning_rate": 4.818694710236505e-05, "loss": 4.0069, "step": 4018 }, { "epoch": 0.036279111753024014, "grad_norm": 4.545173645019531, "learning_rate": 4.818649575735693e-05, "loss": 5.3034, "step": 4019 }, { "epoch": 0.036288138653186496, "grad_norm": 3.2401039600372314, "learning_rate": 4.81860444123488e-05, "loss": 4.8154, "step": 4020 }, { "epoch": 0.03629716555334898, "grad_norm": 6.476132869720459, "learning_rate": 4.818559306734068e-05, "loss": 4.9525, "step": 4021 }, { "epoch": 0.036306192453511466, "grad_norm": 2.940030813217163, "learning_rate": 4.818514172233255e-05, "loss": 4.349, "step": 4022 }, { "epoch": 0.03631521935367395, "grad_norm": 3.396845817565918, "learning_rate": 4.818469037732443e-05, "loss": 4.7552, "step": 4023 }, { "epoch": 0.036324246253836436, "grad_norm": 5.83351993560791, "learning_rate": 4.8184239032316304e-05, "loss": 4.6715, "step": 4024 }, { "epoch": 0.03633327315399892, "grad_norm": 5.752065658569336, "learning_rate": 4.818378768730818e-05, "loss": 4.8871, "step": 4025 }, { "epoch": 0.0363423000541614, "grad_norm": 4.167935848236084, "learning_rate": 4.8183336342300056e-05, "loss": 5.0674, "step": 4026 }, { "epoch": 0.03635132695432389, "grad_norm": 3.2923402786254883, "learning_rate": 4.8182884997291935e-05, "loss": 4.5287, "step": 4027 }, { "epoch": 0.03636035385448637, "grad_norm": 4.375055313110352, "learning_rate": 4.818243365228381e-05, "loss": 5.6521, "step": 4028 }, { "epoch": 0.03636938075464885, "grad_norm": 3.952115297317505, "learning_rate": 4.818198230727569e-05, "loss": 4.5972, "step": 4029 }, { "epoch": 0.03637840765481134, "grad_norm": 3.388779878616333, "learning_rate": 4.818153096226756e-05, "loss": 4.9374, "step": 4030 }, { "epoch": 0.03638743455497382, "grad_norm": 3.3631534576416016, "learning_rate": 4.818107961725943e-05, "loss": 4.8083, "step": 4031 }, { "epoch": 0.03639646145513631, "grad_norm": 3.7617616653442383, "learning_rate": 4.818062827225131e-05, "loss": 5.2693, "step": 4032 }, { "epoch": 0.03640548835529879, "grad_norm": 5.190265655517578, "learning_rate": 4.8180176927243184e-05, "loss": 5.0245, "step": 4033 }, { "epoch": 0.03641451525546127, "grad_norm": 3.0110392570495605, "learning_rate": 4.817972558223506e-05, "loss": 4.8603, "step": 4034 }, { "epoch": 0.03642354215562376, "grad_norm": 4.154257297515869, "learning_rate": 4.817927423722694e-05, "loss": 5.065, "step": 4035 }, { "epoch": 0.03643256905578624, "grad_norm": 2.6042792797088623, "learning_rate": 4.8178822892218815e-05, "loss": 4.5902, "step": 4036 }, { "epoch": 0.03644159595594873, "grad_norm": 3.9058282375335693, "learning_rate": 4.817837154721069e-05, "loss": 4.8018, "step": 4037 }, { "epoch": 0.03645062285611121, "grad_norm": 3.558844566345215, "learning_rate": 4.8177920202202566e-05, "loss": 5.0318, "step": 4038 }, { "epoch": 0.03645964975627369, "grad_norm": 5.148586273193359, "learning_rate": 4.817746885719444e-05, "loss": 4.7226, "step": 4039 }, { "epoch": 0.03646867665643618, "grad_norm": 2.8179540634155273, "learning_rate": 4.817701751218632e-05, "loss": 4.8248, "step": 4040 }, { "epoch": 0.03647770355659866, "grad_norm": 3.836777687072754, "learning_rate": 4.81765661671782e-05, "loss": 5.154, "step": 4041 }, { "epoch": 0.03648673045676115, "grad_norm": 2.6362216472625732, "learning_rate": 4.817611482217007e-05, "loss": 5.1186, "step": 4042 }, { "epoch": 0.03649575735692363, "grad_norm": 2.7209200859069824, "learning_rate": 4.817566347716195e-05, "loss": 4.7065, "step": 4043 }, { "epoch": 0.036504784257086115, "grad_norm": 3.3592689037323, "learning_rate": 4.817521213215382e-05, "loss": 4.9448, "step": 4044 }, { "epoch": 0.0365138111572486, "grad_norm": 3.6993050575256348, "learning_rate": 4.8174760787145694e-05, "loss": 4.5529, "step": 4045 }, { "epoch": 0.036522838057411085, "grad_norm": 3.3714897632598877, "learning_rate": 4.817430944213757e-05, "loss": 4.8074, "step": 4046 }, { "epoch": 0.036531864957573566, "grad_norm": 12.749188423156738, "learning_rate": 4.8173858097129446e-05, "loss": 4.393, "step": 4047 }, { "epoch": 0.036540891857736055, "grad_norm": 3.384935140609741, "learning_rate": 4.8173406752121325e-05, "loss": 5.2758, "step": 4048 }, { "epoch": 0.036549918757898536, "grad_norm": 4.02942419052124, "learning_rate": 4.8172955407113204e-05, "loss": 5.4917, "step": 4049 }, { "epoch": 0.036558945658061025, "grad_norm": 4.108208179473877, "learning_rate": 4.817250406210507e-05, "loss": 5.4729, "step": 4050 }, { "epoch": 0.036567972558223506, "grad_norm": 4.09313440322876, "learning_rate": 4.817205271709695e-05, "loss": 5.5899, "step": 4051 }, { "epoch": 0.03657699945838599, "grad_norm": 4.274714946746826, "learning_rate": 4.817160137208883e-05, "loss": 4.4743, "step": 4052 }, { "epoch": 0.036586026358548476, "grad_norm": 3.7981927394866943, "learning_rate": 4.81711500270807e-05, "loss": 4.695, "step": 4053 }, { "epoch": 0.03659505325871096, "grad_norm": 3.4378082752227783, "learning_rate": 4.817069868207258e-05, "loss": 4.8625, "step": 4054 }, { "epoch": 0.036604080158873446, "grad_norm": 3.625769853591919, "learning_rate": 4.817024733706445e-05, "loss": 5.039, "step": 4055 }, { "epoch": 0.03661310705903593, "grad_norm": 4.801107406616211, "learning_rate": 4.816979599205633e-05, "loss": 4.6936, "step": 4056 }, { "epoch": 0.03662213395919841, "grad_norm": 4.916089057922363, "learning_rate": 4.8169344647048204e-05, "loss": 5.388, "step": 4057 }, { "epoch": 0.0366311608593609, "grad_norm": 3.52996826171875, "learning_rate": 4.8168893302040083e-05, "loss": 5.2486, "step": 4058 }, { "epoch": 0.03664018775952338, "grad_norm": 2.8043806552886963, "learning_rate": 4.8168441957031956e-05, "loss": 5.0708, "step": 4059 }, { "epoch": 0.03664921465968586, "grad_norm": 4.111966609954834, "learning_rate": 4.8167990612023835e-05, "loss": 4.931, "step": 4060 }, { "epoch": 0.03665824155984835, "grad_norm": 3.5825235843658447, "learning_rate": 4.816753926701571e-05, "loss": 3.9204, "step": 4061 }, { "epoch": 0.03666726846001083, "grad_norm": 3.0695278644561768, "learning_rate": 4.816708792200759e-05, "loss": 5.1632, "step": 4062 }, { "epoch": 0.03667629536017332, "grad_norm": 2.4051082134246826, "learning_rate": 4.8166636576999466e-05, "loss": 4.4778, "step": 4063 }, { "epoch": 0.0366853222603358, "grad_norm": 3.6169369220733643, "learning_rate": 4.816618523199133e-05, "loss": 4.3367, "step": 4064 }, { "epoch": 0.03669434916049828, "grad_norm": 3.594041585922241, "learning_rate": 4.816573388698321e-05, "loss": 4.8637, "step": 4065 }, { "epoch": 0.03670337606066077, "grad_norm": 3.806272268295288, "learning_rate": 4.816528254197509e-05, "loss": 4.7603, "step": 4066 }, { "epoch": 0.03671240296082325, "grad_norm": 4.750138759613037, "learning_rate": 4.816483119696696e-05, "loss": 5.0219, "step": 4067 }, { "epoch": 0.03672142986098574, "grad_norm": 3.0000503063201904, "learning_rate": 4.816437985195884e-05, "loss": 4.9759, "step": 4068 }, { "epoch": 0.03673045676114822, "grad_norm": 3.4959070682525635, "learning_rate": 4.8163928506950714e-05, "loss": 5.2014, "step": 4069 }, { "epoch": 0.036739483661310704, "grad_norm": 4.4856767654418945, "learning_rate": 4.8163477161942594e-05, "loss": 4.581, "step": 4070 }, { "epoch": 0.03674851056147319, "grad_norm": 3.8325235843658447, "learning_rate": 4.8163025816934466e-05, "loss": 5.0758, "step": 4071 }, { "epoch": 0.036757537461635674, "grad_norm": 3.1734619140625, "learning_rate": 4.816257447192634e-05, "loss": 3.5946, "step": 4072 }, { "epoch": 0.036766564361798156, "grad_norm": 3.4691970348358154, "learning_rate": 4.816212312691822e-05, "loss": 4.281, "step": 4073 }, { "epoch": 0.036775591261960644, "grad_norm": 4.302907466888428, "learning_rate": 4.81616717819101e-05, "loss": 4.4348, "step": 4074 }, { "epoch": 0.036784618162123126, "grad_norm": 3.6513309478759766, "learning_rate": 4.816122043690197e-05, "loss": 4.6486, "step": 4075 }, { "epoch": 0.036793645062285614, "grad_norm": 3.8445961475372314, "learning_rate": 4.816076909189385e-05, "loss": 4.609, "step": 4076 }, { "epoch": 0.036802671962448096, "grad_norm": 3.730268716812134, "learning_rate": 4.816031774688572e-05, "loss": 4.309, "step": 4077 }, { "epoch": 0.03681169886261058, "grad_norm": 4.4127516746521, "learning_rate": 4.8159866401877594e-05, "loss": 4.8003, "step": 4078 }, { "epoch": 0.036820725762773066, "grad_norm": 4.6143035888671875, "learning_rate": 4.815941505686947e-05, "loss": 4.6543, "step": 4079 }, { "epoch": 0.03682975266293555, "grad_norm": 4.957456588745117, "learning_rate": 4.815896371186135e-05, "loss": 4.3229, "step": 4080 }, { "epoch": 0.036838779563098036, "grad_norm": 2.873077392578125, "learning_rate": 4.8158512366853225e-05, "loss": 4.1859, "step": 4081 }, { "epoch": 0.03684780646326052, "grad_norm": 4.466931343078613, "learning_rate": 4.8158061021845104e-05, "loss": 5.2858, "step": 4082 }, { "epoch": 0.036856833363423, "grad_norm": 3.693155527114868, "learning_rate": 4.8157609676836976e-05, "loss": 4.1792, "step": 4083 }, { "epoch": 0.03686586026358549, "grad_norm": 6.976231098175049, "learning_rate": 4.815715833182885e-05, "loss": 4.594, "step": 4084 }, { "epoch": 0.03687488716374797, "grad_norm": 2.8845126628875732, "learning_rate": 4.815670698682073e-05, "loss": 5.0169, "step": 4085 }, { "epoch": 0.03688391406391045, "grad_norm": 2.998626470565796, "learning_rate": 4.81562556418126e-05, "loss": 5.003, "step": 4086 }, { "epoch": 0.03689294096407294, "grad_norm": 3.974329948425293, "learning_rate": 4.815580429680448e-05, "loss": 5.3876, "step": 4087 }, { "epoch": 0.03690196786423542, "grad_norm": 2.778933525085449, "learning_rate": 4.815535295179636e-05, "loss": 5.1634, "step": 4088 }, { "epoch": 0.03691099476439791, "grad_norm": 3.3482348918914795, "learning_rate": 4.815490160678823e-05, "loss": 4.8076, "step": 4089 }, { "epoch": 0.03692002166456039, "grad_norm": 3.122504234313965, "learning_rate": 4.815445026178011e-05, "loss": 5.3133, "step": 4090 }, { "epoch": 0.03692904856472287, "grad_norm": 3.1034293174743652, "learning_rate": 4.815399891677198e-05, "loss": 4.9642, "step": 4091 }, { "epoch": 0.03693807546488536, "grad_norm": 3.9098989963531494, "learning_rate": 4.8153547571763856e-05, "loss": 4.5885, "step": 4092 }, { "epoch": 0.03694710236504784, "grad_norm": 2.428654670715332, "learning_rate": 4.8153096226755735e-05, "loss": 5.3049, "step": 4093 }, { "epoch": 0.03695612926521033, "grad_norm": 3.404770851135254, "learning_rate": 4.815264488174761e-05, "loss": 4.9756, "step": 4094 }, { "epoch": 0.03696515616537281, "grad_norm": 3.018629550933838, "learning_rate": 4.815219353673949e-05, "loss": 4.6199, "step": 4095 }, { "epoch": 0.03697418306553529, "grad_norm": 2.6878483295440674, "learning_rate": 4.8151742191731366e-05, "loss": 4.3197, "step": 4096 }, { "epoch": 0.03698320996569778, "grad_norm": 2.4445066452026367, "learning_rate": 4.815129084672323e-05, "loss": 4.7712, "step": 4097 }, { "epoch": 0.03699223686586026, "grad_norm": 3.3949966430664062, "learning_rate": 4.815083950171511e-05, "loss": 4.2272, "step": 4098 }, { "epoch": 0.037001263766022745, "grad_norm": 6.3821258544921875, "learning_rate": 4.815038815670699e-05, "loss": 4.5374, "step": 4099 }, { "epoch": 0.03701029066618523, "grad_norm": 3.256580114364624, "learning_rate": 4.814993681169886e-05, "loss": 4.599, "step": 4100 }, { "epoch": 0.037019317566347715, "grad_norm": 2.8938419818878174, "learning_rate": 4.814948546669074e-05, "loss": 4.8809, "step": 4101 }, { "epoch": 0.0370283444665102, "grad_norm": 3.982123613357544, "learning_rate": 4.814903412168262e-05, "loss": 3.8553, "step": 4102 }, { "epoch": 0.037037371366672685, "grad_norm": 3.988692045211792, "learning_rate": 4.8148582776674493e-05, "loss": 4.0741, "step": 4103 }, { "epoch": 0.037046398266835166, "grad_norm": 4.7341156005859375, "learning_rate": 4.8148131431666366e-05, "loss": 4.8208, "step": 4104 }, { "epoch": 0.037055425166997655, "grad_norm": 3.465466260910034, "learning_rate": 4.8147680086658245e-05, "loss": 4.637, "step": 4105 }, { "epoch": 0.037064452067160136, "grad_norm": 4.093832015991211, "learning_rate": 4.814722874165012e-05, "loss": 4.7212, "step": 4106 }, { "epoch": 0.037073478967322625, "grad_norm": 3.2248456478118896, "learning_rate": 4.8146777396642e-05, "loss": 4.0912, "step": 4107 }, { "epoch": 0.037082505867485106, "grad_norm": 3.6867318153381348, "learning_rate": 4.814632605163387e-05, "loss": 4.8778, "step": 4108 }, { "epoch": 0.03709153276764759, "grad_norm": 4.121722221374512, "learning_rate": 4.814587470662575e-05, "loss": 4.5255, "step": 4109 }, { "epoch": 0.037100559667810076, "grad_norm": 3.9002957344055176, "learning_rate": 4.814542336161763e-05, "loss": 4.4606, "step": 4110 }, { "epoch": 0.03710958656797256, "grad_norm": 3.243800163269043, "learning_rate": 4.8144972016609494e-05, "loss": 4.9926, "step": 4111 }, { "epoch": 0.03711861346813504, "grad_norm": 3.535395860671997, "learning_rate": 4.814452067160137e-05, "loss": 4.4974, "step": 4112 }, { "epoch": 0.03712764036829753, "grad_norm": 4.6327996253967285, "learning_rate": 4.814406932659325e-05, "loss": 4.5895, "step": 4113 }, { "epoch": 0.03713666726846001, "grad_norm": 3.1628496646881104, "learning_rate": 4.8143617981585124e-05, "loss": 4.5308, "step": 4114 }, { "epoch": 0.0371456941686225, "grad_norm": 3.163264036178589, "learning_rate": 4.8143166636577004e-05, "loss": 5.0662, "step": 4115 }, { "epoch": 0.03715472106878498, "grad_norm": 3.7826480865478516, "learning_rate": 4.8142715291568876e-05, "loss": 4.4106, "step": 4116 }, { "epoch": 0.03716374796894746, "grad_norm": 3.473163366317749, "learning_rate": 4.8142263946560755e-05, "loss": 4.0799, "step": 4117 }, { "epoch": 0.03717277486910995, "grad_norm": 2.8530242443084717, "learning_rate": 4.814181260155263e-05, "loss": 4.8275, "step": 4118 }, { "epoch": 0.03718180176927243, "grad_norm": 3.5848331451416016, "learning_rate": 4.814136125654451e-05, "loss": 4.4791, "step": 4119 }, { "epoch": 0.03719082866943492, "grad_norm": 4.619282245635986, "learning_rate": 4.814090991153638e-05, "loss": 4.4551, "step": 4120 }, { "epoch": 0.0371998555695974, "grad_norm": 4.052906036376953, "learning_rate": 4.814045856652826e-05, "loss": 4.3792, "step": 4121 }, { "epoch": 0.03720888246975988, "grad_norm": 4.491225719451904, "learning_rate": 4.814000722152013e-05, "loss": 4.623, "step": 4122 }, { "epoch": 0.03721790936992237, "grad_norm": 3.6839327812194824, "learning_rate": 4.813955587651201e-05, "loss": 4.4715, "step": 4123 }, { "epoch": 0.03722693627008485, "grad_norm": 3.688882350921631, "learning_rate": 4.813910453150388e-05, "loss": 4.9039, "step": 4124 }, { "epoch": 0.037235963170247334, "grad_norm": 6.025360107421875, "learning_rate": 4.8138653186495755e-05, "loss": 4.8864, "step": 4125 }, { "epoch": 0.03724499007040982, "grad_norm": 3.7744767665863037, "learning_rate": 4.8138201841487635e-05, "loss": 4.8289, "step": 4126 }, { "epoch": 0.037254016970572304, "grad_norm": 3.226914167404175, "learning_rate": 4.8137750496479514e-05, "loss": 4.539, "step": 4127 }, { "epoch": 0.03726304387073479, "grad_norm": 4.379452228546143, "learning_rate": 4.8137299151471386e-05, "loss": 5.4017, "step": 4128 }, { "epoch": 0.037272070770897274, "grad_norm": 4.561130523681641, "learning_rate": 4.8136847806463266e-05, "loss": 4.8072, "step": 4129 }, { "epoch": 0.037281097671059756, "grad_norm": 3.5770559310913086, "learning_rate": 4.813639646145514e-05, "loss": 4.9606, "step": 4130 }, { "epoch": 0.037290124571222244, "grad_norm": 3.7620317935943604, "learning_rate": 4.813594511644701e-05, "loss": 4.635, "step": 4131 }, { "epoch": 0.037299151471384726, "grad_norm": 3.67903733253479, "learning_rate": 4.813549377143889e-05, "loss": 4.2362, "step": 4132 }, { "epoch": 0.037308178371547214, "grad_norm": 3.6884045600891113, "learning_rate": 4.813504242643076e-05, "loss": 4.7664, "step": 4133 }, { "epoch": 0.037317205271709696, "grad_norm": 3.8539013862609863, "learning_rate": 4.813459108142264e-05, "loss": 4.7515, "step": 4134 }, { "epoch": 0.03732623217187218, "grad_norm": 3.6947896480560303, "learning_rate": 4.813413973641452e-05, "loss": 4.4652, "step": 4135 }, { "epoch": 0.037335259072034666, "grad_norm": 3.2171480655670166, "learning_rate": 4.813368839140639e-05, "loss": 4.3134, "step": 4136 }, { "epoch": 0.03734428597219715, "grad_norm": 3.213747024536133, "learning_rate": 4.813323704639827e-05, "loss": 4.6772, "step": 4137 }, { "epoch": 0.03735331287235963, "grad_norm": 4.40361213684082, "learning_rate": 4.8132785701390145e-05, "loss": 5.1438, "step": 4138 }, { "epoch": 0.03736233977252212, "grad_norm": 4.179299354553223, "learning_rate": 4.813233435638202e-05, "loss": 5.0524, "step": 4139 }, { "epoch": 0.0373713666726846, "grad_norm": 3.345716714859009, "learning_rate": 4.81318830113739e-05, "loss": 4.5352, "step": 4140 }, { "epoch": 0.03738039357284709, "grad_norm": 5.741300582885742, "learning_rate": 4.8131431666365776e-05, "loss": 5.3825, "step": 4141 }, { "epoch": 0.03738942047300957, "grad_norm": 4.109325885772705, "learning_rate": 4.813098032135765e-05, "loss": 4.6758, "step": 4142 }, { "epoch": 0.03739844737317205, "grad_norm": 2.9841384887695312, "learning_rate": 4.813052897634953e-05, "loss": 4.9498, "step": 4143 }, { "epoch": 0.03740747427333454, "grad_norm": 3.6607279777526855, "learning_rate": 4.81300776313414e-05, "loss": 4.744, "step": 4144 }, { "epoch": 0.03741650117349702, "grad_norm": 3.2072675228118896, "learning_rate": 4.812962628633327e-05, "loss": 5.1042, "step": 4145 }, { "epoch": 0.03742552807365951, "grad_norm": 2.7713444232940674, "learning_rate": 4.812917494132515e-05, "loss": 4.4121, "step": 4146 }, { "epoch": 0.03743455497382199, "grad_norm": 3.5300471782684326, "learning_rate": 4.8128723596317024e-05, "loss": 4.456, "step": 4147 }, { "epoch": 0.03744358187398447, "grad_norm": 3.2776503562927246, "learning_rate": 4.8128272251308903e-05, "loss": 4.2897, "step": 4148 }, { "epoch": 0.03745260877414696, "grad_norm": 4.286367893218994, "learning_rate": 4.812782090630078e-05, "loss": 5.0559, "step": 4149 }, { "epoch": 0.03746163567430944, "grad_norm": 4.320574760437012, "learning_rate": 4.8127369561292655e-05, "loss": 4.2557, "step": 4150 }, { "epoch": 0.03747066257447192, "grad_norm": 4.171789169311523, "learning_rate": 4.812691821628453e-05, "loss": 5.0788, "step": 4151 }, { "epoch": 0.03747968947463441, "grad_norm": 3.06679630279541, "learning_rate": 4.812646687127641e-05, "loss": 4.7897, "step": 4152 }, { "epoch": 0.03748871637479689, "grad_norm": 3.1977322101593018, "learning_rate": 4.812601552626828e-05, "loss": 5.1542, "step": 4153 }, { "epoch": 0.03749774327495938, "grad_norm": 3.3648977279663086, "learning_rate": 4.812556418126016e-05, "loss": 4.5481, "step": 4154 }, { "epoch": 0.03750677017512186, "grad_norm": 4.0503458976745605, "learning_rate": 4.812511283625203e-05, "loss": 5.0225, "step": 4155 }, { "epoch": 0.037515797075284345, "grad_norm": 3.4761486053466797, "learning_rate": 4.812466149124391e-05, "loss": 4.6997, "step": 4156 }, { "epoch": 0.03752482397544683, "grad_norm": 3.641814947128296, "learning_rate": 4.812421014623579e-05, "loss": 4.7804, "step": 4157 }, { "epoch": 0.037533850875609315, "grad_norm": 5.038887023925781, "learning_rate": 4.8123758801227655e-05, "loss": 4.9559, "step": 4158 }, { "epoch": 0.0375428777757718, "grad_norm": 2.5169901847839355, "learning_rate": 4.8123307456219534e-05, "loss": 5.1233, "step": 4159 }, { "epoch": 0.037551904675934285, "grad_norm": 3.148465394973755, "learning_rate": 4.8122856111211414e-05, "loss": 5.0219, "step": 4160 }, { "epoch": 0.037560931576096766, "grad_norm": 3.309793710708618, "learning_rate": 4.8122404766203286e-05, "loss": 4.7785, "step": 4161 }, { "epoch": 0.037569958476259255, "grad_norm": 6.29998779296875, "learning_rate": 4.8121953421195165e-05, "loss": 4.3633, "step": 4162 }, { "epoch": 0.037578985376421736, "grad_norm": 3.8378944396972656, "learning_rate": 4.8121502076187045e-05, "loss": 4.6346, "step": 4163 }, { "epoch": 0.03758801227658422, "grad_norm": 3.1096291542053223, "learning_rate": 4.812105073117891e-05, "loss": 4.4477, "step": 4164 }, { "epoch": 0.037597039176746706, "grad_norm": 2.982600212097168, "learning_rate": 4.812059938617079e-05, "loss": 4.9476, "step": 4165 }, { "epoch": 0.03760606607690919, "grad_norm": 3.3592824935913086, "learning_rate": 4.812014804116267e-05, "loss": 4.4679, "step": 4166 }, { "epoch": 0.037615092977071676, "grad_norm": 3.8188469409942627, "learning_rate": 4.811969669615454e-05, "loss": 5.4754, "step": 4167 }, { "epoch": 0.03762411987723416, "grad_norm": 3.465336799621582, "learning_rate": 4.811924535114642e-05, "loss": 4.9736, "step": 4168 }, { "epoch": 0.03763314677739664, "grad_norm": 3.489734649658203, "learning_rate": 4.811879400613829e-05, "loss": 4.3134, "step": 4169 }, { "epoch": 0.03764217367755913, "grad_norm": 4.158909797668457, "learning_rate": 4.811834266113017e-05, "loss": 4.1847, "step": 4170 }, { "epoch": 0.03765120057772161, "grad_norm": 5.423573970794678, "learning_rate": 4.8117891316122045e-05, "loss": 5.1321, "step": 4171 }, { "epoch": 0.0376602274778841, "grad_norm": 3.8907597064971924, "learning_rate": 4.811743997111392e-05, "loss": 5.2054, "step": 4172 }, { "epoch": 0.03766925437804658, "grad_norm": 4.481765270233154, "learning_rate": 4.8116988626105796e-05, "loss": 5.0532, "step": 4173 }, { "epoch": 0.03767828127820906, "grad_norm": 4.319875717163086, "learning_rate": 4.8116537281097676e-05, "loss": 4.314, "step": 4174 }, { "epoch": 0.03768730817837155, "grad_norm": 3.5287907123565674, "learning_rate": 4.811608593608955e-05, "loss": 4.1642, "step": 4175 }, { "epoch": 0.03769633507853403, "grad_norm": 3.2656641006469727, "learning_rate": 4.811563459108143e-05, "loss": 4.6254, "step": 4176 }, { "epoch": 0.03770536197869651, "grad_norm": 4.210569858551025, "learning_rate": 4.81151832460733e-05, "loss": 4.3774, "step": 4177 }, { "epoch": 0.037714388878859, "grad_norm": 5.861791610717773, "learning_rate": 4.811473190106517e-05, "loss": 4.9247, "step": 4178 }, { "epoch": 0.03772341577902148, "grad_norm": 4.04307746887207, "learning_rate": 4.811428055605705e-05, "loss": 4.819, "step": 4179 }, { "epoch": 0.03773244267918397, "grad_norm": 5.072253227233887, "learning_rate": 4.811382921104893e-05, "loss": 4.0047, "step": 4180 }, { "epoch": 0.03774146957934645, "grad_norm": 4.676968097686768, "learning_rate": 4.81133778660408e-05, "loss": 4.7305, "step": 4181 }, { "epoch": 0.037750496479508934, "grad_norm": 3.450491189956665, "learning_rate": 4.811292652103268e-05, "loss": 4.2837, "step": 4182 }, { "epoch": 0.03775952337967142, "grad_norm": 4.291249752044678, "learning_rate": 4.8112475176024555e-05, "loss": 5.4163, "step": 4183 }, { "epoch": 0.037768550279833904, "grad_norm": 2.824199914932251, "learning_rate": 4.8112023831016434e-05, "loss": 4.5283, "step": 4184 }, { "epoch": 0.03777757717999639, "grad_norm": 5.796605587005615, "learning_rate": 4.811157248600831e-05, "loss": 4.7192, "step": 4185 }, { "epoch": 0.037786604080158874, "grad_norm": 2.6043925285339355, "learning_rate": 4.811112114100018e-05, "loss": 5.4953, "step": 4186 }, { "epoch": 0.037795630980321356, "grad_norm": 3.345392942428589, "learning_rate": 4.811066979599206e-05, "loss": 5.635, "step": 4187 }, { "epoch": 0.037804657880483844, "grad_norm": 5.374942302703857, "learning_rate": 4.811021845098394e-05, "loss": 4.4445, "step": 4188 }, { "epoch": 0.037813684780646326, "grad_norm": 3.2097933292388916, "learning_rate": 4.810976710597581e-05, "loss": 5.0673, "step": 4189 }, { "epoch": 0.03782271168080881, "grad_norm": 3.4632437229156494, "learning_rate": 4.810931576096769e-05, "loss": 4.5094, "step": 4190 }, { "epoch": 0.037831738580971296, "grad_norm": 4.0514068603515625, "learning_rate": 4.810886441595956e-05, "loss": 4.4803, "step": 4191 }, { "epoch": 0.03784076548113378, "grad_norm": 4.3911943435668945, "learning_rate": 4.8108413070951434e-05, "loss": 3.327, "step": 4192 }, { "epoch": 0.037849792381296266, "grad_norm": 3.4021430015563965, "learning_rate": 4.8107961725943313e-05, "loss": 4.7268, "step": 4193 }, { "epoch": 0.03785881928145875, "grad_norm": 3.0726723670959473, "learning_rate": 4.8107510380935186e-05, "loss": 5.0631, "step": 4194 }, { "epoch": 0.03786784618162123, "grad_norm": 3.8959598541259766, "learning_rate": 4.8107059035927065e-05, "loss": 4.2944, "step": 4195 }, { "epoch": 0.03787687308178372, "grad_norm": 3.5334560871124268, "learning_rate": 4.8106607690918944e-05, "loss": 4.2851, "step": 4196 }, { "epoch": 0.0378858999819462, "grad_norm": 3.2873265743255615, "learning_rate": 4.810615634591082e-05, "loss": 5.3706, "step": 4197 }, { "epoch": 0.03789492688210869, "grad_norm": 2.820150136947632, "learning_rate": 4.810570500090269e-05, "loss": 5.0244, "step": 4198 }, { "epoch": 0.03790395378227117, "grad_norm": 4.000860214233398, "learning_rate": 4.810525365589457e-05, "loss": 4.4317, "step": 4199 }, { "epoch": 0.03791298068243365, "grad_norm": 3.806933879852295, "learning_rate": 4.810480231088644e-05, "loss": 4.4484, "step": 4200 }, { "epoch": 0.03792200758259614, "grad_norm": 3.1892287731170654, "learning_rate": 4.810435096587832e-05, "loss": 4.7102, "step": 4201 }, { "epoch": 0.03793103448275862, "grad_norm": 3.5709331035614014, "learning_rate": 4.81038996208702e-05, "loss": 4.4977, "step": 4202 }, { "epoch": 0.0379400613829211, "grad_norm": 3.3960790634155273, "learning_rate": 4.810344827586207e-05, "loss": 4.5884, "step": 4203 }, { "epoch": 0.03794908828308359, "grad_norm": 3.878777265548706, "learning_rate": 4.810299693085395e-05, "loss": 4.7976, "step": 4204 }, { "epoch": 0.03795811518324607, "grad_norm": 4.062036991119385, "learning_rate": 4.8102545585845824e-05, "loss": 3.2294, "step": 4205 }, { "epoch": 0.03796714208340856, "grad_norm": 2.840100049972534, "learning_rate": 4.8102094240837696e-05, "loss": 3.7644, "step": 4206 }, { "epoch": 0.03797616898357104, "grad_norm": 3.247023820877075, "learning_rate": 4.8101642895829575e-05, "loss": 4.3138, "step": 4207 }, { "epoch": 0.03798519588373352, "grad_norm": 3.490159273147583, "learning_rate": 4.810119155082145e-05, "loss": 4.7426, "step": 4208 }, { "epoch": 0.03799422278389601, "grad_norm": 4.678101539611816, "learning_rate": 4.810074020581333e-05, "loss": 4.6931, "step": 4209 }, { "epoch": 0.03800324968405849, "grad_norm": 4.5002617835998535, "learning_rate": 4.8100288860805206e-05, "loss": 4.271, "step": 4210 }, { "epoch": 0.03801227658422098, "grad_norm": 3.20916748046875, "learning_rate": 4.809983751579707e-05, "loss": 3.769, "step": 4211 }, { "epoch": 0.03802130348438346, "grad_norm": 4.190361499786377, "learning_rate": 4.809938617078895e-05, "loss": 4.4138, "step": 4212 }, { "epoch": 0.038030330384545945, "grad_norm": 3.6568918228149414, "learning_rate": 4.809893482578083e-05, "loss": 4.2287, "step": 4213 }, { "epoch": 0.03803935728470843, "grad_norm": 3.3639392852783203, "learning_rate": 4.80984834807727e-05, "loss": 4.9972, "step": 4214 }, { "epoch": 0.038048384184870915, "grad_norm": 3.4655938148498535, "learning_rate": 4.809803213576458e-05, "loss": 5.1113, "step": 4215 }, { "epoch": 0.038057411085033396, "grad_norm": 5.32796573638916, "learning_rate": 4.8097580790756455e-05, "loss": 4.7354, "step": 4216 }, { "epoch": 0.038066437985195885, "grad_norm": 3.1464972496032715, "learning_rate": 4.8097129445748334e-05, "loss": 4.3207, "step": 4217 }, { "epoch": 0.038075464885358366, "grad_norm": 3.8956313133239746, "learning_rate": 4.8096678100740206e-05, "loss": 5.2377, "step": 4218 }, { "epoch": 0.038084491785520855, "grad_norm": 3.5521678924560547, "learning_rate": 4.8096226755732086e-05, "loss": 4.4359, "step": 4219 }, { "epoch": 0.038093518685683336, "grad_norm": 3.2156364917755127, "learning_rate": 4.809577541072396e-05, "loss": 4.6617, "step": 4220 }, { "epoch": 0.03810254558584582, "grad_norm": 4.992381572723389, "learning_rate": 4.809532406571584e-05, "loss": 3.4212, "step": 4221 }, { "epoch": 0.038111572486008306, "grad_norm": 3.979001522064209, "learning_rate": 4.809487272070771e-05, "loss": 4.6587, "step": 4222 }, { "epoch": 0.03812059938617079, "grad_norm": 5.5899152755737305, "learning_rate": 4.809442137569959e-05, "loss": 5.2793, "step": 4223 }, { "epoch": 0.038129626286333276, "grad_norm": 3.995206356048584, "learning_rate": 4.809397003069147e-05, "loss": 5.5743, "step": 4224 }, { "epoch": 0.03813865318649576, "grad_norm": 3.5677897930145264, "learning_rate": 4.8093518685683334e-05, "loss": 5.069, "step": 4225 }, { "epoch": 0.03814768008665824, "grad_norm": 2.7867531776428223, "learning_rate": 4.809306734067521e-05, "loss": 4.6645, "step": 4226 }, { "epoch": 0.03815670698682073, "grad_norm": 3.6163101196289062, "learning_rate": 4.809261599566709e-05, "loss": 4.5803, "step": 4227 }, { "epoch": 0.03816573388698321, "grad_norm": 3.631864309310913, "learning_rate": 4.8092164650658965e-05, "loss": 4.8139, "step": 4228 }, { "epoch": 0.03817476078714569, "grad_norm": 3.1159214973449707, "learning_rate": 4.8091713305650844e-05, "loss": 4.8494, "step": 4229 }, { "epoch": 0.03818378768730818, "grad_norm": 3.0306732654571533, "learning_rate": 4.809126196064272e-05, "loss": 4.741, "step": 4230 }, { "epoch": 0.03819281458747066, "grad_norm": 3.338296890258789, "learning_rate": 4.8090810615634596e-05, "loss": 5.0318, "step": 4231 }, { "epoch": 0.03820184148763315, "grad_norm": 4.218885898590088, "learning_rate": 4.809035927062647e-05, "loss": 3.5558, "step": 4232 }, { "epoch": 0.03821086838779563, "grad_norm": 3.890834093093872, "learning_rate": 4.808990792561834e-05, "loss": 4.4971, "step": 4233 }, { "epoch": 0.03821989528795811, "grad_norm": 3.800767421722412, "learning_rate": 4.808945658061022e-05, "loss": 4.7006, "step": 4234 }, { "epoch": 0.0382289221881206, "grad_norm": 7.447356700897217, "learning_rate": 4.80890052356021e-05, "loss": 4.5494, "step": 4235 }, { "epoch": 0.03823794908828308, "grad_norm": 4.764328479766846, "learning_rate": 4.808855389059397e-05, "loss": 3.8582, "step": 4236 }, { "epoch": 0.03824697598844557, "grad_norm": 3.010451316833496, "learning_rate": 4.808810254558585e-05, "loss": 5.2282, "step": 4237 }, { "epoch": 0.03825600288860805, "grad_norm": 2.7604832649230957, "learning_rate": 4.8087651200577723e-05, "loss": 4.9102, "step": 4238 }, { "epoch": 0.038265029788770534, "grad_norm": 3.320434093475342, "learning_rate": 4.8087199855569596e-05, "loss": 5.2048, "step": 4239 }, { "epoch": 0.03827405668893302, "grad_norm": 3.5953547954559326, "learning_rate": 4.8086748510561475e-05, "loss": 4.5971, "step": 4240 }, { "epoch": 0.038283083589095504, "grad_norm": 3.8808815479278564, "learning_rate": 4.8086297165553354e-05, "loss": 5.0142, "step": 4241 }, { "epoch": 0.038292110489257986, "grad_norm": 4.264155864715576, "learning_rate": 4.808584582054523e-05, "loss": 5.3597, "step": 4242 }, { "epoch": 0.038301137389420474, "grad_norm": 3.7712953090667725, "learning_rate": 4.8085394475537106e-05, "loss": 4.899, "step": 4243 }, { "epoch": 0.038310164289582956, "grad_norm": 4.019279956817627, "learning_rate": 4.808494313052898e-05, "loss": 4.986, "step": 4244 }, { "epoch": 0.038319191189745444, "grad_norm": 4.137259483337402, "learning_rate": 4.808449178552085e-05, "loss": 3.4053, "step": 4245 }, { "epoch": 0.038328218089907926, "grad_norm": 3.9429931640625, "learning_rate": 4.808404044051273e-05, "loss": 4.5416, "step": 4246 }, { "epoch": 0.03833724499007041, "grad_norm": 4.241916656494141, "learning_rate": 4.80835890955046e-05, "loss": 4.9958, "step": 4247 }, { "epoch": 0.038346271890232896, "grad_norm": 3.6669399738311768, "learning_rate": 4.808313775049648e-05, "loss": 4.6085, "step": 4248 }, { "epoch": 0.03835529879039538, "grad_norm": 3.275139808654785, "learning_rate": 4.808268640548836e-05, "loss": 4.2184, "step": 4249 }, { "epoch": 0.038364325690557866, "grad_norm": 3.4744677543640137, "learning_rate": 4.8082235060480234e-05, "loss": 4.8555, "step": 4250 }, { "epoch": 0.03837335259072035, "grad_norm": 4.455984592437744, "learning_rate": 4.808178371547211e-05, "loss": 3.7718, "step": 4251 }, { "epoch": 0.03838237949088283, "grad_norm": 3.8333027362823486, "learning_rate": 4.8081332370463985e-05, "loss": 5.0204, "step": 4252 }, { "epoch": 0.03839140639104532, "grad_norm": 4.439959526062012, "learning_rate": 4.808088102545586e-05, "loss": 3.7003, "step": 4253 }, { "epoch": 0.0384004332912078, "grad_norm": 5.506304740905762, "learning_rate": 4.808042968044774e-05, "loss": 4.5861, "step": 4254 }, { "epoch": 0.03840946019137028, "grad_norm": 5.0693864822387695, "learning_rate": 4.807997833543961e-05, "loss": 3.4516, "step": 4255 }, { "epoch": 0.03841848709153277, "grad_norm": 3.9225330352783203, "learning_rate": 4.807952699043149e-05, "loss": 4.8257, "step": 4256 }, { "epoch": 0.03842751399169525, "grad_norm": 4.822421550750732, "learning_rate": 4.807907564542337e-05, "loss": 5.4556, "step": 4257 }, { "epoch": 0.03843654089185774, "grad_norm": 3.0543901920318604, "learning_rate": 4.8078624300415234e-05, "loss": 5.2652, "step": 4258 }, { "epoch": 0.03844556779202022, "grad_norm": 3.0719621181488037, "learning_rate": 4.807817295540711e-05, "loss": 4.8347, "step": 4259 }, { "epoch": 0.0384545946921827, "grad_norm": 3.427523374557495, "learning_rate": 4.807772161039899e-05, "loss": 4.1471, "step": 4260 }, { "epoch": 0.03846362159234519, "grad_norm": 2.7752790451049805, "learning_rate": 4.8077270265390865e-05, "loss": 4.4384, "step": 4261 }, { "epoch": 0.03847264849250767, "grad_norm": 3.439445734024048, "learning_rate": 4.8076818920382744e-05, "loss": 4.9111, "step": 4262 }, { "epoch": 0.03848167539267016, "grad_norm": 2.9340121746063232, "learning_rate": 4.807636757537462e-05, "loss": 4.853, "step": 4263 }, { "epoch": 0.03849070229283264, "grad_norm": 2.653963565826416, "learning_rate": 4.8075916230366496e-05, "loss": 5.0158, "step": 4264 }, { "epoch": 0.03849972919299512, "grad_norm": 4.576376438140869, "learning_rate": 4.807546488535837e-05, "loss": 4.9925, "step": 4265 }, { "epoch": 0.03850875609315761, "grad_norm": 3.775204658508301, "learning_rate": 4.807501354035025e-05, "loss": 4.5672, "step": 4266 }, { "epoch": 0.03851778299332009, "grad_norm": 3.598236560821533, "learning_rate": 4.807456219534212e-05, "loss": 4.9845, "step": 4267 }, { "epoch": 0.038526809893482575, "grad_norm": 4.4263153076171875, "learning_rate": 4.8074110850334e-05, "loss": 5.0116, "step": 4268 }, { "epoch": 0.03853583679364506, "grad_norm": 3.335947275161743, "learning_rate": 4.807365950532587e-05, "loss": 4.7124, "step": 4269 }, { "epoch": 0.038544863693807545, "grad_norm": 3.3982629776000977, "learning_rate": 4.807320816031775e-05, "loss": 5.0047, "step": 4270 }, { "epoch": 0.03855389059397003, "grad_norm": 3.251779794692993, "learning_rate": 4.807275681530963e-05, "loss": 4.7323, "step": 4271 }, { "epoch": 0.038562917494132515, "grad_norm": 3.3910043239593506, "learning_rate": 4.8072305470301496e-05, "loss": 4.46, "step": 4272 }, { "epoch": 0.038571944394294996, "grad_norm": 3.38242506980896, "learning_rate": 4.8071854125293375e-05, "loss": 4.2616, "step": 4273 }, { "epoch": 0.038580971294457485, "grad_norm": 3.316502332687378, "learning_rate": 4.8071402780285254e-05, "loss": 4.9378, "step": 4274 }, { "epoch": 0.038589998194619966, "grad_norm": 3.7815897464752197, "learning_rate": 4.807095143527713e-05, "loss": 4.1148, "step": 4275 }, { "epoch": 0.038599025094782455, "grad_norm": 3.152045965194702, "learning_rate": 4.8070500090269006e-05, "loss": 4.4558, "step": 4276 }, { "epoch": 0.038608051994944936, "grad_norm": 4.017841815948486, "learning_rate": 4.807004874526088e-05, "loss": 4.6163, "step": 4277 }, { "epoch": 0.03861707889510742, "grad_norm": 3.2566041946411133, "learning_rate": 4.806959740025276e-05, "loss": 5.1568, "step": 4278 }, { "epoch": 0.038626105795269906, "grad_norm": 3.0597331523895264, "learning_rate": 4.806914605524463e-05, "loss": 4.2724, "step": 4279 }, { "epoch": 0.03863513269543239, "grad_norm": 3.4191131591796875, "learning_rate": 4.806869471023651e-05, "loss": 4.2072, "step": 4280 }, { "epoch": 0.03864415959559487, "grad_norm": 2.624556541442871, "learning_rate": 4.806824336522838e-05, "loss": 5.0751, "step": 4281 }, { "epoch": 0.03865318649575736, "grad_norm": 5.247943878173828, "learning_rate": 4.806779202022026e-05, "loss": 4.6898, "step": 4282 }, { "epoch": 0.03866221339591984, "grad_norm": 3.689422369003296, "learning_rate": 4.8067340675212133e-05, "loss": 4.299, "step": 4283 }, { "epoch": 0.03867124029608233, "grad_norm": 5.136826992034912, "learning_rate": 4.806688933020401e-05, "loss": 3.394, "step": 4284 }, { "epoch": 0.03868026719624481, "grad_norm": 2.90095853805542, "learning_rate": 4.8066437985195885e-05, "loss": 4.8532, "step": 4285 }, { "epoch": 0.03868929409640729, "grad_norm": 3.8737552165985107, "learning_rate": 4.806598664018776e-05, "loss": 4.9375, "step": 4286 }, { "epoch": 0.03869832099656978, "grad_norm": 3.906104564666748, "learning_rate": 4.806553529517964e-05, "loss": 4.7974, "step": 4287 }, { "epoch": 0.03870734789673226, "grad_norm": 3.344099283218384, "learning_rate": 4.8065083950171516e-05, "loss": 4.2149, "step": 4288 }, { "epoch": 0.03871637479689475, "grad_norm": 4.184983253479004, "learning_rate": 4.806463260516339e-05, "loss": 4.7434, "step": 4289 }, { "epoch": 0.03872540169705723, "grad_norm": 3.3102920055389404, "learning_rate": 4.806418126015527e-05, "loss": 5.3763, "step": 4290 }, { "epoch": 0.03873442859721971, "grad_norm": 3.226306200027466, "learning_rate": 4.806372991514714e-05, "loss": 4.1129, "step": 4291 }, { "epoch": 0.0387434554973822, "grad_norm": 4.953794956207275, "learning_rate": 4.806327857013901e-05, "loss": 4.7547, "step": 4292 }, { "epoch": 0.03875248239754468, "grad_norm": 4.809997081756592, "learning_rate": 4.806282722513089e-05, "loss": 3.9329, "step": 4293 }, { "epoch": 0.038761509297707164, "grad_norm": 2.785442352294922, "learning_rate": 4.8062375880122764e-05, "loss": 4.6794, "step": 4294 }, { "epoch": 0.03877053619786965, "grad_norm": 3.8308441638946533, "learning_rate": 4.8061924535114644e-05, "loss": 4.664, "step": 4295 }, { "epoch": 0.038779563098032134, "grad_norm": 3.639768600463867, "learning_rate": 4.806147319010652e-05, "loss": 4.7662, "step": 4296 }, { "epoch": 0.03878858999819462, "grad_norm": 2.6732709407806396, "learning_rate": 4.8061021845098395e-05, "loss": 4.8371, "step": 4297 }, { "epoch": 0.038797616898357104, "grad_norm": 3.558509588241577, "learning_rate": 4.8060570500090275e-05, "loss": 4.2767, "step": 4298 }, { "epoch": 0.038806643798519586, "grad_norm": 3.104475736618042, "learning_rate": 4.806011915508215e-05, "loss": 4.7494, "step": 4299 }, { "epoch": 0.038815670698682074, "grad_norm": 4.58516263961792, "learning_rate": 4.805966781007402e-05, "loss": 5.0955, "step": 4300 }, { "epoch": 0.038824697598844556, "grad_norm": 3.1816368103027344, "learning_rate": 4.80592164650659e-05, "loss": 4.5996, "step": 4301 }, { "epoch": 0.038833724499007044, "grad_norm": 3.9571189880371094, "learning_rate": 4.805876512005778e-05, "loss": 4.49, "step": 4302 }, { "epoch": 0.038842751399169526, "grad_norm": 6.225006580352783, "learning_rate": 4.805831377504965e-05, "loss": 4.6109, "step": 4303 }, { "epoch": 0.03885177829933201, "grad_norm": 3.4959182739257812, "learning_rate": 4.805786243004153e-05, "loss": 4.7183, "step": 4304 }, { "epoch": 0.038860805199494496, "grad_norm": 4.323953151702881, "learning_rate": 4.80574110850334e-05, "loss": 5.2067, "step": 4305 }, { "epoch": 0.03886983209965698, "grad_norm": 3.838369846343994, "learning_rate": 4.8056959740025275e-05, "loss": 4.1161, "step": 4306 }, { "epoch": 0.03887885899981946, "grad_norm": 3.1593613624572754, "learning_rate": 4.8056508395017154e-05, "loss": 5.4552, "step": 4307 }, { "epoch": 0.03888788589998195, "grad_norm": 3.5478320121765137, "learning_rate": 4.8056057050009026e-05, "loss": 5.2863, "step": 4308 }, { "epoch": 0.03889691280014443, "grad_norm": 3.0497052669525146, "learning_rate": 4.8055605705000906e-05, "loss": 4.6274, "step": 4309 }, { "epoch": 0.03890593970030692, "grad_norm": 3.6080570220947266, "learning_rate": 4.8055154359992785e-05, "loss": 4.6699, "step": 4310 }, { "epoch": 0.0389149666004694, "grad_norm": 3.867964267730713, "learning_rate": 4.805470301498466e-05, "loss": 4.457, "step": 4311 }, { "epoch": 0.03892399350063188, "grad_norm": 2.363957643508911, "learning_rate": 4.805425166997653e-05, "loss": 4.8591, "step": 4312 }, { "epoch": 0.03893302040079437, "grad_norm": 3.0105414390563965, "learning_rate": 4.805380032496841e-05, "loss": 4.9457, "step": 4313 }, { "epoch": 0.03894204730095685, "grad_norm": 2.722428798675537, "learning_rate": 4.805334897996028e-05, "loss": 4.9136, "step": 4314 }, { "epoch": 0.03895107420111934, "grad_norm": 3.223170757293701, "learning_rate": 4.805289763495216e-05, "loss": 4.6193, "step": 4315 }, { "epoch": 0.03896010110128182, "grad_norm": 9.251258850097656, "learning_rate": 4.805244628994403e-05, "loss": 3.9549, "step": 4316 }, { "epoch": 0.0389691280014443, "grad_norm": 3.939375877380371, "learning_rate": 4.805199494493591e-05, "loss": 4.4499, "step": 4317 }, { "epoch": 0.03897815490160679, "grad_norm": 4.083379745483398, "learning_rate": 4.805154359992779e-05, "loss": 4.8989, "step": 4318 }, { "epoch": 0.03898718180176927, "grad_norm": 5.672403335571289, "learning_rate": 4.805109225491966e-05, "loss": 4.9372, "step": 4319 }, { "epoch": 0.03899620870193175, "grad_norm": 3.6248061656951904, "learning_rate": 4.805064090991154e-05, "loss": 4.4332, "step": 4320 }, { "epoch": 0.03900523560209424, "grad_norm": 3.3805041313171387, "learning_rate": 4.8050189564903416e-05, "loss": 4.9954, "step": 4321 }, { "epoch": 0.03901426250225672, "grad_norm": 3.6153295040130615, "learning_rate": 4.804973821989529e-05, "loss": 4.8901, "step": 4322 }, { "epoch": 0.03902328940241921, "grad_norm": 3.2133920192718506, "learning_rate": 4.804928687488717e-05, "loss": 4.9144, "step": 4323 }, { "epoch": 0.03903231630258169, "grad_norm": 3.999835968017578, "learning_rate": 4.804883552987905e-05, "loss": 4.4873, "step": 4324 }, { "epoch": 0.039041343202744175, "grad_norm": 3.4907066822052, "learning_rate": 4.804838418487091e-05, "loss": 5.0649, "step": 4325 }, { "epoch": 0.03905037010290666, "grad_norm": 3.703312873840332, "learning_rate": 4.804793283986279e-05, "loss": 4.9467, "step": 4326 }, { "epoch": 0.039059397003069145, "grad_norm": 4.16528844833374, "learning_rate": 4.804748149485467e-05, "loss": 5.0052, "step": 4327 }, { "epoch": 0.03906842390323163, "grad_norm": 4.488832950592041, "learning_rate": 4.8047030149846543e-05, "loss": 4.8123, "step": 4328 }, { "epoch": 0.039077450803394115, "grad_norm": 3.353274345397949, "learning_rate": 4.804657880483842e-05, "loss": 5.0383, "step": 4329 }, { "epoch": 0.039086477703556596, "grad_norm": 3.187566041946411, "learning_rate": 4.8046127459830295e-05, "loss": 4.9411, "step": 4330 }, { "epoch": 0.039095504603719085, "grad_norm": 3.4960224628448486, "learning_rate": 4.8045676114822174e-05, "loss": 4.6955, "step": 4331 }, { "epoch": 0.039104531503881566, "grad_norm": 4.80142879486084, "learning_rate": 4.804522476981405e-05, "loss": 5.2887, "step": 4332 }, { "epoch": 0.03911355840404405, "grad_norm": 3.461576223373413, "learning_rate": 4.804477342480592e-05, "loss": 3.6293, "step": 4333 }, { "epoch": 0.039122585304206536, "grad_norm": 4.963977336883545, "learning_rate": 4.80443220797978e-05, "loss": 3.9297, "step": 4334 }, { "epoch": 0.03913161220436902, "grad_norm": 3.7416293621063232, "learning_rate": 4.804387073478968e-05, "loss": 4.026, "step": 4335 }, { "epoch": 0.039140639104531506, "grad_norm": 4.6775007247924805, "learning_rate": 4.804341938978155e-05, "loss": 4.5773, "step": 4336 }, { "epoch": 0.03914966600469399, "grad_norm": 5.0157904624938965, "learning_rate": 4.804296804477343e-05, "loss": 5.3079, "step": 4337 }, { "epoch": 0.03915869290485647, "grad_norm": 8.350225448608398, "learning_rate": 4.80425166997653e-05, "loss": 4.8209, "step": 4338 }, { "epoch": 0.03916771980501896, "grad_norm": 3.3143246173858643, "learning_rate": 4.8042065354757174e-05, "loss": 4.7259, "step": 4339 }, { "epoch": 0.03917674670518144, "grad_norm": 4.066913604736328, "learning_rate": 4.8041614009749054e-05, "loss": 4.3349, "step": 4340 }, { "epoch": 0.03918577360534393, "grad_norm": 3.884875535964966, "learning_rate": 4.804116266474093e-05, "loss": 5.125, "step": 4341 }, { "epoch": 0.03919480050550641, "grad_norm": 3.374936819076538, "learning_rate": 4.8040711319732805e-05, "loss": 5.2749, "step": 4342 }, { "epoch": 0.03920382740566889, "grad_norm": 3.560349941253662, "learning_rate": 4.8040259974724685e-05, "loss": 5.0569, "step": 4343 }, { "epoch": 0.03921285430583138, "grad_norm": 6.060695648193359, "learning_rate": 4.803980862971656e-05, "loss": 4.6965, "step": 4344 }, { "epoch": 0.03922188120599386, "grad_norm": 3.9204905033111572, "learning_rate": 4.8039357284708436e-05, "loss": 4.6518, "step": 4345 }, { "epoch": 0.03923090810615634, "grad_norm": 4.127460479736328, "learning_rate": 4.803890593970031e-05, "loss": 5.0555, "step": 4346 }, { "epoch": 0.03923993500631883, "grad_norm": 3.3636984825134277, "learning_rate": 4.803845459469218e-05, "loss": 4.3084, "step": 4347 }, { "epoch": 0.03924896190648131, "grad_norm": 3.734534502029419, "learning_rate": 4.803800324968406e-05, "loss": 5.1294, "step": 4348 }, { "epoch": 0.0392579888066438, "grad_norm": 3.5224456787109375, "learning_rate": 4.803755190467594e-05, "loss": 4.7031, "step": 4349 }, { "epoch": 0.03926701570680628, "grad_norm": 3.132279872894287, "learning_rate": 4.803710055966781e-05, "loss": 3.5147, "step": 4350 }, { "epoch": 0.039276042606968764, "grad_norm": 4.16813325881958, "learning_rate": 4.803664921465969e-05, "loss": 5.064, "step": 4351 }, { "epoch": 0.03928506950713125, "grad_norm": 3.6316094398498535, "learning_rate": 4.8036197869651564e-05, "loss": 4.5924, "step": 4352 }, { "epoch": 0.039294096407293734, "grad_norm": 3.4581189155578613, "learning_rate": 4.8035746524643436e-05, "loss": 4.4822, "step": 4353 }, { "epoch": 0.03930312330745622, "grad_norm": 3.5681381225585938, "learning_rate": 4.8035295179635316e-05, "loss": 4.4534, "step": 4354 }, { "epoch": 0.039312150207618704, "grad_norm": 2.9731104373931885, "learning_rate": 4.803484383462719e-05, "loss": 4.8822, "step": 4355 }, { "epoch": 0.039321177107781186, "grad_norm": 2.8463289737701416, "learning_rate": 4.803439248961907e-05, "loss": 5.0116, "step": 4356 }, { "epoch": 0.039330204007943674, "grad_norm": 4.26508903503418, "learning_rate": 4.8033941144610947e-05, "loss": 4.6988, "step": 4357 }, { "epoch": 0.039339230908106156, "grad_norm": 3.788663864135742, "learning_rate": 4.803348979960282e-05, "loss": 5.0665, "step": 4358 }, { "epoch": 0.03934825780826864, "grad_norm": 3.8999252319335938, "learning_rate": 4.803303845459469e-05, "loss": 4.9997, "step": 4359 }, { "epoch": 0.039357284708431126, "grad_norm": 3.942903757095337, "learning_rate": 4.803258710958657e-05, "loss": 3.6339, "step": 4360 }, { "epoch": 0.03936631160859361, "grad_norm": 3.483783483505249, "learning_rate": 4.803213576457844e-05, "loss": 4.331, "step": 4361 }, { "epoch": 0.039375338508756096, "grad_norm": 7.740213871002197, "learning_rate": 4.803168441957032e-05, "loss": 3.735, "step": 4362 }, { "epoch": 0.03938436540891858, "grad_norm": 6.5436296463012695, "learning_rate": 4.80312330745622e-05, "loss": 4.8325, "step": 4363 }, { "epoch": 0.03939339230908106, "grad_norm": 5.314933776855469, "learning_rate": 4.8030781729554074e-05, "loss": 4.5042, "step": 4364 }, { "epoch": 0.03940241920924355, "grad_norm": 4.093801498413086, "learning_rate": 4.803033038454595e-05, "loss": 4.9126, "step": 4365 }, { "epoch": 0.03941144610940603, "grad_norm": 3.1891868114471436, "learning_rate": 4.8029879039537826e-05, "loss": 3.8521, "step": 4366 }, { "epoch": 0.03942047300956852, "grad_norm": 3.4687609672546387, "learning_rate": 4.80294276945297e-05, "loss": 4.2174, "step": 4367 }, { "epoch": 0.039429499909731, "grad_norm": 3.924067974090576, "learning_rate": 4.802897634952158e-05, "loss": 5.0825, "step": 4368 }, { "epoch": 0.03943852680989348, "grad_norm": 3.821382761001587, "learning_rate": 4.802852500451345e-05, "loss": 4.349, "step": 4369 }, { "epoch": 0.03944755371005597, "grad_norm": 4.187709331512451, "learning_rate": 4.802807365950533e-05, "loss": 4.637, "step": 4370 }, { "epoch": 0.03945658061021845, "grad_norm": 4.931010723114014, "learning_rate": 4.802762231449721e-05, "loss": 4.7257, "step": 4371 }, { "epoch": 0.03946560751038093, "grad_norm": 3.18410325050354, "learning_rate": 4.8027170969489074e-05, "loss": 5.2796, "step": 4372 }, { "epoch": 0.03947463441054342, "grad_norm": 3.7015230655670166, "learning_rate": 4.8026719624480953e-05, "loss": 4.6771, "step": 4373 }, { "epoch": 0.0394836613107059, "grad_norm": 3.8634722232818604, "learning_rate": 4.802626827947283e-05, "loss": 5.0219, "step": 4374 }, { "epoch": 0.03949268821086839, "grad_norm": 3.219417095184326, "learning_rate": 4.8025816934464705e-05, "loss": 5.2328, "step": 4375 }, { "epoch": 0.03950171511103087, "grad_norm": 3.0697455406188965, "learning_rate": 4.8025365589456584e-05, "loss": 4.6542, "step": 4376 }, { "epoch": 0.03951074201119335, "grad_norm": 3.0421011447906494, "learning_rate": 4.802491424444846e-05, "loss": 5.3358, "step": 4377 }, { "epoch": 0.03951976891135584, "grad_norm": 3.4699134826660156, "learning_rate": 4.8024462899440336e-05, "loss": 5.5169, "step": 4378 }, { "epoch": 0.03952879581151832, "grad_norm": 3.6539058685302734, "learning_rate": 4.802401155443221e-05, "loss": 4.8384, "step": 4379 }, { "epoch": 0.03953782271168081, "grad_norm": 5.873090744018555, "learning_rate": 4.802356020942408e-05, "loss": 4.7872, "step": 4380 }, { "epoch": 0.03954684961184329, "grad_norm": 4.219521999359131, "learning_rate": 4.802310886441596e-05, "loss": 5.05, "step": 4381 }, { "epoch": 0.039555876512005775, "grad_norm": 2.86472749710083, "learning_rate": 4.802265751940784e-05, "loss": 4.8117, "step": 4382 }, { "epoch": 0.03956490341216826, "grad_norm": 3.844825029373169, "learning_rate": 4.802220617439971e-05, "loss": 4.5429, "step": 4383 }, { "epoch": 0.039573930312330745, "grad_norm": 10.118697166442871, "learning_rate": 4.802175482939159e-05, "loss": 4.2765, "step": 4384 }, { "epoch": 0.03958295721249323, "grad_norm": 3.517155885696411, "learning_rate": 4.802130348438347e-05, "loss": 4.7569, "step": 4385 }, { "epoch": 0.039591984112655715, "grad_norm": 3.6600406169891357, "learning_rate": 4.8020852139375336e-05, "loss": 4.3093, "step": 4386 }, { "epoch": 0.039601011012818196, "grad_norm": 4.430233001708984, "learning_rate": 4.8020400794367215e-05, "loss": 3.7654, "step": 4387 }, { "epoch": 0.039610037912980685, "grad_norm": 3.592712879180908, "learning_rate": 4.8019949449359095e-05, "loss": 4.4278, "step": 4388 }, { "epoch": 0.039619064813143166, "grad_norm": 2.542684316635132, "learning_rate": 4.801949810435097e-05, "loss": 4.782, "step": 4389 }, { "epoch": 0.03962809171330565, "grad_norm": 4.361973762512207, "learning_rate": 4.8019046759342846e-05, "loss": 4.9773, "step": 4390 }, { "epoch": 0.039637118613468136, "grad_norm": 4.401609420776367, "learning_rate": 4.801859541433472e-05, "loss": 4.8228, "step": 4391 }, { "epoch": 0.03964614551363062, "grad_norm": 3.242934465408325, "learning_rate": 4.80181440693266e-05, "loss": 4.8263, "step": 4392 }, { "epoch": 0.039655172413793106, "grad_norm": 4.091556072235107, "learning_rate": 4.801769272431847e-05, "loss": 4.9449, "step": 4393 }, { "epoch": 0.03966419931395559, "grad_norm": 3.075833320617676, "learning_rate": 4.801724137931034e-05, "loss": 3.769, "step": 4394 }, { "epoch": 0.03967322621411807, "grad_norm": 3.7285561561584473, "learning_rate": 4.801679003430222e-05, "loss": 4.6422, "step": 4395 }, { "epoch": 0.03968225311428056, "grad_norm": 3.3287694454193115, "learning_rate": 4.80163386892941e-05, "loss": 4.5431, "step": 4396 }, { "epoch": 0.03969128001444304, "grad_norm": 5.195852279663086, "learning_rate": 4.8015887344285974e-05, "loss": 5.3168, "step": 4397 }, { "epoch": 0.03970030691460553, "grad_norm": 4.808011054992676, "learning_rate": 4.801543599927785e-05, "loss": 4.9039, "step": 4398 }, { "epoch": 0.03970933381476801, "grad_norm": 3.3488075733184814, "learning_rate": 4.8014984654269726e-05, "loss": 4.9908, "step": 4399 }, { "epoch": 0.03971836071493049, "grad_norm": 2.636690855026245, "learning_rate": 4.80145333092616e-05, "loss": 5.0986, "step": 4400 }, { "epoch": 0.03972738761509298, "grad_norm": 3.2481024265289307, "learning_rate": 4.801408196425348e-05, "loss": 4.6955, "step": 4401 }, { "epoch": 0.03973641451525546, "grad_norm": 5.982296466827393, "learning_rate": 4.8013630619245357e-05, "loss": 4.3757, "step": 4402 }, { "epoch": 0.03974544141541794, "grad_norm": 4.579555511474609, "learning_rate": 4.801317927423723e-05, "loss": 4.1315, "step": 4403 }, { "epoch": 0.03975446831558043, "grad_norm": 3.6733806133270264, "learning_rate": 4.801272792922911e-05, "loss": 4.5293, "step": 4404 }, { "epoch": 0.03976349521574291, "grad_norm": 3.9299023151397705, "learning_rate": 4.801227658422098e-05, "loss": 4.9515, "step": 4405 }, { "epoch": 0.0397725221159054, "grad_norm": 7.488593578338623, "learning_rate": 4.801182523921285e-05, "loss": 5.0811, "step": 4406 }, { "epoch": 0.03978154901606788, "grad_norm": 2.4752449989318848, "learning_rate": 4.801137389420473e-05, "loss": 4.914, "step": 4407 }, { "epoch": 0.039790575916230364, "grad_norm": 4.0495500564575195, "learning_rate": 4.8010922549196605e-05, "loss": 3.5157, "step": 4408 }, { "epoch": 0.03979960281639285, "grad_norm": 3.5661966800689697, "learning_rate": 4.8010471204188484e-05, "loss": 4.6618, "step": 4409 }, { "epoch": 0.039808629716555334, "grad_norm": 3.8008668422698975, "learning_rate": 4.801001985918036e-05, "loss": 4.8362, "step": 4410 }, { "epoch": 0.03981765661671782, "grad_norm": 6.158860206604004, "learning_rate": 4.8009568514172236e-05, "loss": 4.9791, "step": 4411 }, { "epoch": 0.039826683516880304, "grad_norm": 5.072203159332275, "learning_rate": 4.8009117169164115e-05, "loss": 5.4529, "step": 4412 }, { "epoch": 0.039835710417042786, "grad_norm": 3.9058077335357666, "learning_rate": 4.800866582415599e-05, "loss": 5.4944, "step": 4413 }, { "epoch": 0.039844737317205274, "grad_norm": 5.277607440948486, "learning_rate": 4.800821447914786e-05, "loss": 4.8094, "step": 4414 }, { "epoch": 0.039853764217367756, "grad_norm": 3.028406858444214, "learning_rate": 4.800776313413974e-05, "loss": 4.4005, "step": 4415 }, { "epoch": 0.03986279111753024, "grad_norm": 3.9474782943725586, "learning_rate": 4.800731178913161e-05, "loss": 4.2912, "step": 4416 }, { "epoch": 0.039871818017692726, "grad_norm": 2.924403190612793, "learning_rate": 4.800686044412349e-05, "loss": 4.888, "step": 4417 }, { "epoch": 0.03988084491785521, "grad_norm": 3.7609965801239014, "learning_rate": 4.800640909911537e-05, "loss": 5.1069, "step": 4418 }, { "epoch": 0.039889871818017696, "grad_norm": 2.9582788944244385, "learning_rate": 4.8005957754107236e-05, "loss": 4.7268, "step": 4419 }, { "epoch": 0.03989889871818018, "grad_norm": 4.266481876373291, "learning_rate": 4.8005506409099115e-05, "loss": 5.0896, "step": 4420 }, { "epoch": 0.03990792561834266, "grad_norm": 2.81760311126709, "learning_rate": 4.8005055064090994e-05, "loss": 4.3896, "step": 4421 }, { "epoch": 0.03991695251850515, "grad_norm": 3.5440003871917725, "learning_rate": 4.800460371908287e-05, "loss": 4.9508, "step": 4422 }, { "epoch": 0.03992597941866763, "grad_norm": 3.6792657375335693, "learning_rate": 4.8004152374074746e-05, "loss": 4.4855, "step": 4423 }, { "epoch": 0.03993500631883012, "grad_norm": 3.514504909515381, "learning_rate": 4.8003701029066625e-05, "loss": 4.3108, "step": 4424 }, { "epoch": 0.0399440332189926, "grad_norm": 4.191616058349609, "learning_rate": 4.80032496840585e-05, "loss": 5.2678, "step": 4425 }, { "epoch": 0.03995306011915508, "grad_norm": 4.412853717803955, "learning_rate": 4.800279833905037e-05, "loss": 5.7788, "step": 4426 }, { "epoch": 0.03996208701931757, "grad_norm": 3.747305154800415, "learning_rate": 4.800234699404225e-05, "loss": 5.2932, "step": 4427 }, { "epoch": 0.03997111391948005, "grad_norm": 3.7700629234313965, "learning_rate": 4.800189564903412e-05, "loss": 3.8261, "step": 4428 }, { "epoch": 0.03998014081964253, "grad_norm": 8.56903076171875, "learning_rate": 4.8001444304026e-05, "loss": 3.8108, "step": 4429 }, { "epoch": 0.03998916771980502, "grad_norm": 3.339695930480957, "learning_rate": 4.8000992959017874e-05, "loss": 4.9148, "step": 4430 }, { "epoch": 0.0399981946199675, "grad_norm": 6.1025004386901855, "learning_rate": 4.800054161400975e-05, "loss": 4.5624, "step": 4431 }, { "epoch": 0.04000722152012999, "grad_norm": 3.114454507827759, "learning_rate": 4.800009026900163e-05, "loss": 4.9012, "step": 4432 }, { "epoch": 0.04001624842029247, "grad_norm": 3.82395601272583, "learning_rate": 4.79996389239935e-05, "loss": 4.5274, "step": 4433 }, { "epoch": 0.04002527532045495, "grad_norm": 3.5553078651428223, "learning_rate": 4.799918757898538e-05, "loss": 4.1133, "step": 4434 }, { "epoch": 0.04003430222061744, "grad_norm": 4.371072769165039, "learning_rate": 4.7998736233977256e-05, "loss": 5.0522, "step": 4435 }, { "epoch": 0.04004332912077992, "grad_norm": 5.339208602905273, "learning_rate": 4.799828488896913e-05, "loss": 4.6917, "step": 4436 }, { "epoch": 0.04005235602094241, "grad_norm": 2.580592632293701, "learning_rate": 4.799783354396101e-05, "loss": 4.5343, "step": 4437 }, { "epoch": 0.04006138292110489, "grad_norm": 4.2719407081604, "learning_rate": 4.799738219895288e-05, "loss": 4.965, "step": 4438 }, { "epoch": 0.040070409821267375, "grad_norm": 4.16994047164917, "learning_rate": 4.799693085394476e-05, "loss": 4.4031, "step": 4439 }, { "epoch": 0.04007943672142986, "grad_norm": 3.5037944316864014, "learning_rate": 4.799647950893663e-05, "loss": 4.6495, "step": 4440 }, { "epoch": 0.040088463621592345, "grad_norm": 3.128795623779297, "learning_rate": 4.799602816392851e-05, "loss": 4.124, "step": 4441 }, { "epoch": 0.040097490521754826, "grad_norm": 3.5779213905334473, "learning_rate": 4.7995576818920384e-05, "loss": 4.1043, "step": 4442 }, { "epoch": 0.040106517421917315, "grad_norm": 3.8001461029052734, "learning_rate": 4.799512547391226e-05, "loss": 4.9272, "step": 4443 }, { "epoch": 0.040115544322079796, "grad_norm": 3.6154086589813232, "learning_rate": 4.7994674128904136e-05, "loss": 4.5617, "step": 4444 }, { "epoch": 0.040124571222242285, "grad_norm": 4.121683597564697, "learning_rate": 4.7994222783896015e-05, "loss": 5.4349, "step": 4445 }, { "epoch": 0.040133598122404766, "grad_norm": 5.529490947723389, "learning_rate": 4.799377143888789e-05, "loss": 4.5513, "step": 4446 }, { "epoch": 0.04014262502256725, "grad_norm": 3.3819775581359863, "learning_rate": 4.799332009387976e-05, "loss": 4.6273, "step": 4447 }, { "epoch": 0.040151651922729736, "grad_norm": 3.222831964492798, "learning_rate": 4.799286874887164e-05, "loss": 5.1025, "step": 4448 }, { "epoch": 0.04016067882289222, "grad_norm": 4.075422286987305, "learning_rate": 4.799241740386352e-05, "loss": 3.4199, "step": 4449 }, { "epoch": 0.040169705723054706, "grad_norm": 4.073907375335693, "learning_rate": 4.799196605885539e-05, "loss": 4.9058, "step": 4450 }, { "epoch": 0.04017873262321719, "grad_norm": 3.2281606197357178, "learning_rate": 4.799151471384727e-05, "loss": 4.7809, "step": 4451 }, { "epoch": 0.04018775952337967, "grad_norm": 5.291886329650879, "learning_rate": 4.799106336883914e-05, "loss": 4.8443, "step": 4452 }, { "epoch": 0.04019678642354216, "grad_norm": 3.383113384246826, "learning_rate": 4.7990612023831015e-05, "loss": 4.8648, "step": 4453 }, { "epoch": 0.04020581332370464, "grad_norm": 3.8943116664886475, "learning_rate": 4.7990160678822894e-05, "loss": 4.5367, "step": 4454 }, { "epoch": 0.04021484022386712, "grad_norm": 3.75034236907959, "learning_rate": 4.7989709333814767e-05, "loss": 4.7526, "step": 4455 }, { "epoch": 0.04022386712402961, "grad_norm": 5.060700416564941, "learning_rate": 4.7989257988806646e-05, "loss": 5.0863, "step": 4456 }, { "epoch": 0.04023289402419209, "grad_norm": 4.276803016662598, "learning_rate": 4.7988806643798525e-05, "loss": 4.28, "step": 4457 }, { "epoch": 0.04024192092435458, "grad_norm": 5.235600471496582, "learning_rate": 4.79883552987904e-05, "loss": 5.064, "step": 4458 }, { "epoch": 0.04025094782451706, "grad_norm": 6.861544609069824, "learning_rate": 4.798790395378228e-05, "loss": 3.8913, "step": 4459 }, { "epoch": 0.04025997472467954, "grad_norm": 3.9178266525268555, "learning_rate": 4.798745260877415e-05, "loss": 3.9247, "step": 4460 }, { "epoch": 0.04026900162484203, "grad_norm": 4.89644193649292, "learning_rate": 4.798700126376602e-05, "loss": 2.9874, "step": 4461 }, { "epoch": 0.04027802852500451, "grad_norm": 2.8194127082824707, "learning_rate": 4.79865499187579e-05, "loss": 4.8685, "step": 4462 }, { "epoch": 0.040287055425167, "grad_norm": 3.9596798419952393, "learning_rate": 4.798609857374978e-05, "loss": 4.5591, "step": 4463 }, { "epoch": 0.04029608232532948, "grad_norm": 2.7937545776367188, "learning_rate": 4.798564722874165e-05, "loss": 4.1324, "step": 4464 }, { "epoch": 0.040305109225491964, "grad_norm": 3.528787851333618, "learning_rate": 4.798519588373353e-05, "loss": 4.5006, "step": 4465 }, { "epoch": 0.04031413612565445, "grad_norm": 2.544114828109741, "learning_rate": 4.7984744538725404e-05, "loss": 4.9049, "step": 4466 }, { "epoch": 0.040323163025816934, "grad_norm": 4.359383583068848, "learning_rate": 4.798429319371728e-05, "loss": 3.9957, "step": 4467 }, { "epoch": 0.040332189925979416, "grad_norm": 3.880326271057129, "learning_rate": 4.7983841848709156e-05, "loss": 4.5815, "step": 4468 }, { "epoch": 0.040341216826141904, "grad_norm": 3.157484531402588, "learning_rate": 4.798339050370103e-05, "loss": 4.8379, "step": 4469 }, { "epoch": 0.040350243726304386, "grad_norm": 4.848077297210693, "learning_rate": 4.798293915869291e-05, "loss": 3.7717, "step": 4470 }, { "epoch": 0.040359270626466874, "grad_norm": 4.455520153045654, "learning_rate": 4.798248781368479e-05, "loss": 4.8026, "step": 4471 }, { "epoch": 0.040368297526629356, "grad_norm": 3.2079484462738037, "learning_rate": 4.798203646867666e-05, "loss": 4.1614, "step": 4472 }, { "epoch": 0.04037732442679184, "grad_norm": 3.6820061206817627, "learning_rate": 4.798158512366853e-05, "loss": 4.8301, "step": 4473 }, { "epoch": 0.040386351326954326, "grad_norm": 4.336012363433838, "learning_rate": 4.798113377866041e-05, "loss": 4.703, "step": 4474 }, { "epoch": 0.04039537822711681, "grad_norm": 3.0186996459960938, "learning_rate": 4.7980682433652284e-05, "loss": 4.8657, "step": 4475 }, { "epoch": 0.040404405127279296, "grad_norm": 4.0992255210876465, "learning_rate": 4.798023108864416e-05, "loss": 4.3485, "step": 4476 }, { "epoch": 0.04041343202744178, "grad_norm": 3.1427695751190186, "learning_rate": 4.7979779743636035e-05, "loss": 4.7979, "step": 4477 }, { "epoch": 0.04042245892760426, "grad_norm": 4.105617523193359, "learning_rate": 4.7979328398627915e-05, "loss": 4.9772, "step": 4478 }, { "epoch": 0.04043148582776675, "grad_norm": 3.0291049480438232, "learning_rate": 4.7978877053619794e-05, "loss": 4.8065, "step": 4479 }, { "epoch": 0.04044051272792923, "grad_norm": 3.063075065612793, "learning_rate": 4.797842570861166e-05, "loss": 4.9231, "step": 4480 }, { "epoch": 0.04044953962809171, "grad_norm": 3.04530930519104, "learning_rate": 4.797797436360354e-05, "loss": 4.5717, "step": 4481 }, { "epoch": 0.0404585665282542, "grad_norm": 3.798997402191162, "learning_rate": 4.797752301859542e-05, "loss": 4.6826, "step": 4482 }, { "epoch": 0.04046759342841668, "grad_norm": 2.8744847774505615, "learning_rate": 4.797707167358729e-05, "loss": 5.3804, "step": 4483 }, { "epoch": 0.04047662032857917, "grad_norm": 3.253192186355591, "learning_rate": 4.797662032857917e-05, "loss": 4.9598, "step": 4484 }, { "epoch": 0.04048564722874165, "grad_norm": 4.311751365661621, "learning_rate": 4.797616898357105e-05, "loss": 4.7689, "step": 4485 }, { "epoch": 0.04049467412890413, "grad_norm": 2.4765982627868652, "learning_rate": 4.797571763856292e-05, "loss": 5.3992, "step": 4486 }, { "epoch": 0.04050370102906662, "grad_norm": 4.133053302764893, "learning_rate": 4.7975266293554794e-05, "loss": 4.6661, "step": 4487 }, { "epoch": 0.0405127279292291, "grad_norm": 3.3942248821258545, "learning_rate": 4.797481494854667e-05, "loss": 5.1294, "step": 4488 }, { "epoch": 0.04052175482939159, "grad_norm": 4.638483047485352, "learning_rate": 4.7974363603538546e-05, "loss": 4.7577, "step": 4489 }, { "epoch": 0.04053078172955407, "grad_norm": 3.7001614570617676, "learning_rate": 4.7973912258530425e-05, "loss": 4.9888, "step": 4490 }, { "epoch": 0.04053980862971655, "grad_norm": 3.5787599086761475, "learning_rate": 4.79734609135223e-05, "loss": 4.9213, "step": 4491 }, { "epoch": 0.04054883552987904, "grad_norm": 3.329092025756836, "learning_rate": 4.7973009568514177e-05, "loss": 4.7054, "step": 4492 }, { "epoch": 0.04055786243004152, "grad_norm": 3.380418539047241, "learning_rate": 4.797255822350605e-05, "loss": 4.6125, "step": 4493 }, { "epoch": 0.040566889330204005, "grad_norm": 4.432130813598633, "learning_rate": 4.797210687849792e-05, "loss": 4.9121, "step": 4494 }, { "epoch": 0.04057591623036649, "grad_norm": 2.8114407062530518, "learning_rate": 4.79716555334898e-05, "loss": 4.6562, "step": 4495 }, { "epoch": 0.040584943130528975, "grad_norm": 3.570687770843506, "learning_rate": 4.797120418848168e-05, "loss": 4.6416, "step": 4496 }, { "epoch": 0.04059397003069146, "grad_norm": 3.7842679023742676, "learning_rate": 4.797075284347355e-05, "loss": 4.6991, "step": 4497 }, { "epoch": 0.040602996930853945, "grad_norm": 2.3937759399414062, "learning_rate": 4.797030149846543e-05, "loss": 4.7025, "step": 4498 }, { "epoch": 0.040612023831016426, "grad_norm": 3.2268593311309814, "learning_rate": 4.7969850153457304e-05, "loss": 4.5265, "step": 4499 }, { "epoch": 0.040621050731178915, "grad_norm": 4.360003471374512, "learning_rate": 4.7969398808449177e-05, "loss": 4.6045, "step": 4500 }, { "epoch": 0.040630077631341396, "grad_norm": 3.8868391513824463, "learning_rate": 4.7968947463441056e-05, "loss": 4.9202, "step": 4501 }, { "epoch": 0.040639104531503885, "grad_norm": 3.2442104816436768, "learning_rate": 4.7968496118432935e-05, "loss": 5.1102, "step": 4502 }, { "epoch": 0.040648131431666366, "grad_norm": 3.498753547668457, "learning_rate": 4.796804477342481e-05, "loss": 4.0202, "step": 4503 }, { "epoch": 0.04065715833182885, "grad_norm": 3.482520341873169, "learning_rate": 4.796759342841669e-05, "loss": 4.1441, "step": 4504 }, { "epoch": 0.040666185231991336, "grad_norm": 3.3841896057128906, "learning_rate": 4.796714208340856e-05, "loss": 5.0344, "step": 4505 }, { "epoch": 0.04067521213215382, "grad_norm": 3.9162983894348145, "learning_rate": 4.796669073840044e-05, "loss": 3.7189, "step": 4506 }, { "epoch": 0.0406842390323163, "grad_norm": 3.929912805557251, "learning_rate": 4.796623939339231e-05, "loss": 4.4388, "step": 4507 }, { "epoch": 0.04069326593247879, "grad_norm": 3.148411512374878, "learning_rate": 4.7965788048384183e-05, "loss": 5.1608, "step": 4508 }, { "epoch": 0.04070229283264127, "grad_norm": 3.121504783630371, "learning_rate": 4.796533670337606e-05, "loss": 4.8782, "step": 4509 }, { "epoch": 0.04071131973280376, "grad_norm": 5.803809642791748, "learning_rate": 4.796488535836794e-05, "loss": 4.941, "step": 4510 }, { "epoch": 0.04072034663296624, "grad_norm": 6.326407432556152, "learning_rate": 4.7964434013359814e-05, "loss": 5.2056, "step": 4511 }, { "epoch": 0.04072937353312872, "grad_norm": 4.219106674194336, "learning_rate": 4.7963982668351694e-05, "loss": 4.4879, "step": 4512 }, { "epoch": 0.04073840043329121, "grad_norm": 3.524277448654175, "learning_rate": 4.7963531323343566e-05, "loss": 4.3983, "step": 4513 }, { "epoch": 0.04074742733345369, "grad_norm": 3.914003372192383, "learning_rate": 4.796307997833544e-05, "loss": 4.7183, "step": 4514 }, { "epoch": 0.04075645423361618, "grad_norm": 3.5177602767944336, "learning_rate": 4.796262863332732e-05, "loss": 4.5115, "step": 4515 }, { "epoch": 0.04076548113377866, "grad_norm": 4.259546279907227, "learning_rate": 4.796217728831919e-05, "loss": 4.8628, "step": 4516 }, { "epoch": 0.04077450803394114, "grad_norm": 3.856304168701172, "learning_rate": 4.796172594331107e-05, "loss": 5.1936, "step": 4517 }, { "epoch": 0.04078353493410363, "grad_norm": 3.908168077468872, "learning_rate": 4.796127459830295e-05, "loss": 5.2041, "step": 4518 }, { "epoch": 0.04079256183426611, "grad_norm": 2.7807374000549316, "learning_rate": 4.796082325329482e-05, "loss": 4.9428, "step": 4519 }, { "epoch": 0.040801588734428594, "grad_norm": 4.344576835632324, "learning_rate": 4.7960371908286694e-05, "loss": 4.8581, "step": 4520 }, { "epoch": 0.04081061563459108, "grad_norm": 3.897698402404785, "learning_rate": 4.795992056327857e-05, "loss": 4.4829, "step": 4521 }, { "epoch": 0.040819642534753564, "grad_norm": 3.853283643722534, "learning_rate": 4.7959469218270445e-05, "loss": 5.1492, "step": 4522 }, { "epoch": 0.04082866943491605, "grad_norm": 3.7209551334381104, "learning_rate": 4.7959017873262325e-05, "loss": 4.6016, "step": 4523 }, { "epoch": 0.040837696335078534, "grad_norm": 3.6675961017608643, "learning_rate": 4.7958566528254204e-05, "loss": 4.7011, "step": 4524 }, { "epoch": 0.040846723235241016, "grad_norm": 4.59360933303833, "learning_rate": 4.7958115183246076e-05, "loss": 4.6446, "step": 4525 }, { "epoch": 0.040855750135403504, "grad_norm": 4.465760231018066, "learning_rate": 4.7957663838237956e-05, "loss": 4.441, "step": 4526 }, { "epoch": 0.040864777035565986, "grad_norm": 3.3453965187072754, "learning_rate": 4.795721249322983e-05, "loss": 3.7962, "step": 4527 }, { "epoch": 0.040873803935728474, "grad_norm": 4.772029399871826, "learning_rate": 4.79567611482217e-05, "loss": 5.0632, "step": 4528 }, { "epoch": 0.040882830835890956, "grad_norm": 3.7192044258117676, "learning_rate": 4.795630980321358e-05, "loss": 5.4402, "step": 4529 }, { "epoch": 0.04089185773605344, "grad_norm": 3.6032867431640625, "learning_rate": 4.795585845820545e-05, "loss": 4.5801, "step": 4530 }, { "epoch": 0.040900884636215926, "grad_norm": 3.9101529121398926, "learning_rate": 4.795540711319733e-05, "loss": 4.4929, "step": 4531 }, { "epoch": 0.04090991153637841, "grad_norm": 3.7463722229003906, "learning_rate": 4.795495576818921e-05, "loss": 5.0477, "step": 4532 }, { "epoch": 0.04091893843654089, "grad_norm": 3.211094856262207, "learning_rate": 4.7954504423181076e-05, "loss": 4.7224, "step": 4533 }, { "epoch": 0.04092796533670338, "grad_norm": 4.227126598358154, "learning_rate": 4.7954053078172956e-05, "loss": 4.9335, "step": 4534 }, { "epoch": 0.04093699223686586, "grad_norm": 3.9352641105651855, "learning_rate": 4.7953601733164835e-05, "loss": 4.4941, "step": 4535 }, { "epoch": 0.04094601913702835, "grad_norm": 3.0335793495178223, "learning_rate": 4.795315038815671e-05, "loss": 4.7574, "step": 4536 }, { "epoch": 0.04095504603719083, "grad_norm": 2.6140732765197754, "learning_rate": 4.7952699043148587e-05, "loss": 3.6331, "step": 4537 }, { "epoch": 0.04096407293735331, "grad_norm": 3.093252658843994, "learning_rate": 4.795224769814046e-05, "loss": 4.5489, "step": 4538 }, { "epoch": 0.0409730998375158, "grad_norm": 2.97133731842041, "learning_rate": 4.795179635313234e-05, "loss": 5.1786, "step": 4539 }, { "epoch": 0.04098212673767828, "grad_norm": 3.0094430446624756, "learning_rate": 4.795134500812421e-05, "loss": 5.0712, "step": 4540 }, { "epoch": 0.04099115363784077, "grad_norm": 4.464162826538086, "learning_rate": 4.795089366311608e-05, "loss": 3.6817, "step": 4541 }, { "epoch": 0.04100018053800325, "grad_norm": 2.968113422393799, "learning_rate": 4.795044231810796e-05, "loss": 4.5706, "step": 4542 }, { "epoch": 0.04100920743816573, "grad_norm": 4.363295555114746, "learning_rate": 4.794999097309984e-05, "loss": 5.0848, "step": 4543 }, { "epoch": 0.04101823433832822, "grad_norm": 3.4376935958862305, "learning_rate": 4.7949539628091714e-05, "loss": 4.6419, "step": 4544 }, { "epoch": 0.0410272612384907, "grad_norm": 3.6091806888580322, "learning_rate": 4.794908828308359e-05, "loss": 4.9011, "step": 4545 }, { "epoch": 0.04103628813865318, "grad_norm": 4.10672664642334, "learning_rate": 4.794863693807547e-05, "loss": 4.963, "step": 4546 }, { "epoch": 0.04104531503881567, "grad_norm": 3.8495349884033203, "learning_rate": 4.794818559306734e-05, "loss": 4.8086, "step": 4547 }, { "epoch": 0.04105434193897815, "grad_norm": 3.351487398147583, "learning_rate": 4.794773424805922e-05, "loss": 4.669, "step": 4548 }, { "epoch": 0.04106336883914064, "grad_norm": 3.7799301147460938, "learning_rate": 4.79472829030511e-05, "loss": 4.626, "step": 4549 }, { "epoch": 0.04107239573930312, "grad_norm": 4.089585781097412, "learning_rate": 4.794683155804297e-05, "loss": 4.7772, "step": 4550 }, { "epoch": 0.041081422639465605, "grad_norm": 2.43033766746521, "learning_rate": 4.794638021303485e-05, "loss": 4.9492, "step": 4551 }, { "epoch": 0.04109044953962809, "grad_norm": 3.830861806869507, "learning_rate": 4.794592886802672e-05, "loss": 5.1443, "step": 4552 }, { "epoch": 0.041099476439790575, "grad_norm": 2.759425401687622, "learning_rate": 4.79454775230186e-05, "loss": 4.6623, "step": 4553 }, { "epoch": 0.04110850333995306, "grad_norm": 2.766772747039795, "learning_rate": 4.794502617801047e-05, "loss": 4.4964, "step": 4554 }, { "epoch": 0.041117530240115545, "grad_norm": 3.659013032913208, "learning_rate": 4.7944574833002345e-05, "loss": 4.6976, "step": 4555 }, { "epoch": 0.041126557140278026, "grad_norm": 2.622727632522583, "learning_rate": 4.7944123487994224e-05, "loss": 4.9224, "step": 4556 }, { "epoch": 0.041135584040440515, "grad_norm": 3.022163152694702, "learning_rate": 4.7943672142986104e-05, "loss": 5.1625, "step": 4557 }, { "epoch": 0.041144610940602996, "grad_norm": 3.7682907581329346, "learning_rate": 4.7943220797977976e-05, "loss": 4.7284, "step": 4558 }, { "epoch": 0.04115363784076548, "grad_norm": 3.8865387439727783, "learning_rate": 4.7942769452969855e-05, "loss": 4.91, "step": 4559 }, { "epoch": 0.041162664740927966, "grad_norm": 3.6400554180145264, "learning_rate": 4.794231810796173e-05, "loss": 4.7135, "step": 4560 }, { "epoch": 0.04117169164109045, "grad_norm": 3.638627052307129, "learning_rate": 4.79418667629536e-05, "loss": 4.6519, "step": 4561 }, { "epoch": 0.041180718541252936, "grad_norm": 3.9581894874572754, "learning_rate": 4.794141541794548e-05, "loss": 5.0888, "step": 4562 }, { "epoch": 0.04118974544141542, "grad_norm": 3.4052796363830566, "learning_rate": 4.794096407293736e-05, "loss": 4.6816, "step": 4563 }, { "epoch": 0.0411987723415779, "grad_norm": 3.0283992290496826, "learning_rate": 4.794051272792923e-05, "loss": 4.7162, "step": 4564 }, { "epoch": 0.04120779924174039, "grad_norm": 3.0848701000213623, "learning_rate": 4.794006138292111e-05, "loss": 4.6086, "step": 4565 }, { "epoch": 0.04121682614190287, "grad_norm": 3.7582521438598633, "learning_rate": 4.793961003791298e-05, "loss": 4.6263, "step": 4566 }, { "epoch": 0.04122585304206536, "grad_norm": 4.967618942260742, "learning_rate": 4.7939158692904855e-05, "loss": 4.7692, "step": 4567 }, { "epoch": 0.04123487994222784, "grad_norm": 4.292409420013428, "learning_rate": 4.7938707347896735e-05, "loss": 5.1179, "step": 4568 }, { "epoch": 0.04124390684239032, "grad_norm": 2.9233317375183105, "learning_rate": 4.793825600288861e-05, "loss": 4.5732, "step": 4569 }, { "epoch": 0.04125293374255281, "grad_norm": 4.598384857177734, "learning_rate": 4.7937804657880486e-05, "loss": 4.9284, "step": 4570 }, { "epoch": 0.04126196064271529, "grad_norm": 3.728053092956543, "learning_rate": 4.7937353312872366e-05, "loss": 5.0114, "step": 4571 }, { "epoch": 0.04127098754287777, "grad_norm": 3.131488561630249, "learning_rate": 4.793690196786424e-05, "loss": 4.759, "step": 4572 }, { "epoch": 0.04128001444304026, "grad_norm": 2.9205148220062256, "learning_rate": 4.793645062285612e-05, "loss": 4.629, "step": 4573 }, { "epoch": 0.04128904134320274, "grad_norm": 4.17504358291626, "learning_rate": 4.793599927784799e-05, "loss": 4.4212, "step": 4574 }, { "epoch": 0.04129806824336523, "grad_norm": 3.4414916038513184, "learning_rate": 4.793554793283986e-05, "loss": 4.3956, "step": 4575 }, { "epoch": 0.04130709514352771, "grad_norm": 4.036707878112793, "learning_rate": 4.793509658783174e-05, "loss": 5.1269, "step": 4576 }, { "epoch": 0.041316122043690194, "grad_norm": 3.877528667449951, "learning_rate": 4.7934645242823614e-05, "loss": 4.7502, "step": 4577 }, { "epoch": 0.04132514894385268, "grad_norm": 3.5014731884002686, "learning_rate": 4.793419389781549e-05, "loss": 3.7512, "step": 4578 }, { "epoch": 0.041334175844015164, "grad_norm": 3.239046573638916, "learning_rate": 4.793374255280737e-05, "loss": 5.0065, "step": 4579 }, { "epoch": 0.04134320274417765, "grad_norm": 3.591505527496338, "learning_rate": 4.793329120779924e-05, "loss": 4.6137, "step": 4580 }, { "epoch": 0.041352229644340134, "grad_norm": 4.021439552307129, "learning_rate": 4.793283986279112e-05, "loss": 4.7493, "step": 4581 }, { "epoch": 0.041361256544502616, "grad_norm": 2.5705714225769043, "learning_rate": 4.7932388517782997e-05, "loss": 4.9858, "step": 4582 }, { "epoch": 0.041370283444665104, "grad_norm": 4.073733806610107, "learning_rate": 4.793193717277487e-05, "loss": 4.9905, "step": 4583 }, { "epoch": 0.041379310344827586, "grad_norm": 2.3861398696899414, "learning_rate": 4.793148582776675e-05, "loss": 5.0756, "step": 4584 }, { "epoch": 0.04138833724499007, "grad_norm": 3.6058056354522705, "learning_rate": 4.793103448275863e-05, "loss": 4.3235, "step": 4585 }, { "epoch": 0.041397364145152556, "grad_norm": 4.23106575012207, "learning_rate": 4.79305831377505e-05, "loss": 4.9027, "step": 4586 }, { "epoch": 0.04140639104531504, "grad_norm": 3.8641200065612793, "learning_rate": 4.793013179274237e-05, "loss": 4.8117, "step": 4587 }, { "epoch": 0.041415417945477526, "grad_norm": 3.2750167846679688, "learning_rate": 4.792968044773425e-05, "loss": 4.3961, "step": 4588 }, { "epoch": 0.04142444484564001, "grad_norm": 3.2860422134399414, "learning_rate": 4.7929229102726124e-05, "loss": 4.4072, "step": 4589 }, { "epoch": 0.04143347174580249, "grad_norm": 4.7348785400390625, "learning_rate": 4.7928777757718e-05, "loss": 4.6915, "step": 4590 }, { "epoch": 0.04144249864596498, "grad_norm": 3.4109387397766113, "learning_rate": 4.7928326412709876e-05, "loss": 4.7713, "step": 4591 }, { "epoch": 0.04145152554612746, "grad_norm": 3.583218574523926, "learning_rate": 4.7927875067701755e-05, "loss": 5.0982, "step": 4592 }, { "epoch": 0.04146055244628995, "grad_norm": 3.157346487045288, "learning_rate": 4.7927423722693634e-05, "loss": 4.2379, "step": 4593 }, { "epoch": 0.04146957934645243, "grad_norm": 3.6932578086853027, "learning_rate": 4.79269723776855e-05, "loss": 5.0856, "step": 4594 }, { "epoch": 0.04147860624661491, "grad_norm": 3.0923080444335938, "learning_rate": 4.792652103267738e-05, "loss": 4.858, "step": 4595 }, { "epoch": 0.0414876331467774, "grad_norm": 4.095082759857178, "learning_rate": 4.792606968766926e-05, "loss": 4.9453, "step": 4596 }, { "epoch": 0.04149666004693988, "grad_norm": 3.8521299362182617, "learning_rate": 4.792561834266113e-05, "loss": 3.9299, "step": 4597 }, { "epoch": 0.04150568694710236, "grad_norm": 3.3587021827697754, "learning_rate": 4.792516699765301e-05, "loss": 5.1843, "step": 4598 }, { "epoch": 0.04151471384726485, "grad_norm": 2.976309061050415, "learning_rate": 4.792471565264488e-05, "loss": 5.2325, "step": 4599 }, { "epoch": 0.04152374074742733, "grad_norm": 4.74043607711792, "learning_rate": 4.792426430763676e-05, "loss": 4.5906, "step": 4600 }, { "epoch": 0.04153276764758982, "grad_norm": 3.5414528846740723, "learning_rate": 4.7923812962628634e-05, "loss": 4.3466, "step": 4601 }, { "epoch": 0.0415417945477523, "grad_norm": 3.33589243888855, "learning_rate": 4.7923361617620514e-05, "loss": 5.0042, "step": 4602 }, { "epoch": 0.04155082144791478, "grad_norm": 3.774179697036743, "learning_rate": 4.7922910272612386e-05, "loss": 4.8887, "step": 4603 }, { "epoch": 0.04155984834807727, "grad_norm": 3.0098555088043213, "learning_rate": 4.7922458927604265e-05, "loss": 4.5253, "step": 4604 }, { "epoch": 0.04156887524823975, "grad_norm": 6.700408458709717, "learning_rate": 4.792200758259614e-05, "loss": 4.8197, "step": 4605 }, { "epoch": 0.04157790214840224, "grad_norm": 3.448319911956787, "learning_rate": 4.792155623758802e-05, "loss": 4.7578, "step": 4606 }, { "epoch": 0.04158692904856472, "grad_norm": 4.6328277587890625, "learning_rate": 4.792110489257989e-05, "loss": 4.9759, "step": 4607 }, { "epoch": 0.041595955948727205, "grad_norm": 2.92228364944458, "learning_rate": 4.792065354757176e-05, "loss": 4.8722, "step": 4608 }, { "epoch": 0.04160498284888969, "grad_norm": 3.17317533493042, "learning_rate": 4.792020220256364e-05, "loss": 4.5664, "step": 4609 }, { "epoch": 0.041614009749052175, "grad_norm": 2.6429457664489746, "learning_rate": 4.791975085755552e-05, "loss": 4.8384, "step": 4610 }, { "epoch": 0.041623036649214656, "grad_norm": 3.782771348953247, "learning_rate": 4.791929951254739e-05, "loss": 4.9149, "step": 4611 }, { "epoch": 0.041632063549377145, "grad_norm": 4.067492961883545, "learning_rate": 4.791884816753927e-05, "loss": 4.9504, "step": 4612 }, { "epoch": 0.041641090449539626, "grad_norm": 4.457509994506836, "learning_rate": 4.7918396822531145e-05, "loss": 5.1764, "step": 4613 }, { "epoch": 0.041650117349702115, "grad_norm": 3.9244980812072754, "learning_rate": 4.791794547752302e-05, "loss": 4.5022, "step": 4614 }, { "epoch": 0.041659144249864596, "grad_norm": 3.8617160320281982, "learning_rate": 4.7917494132514896e-05, "loss": 4.4312, "step": 4615 }, { "epoch": 0.04166817115002708, "grad_norm": 3.139240264892578, "learning_rate": 4.791704278750677e-05, "loss": 4.7405, "step": 4616 }, { "epoch": 0.041677198050189566, "grad_norm": 4.292646408081055, "learning_rate": 4.791659144249865e-05, "loss": 3.8613, "step": 4617 }, { "epoch": 0.04168622495035205, "grad_norm": 3.5613834857940674, "learning_rate": 4.791614009749053e-05, "loss": 5.0404, "step": 4618 }, { "epoch": 0.041695251850514536, "grad_norm": 5.079935073852539, "learning_rate": 4.79156887524824e-05, "loss": 4.6232, "step": 4619 }, { "epoch": 0.04170427875067702, "grad_norm": 2.7084884643554688, "learning_rate": 4.791523740747428e-05, "loss": 4.6002, "step": 4620 }, { "epoch": 0.0417133056508395, "grad_norm": 2.885418653488159, "learning_rate": 4.791478606246615e-05, "loss": 5.0032, "step": 4621 }, { "epoch": 0.04172233255100199, "grad_norm": 3.2961385250091553, "learning_rate": 4.7914334717458024e-05, "loss": 4.6079, "step": 4622 }, { "epoch": 0.04173135945116447, "grad_norm": 3.93255615234375, "learning_rate": 4.79138833724499e-05, "loss": 5.2232, "step": 4623 }, { "epoch": 0.04174038635132695, "grad_norm": 2.8266305923461914, "learning_rate": 4.791343202744178e-05, "loss": 4.8176, "step": 4624 }, { "epoch": 0.04174941325148944, "grad_norm": 3.343740701675415, "learning_rate": 4.7912980682433655e-05, "loss": 3.4653, "step": 4625 }, { "epoch": 0.04175844015165192, "grad_norm": 4.072131633758545, "learning_rate": 4.7912529337425534e-05, "loss": 5.1952, "step": 4626 }, { "epoch": 0.04176746705181441, "grad_norm": 5.979427814483643, "learning_rate": 4.7912077992417407e-05, "loss": 4.6261, "step": 4627 }, { "epoch": 0.04177649395197689, "grad_norm": 3.2886412143707275, "learning_rate": 4.791162664740928e-05, "loss": 4.5144, "step": 4628 }, { "epoch": 0.04178552085213937, "grad_norm": 3.0794715881347656, "learning_rate": 4.791117530240116e-05, "loss": 4.7876, "step": 4629 }, { "epoch": 0.04179454775230186, "grad_norm": 3.268580913543701, "learning_rate": 4.791072395739303e-05, "loss": 4.813, "step": 4630 }, { "epoch": 0.04180357465246434, "grad_norm": 3.4372570514678955, "learning_rate": 4.791027261238491e-05, "loss": 4.9883, "step": 4631 }, { "epoch": 0.04181260155262683, "grad_norm": 2.912391424179077, "learning_rate": 4.790982126737679e-05, "loss": 3.9805, "step": 4632 }, { "epoch": 0.04182162845278931, "grad_norm": 3.989051342010498, "learning_rate": 4.790936992236866e-05, "loss": 5.0499, "step": 4633 }, { "epoch": 0.041830655352951794, "grad_norm": 3.7658400535583496, "learning_rate": 4.7908918577360534e-05, "loss": 4.4655, "step": 4634 }, { "epoch": 0.04183968225311428, "grad_norm": 3.2644095420837402, "learning_rate": 4.790846723235241e-05, "loss": 4.5268, "step": 4635 }, { "epoch": 0.041848709153276764, "grad_norm": 4.436476707458496, "learning_rate": 4.7908015887344286e-05, "loss": 4.5451, "step": 4636 }, { "epoch": 0.041857736053439246, "grad_norm": 2.9834868907928467, "learning_rate": 4.7907564542336165e-05, "loss": 4.1181, "step": 4637 }, { "epoch": 0.041866762953601734, "grad_norm": 3.4213194847106934, "learning_rate": 4.790711319732804e-05, "loss": 5.1802, "step": 4638 }, { "epoch": 0.041875789853764216, "grad_norm": 2.5178706645965576, "learning_rate": 4.790666185231992e-05, "loss": 4.7975, "step": 4639 }, { "epoch": 0.041884816753926704, "grad_norm": 3.1803927421569824, "learning_rate": 4.7906210507311796e-05, "loss": 4.0658, "step": 4640 }, { "epoch": 0.041893843654089186, "grad_norm": 3.422886371612549, "learning_rate": 4.790575916230366e-05, "loss": 5.0243, "step": 4641 }, { "epoch": 0.04190287055425167, "grad_norm": 3.6153624057769775, "learning_rate": 4.790530781729554e-05, "loss": 5.3803, "step": 4642 }, { "epoch": 0.041911897454414156, "grad_norm": 3.254409074783325, "learning_rate": 4.790485647228742e-05, "loss": 5.1094, "step": 4643 }, { "epoch": 0.04192092435457664, "grad_norm": 4.848958969116211, "learning_rate": 4.790440512727929e-05, "loss": 4.125, "step": 4644 }, { "epoch": 0.041929951254739126, "grad_norm": 3.149606227874756, "learning_rate": 4.790395378227117e-05, "loss": 4.9672, "step": 4645 }, { "epoch": 0.04193897815490161, "grad_norm": 3.0860328674316406, "learning_rate": 4.790350243726305e-05, "loss": 4.1717, "step": 4646 }, { "epoch": 0.04194800505506409, "grad_norm": 4.884387016296387, "learning_rate": 4.7903051092254924e-05, "loss": 3.9447, "step": 4647 }, { "epoch": 0.04195703195522658, "grad_norm": 3.648488759994507, "learning_rate": 4.7902599747246796e-05, "loss": 4.7083, "step": 4648 }, { "epoch": 0.04196605885538906, "grad_norm": 2.642641305923462, "learning_rate": 4.7902148402238675e-05, "loss": 4.8565, "step": 4649 }, { "epoch": 0.04197508575555154, "grad_norm": 4.094734191894531, "learning_rate": 4.790169705723055e-05, "loss": 4.8272, "step": 4650 }, { "epoch": 0.04198411265571403, "grad_norm": 4.005726337432861, "learning_rate": 4.790124571222243e-05, "loss": 5.065, "step": 4651 }, { "epoch": 0.04199313955587651, "grad_norm": 4.233908653259277, "learning_rate": 4.79007943672143e-05, "loss": 5.2748, "step": 4652 }, { "epoch": 0.042002166456039, "grad_norm": 2.8933279514312744, "learning_rate": 4.790034302220618e-05, "loss": 4.797, "step": 4653 }, { "epoch": 0.04201119335620148, "grad_norm": 3.2310664653778076, "learning_rate": 4.789989167719805e-05, "loss": 4.8004, "step": 4654 }, { "epoch": 0.04202022025636396, "grad_norm": 5.420426368713379, "learning_rate": 4.7899440332189924e-05, "loss": 5.1241, "step": 4655 }, { "epoch": 0.04202924715652645, "grad_norm": 4.143190383911133, "learning_rate": 4.78989889871818e-05, "loss": 4.0334, "step": 4656 }, { "epoch": 0.04203827405668893, "grad_norm": 3.725191354751587, "learning_rate": 4.789853764217368e-05, "loss": 4.7726, "step": 4657 }, { "epoch": 0.04204730095685142, "grad_norm": 3.3755481243133545, "learning_rate": 4.7898086297165555e-05, "loss": 5.6735, "step": 4658 }, { "epoch": 0.0420563278570139, "grad_norm": 3.4605531692504883, "learning_rate": 4.7897634952157434e-05, "loss": 4.7126, "step": 4659 }, { "epoch": 0.04206535475717638, "grad_norm": 4.235642910003662, "learning_rate": 4.7897183607149306e-05, "loss": 4.1445, "step": 4660 }, { "epoch": 0.04207438165733887, "grad_norm": 2.9980905055999756, "learning_rate": 4.789673226214118e-05, "loss": 4.2981, "step": 4661 }, { "epoch": 0.04208340855750135, "grad_norm": 3.9594056606292725, "learning_rate": 4.789628091713306e-05, "loss": 4.6729, "step": 4662 }, { "epoch": 0.042092435457663835, "grad_norm": 4.027141571044922, "learning_rate": 4.789582957212494e-05, "loss": 4.3636, "step": 4663 }, { "epoch": 0.04210146235782632, "grad_norm": 3.934460163116455, "learning_rate": 4.789537822711681e-05, "loss": 4.6213, "step": 4664 }, { "epoch": 0.042110489257988805, "grad_norm": 3.6783199310302734, "learning_rate": 4.789492688210869e-05, "loss": 4.6191, "step": 4665 }, { "epoch": 0.04211951615815129, "grad_norm": 2.6769073009490967, "learning_rate": 4.789447553710056e-05, "loss": 4.989, "step": 4666 }, { "epoch": 0.042128543058313775, "grad_norm": 3.7697322368621826, "learning_rate": 4.789402419209244e-05, "loss": 4.66, "step": 4667 }, { "epoch": 0.042137569958476256, "grad_norm": 3.52095365524292, "learning_rate": 4.789357284708431e-05, "loss": 5.061, "step": 4668 }, { "epoch": 0.042146596858638745, "grad_norm": 3.588489532470703, "learning_rate": 4.7893121502076186e-05, "loss": 4.8619, "step": 4669 }, { "epoch": 0.042155623758801226, "grad_norm": 3.8949925899505615, "learning_rate": 4.7892670157068065e-05, "loss": 3.7874, "step": 4670 }, { "epoch": 0.042164650658963715, "grad_norm": 2.8234684467315674, "learning_rate": 4.7892218812059944e-05, "loss": 4.4558, "step": 4671 }, { "epoch": 0.042173677559126196, "grad_norm": 3.4095757007598877, "learning_rate": 4.7891767467051817e-05, "loss": 5.1702, "step": 4672 }, { "epoch": 0.04218270445928868, "grad_norm": 4.036530494689941, "learning_rate": 4.7891316122043696e-05, "loss": 4.6215, "step": 4673 }, { "epoch": 0.042191731359451166, "grad_norm": 4.766295909881592, "learning_rate": 4.789086477703557e-05, "loss": 5.1765, "step": 4674 }, { "epoch": 0.04220075825961365, "grad_norm": 3.686713695526123, "learning_rate": 4.789041343202744e-05, "loss": 4.8415, "step": 4675 }, { "epoch": 0.04220978515977613, "grad_norm": 3.2083821296691895, "learning_rate": 4.788996208701932e-05, "loss": 4.6434, "step": 4676 }, { "epoch": 0.04221881205993862, "grad_norm": 3.0770862102508545, "learning_rate": 4.788951074201119e-05, "loss": 4.709, "step": 4677 }, { "epoch": 0.0422278389601011, "grad_norm": 3.367166519165039, "learning_rate": 4.788905939700307e-05, "loss": 4.833, "step": 4678 }, { "epoch": 0.04223686586026359, "grad_norm": 4.220921993255615, "learning_rate": 4.788860805199495e-05, "loss": 4.8659, "step": 4679 }, { "epoch": 0.04224589276042607, "grad_norm": 2.757277011871338, "learning_rate": 4.788815670698682e-05, "loss": 4.4019, "step": 4680 }, { "epoch": 0.04225491966058855, "grad_norm": 3.212627410888672, "learning_rate": 4.7887705361978696e-05, "loss": 4.533, "step": 4681 }, { "epoch": 0.04226394656075104, "grad_norm": 3.4853405952453613, "learning_rate": 4.7887254016970575e-05, "loss": 4.5054, "step": 4682 }, { "epoch": 0.04227297346091352, "grad_norm": 4.975287914276123, "learning_rate": 4.788680267196245e-05, "loss": 4.0408, "step": 4683 }, { "epoch": 0.04228200036107601, "grad_norm": 2.9647486209869385, "learning_rate": 4.788635132695433e-05, "loss": 4.5138, "step": 4684 }, { "epoch": 0.04229102726123849, "grad_norm": 4.503072738647461, "learning_rate": 4.7885899981946206e-05, "loss": 4.5684, "step": 4685 }, { "epoch": 0.04230005416140097, "grad_norm": 6.532361030578613, "learning_rate": 4.788544863693808e-05, "loss": 4.9352, "step": 4686 }, { "epoch": 0.04230908106156346, "grad_norm": 3.4486026763916016, "learning_rate": 4.788499729192996e-05, "loss": 5.1139, "step": 4687 }, { "epoch": 0.04231810796172594, "grad_norm": 3.9884424209594727, "learning_rate": 4.788454594692183e-05, "loss": 4.0974, "step": 4688 }, { "epoch": 0.042327134861888424, "grad_norm": 3.421302318572998, "learning_rate": 4.78840946019137e-05, "loss": 5.1256, "step": 4689 }, { "epoch": 0.04233616176205091, "grad_norm": 4.493325233459473, "learning_rate": 4.788364325690558e-05, "loss": 4.8724, "step": 4690 }, { "epoch": 0.042345188662213394, "grad_norm": 2.481168270111084, "learning_rate": 4.7883191911897454e-05, "loss": 4.6922, "step": 4691 }, { "epoch": 0.04235421556237588, "grad_norm": 3.7740628719329834, "learning_rate": 4.7882740566889334e-05, "loss": 4.4088, "step": 4692 }, { "epoch": 0.042363242462538364, "grad_norm": 3.4563260078430176, "learning_rate": 4.788228922188121e-05, "loss": 4.66, "step": 4693 }, { "epoch": 0.042372269362700846, "grad_norm": 3.8976314067840576, "learning_rate": 4.788183787687308e-05, "loss": 4.2494, "step": 4694 }, { "epoch": 0.042381296262863334, "grad_norm": 3.3494086265563965, "learning_rate": 4.788138653186496e-05, "loss": 4.7027, "step": 4695 }, { "epoch": 0.042390323163025816, "grad_norm": 3.4121105670928955, "learning_rate": 4.788093518685684e-05, "loss": 4.1859, "step": 4696 }, { "epoch": 0.042399350063188304, "grad_norm": 3.046107530593872, "learning_rate": 4.788048384184871e-05, "loss": 4.6299, "step": 4697 }, { "epoch": 0.042408376963350786, "grad_norm": 4.181197643280029, "learning_rate": 4.788003249684059e-05, "loss": 4.6367, "step": 4698 }, { "epoch": 0.04241740386351327, "grad_norm": 4.529419422149658, "learning_rate": 4.787958115183246e-05, "loss": 5.1939, "step": 4699 }, { "epoch": 0.042426430763675756, "grad_norm": 4.406407356262207, "learning_rate": 4.787912980682434e-05, "loss": 4.5243, "step": 4700 }, { "epoch": 0.04243545766383824, "grad_norm": 3.6789846420288086, "learning_rate": 4.787867846181621e-05, "loss": 4.5666, "step": 4701 }, { "epoch": 0.04244448456400072, "grad_norm": 4.036098480224609, "learning_rate": 4.7878227116808085e-05, "loss": 4.6531, "step": 4702 }, { "epoch": 0.04245351146416321, "grad_norm": 3.6810803413391113, "learning_rate": 4.7877775771799965e-05, "loss": 4.4037, "step": 4703 }, { "epoch": 0.04246253836432569, "grad_norm": 3.0716798305511475, "learning_rate": 4.7877324426791844e-05, "loss": 4.7958, "step": 4704 }, { "epoch": 0.04247156526448818, "grad_norm": 3.4554009437561035, "learning_rate": 4.7876873081783716e-05, "loss": 5.0885, "step": 4705 }, { "epoch": 0.04248059216465066, "grad_norm": 2.8813741207122803, "learning_rate": 4.7876421736775596e-05, "loss": 4.8914, "step": 4706 }, { "epoch": 0.04248961906481314, "grad_norm": 4.07888126373291, "learning_rate": 4.7875970391767475e-05, "loss": 4.853, "step": 4707 }, { "epoch": 0.04249864596497563, "grad_norm": 3.4076333045959473, "learning_rate": 4.787551904675934e-05, "loss": 4.274, "step": 4708 }, { "epoch": 0.04250767286513811, "grad_norm": 4.562119007110596, "learning_rate": 4.787506770175122e-05, "loss": 4.6696, "step": 4709 }, { "epoch": 0.0425166997653006, "grad_norm": 9.682636260986328, "learning_rate": 4.78746163567431e-05, "loss": 4.8174, "step": 4710 }, { "epoch": 0.04252572666546308, "grad_norm": 2.723888874053955, "learning_rate": 4.787416501173497e-05, "loss": 4.8797, "step": 4711 }, { "epoch": 0.04253475356562556, "grad_norm": 3.781066417694092, "learning_rate": 4.787371366672685e-05, "loss": 4.2946, "step": 4712 }, { "epoch": 0.04254378046578805, "grad_norm": 4.438226222991943, "learning_rate": 4.787326232171872e-05, "loss": 5.5299, "step": 4713 }, { "epoch": 0.04255280736595053, "grad_norm": 2.839480400085449, "learning_rate": 4.78728109767106e-05, "loss": 4.7464, "step": 4714 }, { "epoch": 0.04256183426611301, "grad_norm": 3.8488080501556396, "learning_rate": 4.7872359631702475e-05, "loss": 5.0162, "step": 4715 }, { "epoch": 0.0425708611662755, "grad_norm": 2.363065481185913, "learning_rate": 4.787190828669435e-05, "loss": 5.0651, "step": 4716 }, { "epoch": 0.04257988806643798, "grad_norm": 3.7695181369781494, "learning_rate": 4.7871456941686227e-05, "loss": 4.2757, "step": 4717 }, { "epoch": 0.04258891496660047, "grad_norm": 3.9407711029052734, "learning_rate": 4.7871005596678106e-05, "loss": 4.513, "step": 4718 }, { "epoch": 0.04259794186676295, "grad_norm": 2.701441526412964, "learning_rate": 4.787055425166998e-05, "loss": 4.7596, "step": 4719 }, { "epoch": 0.042606968766925435, "grad_norm": 3.00170636177063, "learning_rate": 4.787010290666186e-05, "loss": 4.6016, "step": 4720 }, { "epoch": 0.04261599566708792, "grad_norm": 4.0043745040893555, "learning_rate": 4.786965156165373e-05, "loss": 4.354, "step": 4721 }, { "epoch": 0.042625022567250405, "grad_norm": 3.723184823989868, "learning_rate": 4.78692002166456e-05, "loss": 4.9567, "step": 4722 }, { "epoch": 0.04263404946741289, "grad_norm": 3.0975418090820312, "learning_rate": 4.786874887163748e-05, "loss": 4.5025, "step": 4723 }, { "epoch": 0.042643076367575375, "grad_norm": 3.82940673828125, "learning_rate": 4.786829752662936e-05, "loss": 4.1917, "step": 4724 }, { "epoch": 0.042652103267737856, "grad_norm": 3.433354616165161, "learning_rate": 4.786784618162123e-05, "loss": 4.8736, "step": 4725 }, { "epoch": 0.042661130167900345, "grad_norm": 3.576270580291748, "learning_rate": 4.786739483661311e-05, "loss": 4.9485, "step": 4726 }, { "epoch": 0.042670157068062826, "grad_norm": 3.450174331665039, "learning_rate": 4.7866943491604985e-05, "loss": 4.4579, "step": 4727 }, { "epoch": 0.04267918396822531, "grad_norm": 4.069995403289795, "learning_rate": 4.786649214659686e-05, "loss": 5.419, "step": 4728 }, { "epoch": 0.042688210868387796, "grad_norm": 4.0322723388671875, "learning_rate": 4.786604080158874e-05, "loss": 5.209, "step": 4729 }, { "epoch": 0.04269723776855028, "grad_norm": 3.254912853240967, "learning_rate": 4.786558945658061e-05, "loss": 4.3601, "step": 4730 }, { "epoch": 0.042706264668712766, "grad_norm": 4.525927543640137, "learning_rate": 4.786513811157249e-05, "loss": 4.4122, "step": 4731 }, { "epoch": 0.04271529156887525, "grad_norm": 3.429368019104004, "learning_rate": 4.786468676656437e-05, "loss": 4.5559, "step": 4732 }, { "epoch": 0.04272431846903773, "grad_norm": 2.866947889328003, "learning_rate": 4.786423542155624e-05, "loss": 4.5289, "step": 4733 }, { "epoch": 0.04273334536920022, "grad_norm": 4.12953519821167, "learning_rate": 4.786378407654812e-05, "loss": 4.736, "step": 4734 }, { "epoch": 0.0427423722693627, "grad_norm": 4.433162689208984, "learning_rate": 4.786333273153999e-05, "loss": 4.9763, "step": 4735 }, { "epoch": 0.04275139916952519, "grad_norm": 4.846735000610352, "learning_rate": 4.7862881386531864e-05, "loss": 4.7327, "step": 4736 }, { "epoch": 0.04276042606968767, "grad_norm": 3.6616971492767334, "learning_rate": 4.7862430041523744e-05, "loss": 4.9241, "step": 4737 }, { "epoch": 0.04276945296985015, "grad_norm": 2.9334638118743896, "learning_rate": 4.7861978696515616e-05, "loss": 4.9116, "step": 4738 }, { "epoch": 0.04277847987001264, "grad_norm": 3.964097261428833, "learning_rate": 4.7861527351507495e-05, "loss": 4.1224, "step": 4739 }, { "epoch": 0.04278750677017512, "grad_norm": 3.4809272289276123, "learning_rate": 4.7861076006499375e-05, "loss": 4.5644, "step": 4740 }, { "epoch": 0.04279653367033761, "grad_norm": 6.119604587554932, "learning_rate": 4.786062466149124e-05, "loss": 4.731, "step": 4741 }, { "epoch": 0.04280556057050009, "grad_norm": 4.015038013458252, "learning_rate": 4.786017331648312e-05, "loss": 4.75, "step": 4742 }, { "epoch": 0.04281458747066257, "grad_norm": 3.344067096710205, "learning_rate": 4.7859721971475e-05, "loss": 4.7717, "step": 4743 }, { "epoch": 0.04282361437082506, "grad_norm": 6.373073577880859, "learning_rate": 4.785927062646687e-05, "loss": 4.4101, "step": 4744 }, { "epoch": 0.04283264127098754, "grad_norm": 3.803795337677002, "learning_rate": 4.785881928145875e-05, "loss": 5.106, "step": 4745 }, { "epoch": 0.042841668171150024, "grad_norm": 2.9463911056518555, "learning_rate": 4.785836793645063e-05, "loss": 4.9905, "step": 4746 }, { "epoch": 0.04285069507131251, "grad_norm": 3.9368784427642822, "learning_rate": 4.78579165914425e-05, "loss": 4.2233, "step": 4747 }, { "epoch": 0.042859721971474994, "grad_norm": 3.3811938762664795, "learning_rate": 4.7857465246434375e-05, "loss": 4.7898, "step": 4748 }, { "epoch": 0.04286874887163748, "grad_norm": 3.6358373165130615, "learning_rate": 4.7857013901426254e-05, "loss": 4.1106, "step": 4749 }, { "epoch": 0.042877775771799964, "grad_norm": 4.7955780029296875, "learning_rate": 4.7856562556418126e-05, "loss": 4.5963, "step": 4750 }, { "epoch": 0.042886802671962446, "grad_norm": 3.1258718967437744, "learning_rate": 4.7856111211410006e-05, "loss": 5.0541, "step": 4751 }, { "epoch": 0.042895829572124934, "grad_norm": 3.8079028129577637, "learning_rate": 4.785565986640188e-05, "loss": 4.7182, "step": 4752 }, { "epoch": 0.042904856472287416, "grad_norm": 4.191341876983643, "learning_rate": 4.785520852139376e-05, "loss": 4.7795, "step": 4753 }, { "epoch": 0.042913883372449904, "grad_norm": 4.024031162261963, "learning_rate": 4.7854757176385636e-05, "loss": 4.6904, "step": 4754 }, { "epoch": 0.042922910272612386, "grad_norm": 3.7743165493011475, "learning_rate": 4.78543058313775e-05, "loss": 4.5547, "step": 4755 }, { "epoch": 0.04293193717277487, "grad_norm": 5.422250270843506, "learning_rate": 4.785385448636938e-05, "loss": 4.8409, "step": 4756 }, { "epoch": 0.042940964072937356, "grad_norm": 6.098740100860596, "learning_rate": 4.785340314136126e-05, "loss": 4.9597, "step": 4757 }, { "epoch": 0.04294999097309984, "grad_norm": 4.596536636352539, "learning_rate": 4.785295179635313e-05, "loss": 4.2591, "step": 4758 }, { "epoch": 0.04295901787326232, "grad_norm": 4.131002426147461, "learning_rate": 4.785250045134501e-05, "loss": 4.0622, "step": 4759 }, { "epoch": 0.04296804477342481, "grad_norm": 3.644512176513672, "learning_rate": 4.7852049106336885e-05, "loss": 4.48, "step": 4760 }, { "epoch": 0.04297707167358729, "grad_norm": 2.9870378971099854, "learning_rate": 4.7851597761328764e-05, "loss": 4.5597, "step": 4761 }, { "epoch": 0.04298609857374978, "grad_norm": 4.227797031402588, "learning_rate": 4.7851146416320636e-05, "loss": 4.772, "step": 4762 }, { "epoch": 0.04299512547391226, "grad_norm": 3.600722074508667, "learning_rate": 4.7850695071312516e-05, "loss": 4.4019, "step": 4763 }, { "epoch": 0.04300415237407474, "grad_norm": 4.023229598999023, "learning_rate": 4.785024372630439e-05, "loss": 4.8923, "step": 4764 }, { "epoch": 0.04301317927423723, "grad_norm": 4.358348369598389, "learning_rate": 4.784979238129627e-05, "loss": 5.2859, "step": 4765 }, { "epoch": 0.04302220617439971, "grad_norm": 2.9132235050201416, "learning_rate": 4.784934103628814e-05, "loss": 5.4029, "step": 4766 }, { "epoch": 0.0430312330745622, "grad_norm": 3.943626880645752, "learning_rate": 4.784888969128002e-05, "loss": 4.4886, "step": 4767 }, { "epoch": 0.04304025997472468, "grad_norm": 3.4639737606048584, "learning_rate": 4.784843834627189e-05, "loss": 4.6898, "step": 4768 }, { "epoch": 0.04304928687488716, "grad_norm": 3.79345965385437, "learning_rate": 4.7847987001263764e-05, "loss": 4.4118, "step": 4769 }, { "epoch": 0.04305831377504965, "grad_norm": 3.9894251823425293, "learning_rate": 4.784753565625564e-05, "loss": 4.6994, "step": 4770 }, { "epoch": 0.04306734067521213, "grad_norm": 2.905851364135742, "learning_rate": 4.784708431124752e-05, "loss": 5.0725, "step": 4771 }, { "epoch": 0.04307636757537461, "grad_norm": 3.7991302013397217, "learning_rate": 4.7846632966239395e-05, "loss": 4.4155, "step": 4772 }, { "epoch": 0.0430853944755371, "grad_norm": 3.609490156173706, "learning_rate": 4.7846181621231274e-05, "loss": 4.1776, "step": 4773 }, { "epoch": 0.04309442137569958, "grad_norm": 3.4402761459350586, "learning_rate": 4.784573027622315e-05, "loss": 4.9699, "step": 4774 }, { "epoch": 0.04310344827586207, "grad_norm": 4.317553520202637, "learning_rate": 4.784527893121502e-05, "loss": 5.1514, "step": 4775 }, { "epoch": 0.04311247517602455, "grad_norm": 2.747044563293457, "learning_rate": 4.78448275862069e-05, "loss": 5.2229, "step": 4776 }, { "epoch": 0.043121502076187035, "grad_norm": 3.347416877746582, "learning_rate": 4.784437624119877e-05, "loss": 5.16, "step": 4777 }, { "epoch": 0.04313052897634952, "grad_norm": 3.4821112155914307, "learning_rate": 4.784392489619065e-05, "loss": 4.404, "step": 4778 }, { "epoch": 0.043139555876512005, "grad_norm": 3.9733078479766846, "learning_rate": 4.784347355118253e-05, "loss": 4.7744, "step": 4779 }, { "epoch": 0.04314858277667449, "grad_norm": 3.6451237201690674, "learning_rate": 4.78430222061744e-05, "loss": 4.7969, "step": 4780 }, { "epoch": 0.043157609676836975, "grad_norm": 2.9341986179351807, "learning_rate": 4.784257086116628e-05, "loss": 4.8071, "step": 4781 }, { "epoch": 0.043166636576999456, "grad_norm": 3.846334934234619, "learning_rate": 4.7842119516158154e-05, "loss": 4.5969, "step": 4782 }, { "epoch": 0.043175663477161945, "grad_norm": 3.834216356277466, "learning_rate": 4.7841668171150026e-05, "loss": 3.9011, "step": 4783 }, { "epoch": 0.043184690377324426, "grad_norm": 4.740387439727783, "learning_rate": 4.7841216826141905e-05, "loss": 4.9372, "step": 4784 }, { "epoch": 0.04319371727748691, "grad_norm": 2.742354393005371, "learning_rate": 4.7840765481133784e-05, "loss": 4.6021, "step": 4785 }, { "epoch": 0.043202744177649396, "grad_norm": 3.717724084854126, "learning_rate": 4.784031413612566e-05, "loss": 4.6883, "step": 4786 }, { "epoch": 0.04321177107781188, "grad_norm": 3.2867724895477295, "learning_rate": 4.7839862791117536e-05, "loss": 4.9732, "step": 4787 }, { "epoch": 0.043220797977974366, "grad_norm": 3.772322177886963, "learning_rate": 4.783941144610941e-05, "loss": 4.2754, "step": 4788 }, { "epoch": 0.04322982487813685, "grad_norm": 5.007787227630615, "learning_rate": 4.783896010110128e-05, "loss": 5.2945, "step": 4789 }, { "epoch": 0.04323885177829933, "grad_norm": 3.38468074798584, "learning_rate": 4.783850875609316e-05, "loss": 5.0148, "step": 4790 }, { "epoch": 0.04324787867846182, "grad_norm": 3.8591911792755127, "learning_rate": 4.783805741108503e-05, "loss": 4.9077, "step": 4791 }, { "epoch": 0.0432569055786243, "grad_norm": 3.740009307861328, "learning_rate": 4.783760606607691e-05, "loss": 4.4462, "step": 4792 }, { "epoch": 0.04326593247878679, "grad_norm": 2.744338035583496, "learning_rate": 4.783715472106879e-05, "loss": 4.655, "step": 4793 }, { "epoch": 0.04327495937894927, "grad_norm": 3.695760726928711, "learning_rate": 4.7836703376060664e-05, "loss": 4.9078, "step": 4794 }, { "epoch": 0.04328398627911175, "grad_norm": 4.405290603637695, "learning_rate": 4.7836252031052536e-05, "loss": 5.095, "step": 4795 }, { "epoch": 0.04329301317927424, "grad_norm": 2.5470974445343018, "learning_rate": 4.7835800686044415e-05, "loss": 5.1141, "step": 4796 }, { "epoch": 0.04330204007943672, "grad_norm": 3.8378567695617676, "learning_rate": 4.783534934103629e-05, "loss": 4.5331, "step": 4797 }, { "epoch": 0.0433110669795992, "grad_norm": 3.9598212242126465, "learning_rate": 4.783489799602817e-05, "loss": 4.0671, "step": 4798 }, { "epoch": 0.04332009387976169, "grad_norm": 3.551696538925171, "learning_rate": 4.783444665102004e-05, "loss": 5.1067, "step": 4799 }, { "epoch": 0.04332912077992417, "grad_norm": 4.236985683441162, "learning_rate": 4.783399530601192e-05, "loss": 4.3666, "step": 4800 }, { "epoch": 0.04333814768008666, "grad_norm": 3.289016008377075, "learning_rate": 4.78335439610038e-05, "loss": 4.8169, "step": 4801 }, { "epoch": 0.04334717458024914, "grad_norm": 2.9952540397644043, "learning_rate": 4.7833092615995664e-05, "loss": 4.4437, "step": 4802 }, { "epoch": 0.043356201480411624, "grad_norm": 3.0701990127563477, "learning_rate": 4.783264127098754e-05, "loss": 5.0177, "step": 4803 }, { "epoch": 0.04336522838057411, "grad_norm": 4.067071914672852, "learning_rate": 4.783218992597942e-05, "loss": 3.727, "step": 4804 }, { "epoch": 0.043374255280736594, "grad_norm": 3.8260350227355957, "learning_rate": 4.7831738580971295e-05, "loss": 4.7249, "step": 4805 }, { "epoch": 0.04338328218089908, "grad_norm": 3.630626916885376, "learning_rate": 4.7831287235963174e-05, "loss": 4.5186, "step": 4806 }, { "epoch": 0.043392309081061564, "grad_norm": 2.396045207977295, "learning_rate": 4.783083589095505e-05, "loss": 5.0646, "step": 4807 }, { "epoch": 0.043401335981224046, "grad_norm": 4.19234037399292, "learning_rate": 4.7830384545946926e-05, "loss": 4.8107, "step": 4808 }, { "epoch": 0.043410362881386534, "grad_norm": 2.4242072105407715, "learning_rate": 4.78299332009388e-05, "loss": 4.7866, "step": 4809 }, { "epoch": 0.043419389781549016, "grad_norm": 4.023800373077393, "learning_rate": 4.782948185593068e-05, "loss": 4.5852, "step": 4810 }, { "epoch": 0.0434284166817115, "grad_norm": 3.4793083667755127, "learning_rate": 4.782903051092255e-05, "loss": 4.9665, "step": 4811 }, { "epoch": 0.043437443581873986, "grad_norm": 3.7783844470977783, "learning_rate": 4.782857916591443e-05, "loss": 5.2301, "step": 4812 }, { "epoch": 0.04344647048203647, "grad_norm": 4.399711608886719, "learning_rate": 4.78281278209063e-05, "loss": 4.9184, "step": 4813 }, { "epoch": 0.043455497382198956, "grad_norm": 2.8148484230041504, "learning_rate": 4.782767647589818e-05, "loss": 3.9863, "step": 4814 }, { "epoch": 0.04346452428236144, "grad_norm": 3.0469467639923096, "learning_rate": 4.782722513089005e-05, "loss": 5.1028, "step": 4815 }, { "epoch": 0.04347355118252392, "grad_norm": 4.500543117523193, "learning_rate": 4.7826773785881926e-05, "loss": 4.452, "step": 4816 }, { "epoch": 0.04348257808268641, "grad_norm": 3.588812828063965, "learning_rate": 4.7826322440873805e-05, "loss": 4.8169, "step": 4817 }, { "epoch": 0.04349160498284889, "grad_norm": 3.298532485961914, "learning_rate": 4.7825871095865684e-05, "loss": 4.7656, "step": 4818 }, { "epoch": 0.04350063188301138, "grad_norm": 4.0248026847839355, "learning_rate": 4.782541975085756e-05, "loss": 4.7311, "step": 4819 }, { "epoch": 0.04350965878317386, "grad_norm": 3.1468584537506104, "learning_rate": 4.7824968405849436e-05, "loss": 5.426, "step": 4820 }, { "epoch": 0.04351868568333634, "grad_norm": 4.138670921325684, "learning_rate": 4.782451706084131e-05, "loss": 4.4432, "step": 4821 }, { "epoch": 0.04352771258349883, "grad_norm": 4.234177112579346, "learning_rate": 4.782406571583318e-05, "loss": 4.9206, "step": 4822 }, { "epoch": 0.04353673948366131, "grad_norm": 3.254361629486084, "learning_rate": 4.782361437082506e-05, "loss": 4.6755, "step": 4823 }, { "epoch": 0.04354576638382379, "grad_norm": 2.4812424182891846, "learning_rate": 4.782316302581694e-05, "loss": 4.739, "step": 4824 }, { "epoch": 0.04355479328398628, "grad_norm": 3.5433993339538574, "learning_rate": 4.782271168080881e-05, "loss": 4.3796, "step": 4825 }, { "epoch": 0.04356382018414876, "grad_norm": 3.0883257389068604, "learning_rate": 4.782226033580069e-05, "loss": 4.7135, "step": 4826 }, { "epoch": 0.04357284708431125, "grad_norm": 3.6850974559783936, "learning_rate": 4.7821808990792564e-05, "loss": 4.7818, "step": 4827 }, { "epoch": 0.04358187398447373, "grad_norm": 4.319875240325928, "learning_rate": 4.782135764578444e-05, "loss": 4.9008, "step": 4828 }, { "epoch": 0.04359090088463621, "grad_norm": 4.126944541931152, "learning_rate": 4.7820906300776315e-05, "loss": 4.2996, "step": 4829 }, { "epoch": 0.0435999277847987, "grad_norm": 3.818035840988159, "learning_rate": 4.782045495576819e-05, "loss": 4.8283, "step": 4830 }, { "epoch": 0.04360895468496118, "grad_norm": 3.3065247535705566, "learning_rate": 4.782000361076007e-05, "loss": 4.8645, "step": 4831 }, { "epoch": 0.04361798158512367, "grad_norm": 3.881096124649048, "learning_rate": 4.7819552265751946e-05, "loss": 3.9972, "step": 4832 }, { "epoch": 0.04362700848528615, "grad_norm": 5.733536720275879, "learning_rate": 4.781910092074382e-05, "loss": 4.9088, "step": 4833 }, { "epoch": 0.043636035385448635, "grad_norm": 3.234931707382202, "learning_rate": 4.78186495757357e-05, "loss": 4.6492, "step": 4834 }, { "epoch": 0.04364506228561112, "grad_norm": 3.824125051498413, "learning_rate": 4.781819823072757e-05, "loss": 4.9209, "step": 4835 }, { "epoch": 0.043654089185773605, "grad_norm": 3.440838098526001, "learning_rate": 4.781774688571944e-05, "loss": 4.7154, "step": 4836 }, { "epoch": 0.043663116085936086, "grad_norm": 3.9382176399230957, "learning_rate": 4.781729554071132e-05, "loss": 5.0105, "step": 4837 }, { "epoch": 0.043672142986098575, "grad_norm": 3.7207412719726562, "learning_rate": 4.7816844195703195e-05, "loss": 4.723, "step": 4838 }, { "epoch": 0.043681169886261056, "grad_norm": 4.869022846221924, "learning_rate": 4.7816392850695074e-05, "loss": 4.2773, "step": 4839 }, { "epoch": 0.043690196786423545, "grad_norm": 4.749978542327881, "learning_rate": 4.781594150568695e-05, "loss": 5.1517, "step": 4840 }, { "epoch": 0.043699223686586026, "grad_norm": 3.4445245265960693, "learning_rate": 4.7815490160678825e-05, "loss": 4.8868, "step": 4841 }, { "epoch": 0.04370825058674851, "grad_norm": 4.534551620483398, "learning_rate": 4.78150388156707e-05, "loss": 4.5592, "step": 4842 }, { "epoch": 0.043717277486910996, "grad_norm": 4.251738548278809, "learning_rate": 4.781458747066258e-05, "loss": 4.6861, "step": 4843 }, { "epoch": 0.04372630438707348, "grad_norm": 3.597759962081909, "learning_rate": 4.781413612565445e-05, "loss": 4.396, "step": 4844 }, { "epoch": 0.043735331287235966, "grad_norm": 3.7080559730529785, "learning_rate": 4.781368478064633e-05, "loss": 4.7879, "step": 4845 }, { "epoch": 0.04374435818739845, "grad_norm": 3.3356516361236572, "learning_rate": 4.781323343563821e-05, "loss": 4.3196, "step": 4846 }, { "epoch": 0.04375338508756093, "grad_norm": 4.005497455596924, "learning_rate": 4.781278209063008e-05, "loss": 5.1282, "step": 4847 }, { "epoch": 0.04376241198772342, "grad_norm": 3.8925082683563232, "learning_rate": 4.781233074562196e-05, "loss": 4.9654, "step": 4848 }, { "epoch": 0.0437714388878859, "grad_norm": 3.202054977416992, "learning_rate": 4.781187940061383e-05, "loss": 4.5823, "step": 4849 }, { "epoch": 0.04378046578804838, "grad_norm": 3.894658088684082, "learning_rate": 4.7811428055605705e-05, "loss": 4.9318, "step": 4850 }, { "epoch": 0.04378949268821087, "grad_norm": 4.405616760253906, "learning_rate": 4.7810976710597584e-05, "loss": 5.2063, "step": 4851 }, { "epoch": 0.04379851958837335, "grad_norm": 3.526832103729248, "learning_rate": 4.7810525365589456e-05, "loss": 4.7874, "step": 4852 }, { "epoch": 0.04380754648853584, "grad_norm": 3.1603574752807617, "learning_rate": 4.7810074020581336e-05, "loss": 4.8255, "step": 4853 }, { "epoch": 0.04381657338869832, "grad_norm": 3.811577320098877, "learning_rate": 4.7809622675573215e-05, "loss": 4.96, "step": 4854 }, { "epoch": 0.0438256002888608, "grad_norm": 3.9203953742980957, "learning_rate": 4.780917133056509e-05, "loss": 3.8341, "step": 4855 }, { "epoch": 0.04383462718902329, "grad_norm": 3.472736358642578, "learning_rate": 4.780871998555696e-05, "loss": 4.9363, "step": 4856 }, { "epoch": 0.04384365408918577, "grad_norm": 3.199031352996826, "learning_rate": 4.780826864054884e-05, "loss": 4.9186, "step": 4857 }, { "epoch": 0.04385268098934826, "grad_norm": 3.8914742469787598, "learning_rate": 4.780781729554071e-05, "loss": 4.4698, "step": 4858 }, { "epoch": 0.04386170788951074, "grad_norm": 2.4011850357055664, "learning_rate": 4.780736595053259e-05, "loss": 5.1833, "step": 4859 }, { "epoch": 0.043870734789673224, "grad_norm": 2.8818042278289795, "learning_rate": 4.780691460552446e-05, "loss": 4.7898, "step": 4860 }, { "epoch": 0.04387976168983571, "grad_norm": 2.397477149963379, "learning_rate": 4.780646326051634e-05, "loss": 4.9724, "step": 4861 }, { "epoch": 0.043888788589998194, "grad_norm": 6.808093547821045, "learning_rate": 4.7806011915508215e-05, "loss": 4.9249, "step": 4862 }, { "epoch": 0.043897815490160676, "grad_norm": 3.8831021785736084, "learning_rate": 4.780556057050009e-05, "loss": 4.3793, "step": 4863 }, { "epoch": 0.043906842390323164, "grad_norm": 3.0992655754089355, "learning_rate": 4.780510922549197e-05, "loss": 4.6124, "step": 4864 }, { "epoch": 0.043915869290485646, "grad_norm": 4.5812201499938965, "learning_rate": 4.7804657880483846e-05, "loss": 2.958, "step": 4865 }, { "epoch": 0.043924896190648134, "grad_norm": 4.564762115478516, "learning_rate": 4.780420653547572e-05, "loss": 5.0436, "step": 4866 }, { "epoch": 0.043933923090810616, "grad_norm": 4.500792503356934, "learning_rate": 4.78037551904676e-05, "loss": 5.1468, "step": 4867 }, { "epoch": 0.0439429499909731, "grad_norm": 2.919541835784912, "learning_rate": 4.780330384545948e-05, "loss": 5.0861, "step": 4868 }, { "epoch": 0.043951976891135586, "grad_norm": 3.5180718898773193, "learning_rate": 4.780285250045134e-05, "loss": 4.7916, "step": 4869 }, { "epoch": 0.04396100379129807, "grad_norm": 3.6458780765533447, "learning_rate": 4.780240115544322e-05, "loss": 4.4652, "step": 4870 }, { "epoch": 0.043970030691460556, "grad_norm": 3.1121461391448975, "learning_rate": 4.78019498104351e-05, "loss": 4.3912, "step": 4871 }, { "epoch": 0.04397905759162304, "grad_norm": 2.5612130165100098, "learning_rate": 4.7801498465426974e-05, "loss": 4.8518, "step": 4872 }, { "epoch": 0.04398808449178552, "grad_norm": 3.149946928024292, "learning_rate": 4.780104712041885e-05, "loss": 4.9431, "step": 4873 }, { "epoch": 0.04399711139194801, "grad_norm": 2.6908421516418457, "learning_rate": 4.7800595775410725e-05, "loss": 4.2078, "step": 4874 }, { "epoch": 0.04400613829211049, "grad_norm": 3.005645751953125, "learning_rate": 4.7800144430402604e-05, "loss": 4.7586, "step": 4875 }, { "epoch": 0.04401516519227297, "grad_norm": 5.412473678588867, "learning_rate": 4.779969308539448e-05, "loss": 4.82, "step": 4876 }, { "epoch": 0.04402419209243546, "grad_norm": 3.8978559970855713, "learning_rate": 4.779924174038635e-05, "loss": 4.371, "step": 4877 }, { "epoch": 0.04403321899259794, "grad_norm": 2.9038925170898438, "learning_rate": 4.779879039537823e-05, "loss": 4.9534, "step": 4878 }, { "epoch": 0.04404224589276043, "grad_norm": 4.1816630363464355, "learning_rate": 4.779833905037011e-05, "loss": 3.9069, "step": 4879 }, { "epoch": 0.04405127279292291, "grad_norm": 3.3781790733337402, "learning_rate": 4.779788770536198e-05, "loss": 4.4871, "step": 4880 }, { "epoch": 0.04406029969308539, "grad_norm": 3.7975001335144043, "learning_rate": 4.779743636035386e-05, "loss": 4.1835, "step": 4881 }, { "epoch": 0.04406932659324788, "grad_norm": 3.8070709705352783, "learning_rate": 4.779698501534573e-05, "loss": 4.6095, "step": 4882 }, { "epoch": 0.04407835349341036, "grad_norm": 3.357011079788208, "learning_rate": 4.7796533670337605e-05, "loss": 4.6035, "step": 4883 }, { "epoch": 0.04408738039357285, "grad_norm": 3.2323029041290283, "learning_rate": 4.7796082325329484e-05, "loss": 4.4322, "step": 4884 }, { "epoch": 0.04409640729373533, "grad_norm": 4.984407424926758, "learning_rate": 4.779563098032136e-05, "loss": 3.5345, "step": 4885 }, { "epoch": 0.04410543419389781, "grad_norm": 6.063580513000488, "learning_rate": 4.7795179635313235e-05, "loss": 4.6536, "step": 4886 }, { "epoch": 0.0441144610940603, "grad_norm": 4.439483165740967, "learning_rate": 4.7794728290305115e-05, "loss": 5.0776, "step": 4887 }, { "epoch": 0.04412348799422278, "grad_norm": 4.191533088684082, "learning_rate": 4.779427694529699e-05, "loss": 4.5785, "step": 4888 }, { "epoch": 0.044132514894385265, "grad_norm": 2.935789108276367, "learning_rate": 4.779382560028886e-05, "loss": 4.5473, "step": 4889 }, { "epoch": 0.04414154179454775, "grad_norm": 4.310101509094238, "learning_rate": 4.779337425528074e-05, "loss": 5.1858, "step": 4890 }, { "epoch": 0.044150568694710235, "grad_norm": 3.188997507095337, "learning_rate": 4.779292291027261e-05, "loss": 4.9624, "step": 4891 }, { "epoch": 0.04415959559487272, "grad_norm": 6.1987128257751465, "learning_rate": 4.779247156526449e-05, "loss": 4.4251, "step": 4892 }, { "epoch": 0.044168622495035205, "grad_norm": 3.636007308959961, "learning_rate": 4.779202022025637e-05, "loss": 5.1348, "step": 4893 }, { "epoch": 0.044177649395197686, "grad_norm": 3.657756805419922, "learning_rate": 4.779156887524824e-05, "loss": 4.1019, "step": 4894 }, { "epoch": 0.044186676295360175, "grad_norm": 2.7832841873168945, "learning_rate": 4.779111753024012e-05, "loss": 4.7079, "step": 4895 }, { "epoch": 0.044195703195522656, "grad_norm": 5.004985332489014, "learning_rate": 4.7790666185231994e-05, "loss": 4.2057, "step": 4896 }, { "epoch": 0.044204730095685145, "grad_norm": 4.001626491546631, "learning_rate": 4.7790214840223866e-05, "loss": 4.8625, "step": 4897 }, { "epoch": 0.044213756995847626, "grad_norm": 3.4701087474823, "learning_rate": 4.7789763495215746e-05, "loss": 4.5999, "step": 4898 }, { "epoch": 0.04422278389601011, "grad_norm": 3.0299015045166016, "learning_rate": 4.778931215020762e-05, "loss": 4.911, "step": 4899 }, { "epoch": 0.044231810796172596, "grad_norm": 4.076066017150879, "learning_rate": 4.77888608051995e-05, "loss": 4.6185, "step": 4900 }, { "epoch": 0.04424083769633508, "grad_norm": 3.7186505794525146, "learning_rate": 4.778840946019138e-05, "loss": 5.3193, "step": 4901 }, { "epoch": 0.04424986459649756, "grad_norm": 4.156846046447754, "learning_rate": 4.778795811518324e-05, "loss": 4.6854, "step": 4902 }, { "epoch": 0.04425889149666005, "grad_norm": 3.508045196533203, "learning_rate": 4.778750677017512e-05, "loss": 4.8546, "step": 4903 }, { "epoch": 0.04426791839682253, "grad_norm": 7.056553363800049, "learning_rate": 4.7787055425167e-05, "loss": 5.2545, "step": 4904 }, { "epoch": 0.04427694529698502, "grad_norm": 3.413372039794922, "learning_rate": 4.778660408015887e-05, "loss": 4.3762, "step": 4905 }, { "epoch": 0.0442859721971475, "grad_norm": 3.370190382003784, "learning_rate": 4.778615273515075e-05, "loss": 4.3375, "step": 4906 }, { "epoch": 0.04429499909730998, "grad_norm": 5.47584867477417, "learning_rate": 4.778570139014263e-05, "loss": 4.6331, "step": 4907 }, { "epoch": 0.04430402599747247, "grad_norm": 3.811279773712158, "learning_rate": 4.7785250045134504e-05, "loss": 5.0279, "step": 4908 }, { "epoch": 0.04431305289763495, "grad_norm": 3.742767572402954, "learning_rate": 4.778479870012638e-05, "loss": 4.4861, "step": 4909 }, { "epoch": 0.04432207979779744, "grad_norm": 4.3819427490234375, "learning_rate": 4.7784347355118256e-05, "loss": 3.7745, "step": 4910 }, { "epoch": 0.04433110669795992, "grad_norm": 3.828709125518799, "learning_rate": 4.778389601011013e-05, "loss": 4.9013, "step": 4911 }, { "epoch": 0.0443401335981224, "grad_norm": 2.958676338195801, "learning_rate": 4.778344466510201e-05, "loss": 4.8388, "step": 4912 }, { "epoch": 0.04434916049828489, "grad_norm": 4.511913776397705, "learning_rate": 4.778299332009388e-05, "loss": 5.4436, "step": 4913 }, { "epoch": 0.04435818739844737, "grad_norm": 6.008811950683594, "learning_rate": 4.778254197508576e-05, "loss": 4.9383, "step": 4914 }, { "epoch": 0.044367214298609854, "grad_norm": 3.819075345993042, "learning_rate": 4.778209063007764e-05, "loss": 4.9993, "step": 4915 }, { "epoch": 0.04437624119877234, "grad_norm": 2.5517263412475586, "learning_rate": 4.7781639285069504e-05, "loss": 3.9387, "step": 4916 }, { "epoch": 0.044385268098934824, "grad_norm": 4.242903709411621, "learning_rate": 4.7781187940061384e-05, "loss": 4.6537, "step": 4917 }, { "epoch": 0.04439429499909731, "grad_norm": 3.5228161811828613, "learning_rate": 4.778073659505326e-05, "loss": 4.5941, "step": 4918 }, { "epoch": 0.044403321899259794, "grad_norm": 3.8852055072784424, "learning_rate": 4.7780285250045135e-05, "loss": 4.8987, "step": 4919 }, { "epoch": 0.044412348799422276, "grad_norm": 3.7832276821136475, "learning_rate": 4.7779833905037014e-05, "loss": 4.5182, "step": 4920 }, { "epoch": 0.044421375699584764, "grad_norm": 3.708751916885376, "learning_rate": 4.777938256002889e-05, "loss": 4.5482, "step": 4921 }, { "epoch": 0.044430402599747246, "grad_norm": 4.024532318115234, "learning_rate": 4.7778931215020766e-05, "loss": 4.7881, "step": 4922 }, { "epoch": 0.044439429499909734, "grad_norm": 3.100436210632324, "learning_rate": 4.777847987001264e-05, "loss": 4.7539, "step": 4923 }, { "epoch": 0.044448456400072216, "grad_norm": 3.626936912536621, "learning_rate": 4.777802852500452e-05, "loss": 4.3262, "step": 4924 }, { "epoch": 0.0444574833002347, "grad_norm": 3.714813232421875, "learning_rate": 4.777757717999639e-05, "loss": 4.8858, "step": 4925 }, { "epoch": 0.044466510200397186, "grad_norm": 6.4364542961120605, "learning_rate": 4.777712583498827e-05, "loss": 4.3031, "step": 4926 }, { "epoch": 0.04447553710055967, "grad_norm": 3.7806615829467773, "learning_rate": 4.777667448998014e-05, "loss": 5.4077, "step": 4927 }, { "epoch": 0.04448456400072215, "grad_norm": 4.430490016937256, "learning_rate": 4.777622314497202e-05, "loss": 3.4251, "step": 4928 }, { "epoch": 0.04449359090088464, "grad_norm": 3.6085779666900635, "learning_rate": 4.7775771799963894e-05, "loss": 4.6998, "step": 4929 }, { "epoch": 0.04450261780104712, "grad_norm": 3.095247507095337, "learning_rate": 4.7775320454955766e-05, "loss": 4.2159, "step": 4930 }, { "epoch": 0.04451164470120961, "grad_norm": 3.2801804542541504, "learning_rate": 4.7774869109947645e-05, "loss": 4.5239, "step": 4931 }, { "epoch": 0.04452067160137209, "grad_norm": 2.92525577545166, "learning_rate": 4.7774417764939525e-05, "loss": 4.7912, "step": 4932 }, { "epoch": 0.04452969850153457, "grad_norm": 4.430853366851807, "learning_rate": 4.77739664199314e-05, "loss": 5.1199, "step": 4933 }, { "epoch": 0.04453872540169706, "grad_norm": 4.674463748931885, "learning_rate": 4.7773515074923276e-05, "loss": 4.6373, "step": 4934 }, { "epoch": 0.04454775230185954, "grad_norm": 4.120530128479004, "learning_rate": 4.777306372991515e-05, "loss": 5.3035, "step": 4935 }, { "epoch": 0.04455677920202203, "grad_norm": 3.0853824615478516, "learning_rate": 4.777261238490702e-05, "loss": 5.2564, "step": 4936 }, { "epoch": 0.04456580610218451, "grad_norm": 4.637140274047852, "learning_rate": 4.77721610398989e-05, "loss": 4.7205, "step": 4937 }, { "epoch": 0.04457483300234699, "grad_norm": 3.4837663173675537, "learning_rate": 4.777170969489077e-05, "loss": 4.8569, "step": 4938 }, { "epoch": 0.04458385990250948, "grad_norm": 5.277425765991211, "learning_rate": 4.777125834988265e-05, "loss": 3.7531, "step": 4939 }, { "epoch": 0.04459288680267196, "grad_norm": 4.118377685546875, "learning_rate": 4.777080700487453e-05, "loss": 4.809, "step": 4940 }, { "epoch": 0.04460191370283444, "grad_norm": 3.9472973346710205, "learning_rate": 4.7770355659866404e-05, "loss": 5.199, "step": 4941 }, { "epoch": 0.04461094060299693, "grad_norm": 3.7721898555755615, "learning_rate": 4.776990431485828e-05, "loss": 4.6581, "step": 4942 }, { "epoch": 0.04461996750315941, "grad_norm": 3.1755363941192627, "learning_rate": 4.7769452969850156e-05, "loss": 3.9318, "step": 4943 }, { "epoch": 0.0446289944033219, "grad_norm": 4.687845230102539, "learning_rate": 4.776900162484203e-05, "loss": 4.6492, "step": 4944 }, { "epoch": 0.04463802130348438, "grad_norm": 2.9942591190338135, "learning_rate": 4.776855027983391e-05, "loss": 5.0983, "step": 4945 }, { "epoch": 0.044647048203646865, "grad_norm": 3.27264666557312, "learning_rate": 4.776809893482579e-05, "loss": 4.8805, "step": 4946 }, { "epoch": 0.04465607510380935, "grad_norm": 3.381481409072876, "learning_rate": 4.776764758981766e-05, "loss": 4.9836, "step": 4947 }, { "epoch": 0.044665102003971835, "grad_norm": 3.872251033782959, "learning_rate": 4.776719624480954e-05, "loss": 4.1193, "step": 4948 }, { "epoch": 0.04467412890413432, "grad_norm": 3.0358667373657227, "learning_rate": 4.776674489980141e-05, "loss": 3.9011, "step": 4949 }, { "epoch": 0.044683155804296805, "grad_norm": 3.3345017433166504, "learning_rate": 4.776629355479328e-05, "loss": 4.2591, "step": 4950 }, { "epoch": 0.044692182704459286, "grad_norm": 4.292920112609863, "learning_rate": 4.776584220978516e-05, "loss": 5.3563, "step": 4951 }, { "epoch": 0.044701209604621775, "grad_norm": 3.0928707122802734, "learning_rate": 4.7765390864777035e-05, "loss": 4.9696, "step": 4952 }, { "epoch": 0.044710236504784256, "grad_norm": 3.98934006690979, "learning_rate": 4.7764939519768914e-05, "loss": 5.2478, "step": 4953 }, { "epoch": 0.04471926340494674, "grad_norm": 2.8504951000213623, "learning_rate": 4.7764488174760793e-05, "loss": 5.333, "step": 4954 }, { "epoch": 0.044728290305109226, "grad_norm": 3.2884602546691895, "learning_rate": 4.7764036829752666e-05, "loss": 4.6251, "step": 4955 }, { "epoch": 0.04473731720527171, "grad_norm": 3.138282299041748, "learning_rate": 4.776358548474454e-05, "loss": 4.3391, "step": 4956 }, { "epoch": 0.044746344105434196, "grad_norm": 3.1133127212524414, "learning_rate": 4.776313413973642e-05, "loss": 5.0986, "step": 4957 }, { "epoch": 0.04475537100559668, "grad_norm": 3.6982293128967285, "learning_rate": 4.776268279472829e-05, "loss": 4.2173, "step": 4958 }, { "epoch": 0.04476439790575916, "grad_norm": 4.65761137008667, "learning_rate": 4.776223144972017e-05, "loss": 4.6874, "step": 4959 }, { "epoch": 0.04477342480592165, "grad_norm": 3.618288516998291, "learning_rate": 4.776178010471204e-05, "loss": 4.5718, "step": 4960 }, { "epoch": 0.04478245170608413, "grad_norm": 3.4906375408172607, "learning_rate": 4.776132875970392e-05, "loss": 4.4909, "step": 4961 }, { "epoch": 0.04479147860624662, "grad_norm": 2.9560530185699463, "learning_rate": 4.77608774146958e-05, "loss": 4.7897, "step": 4962 }, { "epoch": 0.0448005055064091, "grad_norm": 3.516246795654297, "learning_rate": 4.7760426069687666e-05, "loss": 4.5187, "step": 4963 }, { "epoch": 0.04480953240657158, "grad_norm": 3.488327741622925, "learning_rate": 4.7759974724679545e-05, "loss": 5.0909, "step": 4964 }, { "epoch": 0.04481855930673407, "grad_norm": 2.9749507904052734, "learning_rate": 4.7759523379671424e-05, "loss": 4.9274, "step": 4965 }, { "epoch": 0.04482758620689655, "grad_norm": 3.6972451210021973, "learning_rate": 4.77590720346633e-05, "loss": 4.9571, "step": 4966 }, { "epoch": 0.04483661310705903, "grad_norm": 2.6082987785339355, "learning_rate": 4.7758620689655176e-05, "loss": 4.9502, "step": 4967 }, { "epoch": 0.04484564000722152, "grad_norm": 3.7375786304473877, "learning_rate": 4.7758169344647055e-05, "loss": 4.933, "step": 4968 }, { "epoch": 0.044854666907384, "grad_norm": 4.1620097160339355, "learning_rate": 4.775771799963893e-05, "loss": 4.658, "step": 4969 }, { "epoch": 0.04486369380754649, "grad_norm": 4.166549205780029, "learning_rate": 4.77572666546308e-05, "loss": 3.0356, "step": 4970 }, { "epoch": 0.04487272070770897, "grad_norm": 3.578346014022827, "learning_rate": 4.775681530962268e-05, "loss": 4.3847, "step": 4971 }, { "epoch": 0.044881747607871454, "grad_norm": 3.8035738468170166, "learning_rate": 4.775636396461455e-05, "loss": 4.4667, "step": 4972 }, { "epoch": 0.04489077450803394, "grad_norm": 4.655449867248535, "learning_rate": 4.775591261960643e-05, "loss": 4.8934, "step": 4973 }, { "epoch": 0.044899801408196424, "grad_norm": 3.8851895332336426, "learning_rate": 4.7755461274598304e-05, "loss": 4.5843, "step": 4974 }, { "epoch": 0.04490882830835891, "grad_norm": 3.1410064697265625, "learning_rate": 4.775500992959018e-05, "loss": 4.6606, "step": 4975 }, { "epoch": 0.044917855208521394, "grad_norm": 3.383068084716797, "learning_rate": 4.7754558584582055e-05, "loss": 5.0636, "step": 4976 }, { "epoch": 0.044926882108683876, "grad_norm": 3.9749209880828857, "learning_rate": 4.775410723957393e-05, "loss": 5.0597, "step": 4977 }, { "epoch": 0.044935909008846364, "grad_norm": 3.987271785736084, "learning_rate": 4.775365589456581e-05, "loss": 5.1195, "step": 4978 }, { "epoch": 0.044944935909008846, "grad_norm": 3.863375663757324, "learning_rate": 4.7753204549557686e-05, "loss": 4.8892, "step": 4979 }, { "epoch": 0.04495396280917133, "grad_norm": 3.8785338401794434, "learning_rate": 4.775275320454956e-05, "loss": 5.5781, "step": 4980 }, { "epoch": 0.044962989709333816, "grad_norm": 3.919130325317383, "learning_rate": 4.775230185954144e-05, "loss": 4.8616, "step": 4981 }, { "epoch": 0.0449720166094963, "grad_norm": 4.63123893737793, "learning_rate": 4.775185051453331e-05, "loss": 4.9969, "step": 4982 }, { "epoch": 0.044981043509658786, "grad_norm": 3.6930153369903564, "learning_rate": 4.775139916952518e-05, "loss": 4.7802, "step": 4983 }, { "epoch": 0.04499007040982127, "grad_norm": 3.682123899459839, "learning_rate": 4.775094782451706e-05, "loss": 5.0697, "step": 4984 }, { "epoch": 0.04499909730998375, "grad_norm": 3.580629587173462, "learning_rate": 4.775049647950894e-05, "loss": 4.4278, "step": 4985 }, { "epoch": 0.04500812421014624, "grad_norm": 3.7862772941589355, "learning_rate": 4.7750045134500814e-05, "loss": 4.7837, "step": 4986 }, { "epoch": 0.04501715111030872, "grad_norm": 3.1781005859375, "learning_rate": 4.774959378949269e-05, "loss": 4.4462, "step": 4987 }, { "epoch": 0.04502617801047121, "grad_norm": 2.4100921154022217, "learning_rate": 4.7749142444484566e-05, "loss": 4.4607, "step": 4988 }, { "epoch": 0.04503520491063369, "grad_norm": 3.8346288204193115, "learning_rate": 4.7748691099476445e-05, "loss": 4.4272, "step": 4989 }, { "epoch": 0.04504423181079617, "grad_norm": 3.8583438396453857, "learning_rate": 4.774823975446832e-05, "loss": 4.8766, "step": 4990 }, { "epoch": 0.04505325871095866, "grad_norm": 3.790656089782715, "learning_rate": 4.774778840946019e-05, "loss": 4.3326, "step": 4991 }, { "epoch": 0.04506228561112114, "grad_norm": 2.953552007675171, "learning_rate": 4.774733706445207e-05, "loss": 4.9238, "step": 4992 }, { "epoch": 0.04507131251128362, "grad_norm": 4.180334091186523, "learning_rate": 4.774688571944395e-05, "loss": 4.0806, "step": 4993 }, { "epoch": 0.04508033941144611, "grad_norm": 3.810128688812256, "learning_rate": 4.774643437443582e-05, "loss": 4.1064, "step": 4994 }, { "epoch": 0.04508936631160859, "grad_norm": 3.301218032836914, "learning_rate": 4.77459830294277e-05, "loss": 4.7672, "step": 4995 }, { "epoch": 0.04509839321177108, "grad_norm": 2.8940093517303467, "learning_rate": 4.774553168441957e-05, "loss": 5.0771, "step": 4996 }, { "epoch": 0.04510742011193356, "grad_norm": 4.10625696182251, "learning_rate": 4.7745080339411445e-05, "loss": 5.5759, "step": 4997 }, { "epoch": 0.04511644701209604, "grad_norm": 4.732582092285156, "learning_rate": 4.7744628994403324e-05, "loss": 5.1823, "step": 4998 }, { "epoch": 0.04512547391225853, "grad_norm": 3.3453896045684814, "learning_rate": 4.77441776493952e-05, "loss": 4.5024, "step": 4999 }, { "epoch": 0.04513450081242101, "grad_norm": 2.9820189476013184, "learning_rate": 4.7743726304387076e-05, "loss": 4.8382, "step": 5000 }, { "epoch": 0.0451435277125835, "grad_norm": 3.2561333179473877, "learning_rate": 4.7743274959378955e-05, "loss": 4.179, "step": 5001 }, { "epoch": 0.04515255461274598, "grad_norm": 3.309999704360962, "learning_rate": 4.774282361437083e-05, "loss": 4.4169, "step": 5002 }, { "epoch": 0.045161581512908465, "grad_norm": 4.588602542877197, "learning_rate": 4.77423722693627e-05, "loss": 5.3687, "step": 5003 }, { "epoch": 0.04517060841307095, "grad_norm": 5.6903157234191895, "learning_rate": 4.774192092435458e-05, "loss": 3.6535, "step": 5004 }, { "epoch": 0.045179635313233435, "grad_norm": 3.0954091548919678, "learning_rate": 4.774146957934645e-05, "loss": 4.7516, "step": 5005 }, { "epoch": 0.045188662213395916, "grad_norm": 3.298279047012329, "learning_rate": 4.774101823433833e-05, "loss": 4.2394, "step": 5006 }, { "epoch": 0.045197689113558405, "grad_norm": 4.8819169998168945, "learning_rate": 4.774056688933021e-05, "loss": 3.2131, "step": 5007 }, { "epoch": 0.045206716013720886, "grad_norm": 3.55344295501709, "learning_rate": 4.774011554432208e-05, "loss": 4.9834, "step": 5008 }, { "epoch": 0.045215742913883375, "grad_norm": 4.355927467346191, "learning_rate": 4.773966419931396e-05, "loss": 5.218, "step": 5009 }, { "epoch": 0.045224769814045856, "grad_norm": 2.9401772022247314, "learning_rate": 4.7739212854305834e-05, "loss": 5.1254, "step": 5010 }, { "epoch": 0.04523379671420834, "grad_norm": 3.2130303382873535, "learning_rate": 4.773876150929771e-05, "loss": 4.1456, "step": 5011 }, { "epoch": 0.045242823614370827, "grad_norm": 3.3984501361846924, "learning_rate": 4.7738310164289586e-05, "loss": 5.1051, "step": 5012 }, { "epoch": 0.04525185051453331, "grad_norm": 3.7179183959960938, "learning_rate": 4.773785881928146e-05, "loss": 4.1078, "step": 5013 }, { "epoch": 0.045260877414695797, "grad_norm": 3.2927515506744385, "learning_rate": 4.773740747427334e-05, "loss": 3.76, "step": 5014 }, { "epoch": 0.04526990431485828, "grad_norm": 3.631371021270752, "learning_rate": 4.773695612926522e-05, "loss": 5.1111, "step": 5015 }, { "epoch": 0.04527893121502076, "grad_norm": 4.075905799865723, "learning_rate": 4.773650478425709e-05, "loss": 4.3826, "step": 5016 }, { "epoch": 0.04528795811518325, "grad_norm": 3.3886072635650635, "learning_rate": 4.773605343924896e-05, "loss": 5.1511, "step": 5017 }, { "epoch": 0.04529698501534573, "grad_norm": 3.7511708736419678, "learning_rate": 4.773560209424084e-05, "loss": 4.3752, "step": 5018 }, { "epoch": 0.04530601191550821, "grad_norm": 2.8630006313323975, "learning_rate": 4.7735150749232714e-05, "loss": 4.1948, "step": 5019 }, { "epoch": 0.0453150388156707, "grad_norm": 3.470426082611084, "learning_rate": 4.773469940422459e-05, "loss": 4.6653, "step": 5020 }, { "epoch": 0.04532406571583318, "grad_norm": 3.310028314590454, "learning_rate": 4.7734248059216465e-05, "loss": 4.5148, "step": 5021 }, { "epoch": 0.04533309261599567, "grad_norm": 3.8053507804870605, "learning_rate": 4.7733796714208345e-05, "loss": 4.4552, "step": 5022 }, { "epoch": 0.04534211951615815, "grad_norm": 3.6311957836151123, "learning_rate": 4.773334536920022e-05, "loss": 4.8017, "step": 5023 }, { "epoch": 0.04535114641632063, "grad_norm": 3.6809427738189697, "learning_rate": 4.773289402419209e-05, "loss": 5.2251, "step": 5024 }, { "epoch": 0.04536017331648312, "grad_norm": 3.0517525672912598, "learning_rate": 4.773244267918397e-05, "loss": 4.7514, "step": 5025 }, { "epoch": 0.0453692002166456, "grad_norm": 3.8027312755584717, "learning_rate": 4.773199133417585e-05, "loss": 5.6883, "step": 5026 }, { "epoch": 0.04537822711680809, "grad_norm": 3.956774950027466, "learning_rate": 4.773153998916772e-05, "loss": 4.8105, "step": 5027 }, { "epoch": 0.04538725401697057, "grad_norm": 3.331444501876831, "learning_rate": 4.77310886441596e-05, "loss": 5.3217, "step": 5028 }, { "epoch": 0.045396280917133054, "grad_norm": 4.142138481140137, "learning_rate": 4.773063729915148e-05, "loss": 5.1125, "step": 5029 }, { "epoch": 0.04540530781729554, "grad_norm": 3.6328489780426025, "learning_rate": 4.7730185954143345e-05, "loss": 5.2322, "step": 5030 }, { "epoch": 0.045414334717458024, "grad_norm": 3.776442766189575, "learning_rate": 4.7729734609135224e-05, "loss": 4.0723, "step": 5031 }, { "epoch": 0.045423361617620506, "grad_norm": 4.501509666442871, "learning_rate": 4.77292832641271e-05, "loss": 3.4323, "step": 5032 }, { "epoch": 0.045432388517782994, "grad_norm": 4.199761390686035, "learning_rate": 4.7728831919118976e-05, "loss": 5.477, "step": 5033 }, { "epoch": 0.045441415417945476, "grad_norm": 2.2894339561462402, "learning_rate": 4.7728380574110855e-05, "loss": 4.7429, "step": 5034 }, { "epoch": 0.045450442318107964, "grad_norm": 4.746763229370117, "learning_rate": 4.772792922910273e-05, "loss": 4.072, "step": 5035 }, { "epoch": 0.045459469218270446, "grad_norm": 4.184184551239014, "learning_rate": 4.772747788409461e-05, "loss": 4.2554, "step": 5036 }, { "epoch": 0.04546849611843293, "grad_norm": 3.728641986846924, "learning_rate": 4.772702653908648e-05, "loss": 4.5156, "step": 5037 }, { "epoch": 0.045477523018595416, "grad_norm": 5.549261569976807, "learning_rate": 4.772657519407835e-05, "loss": 4.7124, "step": 5038 }, { "epoch": 0.0454865499187579, "grad_norm": 4.294014930725098, "learning_rate": 4.772612384907023e-05, "loss": 3.6253, "step": 5039 }, { "epoch": 0.045495576818920386, "grad_norm": 3.8534834384918213, "learning_rate": 4.772567250406211e-05, "loss": 4.8851, "step": 5040 }, { "epoch": 0.04550460371908287, "grad_norm": 4.411771297454834, "learning_rate": 4.772522115905398e-05, "loss": 3.7924, "step": 5041 }, { "epoch": 0.04551363061924535, "grad_norm": 4.236968517303467, "learning_rate": 4.772476981404586e-05, "loss": 4.3267, "step": 5042 }, { "epoch": 0.04552265751940784, "grad_norm": 3.47987699508667, "learning_rate": 4.7724318469037734e-05, "loss": 4.9373, "step": 5043 }, { "epoch": 0.04553168441957032, "grad_norm": 3.6703739166259766, "learning_rate": 4.772386712402961e-05, "loss": 5.0293, "step": 5044 }, { "epoch": 0.0455407113197328, "grad_norm": 2.965055227279663, "learning_rate": 4.7723415779021486e-05, "loss": 4.1855, "step": 5045 }, { "epoch": 0.04554973821989529, "grad_norm": 6.501061916351318, "learning_rate": 4.7722964434013365e-05, "loss": 4.9507, "step": 5046 }, { "epoch": 0.04555876512005777, "grad_norm": 4.6574883460998535, "learning_rate": 4.772251308900524e-05, "loss": 5.2521, "step": 5047 }, { "epoch": 0.04556779202022026, "grad_norm": 3.739501953125, "learning_rate": 4.772206174399712e-05, "loss": 4.4057, "step": 5048 }, { "epoch": 0.04557681892038274, "grad_norm": 4.2252068519592285, "learning_rate": 4.772161039898899e-05, "loss": 4.8987, "step": 5049 }, { "epoch": 0.04558584582054522, "grad_norm": 3.386514663696289, "learning_rate": 4.772115905398086e-05, "loss": 4.2176, "step": 5050 }, { "epoch": 0.04559487272070771, "grad_norm": 3.131146192550659, "learning_rate": 4.772070770897274e-05, "loss": 5.0522, "step": 5051 }, { "epoch": 0.04560389962087019, "grad_norm": 4.360264778137207, "learning_rate": 4.7720256363964614e-05, "loss": 4.8652, "step": 5052 }, { "epoch": 0.04561292652103268, "grad_norm": 3.7042691707611084, "learning_rate": 4.771980501895649e-05, "loss": 4.3954, "step": 5053 }, { "epoch": 0.04562195342119516, "grad_norm": 2.684730052947998, "learning_rate": 4.771935367394837e-05, "loss": 5.1612, "step": 5054 }, { "epoch": 0.04563098032135764, "grad_norm": 4.801504611968994, "learning_rate": 4.7718902328940244e-05, "loss": 4.3654, "step": 5055 }, { "epoch": 0.04564000722152013, "grad_norm": 3.2595462799072266, "learning_rate": 4.7718450983932124e-05, "loss": 4.9618, "step": 5056 }, { "epoch": 0.04564903412168261, "grad_norm": 3.6095099449157715, "learning_rate": 4.7717999638923996e-05, "loss": 4.9377, "step": 5057 }, { "epoch": 0.045658061021845095, "grad_norm": 3.226149797439575, "learning_rate": 4.771754829391587e-05, "loss": 4.7589, "step": 5058 }, { "epoch": 0.04566708792200758, "grad_norm": 3.7641966342926025, "learning_rate": 4.771709694890775e-05, "loss": 4.5534, "step": 5059 }, { "epoch": 0.045676114822170065, "grad_norm": 3.017996311187744, "learning_rate": 4.771664560389962e-05, "loss": 4.8923, "step": 5060 }, { "epoch": 0.04568514172233255, "grad_norm": 4.075728416442871, "learning_rate": 4.77161942588915e-05, "loss": 5.408, "step": 5061 }, { "epoch": 0.045694168622495035, "grad_norm": 2.9722628593444824, "learning_rate": 4.771574291388338e-05, "loss": 4.6014, "step": 5062 }, { "epoch": 0.045703195522657517, "grad_norm": 4.065033912658691, "learning_rate": 4.7715291568875245e-05, "loss": 4.3348, "step": 5063 }, { "epoch": 0.045712222422820005, "grad_norm": 3.4290013313293457, "learning_rate": 4.7714840223867124e-05, "loss": 5.16, "step": 5064 }, { "epoch": 0.045721249322982487, "grad_norm": 2.928220510482788, "learning_rate": 4.7714388878859e-05, "loss": 4.5387, "step": 5065 }, { "epoch": 0.045730276223144975, "grad_norm": 3.6894350051879883, "learning_rate": 4.7713937533850875e-05, "loss": 5.4054, "step": 5066 }, { "epoch": 0.045739303123307457, "grad_norm": 3.1101717948913574, "learning_rate": 4.7713486188842755e-05, "loss": 4.7997, "step": 5067 }, { "epoch": 0.04574833002346994, "grad_norm": 3.065674066543579, "learning_rate": 4.7713034843834634e-05, "loss": 5.052, "step": 5068 }, { "epoch": 0.045757356923632427, "grad_norm": 3.6480960845947266, "learning_rate": 4.7712583498826506e-05, "loss": 4.6907, "step": 5069 }, { "epoch": 0.04576638382379491, "grad_norm": 3.581719398498535, "learning_rate": 4.771213215381838e-05, "loss": 4.4821, "step": 5070 }, { "epoch": 0.04577541072395739, "grad_norm": 2.969400405883789, "learning_rate": 4.771168080881026e-05, "loss": 4.5557, "step": 5071 }, { "epoch": 0.04578443762411988, "grad_norm": 2.9545915126800537, "learning_rate": 4.771122946380213e-05, "loss": 5.0487, "step": 5072 }, { "epoch": 0.04579346452428236, "grad_norm": 3.745962619781494, "learning_rate": 4.771077811879401e-05, "loss": 4.6721, "step": 5073 }, { "epoch": 0.04580249142444485, "grad_norm": 4.1580071449279785, "learning_rate": 4.771032677378588e-05, "loss": 4.8788, "step": 5074 }, { "epoch": 0.04581151832460733, "grad_norm": 3.0749099254608154, "learning_rate": 4.770987542877776e-05, "loss": 4.8711, "step": 5075 }, { "epoch": 0.04582054522476981, "grad_norm": 3.316472053527832, "learning_rate": 4.770942408376964e-05, "loss": 4.7357, "step": 5076 }, { "epoch": 0.0458295721249323, "grad_norm": 3.3936688899993896, "learning_rate": 4.7708972738761506e-05, "loss": 4.5714, "step": 5077 }, { "epoch": 0.04583859902509478, "grad_norm": 3.3571279048919678, "learning_rate": 4.7708521393753386e-05, "loss": 4.4824, "step": 5078 }, { "epoch": 0.04584762592525727, "grad_norm": 3.259596586227417, "learning_rate": 4.7708070048745265e-05, "loss": 4.5635, "step": 5079 }, { "epoch": 0.04585665282541975, "grad_norm": 4.066307544708252, "learning_rate": 4.770761870373714e-05, "loss": 3.3774, "step": 5080 }, { "epoch": 0.04586567972558223, "grad_norm": 3.022252082824707, "learning_rate": 4.770716735872902e-05, "loss": 4.8076, "step": 5081 }, { "epoch": 0.04587470662574472, "grad_norm": 2.817779302597046, "learning_rate": 4.770671601372089e-05, "loss": 4.3849, "step": 5082 }, { "epoch": 0.0458837335259072, "grad_norm": 3.1523525714874268, "learning_rate": 4.770626466871277e-05, "loss": 4.6373, "step": 5083 }, { "epoch": 0.04589276042606969, "grad_norm": 2.971147298812866, "learning_rate": 4.770581332370464e-05, "loss": 5.355, "step": 5084 }, { "epoch": 0.04590178732623217, "grad_norm": 3.3712832927703857, "learning_rate": 4.770536197869652e-05, "loss": 5.1315, "step": 5085 }, { "epoch": 0.045910814226394654, "grad_norm": 4.9401702880859375, "learning_rate": 4.770491063368839e-05, "loss": 4.319, "step": 5086 }, { "epoch": 0.04591984112655714, "grad_norm": 2.607349395751953, "learning_rate": 4.770445928868027e-05, "loss": 4.4428, "step": 5087 }, { "epoch": 0.045928868026719624, "grad_norm": 3.1599323749542236, "learning_rate": 4.7704007943672144e-05, "loss": 4.5187, "step": 5088 }, { "epoch": 0.045937894926882106, "grad_norm": 3.558528184890747, "learning_rate": 4.7703556598664023e-05, "loss": 4.6131, "step": 5089 }, { "epoch": 0.045946921827044594, "grad_norm": 4.450822830200195, "learning_rate": 4.7703105253655896e-05, "loss": 4.7715, "step": 5090 }, { "epoch": 0.045955948727207076, "grad_norm": 3.2851853370666504, "learning_rate": 4.770265390864777e-05, "loss": 4.718, "step": 5091 }, { "epoch": 0.045964975627369564, "grad_norm": 3.1391654014587402, "learning_rate": 4.770220256363965e-05, "loss": 4.6422, "step": 5092 }, { "epoch": 0.045974002527532046, "grad_norm": 4.0761823654174805, "learning_rate": 4.770175121863153e-05, "loss": 5.0628, "step": 5093 }, { "epoch": 0.04598302942769453, "grad_norm": 4.6555867195129395, "learning_rate": 4.77012998736234e-05, "loss": 4.6044, "step": 5094 }, { "epoch": 0.045992056327857016, "grad_norm": 3.4405100345611572, "learning_rate": 4.770084852861528e-05, "loss": 5.0234, "step": 5095 }, { "epoch": 0.0460010832280195, "grad_norm": 2.8940746784210205, "learning_rate": 4.770039718360715e-05, "loss": 5.044, "step": 5096 }, { "epoch": 0.046010110128181986, "grad_norm": 3.8011562824249268, "learning_rate": 4.7699945838599023e-05, "loss": 5.1063, "step": 5097 }, { "epoch": 0.04601913702834447, "grad_norm": 3.681957483291626, "learning_rate": 4.76994944935909e-05, "loss": 4.7561, "step": 5098 }, { "epoch": 0.04602816392850695, "grad_norm": 2.761226177215576, "learning_rate": 4.7699043148582775e-05, "loss": 4.8898, "step": 5099 }, { "epoch": 0.04603719082866944, "grad_norm": 3.0009844303131104, "learning_rate": 4.7698591803574654e-05, "loss": 4.7706, "step": 5100 }, { "epoch": 0.04604621772883192, "grad_norm": 4.447072505950928, "learning_rate": 4.7698140458566534e-05, "loss": 4.4612, "step": 5101 }, { "epoch": 0.0460552446289944, "grad_norm": 4.101443290710449, "learning_rate": 4.7697689113558406e-05, "loss": 4.3299, "step": 5102 }, { "epoch": 0.04606427152915689, "grad_norm": 3.1199493408203125, "learning_rate": 4.7697237768550285e-05, "loss": 4.4241, "step": 5103 }, { "epoch": 0.04607329842931937, "grad_norm": 3.4484810829162598, "learning_rate": 4.769678642354216e-05, "loss": 4.562, "step": 5104 }, { "epoch": 0.04608232532948186, "grad_norm": 2.68536114692688, "learning_rate": 4.769633507853403e-05, "loss": 5.192, "step": 5105 }, { "epoch": 0.04609135222964434, "grad_norm": 3.55033278465271, "learning_rate": 4.769588373352591e-05, "loss": 4.3727, "step": 5106 }, { "epoch": 0.04610037912980682, "grad_norm": 4.8742780685424805, "learning_rate": 4.769543238851779e-05, "loss": 4.3262, "step": 5107 }, { "epoch": 0.04610940602996931, "grad_norm": 2.7984492778778076, "learning_rate": 4.769498104350966e-05, "loss": 4.768, "step": 5108 }, { "epoch": 0.04611843293013179, "grad_norm": 4.423396587371826, "learning_rate": 4.769452969850154e-05, "loss": 4.581, "step": 5109 }, { "epoch": 0.04612745983029428, "grad_norm": 4.7924370765686035, "learning_rate": 4.769407835349341e-05, "loss": 3.1931, "step": 5110 }, { "epoch": 0.04613648673045676, "grad_norm": 3.5781986713409424, "learning_rate": 4.7693627008485285e-05, "loss": 5.0303, "step": 5111 }, { "epoch": 0.04614551363061924, "grad_norm": 3.3557703495025635, "learning_rate": 4.7693175663477165e-05, "loss": 4.9523, "step": 5112 }, { "epoch": 0.04615454053078173, "grad_norm": 2.254403591156006, "learning_rate": 4.769272431846904e-05, "loss": 5.2796, "step": 5113 }, { "epoch": 0.04616356743094421, "grad_norm": 4.250800132751465, "learning_rate": 4.7692272973460916e-05, "loss": 4.0799, "step": 5114 }, { "epoch": 0.046172594331106695, "grad_norm": 3.9323134422302246, "learning_rate": 4.7691821628452796e-05, "loss": 4.8192, "step": 5115 }, { "epoch": 0.04618162123126918, "grad_norm": 3.8565940856933594, "learning_rate": 4.769137028344467e-05, "loss": 4.1668, "step": 5116 }, { "epoch": 0.046190648131431665, "grad_norm": 3.925431251525879, "learning_rate": 4.769091893843654e-05, "loss": 4.0701, "step": 5117 }, { "epoch": 0.04619967503159415, "grad_norm": 5.025015830993652, "learning_rate": 4.769046759342842e-05, "loss": 3.6738, "step": 5118 }, { "epoch": 0.046208701931756635, "grad_norm": 6.141086101531982, "learning_rate": 4.769001624842029e-05, "loss": 4.0925, "step": 5119 }, { "epoch": 0.046217728831919117, "grad_norm": 2.923752546310425, "learning_rate": 4.768956490341217e-05, "loss": 4.4888, "step": 5120 }, { "epoch": 0.046226755732081605, "grad_norm": 5.434839248657227, "learning_rate": 4.7689113558404044e-05, "loss": 5.0761, "step": 5121 }, { "epoch": 0.046235782632244087, "grad_norm": 2.9049875736236572, "learning_rate": 4.768866221339592e-05, "loss": 4.8299, "step": 5122 }, { "epoch": 0.046244809532406575, "grad_norm": 4.713476657867432, "learning_rate": 4.76882108683878e-05, "loss": 4.7068, "step": 5123 }, { "epoch": 0.046253836432569057, "grad_norm": 3.92314076423645, "learning_rate": 4.768775952337967e-05, "loss": 5.1037, "step": 5124 }, { "epoch": 0.04626286333273154, "grad_norm": 4.7369184494018555, "learning_rate": 4.768730817837155e-05, "loss": 4.5211, "step": 5125 }, { "epoch": 0.046271890232894027, "grad_norm": 2.3683550357818604, "learning_rate": 4.768685683336343e-05, "loss": 3.9392, "step": 5126 }, { "epoch": 0.04628091713305651, "grad_norm": 3.3764495849609375, "learning_rate": 4.76864054883553e-05, "loss": 4.4488, "step": 5127 }, { "epoch": 0.04628994403321899, "grad_norm": 4.040072441101074, "learning_rate": 4.768595414334718e-05, "loss": 4.546, "step": 5128 }, { "epoch": 0.04629897093338148, "grad_norm": 2.9948697090148926, "learning_rate": 4.768550279833906e-05, "loss": 4.6027, "step": 5129 }, { "epoch": 0.04630799783354396, "grad_norm": 3.09934663772583, "learning_rate": 4.768505145333093e-05, "loss": 4.5527, "step": 5130 }, { "epoch": 0.04631702473370645, "grad_norm": 3.9233222007751465, "learning_rate": 4.76846001083228e-05, "loss": 4.0137, "step": 5131 }, { "epoch": 0.04632605163386893, "grad_norm": 3.3657455444335938, "learning_rate": 4.768414876331468e-05, "loss": 4.8626, "step": 5132 }, { "epoch": 0.04633507853403141, "grad_norm": 3.1662652492523193, "learning_rate": 4.7683697418306554e-05, "loss": 5.0949, "step": 5133 }, { "epoch": 0.0463441054341939, "grad_norm": 3.5861072540283203, "learning_rate": 4.7683246073298433e-05, "loss": 4.6047, "step": 5134 }, { "epoch": 0.04635313233435638, "grad_norm": 3.920414686203003, "learning_rate": 4.7682794728290306e-05, "loss": 5.0076, "step": 5135 }, { "epoch": 0.04636215923451887, "grad_norm": 2.765810966491699, "learning_rate": 4.7682343383282185e-05, "loss": 5.0022, "step": 5136 }, { "epoch": 0.04637118613468135, "grad_norm": 3.168856143951416, "learning_rate": 4.768189203827406e-05, "loss": 4.7054, "step": 5137 }, { "epoch": 0.04638021303484383, "grad_norm": 3.2771146297454834, "learning_rate": 4.768144069326593e-05, "loss": 4.918, "step": 5138 }, { "epoch": 0.04638923993500632, "grad_norm": 2.673640012741089, "learning_rate": 4.768098934825781e-05, "loss": 5.0791, "step": 5139 }, { "epoch": 0.0463982668351688, "grad_norm": 3.2048397064208984, "learning_rate": 4.768053800324969e-05, "loss": 4.3625, "step": 5140 }, { "epoch": 0.046407293735331284, "grad_norm": 4.326404571533203, "learning_rate": 4.768008665824156e-05, "loss": 4.551, "step": 5141 }, { "epoch": 0.04641632063549377, "grad_norm": 2.320584297180176, "learning_rate": 4.767963531323344e-05, "loss": 5.0479, "step": 5142 }, { "epoch": 0.046425347535656254, "grad_norm": 3.1983425617218018, "learning_rate": 4.767918396822531e-05, "loss": 4.9485, "step": 5143 }, { "epoch": 0.04643437443581874, "grad_norm": 3.317584276199341, "learning_rate": 4.7678732623217185e-05, "loss": 4.3832, "step": 5144 }, { "epoch": 0.046443401335981224, "grad_norm": 4.087357521057129, "learning_rate": 4.7678281278209064e-05, "loss": 4.7463, "step": 5145 }, { "epoch": 0.046452428236143706, "grad_norm": 4.442031383514404, "learning_rate": 4.7677829933200944e-05, "loss": 3.8292, "step": 5146 }, { "epoch": 0.046461455136306194, "grad_norm": 3.9371707439422607, "learning_rate": 4.7677378588192816e-05, "loss": 4.5944, "step": 5147 }, { "epoch": 0.046470482036468676, "grad_norm": 2.979633092880249, "learning_rate": 4.7676927243184695e-05, "loss": 5.2078, "step": 5148 }, { "epoch": 0.046479508936631164, "grad_norm": 3.4129834175109863, "learning_rate": 4.767647589817657e-05, "loss": 5.3151, "step": 5149 }, { "epoch": 0.046488535836793646, "grad_norm": 4.018768310546875, "learning_rate": 4.767602455316845e-05, "loss": 4.8363, "step": 5150 }, { "epoch": 0.04649756273695613, "grad_norm": 2.6706416606903076, "learning_rate": 4.767557320816032e-05, "loss": 4.1964, "step": 5151 }, { "epoch": 0.046506589637118616, "grad_norm": 4.695626258850098, "learning_rate": 4.767512186315219e-05, "loss": 4.7997, "step": 5152 }, { "epoch": 0.0465156165372811, "grad_norm": 3.104442834854126, "learning_rate": 4.767467051814407e-05, "loss": 4.4712, "step": 5153 }, { "epoch": 0.04652464343744358, "grad_norm": 3.2195560932159424, "learning_rate": 4.767421917313595e-05, "loss": 4.5265, "step": 5154 }, { "epoch": 0.04653367033760607, "grad_norm": 3.251591444015503, "learning_rate": 4.767376782812782e-05, "loss": 4.5074, "step": 5155 }, { "epoch": 0.04654269723776855, "grad_norm": 4.005561828613281, "learning_rate": 4.76733164831197e-05, "loss": 4.6622, "step": 5156 }, { "epoch": 0.04655172413793104, "grad_norm": 3.544058084487915, "learning_rate": 4.7672865138111575e-05, "loss": 4.5223, "step": 5157 }, { "epoch": 0.04656075103809352, "grad_norm": 3.2680697441101074, "learning_rate": 4.767241379310345e-05, "loss": 4.3853, "step": 5158 }, { "epoch": 0.046569777938256, "grad_norm": 3.7151474952697754, "learning_rate": 4.7671962448095326e-05, "loss": 4.7101, "step": 5159 }, { "epoch": 0.04657880483841849, "grad_norm": 3.663386344909668, "learning_rate": 4.76715111030872e-05, "loss": 5.0622, "step": 5160 }, { "epoch": 0.04658783173858097, "grad_norm": 5.53399658203125, "learning_rate": 4.767105975807908e-05, "loss": 3.5569, "step": 5161 }, { "epoch": 0.04659685863874346, "grad_norm": 3.0201332569122314, "learning_rate": 4.767060841307096e-05, "loss": 4.9441, "step": 5162 }, { "epoch": 0.04660588553890594, "grad_norm": 4.913603782653809, "learning_rate": 4.767015706806283e-05, "loss": 3.2891, "step": 5163 }, { "epoch": 0.04661491243906842, "grad_norm": 4.010885715484619, "learning_rate": 4.76697057230547e-05, "loss": 4.6548, "step": 5164 }, { "epoch": 0.04662393933923091, "grad_norm": 3.1660003662109375, "learning_rate": 4.766925437804658e-05, "loss": 4.4679, "step": 5165 }, { "epoch": 0.04663296623939339, "grad_norm": 3.941168785095215, "learning_rate": 4.7668803033038454e-05, "loss": 4.6946, "step": 5166 }, { "epoch": 0.04664199313955587, "grad_norm": 3.8105390071868896, "learning_rate": 4.766835168803033e-05, "loss": 4.3853, "step": 5167 }, { "epoch": 0.04665102003971836, "grad_norm": 4.861116886138916, "learning_rate": 4.766790034302221e-05, "loss": 4.9048, "step": 5168 }, { "epoch": 0.04666004693988084, "grad_norm": 3.0058631896972656, "learning_rate": 4.7667448998014085e-05, "loss": 4.2886, "step": 5169 }, { "epoch": 0.04666907384004333, "grad_norm": 3.826143741607666, "learning_rate": 4.7666997653005964e-05, "loss": 4.8097, "step": 5170 }, { "epoch": 0.04667810074020581, "grad_norm": 2.973891019821167, "learning_rate": 4.766654630799784e-05, "loss": 5.0761, "step": 5171 }, { "epoch": 0.046687127640368295, "grad_norm": 3.4722352027893066, "learning_rate": 4.766609496298971e-05, "loss": 4.1991, "step": 5172 }, { "epoch": 0.04669615454053078, "grad_norm": 3.3169808387756348, "learning_rate": 4.766564361798159e-05, "loss": 4.1658, "step": 5173 }, { "epoch": 0.046705181440693265, "grad_norm": 3.7638869285583496, "learning_rate": 4.766519227297346e-05, "loss": 4.7221, "step": 5174 }, { "epoch": 0.04671420834085575, "grad_norm": 3.3451905250549316, "learning_rate": 4.766474092796534e-05, "loss": 5.2514, "step": 5175 }, { "epoch": 0.046723235241018235, "grad_norm": 2.3200130462646484, "learning_rate": 4.766428958295722e-05, "loss": 5.5395, "step": 5176 }, { "epoch": 0.046732262141180717, "grad_norm": 4.216827869415283, "learning_rate": 4.766383823794909e-05, "loss": 4.6863, "step": 5177 }, { "epoch": 0.046741289041343205, "grad_norm": 3.1970889568328857, "learning_rate": 4.7663386892940964e-05, "loss": 4.1584, "step": 5178 }, { "epoch": 0.04675031594150569, "grad_norm": 4.177265644073486, "learning_rate": 4.7662935547932843e-05, "loss": 4.2777, "step": 5179 }, { "epoch": 0.04675934284166817, "grad_norm": 3.1985490322113037, "learning_rate": 4.7662484202924716e-05, "loss": 5.0932, "step": 5180 }, { "epoch": 0.04676836974183066, "grad_norm": 3.4974145889282227, "learning_rate": 4.7662032857916595e-05, "loss": 4.2046, "step": 5181 }, { "epoch": 0.04677739664199314, "grad_norm": 2.788702964782715, "learning_rate": 4.766158151290847e-05, "loss": 4.6093, "step": 5182 }, { "epoch": 0.04678642354215563, "grad_norm": 3.518373489379883, "learning_rate": 4.766113016790035e-05, "loss": 4.7327, "step": 5183 }, { "epoch": 0.04679545044231811, "grad_norm": 3.3480348587036133, "learning_rate": 4.766067882289222e-05, "loss": 4.577, "step": 5184 }, { "epoch": 0.04680447734248059, "grad_norm": 3.312741756439209, "learning_rate": 4.766022747788409e-05, "loss": 4.3161, "step": 5185 }, { "epoch": 0.04681350424264308, "grad_norm": 4.439831733703613, "learning_rate": 4.765977613287597e-05, "loss": 4.7373, "step": 5186 }, { "epoch": 0.04682253114280556, "grad_norm": 3.0575461387634277, "learning_rate": 4.765932478786785e-05, "loss": 5.0548, "step": 5187 }, { "epoch": 0.04683155804296805, "grad_norm": 3.7373266220092773, "learning_rate": 4.765887344285972e-05, "loss": 4.5888, "step": 5188 }, { "epoch": 0.04684058494313053, "grad_norm": 5.579893589019775, "learning_rate": 4.76584220978516e-05, "loss": 4.7593, "step": 5189 }, { "epoch": 0.04684961184329301, "grad_norm": 3.8645598888397217, "learning_rate": 4.765797075284348e-05, "loss": 4.6606, "step": 5190 }, { "epoch": 0.0468586387434555, "grad_norm": 3.8974416255950928, "learning_rate": 4.765751940783535e-05, "loss": 5.1168, "step": 5191 }, { "epoch": 0.04686766564361798, "grad_norm": 3.778327226638794, "learning_rate": 4.7657068062827226e-05, "loss": 4.8555, "step": 5192 }, { "epoch": 0.04687669254378046, "grad_norm": 2.935554265975952, "learning_rate": 4.7656616717819105e-05, "loss": 5.0014, "step": 5193 }, { "epoch": 0.04688571944394295, "grad_norm": 4.325032711029053, "learning_rate": 4.765616537281098e-05, "loss": 5.0669, "step": 5194 }, { "epoch": 0.04689474634410543, "grad_norm": 3.5779449939727783, "learning_rate": 4.765571402780286e-05, "loss": 5.0529, "step": 5195 }, { "epoch": 0.04690377324426792, "grad_norm": 3.7575745582580566, "learning_rate": 4.765526268279473e-05, "loss": 4.3081, "step": 5196 }, { "epoch": 0.0469128001444304, "grad_norm": 3.104260206222534, "learning_rate": 4.765481133778661e-05, "loss": 4.8167, "step": 5197 }, { "epoch": 0.046921827044592884, "grad_norm": 3.2845704555511475, "learning_rate": 4.765435999277848e-05, "loss": 4.9348, "step": 5198 }, { "epoch": 0.04693085394475537, "grad_norm": 3.292487144470215, "learning_rate": 4.7653908647770354e-05, "loss": 4.6745, "step": 5199 }, { "epoch": 0.046939880844917854, "grad_norm": 2.792527675628662, "learning_rate": 4.765345730276223e-05, "loss": 5.1052, "step": 5200 }, { "epoch": 0.04694890774508034, "grad_norm": 3.209451913833618, "learning_rate": 4.765300595775411e-05, "loss": 4.6614, "step": 5201 }, { "epoch": 0.046957934645242824, "grad_norm": 3.7016756534576416, "learning_rate": 4.7652554612745985e-05, "loss": 4.6285, "step": 5202 }, { "epoch": 0.046966961545405306, "grad_norm": 3.5542514324188232, "learning_rate": 4.7652103267737864e-05, "loss": 4.3837, "step": 5203 }, { "epoch": 0.046975988445567794, "grad_norm": 3.56548810005188, "learning_rate": 4.7651651922729736e-05, "loss": 4.4428, "step": 5204 }, { "epoch": 0.046985015345730276, "grad_norm": 2.8539812564849854, "learning_rate": 4.765120057772161e-05, "loss": 4.3935, "step": 5205 }, { "epoch": 0.04699404224589276, "grad_norm": 3.2409205436706543, "learning_rate": 4.765074923271349e-05, "loss": 4.6907, "step": 5206 }, { "epoch": 0.047003069146055246, "grad_norm": 2.995880126953125, "learning_rate": 4.765029788770537e-05, "loss": 5.0302, "step": 5207 }, { "epoch": 0.04701209604621773, "grad_norm": 3.223555564880371, "learning_rate": 4.764984654269724e-05, "loss": 4.3055, "step": 5208 }, { "epoch": 0.047021122946380216, "grad_norm": 3.0158884525299072, "learning_rate": 4.764939519768912e-05, "loss": 4.6377, "step": 5209 }, { "epoch": 0.0470301498465427, "grad_norm": 6.016134738922119, "learning_rate": 4.764894385268099e-05, "loss": 4.9069, "step": 5210 }, { "epoch": 0.04703917674670518, "grad_norm": 3.41780686378479, "learning_rate": 4.7648492507672864e-05, "loss": 5.2243, "step": 5211 }, { "epoch": 0.04704820364686767, "grad_norm": 2.6885712146759033, "learning_rate": 4.764804116266474e-05, "loss": 4.7524, "step": 5212 }, { "epoch": 0.04705723054703015, "grad_norm": 3.955226421356201, "learning_rate": 4.7647589817656616e-05, "loss": 4.3714, "step": 5213 }, { "epoch": 0.04706625744719264, "grad_norm": 3.4704911708831787, "learning_rate": 4.7647138472648495e-05, "loss": 4.6697, "step": 5214 }, { "epoch": 0.04707528434735512, "grad_norm": 2.7729063034057617, "learning_rate": 4.7646687127640374e-05, "loss": 5.0132, "step": 5215 }, { "epoch": 0.0470843112475176, "grad_norm": 3.28059458732605, "learning_rate": 4.7646235782632247e-05, "loss": 5.0658, "step": 5216 }, { "epoch": 0.04709333814768009, "grad_norm": 2.8124806880950928, "learning_rate": 4.7645784437624126e-05, "loss": 5.0264, "step": 5217 }, { "epoch": 0.04710236504784257, "grad_norm": 3.5372555255889893, "learning_rate": 4.7645333092616e-05, "loss": 5.1487, "step": 5218 }, { "epoch": 0.04711139194800505, "grad_norm": 2.998683452606201, "learning_rate": 4.764488174760787e-05, "loss": 4.5696, "step": 5219 }, { "epoch": 0.04712041884816754, "grad_norm": 2.617048978805542, "learning_rate": 4.764443040259975e-05, "loss": 4.5873, "step": 5220 }, { "epoch": 0.04712944574833002, "grad_norm": 3.0873942375183105, "learning_rate": 4.764397905759162e-05, "loss": 4.779, "step": 5221 }, { "epoch": 0.04713847264849251, "grad_norm": 3.493433952331543, "learning_rate": 4.76435277125835e-05, "loss": 4.1779, "step": 5222 }, { "epoch": 0.04714749954865499, "grad_norm": 3.3142242431640625, "learning_rate": 4.764307636757538e-05, "loss": 4.6166, "step": 5223 }, { "epoch": 0.04715652644881747, "grad_norm": 5.151952743530273, "learning_rate": 4.7642625022567253e-05, "loss": 4.2877, "step": 5224 }, { "epoch": 0.04716555334897996, "grad_norm": 3.186612129211426, "learning_rate": 4.7642173677559126e-05, "loss": 5.0613, "step": 5225 }, { "epoch": 0.04717458024914244, "grad_norm": 3.481349468231201, "learning_rate": 4.7641722332551005e-05, "loss": 5.1135, "step": 5226 }, { "epoch": 0.04718360714930493, "grad_norm": 2.837677478790283, "learning_rate": 4.764127098754288e-05, "loss": 4.7113, "step": 5227 }, { "epoch": 0.047192634049467413, "grad_norm": 3.25327205657959, "learning_rate": 4.764081964253476e-05, "loss": 4.2514, "step": 5228 }, { "epoch": 0.047201660949629895, "grad_norm": 2.871187686920166, "learning_rate": 4.7640368297526636e-05, "loss": 5.0598, "step": 5229 }, { "epoch": 0.047210687849792383, "grad_norm": 3.319019317626953, "learning_rate": 4.763991695251851e-05, "loss": 4.7376, "step": 5230 }, { "epoch": 0.047219714749954865, "grad_norm": 3.0607738494873047, "learning_rate": 4.763946560751038e-05, "loss": 4.3564, "step": 5231 }, { "epoch": 0.04722874165011735, "grad_norm": 3.924424648284912, "learning_rate": 4.763901426250226e-05, "loss": 4.2398, "step": 5232 }, { "epoch": 0.047237768550279835, "grad_norm": 3.5318920612335205, "learning_rate": 4.763856291749413e-05, "loss": 4.2987, "step": 5233 }, { "epoch": 0.04724679545044232, "grad_norm": 3.496119499206543, "learning_rate": 4.763811157248601e-05, "loss": 4.8082, "step": 5234 }, { "epoch": 0.047255822350604805, "grad_norm": 3.786328077316284, "learning_rate": 4.7637660227477884e-05, "loss": 4.7051, "step": 5235 }, { "epoch": 0.04726484925076729, "grad_norm": 3.665433168411255, "learning_rate": 4.7637208882469764e-05, "loss": 4.3643, "step": 5236 }, { "epoch": 0.04727387615092977, "grad_norm": 3.5293309688568115, "learning_rate": 4.763675753746164e-05, "loss": 4.9774, "step": 5237 }, { "epoch": 0.04728290305109226, "grad_norm": 3.069836378097534, "learning_rate": 4.763630619245351e-05, "loss": 4.8612, "step": 5238 }, { "epoch": 0.04729192995125474, "grad_norm": 4.64552116394043, "learning_rate": 4.763585484744539e-05, "loss": 3.5733, "step": 5239 }, { "epoch": 0.04730095685141723, "grad_norm": 3.5115532875061035, "learning_rate": 4.763540350243727e-05, "loss": 4.9437, "step": 5240 }, { "epoch": 0.04730998375157971, "grad_norm": 3.7635653018951416, "learning_rate": 4.763495215742914e-05, "loss": 4.3875, "step": 5241 }, { "epoch": 0.04731901065174219, "grad_norm": 3.017598867416382, "learning_rate": 4.763450081242102e-05, "loss": 5.0199, "step": 5242 }, { "epoch": 0.04732803755190468, "grad_norm": 2.846165657043457, "learning_rate": 4.763404946741289e-05, "loss": 4.9578, "step": 5243 }, { "epoch": 0.04733706445206716, "grad_norm": 5.806415557861328, "learning_rate": 4.763359812240477e-05, "loss": 5.0843, "step": 5244 }, { "epoch": 0.04734609135222964, "grad_norm": 4.336546421051025, "learning_rate": 4.763314677739664e-05, "loss": 4.3011, "step": 5245 }, { "epoch": 0.04735511825239213, "grad_norm": 3.86142897605896, "learning_rate": 4.763269543238852e-05, "loss": 4.6621, "step": 5246 }, { "epoch": 0.04736414515255461, "grad_norm": 4.492775917053223, "learning_rate": 4.7632244087380395e-05, "loss": 4.2817, "step": 5247 }, { "epoch": 0.0473731720527171, "grad_norm": 2.7706220149993896, "learning_rate": 4.7631792742372274e-05, "loss": 5.2372, "step": 5248 }, { "epoch": 0.04738219895287958, "grad_norm": 3.930478811264038, "learning_rate": 4.7631341397364146e-05, "loss": 4.3615, "step": 5249 }, { "epoch": 0.04739122585304206, "grad_norm": 5.161154747009277, "learning_rate": 4.7630890052356026e-05, "loss": 5.2898, "step": 5250 }, { "epoch": 0.04740025275320455, "grad_norm": 4.905542373657227, "learning_rate": 4.76304387073479e-05, "loss": 4.4589, "step": 5251 }, { "epoch": 0.04740927965336703, "grad_norm": 3.588484525680542, "learning_rate": 4.762998736233977e-05, "loss": 5.0065, "step": 5252 }, { "epoch": 0.04741830655352952, "grad_norm": 4.519234657287598, "learning_rate": 4.762953601733165e-05, "loss": 4.1003, "step": 5253 }, { "epoch": 0.047427333453692, "grad_norm": 3.730865478515625, "learning_rate": 4.762908467232353e-05, "loss": 4.9961, "step": 5254 }, { "epoch": 0.047436360353854484, "grad_norm": 3.689013957977295, "learning_rate": 4.76286333273154e-05, "loss": 4.7244, "step": 5255 }, { "epoch": 0.04744538725401697, "grad_norm": 3.4097800254821777, "learning_rate": 4.762818198230728e-05, "loss": 4.3104, "step": 5256 }, { "epoch": 0.047454414154179454, "grad_norm": 4.2849249839782715, "learning_rate": 4.762773063729915e-05, "loss": 4.4461, "step": 5257 }, { "epoch": 0.047463441054341936, "grad_norm": 3.8423874378204346, "learning_rate": 4.7627279292291026e-05, "loss": 4.6252, "step": 5258 }, { "epoch": 0.047472467954504424, "grad_norm": 3.200942039489746, "learning_rate": 4.7626827947282905e-05, "loss": 4.7813, "step": 5259 }, { "epoch": 0.047481494854666906, "grad_norm": 4.181126117706299, "learning_rate": 4.762637660227478e-05, "loss": 4.4696, "step": 5260 }, { "epoch": 0.047490521754829394, "grad_norm": 3.352766275405884, "learning_rate": 4.7625925257266657e-05, "loss": 4.107, "step": 5261 }, { "epoch": 0.047499548654991876, "grad_norm": 4.408574104309082, "learning_rate": 4.7625473912258536e-05, "loss": 5.018, "step": 5262 }, { "epoch": 0.04750857555515436, "grad_norm": 2.1460323333740234, "learning_rate": 4.762502256725041e-05, "loss": 5.0153, "step": 5263 }, { "epoch": 0.047517602455316846, "grad_norm": 3.7917559146881104, "learning_rate": 4.762457122224229e-05, "loss": 4.6645, "step": 5264 }, { "epoch": 0.04752662935547933, "grad_norm": 3.848240852355957, "learning_rate": 4.762411987723416e-05, "loss": 5.3428, "step": 5265 }, { "epoch": 0.047535656255641816, "grad_norm": 3.8002381324768066, "learning_rate": 4.762366853222603e-05, "loss": 4.7139, "step": 5266 }, { "epoch": 0.0475446831558043, "grad_norm": 2.51914119720459, "learning_rate": 4.762321718721791e-05, "loss": 5.288, "step": 5267 }, { "epoch": 0.04755371005596678, "grad_norm": 3.1479458808898926, "learning_rate": 4.762276584220979e-05, "loss": 4.1616, "step": 5268 }, { "epoch": 0.04756273695612927, "grad_norm": 3.811283588409424, "learning_rate": 4.7622314497201663e-05, "loss": 4.3688, "step": 5269 }, { "epoch": 0.04757176385629175, "grad_norm": 3.5203654766082764, "learning_rate": 4.762186315219354e-05, "loss": 5.0854, "step": 5270 }, { "epoch": 0.04758079075645423, "grad_norm": 3.523578405380249, "learning_rate": 4.7621411807185415e-05, "loss": 4.7235, "step": 5271 }, { "epoch": 0.04758981765661672, "grad_norm": 3.2775306701660156, "learning_rate": 4.762096046217729e-05, "loss": 4.917, "step": 5272 }, { "epoch": 0.0475988445567792, "grad_norm": 4.0261335372924805, "learning_rate": 4.762050911716917e-05, "loss": 3.7299, "step": 5273 }, { "epoch": 0.04760787145694169, "grad_norm": 3.8927996158599854, "learning_rate": 4.762005777216104e-05, "loss": 4.428, "step": 5274 }, { "epoch": 0.04761689835710417, "grad_norm": 3.28279972076416, "learning_rate": 4.761960642715292e-05, "loss": 4.942, "step": 5275 }, { "epoch": 0.04762592525726665, "grad_norm": 4.085088729858398, "learning_rate": 4.76191550821448e-05, "loss": 4.4841, "step": 5276 }, { "epoch": 0.04763495215742914, "grad_norm": 4.248687744140625, "learning_rate": 4.761870373713667e-05, "loss": 5.3285, "step": 5277 }, { "epoch": 0.04764397905759162, "grad_norm": 3.811330795288086, "learning_rate": 4.761825239212854e-05, "loss": 4.9506, "step": 5278 }, { "epoch": 0.04765300595775411, "grad_norm": 2.898952007293701, "learning_rate": 4.761780104712042e-05, "loss": 4.6125, "step": 5279 }, { "epoch": 0.04766203285791659, "grad_norm": 3.540722131729126, "learning_rate": 4.7617349702112294e-05, "loss": 4.5411, "step": 5280 }, { "epoch": 0.047671059758079073, "grad_norm": 3.97054123878479, "learning_rate": 4.7616898357104174e-05, "loss": 4.8316, "step": 5281 }, { "epoch": 0.04768008665824156, "grad_norm": 3.68467378616333, "learning_rate": 4.7616447012096046e-05, "loss": 4.3323, "step": 5282 }, { "epoch": 0.047689113558404043, "grad_norm": 3.467081069946289, "learning_rate": 4.7615995667087925e-05, "loss": 4.382, "step": 5283 }, { "epoch": 0.047698140458566525, "grad_norm": 3.655292510986328, "learning_rate": 4.7615544322079805e-05, "loss": 4.6274, "step": 5284 }, { "epoch": 0.047707167358729013, "grad_norm": 3.6495258808135986, "learning_rate": 4.761509297707167e-05, "loss": 4.4039, "step": 5285 }, { "epoch": 0.047716194258891495, "grad_norm": 3.0548079013824463, "learning_rate": 4.761464163206355e-05, "loss": 4.0698, "step": 5286 }, { "epoch": 0.047725221159053983, "grad_norm": 2.9850776195526123, "learning_rate": 4.761419028705543e-05, "loss": 4.9589, "step": 5287 }, { "epoch": 0.047734248059216465, "grad_norm": 4.448087692260742, "learning_rate": 4.76137389420473e-05, "loss": 4.3936, "step": 5288 }, { "epoch": 0.04774327495937895, "grad_norm": 3.6273937225341797, "learning_rate": 4.761328759703918e-05, "loss": 4.7985, "step": 5289 }, { "epoch": 0.047752301859541435, "grad_norm": 3.9163265228271484, "learning_rate": 4.761283625203106e-05, "loss": 4.6668, "step": 5290 }, { "epoch": 0.04776132875970392, "grad_norm": 3.9032490253448486, "learning_rate": 4.761238490702293e-05, "loss": 5.0094, "step": 5291 }, { "epoch": 0.047770355659866405, "grad_norm": 3.671234130859375, "learning_rate": 4.7611933562014805e-05, "loss": 4.4538, "step": 5292 }, { "epoch": 0.04777938256002889, "grad_norm": 3.8829686641693115, "learning_rate": 4.7611482217006684e-05, "loss": 4.8892, "step": 5293 }, { "epoch": 0.04778840946019137, "grad_norm": 4.542286396026611, "learning_rate": 4.7611030871998556e-05, "loss": 4.6774, "step": 5294 }, { "epoch": 0.04779743636035386, "grad_norm": 3.1168110370635986, "learning_rate": 4.7610579526990436e-05, "loss": 4.6202, "step": 5295 }, { "epoch": 0.04780646326051634, "grad_norm": 3.4209487438201904, "learning_rate": 4.761012818198231e-05, "loss": 4.3258, "step": 5296 }, { "epoch": 0.04781549016067882, "grad_norm": 3.931180000305176, "learning_rate": 4.760967683697419e-05, "loss": 4.1448, "step": 5297 }, { "epoch": 0.04782451706084131, "grad_norm": 2.830533266067505, "learning_rate": 4.760922549196606e-05, "loss": 5.594, "step": 5298 }, { "epoch": 0.04783354396100379, "grad_norm": 3.4221527576446533, "learning_rate": 4.760877414695793e-05, "loss": 5.4244, "step": 5299 }, { "epoch": 0.04784257086116628, "grad_norm": 3.9213740825653076, "learning_rate": 4.760832280194981e-05, "loss": 4.5442, "step": 5300 }, { "epoch": 0.04785159776132876, "grad_norm": 3.7119131088256836, "learning_rate": 4.760787145694169e-05, "loss": 4.3712, "step": 5301 }, { "epoch": 0.04786062466149124, "grad_norm": 2.5746452808380127, "learning_rate": 4.760742011193356e-05, "loss": 4.6796, "step": 5302 }, { "epoch": 0.04786965156165373, "grad_norm": 4.177112102508545, "learning_rate": 4.760696876692544e-05, "loss": 3.2852, "step": 5303 }, { "epoch": 0.04787867846181621, "grad_norm": 2.3417611122131348, "learning_rate": 4.7606517421917315e-05, "loss": 4.8595, "step": 5304 }, { "epoch": 0.0478877053619787, "grad_norm": 3.5023915767669678, "learning_rate": 4.760606607690919e-05, "loss": 4.7861, "step": 5305 }, { "epoch": 0.04789673226214118, "grad_norm": 3.3000876903533936, "learning_rate": 4.7605614731901067e-05, "loss": 5.4849, "step": 5306 }, { "epoch": 0.04790575916230366, "grad_norm": 3.3507723808288574, "learning_rate": 4.7605163386892946e-05, "loss": 4.96, "step": 5307 }, { "epoch": 0.04791478606246615, "grad_norm": 3.213132858276367, "learning_rate": 4.760471204188482e-05, "loss": 5.4012, "step": 5308 }, { "epoch": 0.04792381296262863, "grad_norm": 3.175816059112549, "learning_rate": 4.76042606968767e-05, "loss": 4.7167, "step": 5309 }, { "epoch": 0.047932839862791114, "grad_norm": 2.4434762001037598, "learning_rate": 4.760380935186857e-05, "loss": 4.9753, "step": 5310 }, { "epoch": 0.0479418667629536, "grad_norm": 3.6150166988372803, "learning_rate": 4.760335800686045e-05, "loss": 3.907, "step": 5311 }, { "epoch": 0.047950893663116084, "grad_norm": 3.8180174827575684, "learning_rate": 4.760290666185232e-05, "loss": 4.9079, "step": 5312 }, { "epoch": 0.04795992056327857, "grad_norm": 3.900405168533325, "learning_rate": 4.7602455316844194e-05, "loss": 5.1002, "step": 5313 }, { "epoch": 0.047968947463441054, "grad_norm": 3.659144878387451, "learning_rate": 4.7602003971836073e-05, "loss": 4.5709, "step": 5314 }, { "epoch": 0.047977974363603536, "grad_norm": 5.345127582550049, "learning_rate": 4.760155262682795e-05, "loss": 4.5048, "step": 5315 }, { "epoch": 0.047987001263766024, "grad_norm": 3.3457369804382324, "learning_rate": 4.7601101281819825e-05, "loss": 4.7071, "step": 5316 }, { "epoch": 0.047996028163928506, "grad_norm": 2.3319804668426514, "learning_rate": 4.7600649936811704e-05, "loss": 4.1575, "step": 5317 }, { "epoch": 0.048005055064090994, "grad_norm": 4.7024431228637695, "learning_rate": 4.760019859180358e-05, "loss": 3.7295, "step": 5318 }, { "epoch": 0.048014081964253476, "grad_norm": 4.268726825714111, "learning_rate": 4.759974724679545e-05, "loss": 4.6119, "step": 5319 }, { "epoch": 0.04802310886441596, "grad_norm": 3.919654130935669, "learning_rate": 4.759929590178733e-05, "loss": 4.7292, "step": 5320 }, { "epoch": 0.048032135764578446, "grad_norm": 4.176498889923096, "learning_rate": 4.75988445567792e-05, "loss": 5.2396, "step": 5321 }, { "epoch": 0.04804116266474093, "grad_norm": 4.627944469451904, "learning_rate": 4.759839321177108e-05, "loss": 3.744, "step": 5322 }, { "epoch": 0.04805018956490341, "grad_norm": 3.1411867141723633, "learning_rate": 4.759794186676296e-05, "loss": 4.7752, "step": 5323 }, { "epoch": 0.0480592164650659, "grad_norm": 3.1730098724365234, "learning_rate": 4.759749052175483e-05, "loss": 4.566, "step": 5324 }, { "epoch": 0.04806824336522838, "grad_norm": 3.5002293586730957, "learning_rate": 4.7597039176746704e-05, "loss": 4.4793, "step": 5325 }, { "epoch": 0.04807727026539087, "grad_norm": 3.039681911468506, "learning_rate": 4.7596587831738584e-05, "loss": 4.5737, "step": 5326 }, { "epoch": 0.04808629716555335, "grad_norm": 3.4465231895446777, "learning_rate": 4.7596136486730456e-05, "loss": 4.4574, "step": 5327 }, { "epoch": 0.04809532406571583, "grad_norm": 5.554802417755127, "learning_rate": 4.7595685141722335e-05, "loss": 3.8041, "step": 5328 }, { "epoch": 0.04810435096587832, "grad_norm": 3.341740846633911, "learning_rate": 4.7595233796714215e-05, "loss": 4.4429, "step": 5329 }, { "epoch": 0.0481133778660408, "grad_norm": 3.1533565521240234, "learning_rate": 4.759478245170609e-05, "loss": 4.9704, "step": 5330 }, { "epoch": 0.04812240476620329, "grad_norm": 4.055233955383301, "learning_rate": 4.7594331106697966e-05, "loss": 4.4749, "step": 5331 }, { "epoch": 0.04813143166636577, "grad_norm": 3.8778793811798096, "learning_rate": 4.759387976168984e-05, "loss": 5.0277, "step": 5332 }, { "epoch": 0.04814045856652825, "grad_norm": 4.417270183563232, "learning_rate": 4.759342841668171e-05, "loss": 5.2651, "step": 5333 }, { "epoch": 0.04814948546669074, "grad_norm": 2.8421831130981445, "learning_rate": 4.759297707167359e-05, "loss": 4.6567, "step": 5334 }, { "epoch": 0.04815851236685322, "grad_norm": 3.87196946144104, "learning_rate": 4.759252572666546e-05, "loss": 4.8673, "step": 5335 }, { "epoch": 0.048167539267015703, "grad_norm": 3.6151065826416016, "learning_rate": 4.759207438165734e-05, "loss": 5.0428, "step": 5336 }, { "epoch": 0.04817656616717819, "grad_norm": 3.445810079574585, "learning_rate": 4.759162303664922e-05, "loss": 4.2218, "step": 5337 }, { "epoch": 0.048185593067340673, "grad_norm": 3.4890780448913574, "learning_rate": 4.7591171691641094e-05, "loss": 4.3847, "step": 5338 }, { "epoch": 0.04819461996750316, "grad_norm": 3.8033578395843506, "learning_rate": 4.7590720346632966e-05, "loss": 4.8578, "step": 5339 }, { "epoch": 0.048203646867665643, "grad_norm": 3.9496312141418457, "learning_rate": 4.7590269001624846e-05, "loss": 4.6022, "step": 5340 }, { "epoch": 0.048212673767828125, "grad_norm": 3.801774024963379, "learning_rate": 4.758981765661672e-05, "loss": 4.9219, "step": 5341 }, { "epoch": 0.048221700667990613, "grad_norm": 5.957233428955078, "learning_rate": 4.75893663116086e-05, "loss": 3.7464, "step": 5342 }, { "epoch": 0.048230727568153095, "grad_norm": 4.729247093200684, "learning_rate": 4.758891496660047e-05, "loss": 3.8133, "step": 5343 }, { "epoch": 0.048239754468315584, "grad_norm": 3.823596239089966, "learning_rate": 4.758846362159235e-05, "loss": 4.8544, "step": 5344 }, { "epoch": 0.048248781368478065, "grad_norm": 2.75580096244812, "learning_rate": 4.758801227658422e-05, "loss": 4.6587, "step": 5345 }, { "epoch": 0.04825780826864055, "grad_norm": 3.931483745574951, "learning_rate": 4.7587560931576094e-05, "loss": 4.5012, "step": 5346 }, { "epoch": 0.048266835168803035, "grad_norm": 4.530682563781738, "learning_rate": 4.758710958656797e-05, "loss": 3.3778, "step": 5347 }, { "epoch": 0.04827586206896552, "grad_norm": 3.340167760848999, "learning_rate": 4.758665824155985e-05, "loss": 4.3532, "step": 5348 }, { "epoch": 0.048284888969128, "grad_norm": 3.87874436378479, "learning_rate": 4.7586206896551725e-05, "loss": 4.3681, "step": 5349 }, { "epoch": 0.04829391586929049, "grad_norm": 3.8113973140716553, "learning_rate": 4.7585755551543604e-05, "loss": 5.0757, "step": 5350 }, { "epoch": 0.04830294276945297, "grad_norm": 3.7492494583129883, "learning_rate": 4.758530420653548e-05, "loss": 3.2946, "step": 5351 }, { "epoch": 0.04831196966961546, "grad_norm": 3.381476879119873, "learning_rate": 4.758485286152735e-05, "loss": 5.1401, "step": 5352 }, { "epoch": 0.04832099656977794, "grad_norm": 2.849048614501953, "learning_rate": 4.758440151651923e-05, "loss": 4.6459, "step": 5353 }, { "epoch": 0.04833002346994042, "grad_norm": 3.3600480556488037, "learning_rate": 4.758395017151111e-05, "loss": 4.7466, "step": 5354 }, { "epoch": 0.04833905037010291, "grad_norm": 2.64221453666687, "learning_rate": 4.758349882650298e-05, "loss": 5.0018, "step": 5355 }, { "epoch": 0.04834807727026539, "grad_norm": 4.873443126678467, "learning_rate": 4.758304748149486e-05, "loss": 5.1144, "step": 5356 }, { "epoch": 0.04835710417042788, "grad_norm": 3.397275447845459, "learning_rate": 4.758259613648673e-05, "loss": 4.4944, "step": 5357 }, { "epoch": 0.04836613107059036, "grad_norm": 2.8260838985443115, "learning_rate": 4.758214479147861e-05, "loss": 4.5264, "step": 5358 }, { "epoch": 0.04837515797075284, "grad_norm": 3.0608909130096436, "learning_rate": 4.7581693446470483e-05, "loss": 5.3968, "step": 5359 }, { "epoch": 0.04838418487091533, "grad_norm": 3.9760189056396484, "learning_rate": 4.7581242101462356e-05, "loss": 4.5598, "step": 5360 }, { "epoch": 0.04839321177107781, "grad_norm": 6.748666286468506, "learning_rate": 4.7580790756454235e-05, "loss": 5.313, "step": 5361 }, { "epoch": 0.04840223867124029, "grad_norm": 3.893277883529663, "learning_rate": 4.7580339411446114e-05, "loss": 5.2514, "step": 5362 }, { "epoch": 0.04841126557140278, "grad_norm": 3.52070951461792, "learning_rate": 4.757988806643799e-05, "loss": 3.9305, "step": 5363 }, { "epoch": 0.04842029247156526, "grad_norm": 4.4975433349609375, "learning_rate": 4.7579436721429866e-05, "loss": 4.9183, "step": 5364 }, { "epoch": 0.04842931937172775, "grad_norm": 3.3844499588012695, "learning_rate": 4.757898537642174e-05, "loss": 4.3236, "step": 5365 }, { "epoch": 0.04843834627189023, "grad_norm": 2.9451026916503906, "learning_rate": 4.757853403141361e-05, "loss": 4.4127, "step": 5366 }, { "epoch": 0.048447373172052714, "grad_norm": 3.556730270385742, "learning_rate": 4.757808268640549e-05, "loss": 4.8872, "step": 5367 }, { "epoch": 0.0484564000722152, "grad_norm": 3.2235703468322754, "learning_rate": 4.757763134139737e-05, "loss": 4.7963, "step": 5368 }, { "epoch": 0.048465426972377684, "grad_norm": 3.173337936401367, "learning_rate": 4.757717999638924e-05, "loss": 4.6001, "step": 5369 }, { "epoch": 0.04847445387254017, "grad_norm": 3.7076709270477295, "learning_rate": 4.757672865138112e-05, "loss": 4.7637, "step": 5370 }, { "epoch": 0.048483480772702654, "grad_norm": 2.9586758613586426, "learning_rate": 4.7576277306372994e-05, "loss": 6.1224, "step": 5371 }, { "epoch": 0.048492507672865136, "grad_norm": 5.007355213165283, "learning_rate": 4.7575825961364866e-05, "loss": 4.75, "step": 5372 }, { "epoch": 0.048501534573027624, "grad_norm": 3.433591842651367, "learning_rate": 4.7575374616356745e-05, "loss": 4.1594, "step": 5373 }, { "epoch": 0.048510561473190106, "grad_norm": 4.404481410980225, "learning_rate": 4.757492327134862e-05, "loss": 5.2186, "step": 5374 }, { "epoch": 0.04851958837335259, "grad_norm": 3.3470630645751953, "learning_rate": 4.75744719263405e-05, "loss": 4.1687, "step": 5375 }, { "epoch": 0.048528615273515076, "grad_norm": 3.8114309310913086, "learning_rate": 4.7574020581332376e-05, "loss": 4.5333, "step": 5376 }, { "epoch": 0.04853764217367756, "grad_norm": 3.694117307662964, "learning_rate": 4.757356923632425e-05, "loss": 4.4705, "step": 5377 }, { "epoch": 0.048546669073840046, "grad_norm": 3.0528576374053955, "learning_rate": 4.757311789131613e-05, "loss": 4.9614, "step": 5378 }, { "epoch": 0.04855569597400253, "grad_norm": 3.7378640174865723, "learning_rate": 4.7572666546308e-05, "loss": 5.2415, "step": 5379 }, { "epoch": 0.04856472287416501, "grad_norm": 4.205637454986572, "learning_rate": 4.757221520129987e-05, "loss": 4.181, "step": 5380 }, { "epoch": 0.0485737497743275, "grad_norm": 3.7443647384643555, "learning_rate": 4.757176385629175e-05, "loss": 5.1956, "step": 5381 }, { "epoch": 0.04858277667448998, "grad_norm": 2.6622583866119385, "learning_rate": 4.7571312511283625e-05, "loss": 4.6021, "step": 5382 }, { "epoch": 0.04859180357465247, "grad_norm": 3.8953518867492676, "learning_rate": 4.7570861166275504e-05, "loss": 4.9973, "step": 5383 }, { "epoch": 0.04860083047481495, "grad_norm": 2.9746828079223633, "learning_rate": 4.757040982126738e-05, "loss": 5.3883, "step": 5384 }, { "epoch": 0.04860985737497743, "grad_norm": 4.397190093994141, "learning_rate": 4.7569958476259256e-05, "loss": 4.0055, "step": 5385 }, { "epoch": 0.04861888427513992, "grad_norm": 3.6243488788604736, "learning_rate": 4.756950713125113e-05, "loss": 4.6843, "step": 5386 }, { "epoch": 0.0486279111753024, "grad_norm": 3.5393192768096924, "learning_rate": 4.756905578624301e-05, "loss": 4.433, "step": 5387 }, { "epoch": 0.04863693807546488, "grad_norm": 3.0904629230499268, "learning_rate": 4.756860444123488e-05, "loss": 4.9151, "step": 5388 }, { "epoch": 0.04864596497562737, "grad_norm": 3.351759672164917, "learning_rate": 4.756815309622676e-05, "loss": 4.5494, "step": 5389 }, { "epoch": 0.04865499187578985, "grad_norm": 2.9754257202148438, "learning_rate": 4.756770175121864e-05, "loss": 4.6766, "step": 5390 }, { "epoch": 0.04866401877595234, "grad_norm": 4.512895584106445, "learning_rate": 4.756725040621051e-05, "loss": 4.639, "step": 5391 }, { "epoch": 0.04867304567611482, "grad_norm": 3.8101234436035156, "learning_rate": 4.756679906120238e-05, "loss": 5.2148, "step": 5392 }, { "epoch": 0.048682072576277303, "grad_norm": 5.043812274932861, "learning_rate": 4.756634771619426e-05, "loss": 2.8659, "step": 5393 }, { "epoch": 0.04869109947643979, "grad_norm": 3.7793490886688232, "learning_rate": 4.7565896371186135e-05, "loss": 5.6783, "step": 5394 }, { "epoch": 0.048700126376602274, "grad_norm": 2.5115694999694824, "learning_rate": 4.7565445026178014e-05, "loss": 4.9389, "step": 5395 }, { "epoch": 0.04870915327676476, "grad_norm": 3.8789327144622803, "learning_rate": 4.7564993681169887e-05, "loss": 5.1549, "step": 5396 }, { "epoch": 0.048718180176927244, "grad_norm": 3.491077423095703, "learning_rate": 4.7564542336161766e-05, "loss": 4.5415, "step": 5397 }, { "epoch": 0.048727207077089725, "grad_norm": 2.9002907276153564, "learning_rate": 4.7564090991153645e-05, "loss": 5.1281, "step": 5398 }, { "epoch": 0.048736233977252214, "grad_norm": 2.8766262531280518, "learning_rate": 4.756363964614551e-05, "loss": 5.4957, "step": 5399 }, { "epoch": 0.048745260877414695, "grad_norm": 4.5441412925720215, "learning_rate": 4.756318830113739e-05, "loss": 3.3711, "step": 5400 }, { "epoch": 0.04875428777757718, "grad_norm": 2.9213006496429443, "learning_rate": 4.756273695612927e-05, "loss": 4.4459, "step": 5401 }, { "epoch": 0.048763314677739665, "grad_norm": 2.752117156982422, "learning_rate": 4.756228561112114e-05, "loss": 4.9246, "step": 5402 }, { "epoch": 0.04877234157790215, "grad_norm": 4.364717960357666, "learning_rate": 4.756183426611302e-05, "loss": 4.7297, "step": 5403 }, { "epoch": 0.048781368478064635, "grad_norm": 3.945502996444702, "learning_rate": 4.7561382921104893e-05, "loss": 4.7366, "step": 5404 }, { "epoch": 0.04879039537822712, "grad_norm": 4.605551242828369, "learning_rate": 4.756093157609677e-05, "loss": 4.2752, "step": 5405 }, { "epoch": 0.0487994222783896, "grad_norm": 3.5301005840301514, "learning_rate": 4.7560480231088645e-05, "loss": 4.3132, "step": 5406 }, { "epoch": 0.04880844917855209, "grad_norm": 3.7758028507232666, "learning_rate": 4.7560028886080524e-05, "loss": 4.4217, "step": 5407 }, { "epoch": 0.04881747607871457, "grad_norm": 5.2274394035339355, "learning_rate": 4.75595775410724e-05, "loss": 3.7668, "step": 5408 }, { "epoch": 0.04882650297887706, "grad_norm": 2.846381425857544, "learning_rate": 4.7559126196064276e-05, "loss": 4.2731, "step": 5409 }, { "epoch": 0.04883552987903954, "grad_norm": 6.458004951477051, "learning_rate": 4.755867485105615e-05, "loss": 4.9502, "step": 5410 }, { "epoch": 0.04884455677920202, "grad_norm": 4.187428951263428, "learning_rate": 4.755822350604803e-05, "loss": 4.8624, "step": 5411 }, { "epoch": 0.04885358367936451, "grad_norm": 4.649775981903076, "learning_rate": 4.75577721610399e-05, "loss": 3.9786, "step": 5412 }, { "epoch": 0.04886261057952699, "grad_norm": 3.1388072967529297, "learning_rate": 4.755732081603177e-05, "loss": 4.8131, "step": 5413 }, { "epoch": 0.04887163747968947, "grad_norm": 2.9994664192199707, "learning_rate": 4.755686947102365e-05, "loss": 4.9934, "step": 5414 }, { "epoch": 0.04888066437985196, "grad_norm": 3.031088352203369, "learning_rate": 4.755641812601553e-05, "loss": 4.4269, "step": 5415 }, { "epoch": 0.04888969128001444, "grad_norm": 4.729447364807129, "learning_rate": 4.7555966781007404e-05, "loss": 4.0332, "step": 5416 }, { "epoch": 0.04889871818017693, "grad_norm": 4.241055965423584, "learning_rate": 4.755551543599928e-05, "loss": 3.8654, "step": 5417 }, { "epoch": 0.04890774508033941, "grad_norm": 3.2206203937530518, "learning_rate": 4.7555064090991155e-05, "loss": 4.6883, "step": 5418 }, { "epoch": 0.04891677198050189, "grad_norm": 4.232078552246094, "learning_rate": 4.755461274598303e-05, "loss": 3.3158, "step": 5419 }, { "epoch": 0.04892579888066438, "grad_norm": 4.228814125061035, "learning_rate": 4.755416140097491e-05, "loss": 4.8189, "step": 5420 }, { "epoch": 0.04893482578082686, "grad_norm": 4.015137195587158, "learning_rate": 4.755371005596678e-05, "loss": 5.0244, "step": 5421 }, { "epoch": 0.04894385268098935, "grad_norm": 4.241273403167725, "learning_rate": 4.755325871095866e-05, "loss": 3.7202, "step": 5422 }, { "epoch": 0.04895287958115183, "grad_norm": 3.1897430419921875, "learning_rate": 4.755280736595054e-05, "loss": 4.9412, "step": 5423 }, { "epoch": 0.048961906481314314, "grad_norm": 3.990097999572754, "learning_rate": 4.755235602094241e-05, "loss": 4.8736, "step": 5424 }, { "epoch": 0.0489709333814768, "grad_norm": 3.921219825744629, "learning_rate": 4.755190467593429e-05, "loss": 4.9012, "step": 5425 }, { "epoch": 0.048979960281639284, "grad_norm": 4.098515033721924, "learning_rate": 4.755145333092616e-05, "loss": 4.6494, "step": 5426 }, { "epoch": 0.048988987181801766, "grad_norm": 3.0678157806396484, "learning_rate": 4.7551001985918035e-05, "loss": 4.6436, "step": 5427 }, { "epoch": 0.048998014081964254, "grad_norm": 3.3182742595672607, "learning_rate": 4.7550550640909914e-05, "loss": 4.6078, "step": 5428 }, { "epoch": 0.049007040982126736, "grad_norm": 3.9316861629486084, "learning_rate": 4.755009929590179e-05, "loss": 4.4293, "step": 5429 }, { "epoch": 0.049016067882289224, "grad_norm": 3.3324899673461914, "learning_rate": 4.7549647950893666e-05, "loss": 3.4983, "step": 5430 }, { "epoch": 0.049025094782451706, "grad_norm": 3.438051462173462, "learning_rate": 4.7549196605885545e-05, "loss": 4.7503, "step": 5431 }, { "epoch": 0.04903412168261419, "grad_norm": 4.162927150726318, "learning_rate": 4.754874526087742e-05, "loss": 4.6745, "step": 5432 }, { "epoch": 0.049043148582776676, "grad_norm": 3.9575586318969727, "learning_rate": 4.754829391586929e-05, "loss": 4.727, "step": 5433 }, { "epoch": 0.04905217548293916, "grad_norm": 4.364024639129639, "learning_rate": 4.754784257086117e-05, "loss": 4.366, "step": 5434 }, { "epoch": 0.049061202383101646, "grad_norm": 3.921276569366455, "learning_rate": 4.754739122585304e-05, "loss": 4.1033, "step": 5435 }, { "epoch": 0.04907022928326413, "grad_norm": 3.1691834926605225, "learning_rate": 4.754693988084492e-05, "loss": 5.4175, "step": 5436 }, { "epoch": 0.04907925618342661, "grad_norm": 4.385293006896973, "learning_rate": 4.75464885358368e-05, "loss": 4.1275, "step": 5437 }, { "epoch": 0.0490882830835891, "grad_norm": 3.9865176677703857, "learning_rate": 4.754603719082867e-05, "loss": 4.7744, "step": 5438 }, { "epoch": 0.04909730998375158, "grad_norm": 3.6849427223205566, "learning_rate": 4.7545585845820545e-05, "loss": 4.4139, "step": 5439 }, { "epoch": 0.04910633688391407, "grad_norm": 4.346451282501221, "learning_rate": 4.7545134500812424e-05, "loss": 5.6391, "step": 5440 }, { "epoch": 0.04911536378407655, "grad_norm": 3.7564573287963867, "learning_rate": 4.7544683155804297e-05, "loss": 4.483, "step": 5441 }, { "epoch": 0.04912439068423903, "grad_norm": 2.970789909362793, "learning_rate": 4.7544231810796176e-05, "loss": 4.7779, "step": 5442 }, { "epoch": 0.04913341758440152, "grad_norm": 3.5657832622528076, "learning_rate": 4.754378046578805e-05, "loss": 4.3643, "step": 5443 }, { "epoch": 0.049142444484564, "grad_norm": 3.524226188659668, "learning_rate": 4.754332912077993e-05, "loss": 4.6648, "step": 5444 }, { "epoch": 0.04915147138472648, "grad_norm": 4.10951566696167, "learning_rate": 4.754287777577181e-05, "loss": 4.2457, "step": 5445 }, { "epoch": 0.04916049828488897, "grad_norm": 3.841463327407837, "learning_rate": 4.754242643076367e-05, "loss": 4.3043, "step": 5446 }, { "epoch": 0.04916952518505145, "grad_norm": 3.5850303173065186, "learning_rate": 4.754197508575555e-05, "loss": 4.6752, "step": 5447 }, { "epoch": 0.04917855208521394, "grad_norm": 5.5818376541137695, "learning_rate": 4.754152374074743e-05, "loss": 4.0446, "step": 5448 }, { "epoch": 0.04918757898537642, "grad_norm": 3.433581829071045, "learning_rate": 4.7541072395739303e-05, "loss": 4.7526, "step": 5449 }, { "epoch": 0.049196605885538904, "grad_norm": 4.6243157386779785, "learning_rate": 4.754062105073118e-05, "loss": 4.6024, "step": 5450 }, { "epoch": 0.04920563278570139, "grad_norm": 3.1235244274139404, "learning_rate": 4.754016970572306e-05, "loss": 4.9235, "step": 5451 }, { "epoch": 0.049214659685863874, "grad_norm": 3.7603933811187744, "learning_rate": 4.7539718360714934e-05, "loss": 4.977, "step": 5452 }, { "epoch": 0.04922368658602636, "grad_norm": 2.538417339324951, "learning_rate": 4.753926701570681e-05, "loss": 4.6039, "step": 5453 }, { "epoch": 0.049232713486188844, "grad_norm": 4.495699405670166, "learning_rate": 4.7538815670698686e-05, "loss": 4.9295, "step": 5454 }, { "epoch": 0.049241740386351325, "grad_norm": 4.2494635581970215, "learning_rate": 4.753836432569056e-05, "loss": 5.0467, "step": 5455 }, { "epoch": 0.049250767286513814, "grad_norm": 3.287144184112549, "learning_rate": 4.753791298068244e-05, "loss": 4.6609, "step": 5456 }, { "epoch": 0.049259794186676295, "grad_norm": 3.2885937690734863, "learning_rate": 4.753746163567431e-05, "loss": 4.8252, "step": 5457 }, { "epoch": 0.04926882108683878, "grad_norm": 2.925724744796753, "learning_rate": 4.753701029066619e-05, "loss": 4.5854, "step": 5458 }, { "epoch": 0.049277847987001265, "grad_norm": 3.04756498336792, "learning_rate": 4.753655894565806e-05, "loss": 4.4238, "step": 5459 }, { "epoch": 0.04928687488716375, "grad_norm": 5.505599021911621, "learning_rate": 4.7536107600649934e-05, "loss": 3.5419, "step": 5460 }, { "epoch": 0.049295901787326235, "grad_norm": 3.7860958576202393, "learning_rate": 4.7535656255641814e-05, "loss": 4.6744, "step": 5461 }, { "epoch": 0.04930492868748872, "grad_norm": 3.141843557357788, "learning_rate": 4.753520491063369e-05, "loss": 4.917, "step": 5462 }, { "epoch": 0.0493139555876512, "grad_norm": 4.159870147705078, "learning_rate": 4.7534753565625565e-05, "loss": 3.5487, "step": 5463 }, { "epoch": 0.04932298248781369, "grad_norm": 5.561317443847656, "learning_rate": 4.7534302220617445e-05, "loss": 4.2936, "step": 5464 }, { "epoch": 0.04933200938797617, "grad_norm": 3.5657150745391846, "learning_rate": 4.753385087560932e-05, "loss": 4.8893, "step": 5465 }, { "epoch": 0.04934103628813866, "grad_norm": 3.6961240768432617, "learning_rate": 4.753339953060119e-05, "loss": 4.6025, "step": 5466 }, { "epoch": 0.04935006318830114, "grad_norm": 3.283890724182129, "learning_rate": 4.753294818559307e-05, "loss": 4.9237, "step": 5467 }, { "epoch": 0.04935909008846362, "grad_norm": 3.897824764251709, "learning_rate": 4.753249684058495e-05, "loss": 4.3129, "step": 5468 }, { "epoch": 0.04936811698862611, "grad_norm": 4.345508098602295, "learning_rate": 4.753204549557682e-05, "loss": 4.7049, "step": 5469 }, { "epoch": 0.04937714388878859, "grad_norm": 3.1502130031585693, "learning_rate": 4.75315941505687e-05, "loss": 4.7172, "step": 5470 }, { "epoch": 0.04938617078895107, "grad_norm": 3.8951263427734375, "learning_rate": 4.753114280556057e-05, "loss": 4.2149, "step": 5471 }, { "epoch": 0.04939519768911356, "grad_norm": 2.9085195064544678, "learning_rate": 4.753069146055245e-05, "loss": 4.489, "step": 5472 }, { "epoch": 0.04940422458927604, "grad_norm": 3.557996988296509, "learning_rate": 4.7530240115544324e-05, "loss": 5.1957, "step": 5473 }, { "epoch": 0.04941325148943853, "grad_norm": 3.722235679626465, "learning_rate": 4.7529788770536196e-05, "loss": 4.7724, "step": 5474 }, { "epoch": 0.04942227838960101, "grad_norm": 3.784684419631958, "learning_rate": 4.7529337425528076e-05, "loss": 4.6016, "step": 5475 }, { "epoch": 0.04943130528976349, "grad_norm": 3.4906604290008545, "learning_rate": 4.7528886080519955e-05, "loss": 5.0955, "step": 5476 }, { "epoch": 0.04944033218992598, "grad_norm": 2.724008321762085, "learning_rate": 4.752843473551183e-05, "loss": 4.5674, "step": 5477 }, { "epoch": 0.04944935909008846, "grad_norm": 4.119565486907959, "learning_rate": 4.7527983390503707e-05, "loss": 5.2727, "step": 5478 }, { "epoch": 0.04945838599025095, "grad_norm": 3.742647171020508, "learning_rate": 4.752753204549558e-05, "loss": 4.4802, "step": 5479 }, { "epoch": 0.04946741289041343, "grad_norm": 2.865264654159546, "learning_rate": 4.752708070048745e-05, "loss": 4.8695, "step": 5480 }, { "epoch": 0.049476439790575914, "grad_norm": 3.238933801651001, "learning_rate": 4.752662935547933e-05, "loss": 4.1739, "step": 5481 }, { "epoch": 0.0494854666907384, "grad_norm": 3.3702661991119385, "learning_rate": 4.75261780104712e-05, "loss": 4.5758, "step": 5482 }, { "epoch": 0.049494493590900884, "grad_norm": 3.3124046325683594, "learning_rate": 4.752572666546308e-05, "loss": 4.7654, "step": 5483 }, { "epoch": 0.049503520491063366, "grad_norm": 2.9131524562835693, "learning_rate": 4.752527532045496e-05, "loss": 4.197, "step": 5484 }, { "epoch": 0.049512547391225854, "grad_norm": 7.864310264587402, "learning_rate": 4.7524823975446834e-05, "loss": 4.7581, "step": 5485 }, { "epoch": 0.049521574291388336, "grad_norm": 2.8732008934020996, "learning_rate": 4.7524372630438707e-05, "loss": 4.731, "step": 5486 }, { "epoch": 0.049530601191550824, "grad_norm": 2.7731664180755615, "learning_rate": 4.7523921285430586e-05, "loss": 4.7422, "step": 5487 }, { "epoch": 0.049539628091713306, "grad_norm": 3.7796378135681152, "learning_rate": 4.752346994042246e-05, "loss": 4.4389, "step": 5488 }, { "epoch": 0.04954865499187579, "grad_norm": 4.009249687194824, "learning_rate": 4.752301859541434e-05, "loss": 5.1868, "step": 5489 }, { "epoch": 0.049557681892038276, "grad_norm": 3.4763400554656982, "learning_rate": 4.752256725040622e-05, "loss": 3.7259, "step": 5490 }, { "epoch": 0.04956670879220076, "grad_norm": 6.596714973449707, "learning_rate": 4.752211590539809e-05, "loss": 5.2737, "step": 5491 }, { "epoch": 0.049575735692363246, "grad_norm": 3.6777055263519287, "learning_rate": 4.752166456038997e-05, "loss": 4.807, "step": 5492 }, { "epoch": 0.04958476259252573, "grad_norm": 3.3901045322418213, "learning_rate": 4.752121321538184e-05, "loss": 4.9988, "step": 5493 }, { "epoch": 0.04959378949268821, "grad_norm": 3.803833484649658, "learning_rate": 4.752076187037371e-05, "loss": 5.0528, "step": 5494 }, { "epoch": 0.0496028163928507, "grad_norm": 4.162907600402832, "learning_rate": 4.752031052536559e-05, "loss": 4.5815, "step": 5495 }, { "epoch": 0.04961184329301318, "grad_norm": 3.4041190147399902, "learning_rate": 4.7519859180357465e-05, "loss": 4.3502, "step": 5496 }, { "epoch": 0.04962087019317566, "grad_norm": 4.041830539703369, "learning_rate": 4.7519407835349344e-05, "loss": 3.8093, "step": 5497 }, { "epoch": 0.04962989709333815, "grad_norm": 3.979121208190918, "learning_rate": 4.7518956490341224e-05, "loss": 4.9806, "step": 5498 }, { "epoch": 0.04963892399350063, "grad_norm": 3.970928430557251, "learning_rate": 4.7518505145333096e-05, "loss": 4.3116, "step": 5499 }, { "epoch": 0.04964795089366312, "grad_norm": 3.4133145809173584, "learning_rate": 4.751805380032497e-05, "loss": 4.8864, "step": 5500 }, { "epoch": 0.0496569777938256, "grad_norm": 3.255948305130005, "learning_rate": 4.751760245531685e-05, "loss": 4.4724, "step": 5501 }, { "epoch": 0.04966600469398808, "grad_norm": 3.514039993286133, "learning_rate": 4.751715111030872e-05, "loss": 4.999, "step": 5502 }, { "epoch": 0.04967503159415057, "grad_norm": 3.634982109069824, "learning_rate": 4.75166997653006e-05, "loss": 4.6677, "step": 5503 }, { "epoch": 0.04968405849431305, "grad_norm": 3.1634140014648438, "learning_rate": 4.751624842029247e-05, "loss": 4.8528, "step": 5504 }, { "epoch": 0.04969308539447554, "grad_norm": 3.8060266971588135, "learning_rate": 4.751579707528435e-05, "loss": 4.7382, "step": 5505 }, { "epoch": 0.04970211229463802, "grad_norm": 3.5153160095214844, "learning_rate": 4.7515345730276224e-05, "loss": 4.261, "step": 5506 }, { "epoch": 0.049711139194800504, "grad_norm": 4.8821330070495605, "learning_rate": 4.7514894385268096e-05, "loss": 4.3632, "step": 5507 }, { "epoch": 0.04972016609496299, "grad_norm": 3.260535478591919, "learning_rate": 4.7514443040259975e-05, "loss": 4.9809, "step": 5508 }, { "epoch": 0.049729192995125474, "grad_norm": 2.846372604370117, "learning_rate": 4.7513991695251855e-05, "loss": 4.8511, "step": 5509 }, { "epoch": 0.049738219895287955, "grad_norm": 3.9902830123901367, "learning_rate": 4.751354035024373e-05, "loss": 3.8918, "step": 5510 }, { "epoch": 0.049747246795450444, "grad_norm": 3.76947021484375, "learning_rate": 4.7513089005235606e-05, "loss": 5.2385, "step": 5511 }, { "epoch": 0.049756273695612925, "grad_norm": 3.703295946121216, "learning_rate": 4.7512637660227486e-05, "loss": 4.9159, "step": 5512 }, { "epoch": 0.049765300595775414, "grad_norm": 3.1939783096313477, "learning_rate": 4.751218631521935e-05, "loss": 4.6693, "step": 5513 }, { "epoch": 0.049774327495937895, "grad_norm": 3.2515816688537598, "learning_rate": 4.751173497021123e-05, "loss": 5.2534, "step": 5514 }, { "epoch": 0.04978335439610038, "grad_norm": 3.5088837146759033, "learning_rate": 4.751128362520311e-05, "loss": 4.451, "step": 5515 }, { "epoch": 0.049792381296262865, "grad_norm": 3.765716075897217, "learning_rate": 4.751083228019498e-05, "loss": 5.2695, "step": 5516 }, { "epoch": 0.04980140819642535, "grad_norm": 3.2307684421539307, "learning_rate": 4.751038093518686e-05, "loss": 4.5169, "step": 5517 }, { "epoch": 0.049810435096587835, "grad_norm": 3.3339648246765137, "learning_rate": 4.7509929590178734e-05, "loss": 4.645, "step": 5518 }, { "epoch": 0.04981946199675032, "grad_norm": 3.2895050048828125, "learning_rate": 4.750947824517061e-05, "loss": 4.7086, "step": 5519 }, { "epoch": 0.0498284888969128, "grad_norm": 3.4161741733551025, "learning_rate": 4.7509026900162486e-05, "loss": 4.8782, "step": 5520 }, { "epoch": 0.04983751579707529, "grad_norm": 3.9906132221221924, "learning_rate": 4.750857555515436e-05, "loss": 3.9749, "step": 5521 }, { "epoch": 0.04984654269723777, "grad_norm": 3.1016907691955566, "learning_rate": 4.750812421014624e-05, "loss": 4.2359, "step": 5522 }, { "epoch": 0.04985556959740025, "grad_norm": 3.5932159423828125, "learning_rate": 4.7507672865138117e-05, "loss": 4.8356, "step": 5523 }, { "epoch": 0.04986459649756274, "grad_norm": 2.951899766921997, "learning_rate": 4.750722152012999e-05, "loss": 4.6994, "step": 5524 }, { "epoch": 0.04987362339772522, "grad_norm": 3.185938835144043, "learning_rate": 4.750677017512187e-05, "loss": 4.1242, "step": 5525 }, { "epoch": 0.04988265029788771, "grad_norm": 4.094841957092285, "learning_rate": 4.750631883011374e-05, "loss": 3.9041, "step": 5526 }, { "epoch": 0.04989167719805019, "grad_norm": 3.038675308227539, "learning_rate": 4.750586748510561e-05, "loss": 4.2301, "step": 5527 }, { "epoch": 0.04990070409821267, "grad_norm": 4.18644380569458, "learning_rate": 4.750541614009749e-05, "loss": 4.0759, "step": 5528 }, { "epoch": 0.04990973099837516, "grad_norm": 3.4422078132629395, "learning_rate": 4.750496479508937e-05, "loss": 5.227, "step": 5529 }, { "epoch": 0.04991875789853764, "grad_norm": 3.022109270095825, "learning_rate": 4.7504513450081244e-05, "loss": 4.8641, "step": 5530 }, { "epoch": 0.04992778479870013, "grad_norm": 2.900071144104004, "learning_rate": 4.750406210507312e-05, "loss": 5.1647, "step": 5531 }, { "epoch": 0.04993681169886261, "grad_norm": 5.333555221557617, "learning_rate": 4.7503610760064996e-05, "loss": 4.5629, "step": 5532 }, { "epoch": 0.04994583859902509, "grad_norm": 5.014530658721924, "learning_rate": 4.750315941505687e-05, "loss": 5.2098, "step": 5533 }, { "epoch": 0.04995486549918758, "grad_norm": 3.0446388721466064, "learning_rate": 4.750270807004875e-05, "loss": 5.0105, "step": 5534 }, { "epoch": 0.04996389239935006, "grad_norm": 11.606042861938477, "learning_rate": 4.750225672504062e-05, "loss": 4.4947, "step": 5535 }, { "epoch": 0.049972919299512544, "grad_norm": 4.268584728240967, "learning_rate": 4.75018053800325e-05, "loss": 5.3452, "step": 5536 }, { "epoch": 0.04998194619967503, "grad_norm": 3.7238240242004395, "learning_rate": 4.750135403502438e-05, "loss": 3.8038, "step": 5537 }, { "epoch": 0.049990973099837514, "grad_norm": 3.3648531436920166, "learning_rate": 4.750090269001625e-05, "loss": 4.9168, "step": 5538 }, { "epoch": 0.05, "grad_norm": 3.3614046573638916, "learning_rate": 4.750045134500813e-05, "loss": 5.0797, "step": 5539 }, { "epoch": 0.050009026900162484, "grad_norm": 2.989983320236206, "learning_rate": 4.75e-05, "loss": 4.5599, "step": 5540 }, { "epoch": 0.050018053800324966, "grad_norm": 3.623171806335449, "learning_rate": 4.7499548654991875e-05, "loss": 4.3359, "step": 5541 }, { "epoch": 0.050027080700487454, "grad_norm": 3.125476598739624, "learning_rate": 4.7499097309983754e-05, "loss": 4.2392, "step": 5542 }, { "epoch": 0.050036107600649936, "grad_norm": 3.4036924839019775, "learning_rate": 4.749864596497563e-05, "loss": 4.749, "step": 5543 }, { "epoch": 0.050045134500812424, "grad_norm": 4.133789539337158, "learning_rate": 4.7498194619967506e-05, "loss": 4.5599, "step": 5544 }, { "epoch": 0.050054161400974906, "grad_norm": 3.7860214710235596, "learning_rate": 4.7497743274959385e-05, "loss": 5.4816, "step": 5545 }, { "epoch": 0.05006318830113739, "grad_norm": 3.6657774448394775, "learning_rate": 4.749729192995126e-05, "loss": 4.4808, "step": 5546 }, { "epoch": 0.050072215201299876, "grad_norm": 3.2856757640838623, "learning_rate": 4.749684058494313e-05, "loss": 4.5163, "step": 5547 }, { "epoch": 0.05008124210146236, "grad_norm": 3.06716251373291, "learning_rate": 4.749638923993501e-05, "loss": 4.3061, "step": 5548 }, { "epoch": 0.05009026900162484, "grad_norm": 4.171052932739258, "learning_rate": 4.749593789492688e-05, "loss": 4.4929, "step": 5549 }, { "epoch": 0.05009929590178733, "grad_norm": 4.665515899658203, "learning_rate": 4.749548654991876e-05, "loss": 5.3143, "step": 5550 }, { "epoch": 0.05010832280194981, "grad_norm": 4.6780900955200195, "learning_rate": 4.749503520491064e-05, "loss": 4.0451, "step": 5551 }, { "epoch": 0.0501173497021123, "grad_norm": 2.801486015319824, "learning_rate": 4.749458385990251e-05, "loss": 5.0421, "step": 5552 }, { "epoch": 0.05012637660227478, "grad_norm": 3.216201066970825, "learning_rate": 4.7494132514894385e-05, "loss": 5.0601, "step": 5553 }, { "epoch": 0.05013540350243726, "grad_norm": 4.213660717010498, "learning_rate": 4.7493681169886265e-05, "loss": 4.8936, "step": 5554 }, { "epoch": 0.05014443040259975, "grad_norm": 3.423809766769409, "learning_rate": 4.749322982487814e-05, "loss": 4.8615, "step": 5555 }, { "epoch": 0.05015345730276223, "grad_norm": 3.200859308242798, "learning_rate": 4.7492778479870016e-05, "loss": 5.1361, "step": 5556 }, { "epoch": 0.05016248420292472, "grad_norm": 2.873002290725708, "learning_rate": 4.749232713486189e-05, "loss": 4.6658, "step": 5557 }, { "epoch": 0.0501715111030872, "grad_norm": 2.9542107582092285, "learning_rate": 4.749187578985377e-05, "loss": 4.8787, "step": 5558 }, { "epoch": 0.05018053800324968, "grad_norm": 3.5376064777374268, "learning_rate": 4.749142444484565e-05, "loss": 4.5565, "step": 5559 }, { "epoch": 0.05018956490341217, "grad_norm": 3.663639545440674, "learning_rate": 4.749097309983751e-05, "loss": 4.6471, "step": 5560 }, { "epoch": 0.05019859180357465, "grad_norm": 3.273286819458008, "learning_rate": 4.749052175482939e-05, "loss": 4.3656, "step": 5561 }, { "epoch": 0.050207618703737134, "grad_norm": 3.648242712020874, "learning_rate": 4.749007040982127e-05, "loss": 4.8175, "step": 5562 }, { "epoch": 0.05021664560389962, "grad_norm": 3.1904284954071045, "learning_rate": 4.7489619064813144e-05, "loss": 4.3186, "step": 5563 }, { "epoch": 0.050225672504062104, "grad_norm": 3.494744062423706, "learning_rate": 4.748916771980502e-05, "loss": 4.4637, "step": 5564 }, { "epoch": 0.05023469940422459, "grad_norm": 3.923457145690918, "learning_rate": 4.7488716374796896e-05, "loss": 4.7013, "step": 5565 }, { "epoch": 0.050243726304387074, "grad_norm": 3.0538346767425537, "learning_rate": 4.7488265029788775e-05, "loss": 5.0161, "step": 5566 }, { "epoch": 0.050252753204549555, "grad_norm": 3.3776698112487793, "learning_rate": 4.748781368478065e-05, "loss": 4.4331, "step": 5567 }, { "epoch": 0.050261780104712044, "grad_norm": 4.288001537322998, "learning_rate": 4.7487362339772526e-05, "loss": 4.2314, "step": 5568 }, { "epoch": 0.050270807004874525, "grad_norm": 7.164994239807129, "learning_rate": 4.74869109947644e-05, "loss": 4.7586, "step": 5569 }, { "epoch": 0.050279833905037014, "grad_norm": 2.86692214012146, "learning_rate": 4.748645964975628e-05, "loss": 4.7184, "step": 5570 }, { "epoch": 0.050288860805199495, "grad_norm": 4.227648735046387, "learning_rate": 4.748600830474815e-05, "loss": 4.9075, "step": 5571 }, { "epoch": 0.05029788770536198, "grad_norm": 2.3632030487060547, "learning_rate": 4.748555695974003e-05, "loss": 4.7276, "step": 5572 }, { "epoch": 0.050306914605524465, "grad_norm": 3.08657169342041, "learning_rate": 4.74851056147319e-05, "loss": 4.8757, "step": 5573 }, { "epoch": 0.05031594150568695, "grad_norm": 4.418557643890381, "learning_rate": 4.7484654269723775e-05, "loss": 5.2923, "step": 5574 }, { "epoch": 0.05032496840584943, "grad_norm": 2.3772530555725098, "learning_rate": 4.7484202924715654e-05, "loss": 4.8793, "step": 5575 }, { "epoch": 0.05033399530601192, "grad_norm": 3.9680399894714355, "learning_rate": 4.748375157970753e-05, "loss": 5.0175, "step": 5576 }, { "epoch": 0.0503430222061744, "grad_norm": 3.1971395015716553, "learning_rate": 4.7483300234699406e-05, "loss": 4.8178, "step": 5577 }, { "epoch": 0.05035204910633689, "grad_norm": 3.4954869747161865, "learning_rate": 4.7482848889691285e-05, "loss": 4.6459, "step": 5578 }, { "epoch": 0.05036107600649937, "grad_norm": 4.476620674133301, "learning_rate": 4.748239754468316e-05, "loss": 4.286, "step": 5579 }, { "epoch": 0.05037010290666185, "grad_norm": 2.7983806133270264, "learning_rate": 4.748194619967503e-05, "loss": 5.0231, "step": 5580 }, { "epoch": 0.05037912980682434, "grad_norm": 4.0810089111328125, "learning_rate": 4.748149485466691e-05, "loss": 4.9648, "step": 5581 }, { "epoch": 0.05038815670698682, "grad_norm": 4.041627407073975, "learning_rate": 4.748104350965878e-05, "loss": 3.7773, "step": 5582 }, { "epoch": 0.05039718360714931, "grad_norm": 3.2339346408843994, "learning_rate": 4.748059216465066e-05, "loss": 4.4591, "step": 5583 }, { "epoch": 0.05040621050731179, "grad_norm": 3.4308860301971436, "learning_rate": 4.748014081964254e-05, "loss": 4.8608, "step": 5584 }, { "epoch": 0.05041523740747427, "grad_norm": 3.6362876892089844, "learning_rate": 4.747968947463441e-05, "loss": 5.0328, "step": 5585 }, { "epoch": 0.05042426430763676, "grad_norm": 3.704942226409912, "learning_rate": 4.747923812962629e-05, "loss": 3.6161, "step": 5586 }, { "epoch": 0.05043329120779924, "grad_norm": 3.5031700134277344, "learning_rate": 4.7478786784618164e-05, "loss": 4.5478, "step": 5587 }, { "epoch": 0.05044231810796172, "grad_norm": 3.286733865737915, "learning_rate": 4.747833543961004e-05, "loss": 4.7107, "step": 5588 }, { "epoch": 0.05045134500812421, "grad_norm": 3.2827067375183105, "learning_rate": 4.7477884094601916e-05, "loss": 5.636, "step": 5589 }, { "epoch": 0.05046037190828669, "grad_norm": 4.350510597229004, "learning_rate": 4.7477432749593795e-05, "loss": 4.8848, "step": 5590 }, { "epoch": 0.05046939880844918, "grad_norm": 4.291882038116455, "learning_rate": 4.747698140458567e-05, "loss": 4.3349, "step": 5591 }, { "epoch": 0.05047842570861166, "grad_norm": 3.402448892593384, "learning_rate": 4.747653005957755e-05, "loss": 4.7821, "step": 5592 }, { "epoch": 0.050487452608774144, "grad_norm": 4.946234703063965, "learning_rate": 4.747607871456942e-05, "loss": 4.0179, "step": 5593 }, { "epoch": 0.05049647950893663, "grad_norm": 3.8502285480499268, "learning_rate": 4.747562736956129e-05, "loss": 4.9562, "step": 5594 }, { "epoch": 0.050505506409099114, "grad_norm": 3.7885632514953613, "learning_rate": 4.747517602455317e-05, "loss": 4.9372, "step": 5595 }, { "epoch": 0.0505145333092616, "grad_norm": 3.3137967586517334, "learning_rate": 4.7474724679545044e-05, "loss": 4.3105, "step": 5596 }, { "epoch": 0.050523560209424084, "grad_norm": 3.9918222427368164, "learning_rate": 4.747427333453692e-05, "loss": 4.6001, "step": 5597 }, { "epoch": 0.050532587109586566, "grad_norm": 4.089661121368408, "learning_rate": 4.74738219895288e-05, "loss": 4.4909, "step": 5598 }, { "epoch": 0.050541614009749054, "grad_norm": 3.2212979793548584, "learning_rate": 4.7473370644520675e-05, "loss": 5.0143, "step": 5599 }, { "epoch": 0.050550640909911536, "grad_norm": 3.067023515701294, "learning_rate": 4.747291929951255e-05, "loss": 4.2756, "step": 5600 }, { "epoch": 0.05055966781007402, "grad_norm": 3.158184766769409, "learning_rate": 4.7472467954504426e-05, "loss": 5.141, "step": 5601 }, { "epoch": 0.050568694710236506, "grad_norm": 2.6664602756500244, "learning_rate": 4.74720166094963e-05, "loss": 4.8847, "step": 5602 }, { "epoch": 0.05057772161039899, "grad_norm": 3.605046510696411, "learning_rate": 4.747156526448818e-05, "loss": 4.6107, "step": 5603 }, { "epoch": 0.050586748510561476, "grad_norm": 3.4412097930908203, "learning_rate": 4.747111391948005e-05, "loss": 4.488, "step": 5604 }, { "epoch": 0.05059577541072396, "grad_norm": 3.0582475662231445, "learning_rate": 4.747066257447193e-05, "loss": 5.1105, "step": 5605 }, { "epoch": 0.05060480231088644, "grad_norm": 3.8657736778259277, "learning_rate": 4.747021122946381e-05, "loss": 3.3927, "step": 5606 }, { "epoch": 0.05061382921104893, "grad_norm": 3.2140841484069824, "learning_rate": 4.7469759884455675e-05, "loss": 4.8307, "step": 5607 }, { "epoch": 0.05062285611121141, "grad_norm": 3.127736806869507, "learning_rate": 4.7469308539447554e-05, "loss": 4.9423, "step": 5608 }, { "epoch": 0.0506318830113739, "grad_norm": 3.323141098022461, "learning_rate": 4.746885719443943e-05, "loss": 4.7477, "step": 5609 }, { "epoch": 0.05064090991153638, "grad_norm": 4.472989559173584, "learning_rate": 4.7468405849431306e-05, "loss": 5.2442, "step": 5610 }, { "epoch": 0.05064993681169886, "grad_norm": 3.234394073486328, "learning_rate": 4.7467954504423185e-05, "loss": 4.4177, "step": 5611 }, { "epoch": 0.05065896371186135, "grad_norm": 3.298283815383911, "learning_rate": 4.7467503159415064e-05, "loss": 4.861, "step": 5612 }, { "epoch": 0.05066799061202383, "grad_norm": 4.202398777008057, "learning_rate": 4.7467051814406936e-05, "loss": 4.277, "step": 5613 }, { "epoch": 0.05067701751218631, "grad_norm": 4.402309417724609, "learning_rate": 4.746660046939881e-05, "loss": 4.1739, "step": 5614 }, { "epoch": 0.0506860444123488, "grad_norm": 3.0960302352905273, "learning_rate": 4.746614912439069e-05, "loss": 4.4616, "step": 5615 }, { "epoch": 0.05069507131251128, "grad_norm": 3.371363639831543, "learning_rate": 4.746569777938256e-05, "loss": 4.4914, "step": 5616 }, { "epoch": 0.05070409821267377, "grad_norm": 2.9373462200164795, "learning_rate": 4.746524643437444e-05, "loss": 4.6006, "step": 5617 }, { "epoch": 0.05071312511283625, "grad_norm": 3.3491477966308594, "learning_rate": 4.746479508936631e-05, "loss": 4.6337, "step": 5618 }, { "epoch": 0.050722152012998734, "grad_norm": 3.843020439147949, "learning_rate": 4.746434374435819e-05, "loss": 4.552, "step": 5619 }, { "epoch": 0.05073117891316122, "grad_norm": 3.056244134902954, "learning_rate": 4.7463892399350064e-05, "loss": 5.0467, "step": 5620 }, { "epoch": 0.050740205813323704, "grad_norm": 4.1651291847229, "learning_rate": 4.7463441054341937e-05, "loss": 4.6474, "step": 5621 }, { "epoch": 0.05074923271348619, "grad_norm": 3.845198154449463, "learning_rate": 4.7462989709333816e-05, "loss": 4.2754, "step": 5622 }, { "epoch": 0.050758259613648674, "grad_norm": 4.009231090545654, "learning_rate": 4.7462538364325695e-05, "loss": 4.3972, "step": 5623 }, { "epoch": 0.050767286513811155, "grad_norm": 2.6869633197784424, "learning_rate": 4.746208701931757e-05, "loss": 5.0029, "step": 5624 }, { "epoch": 0.050776313413973644, "grad_norm": 5.243844509124756, "learning_rate": 4.746163567430945e-05, "loss": 3.8592, "step": 5625 }, { "epoch": 0.050785340314136125, "grad_norm": 3.1214141845703125, "learning_rate": 4.746118432930132e-05, "loss": 4.707, "step": 5626 }, { "epoch": 0.05079436721429861, "grad_norm": 3.391693592071533, "learning_rate": 4.746073298429319e-05, "loss": 4.9065, "step": 5627 }, { "epoch": 0.050803394114461095, "grad_norm": 2.7699058055877686, "learning_rate": 4.746028163928507e-05, "loss": 4.9091, "step": 5628 }, { "epoch": 0.05081242101462358, "grad_norm": 4.102895259857178, "learning_rate": 4.745983029427695e-05, "loss": 4.9393, "step": 5629 }, { "epoch": 0.050821447914786065, "grad_norm": 3.684999704360962, "learning_rate": 4.745937894926882e-05, "loss": 4.3637, "step": 5630 }, { "epoch": 0.05083047481494855, "grad_norm": 3.699279308319092, "learning_rate": 4.74589276042607e-05, "loss": 5.0366, "step": 5631 }, { "epoch": 0.05083950171511103, "grad_norm": 5.019641876220703, "learning_rate": 4.7458476259252574e-05, "loss": 3.5367, "step": 5632 }, { "epoch": 0.05084852861527352, "grad_norm": 2.4747884273529053, "learning_rate": 4.7458024914244454e-05, "loss": 4.7677, "step": 5633 }, { "epoch": 0.050857555515436, "grad_norm": 3.9463658332824707, "learning_rate": 4.7457573569236326e-05, "loss": 5.0158, "step": 5634 }, { "epoch": 0.05086658241559849, "grad_norm": 4.065112113952637, "learning_rate": 4.74571222242282e-05, "loss": 3.9999, "step": 5635 }, { "epoch": 0.05087560931576097, "grad_norm": 4.155209064483643, "learning_rate": 4.745667087922008e-05, "loss": 4.5389, "step": 5636 }, { "epoch": 0.05088463621592345, "grad_norm": 3.802983283996582, "learning_rate": 4.745621953421196e-05, "loss": 4.4312, "step": 5637 }, { "epoch": 0.05089366311608594, "grad_norm": 3.3273584842681885, "learning_rate": 4.745576818920383e-05, "loss": 4.54, "step": 5638 }, { "epoch": 0.05090269001624842, "grad_norm": 3.401994466781616, "learning_rate": 4.745531684419571e-05, "loss": 4.5997, "step": 5639 }, { "epoch": 0.0509117169164109, "grad_norm": 3.045050621032715, "learning_rate": 4.745486549918758e-05, "loss": 4.8156, "step": 5640 }, { "epoch": 0.05092074381657339, "grad_norm": 3.648642063140869, "learning_rate": 4.7454414154179454e-05, "loss": 4.1099, "step": 5641 }, { "epoch": 0.05092977071673587, "grad_norm": 3.547431230545044, "learning_rate": 4.745396280917133e-05, "loss": 4.318, "step": 5642 }, { "epoch": 0.05093879761689836, "grad_norm": 3.134742259979248, "learning_rate": 4.7453511464163205e-05, "loss": 5.2999, "step": 5643 }, { "epoch": 0.05094782451706084, "grad_norm": 3.8899710178375244, "learning_rate": 4.7453060119155085e-05, "loss": 4.1528, "step": 5644 }, { "epoch": 0.05095685141722332, "grad_norm": 2.9915101528167725, "learning_rate": 4.7452608774146964e-05, "loss": 4.8754, "step": 5645 }, { "epoch": 0.05096587831738581, "grad_norm": 4.213393688201904, "learning_rate": 4.7452157429138836e-05, "loss": 4.1107, "step": 5646 }, { "epoch": 0.05097490521754829, "grad_norm": 3.3605804443359375, "learning_rate": 4.745170608413071e-05, "loss": 4.6189, "step": 5647 }, { "epoch": 0.05098393211771078, "grad_norm": 3.5600838661193848, "learning_rate": 4.745125473912259e-05, "loss": 4.725, "step": 5648 }, { "epoch": 0.05099295901787326, "grad_norm": 4.00744104385376, "learning_rate": 4.745080339411446e-05, "loss": 5.2436, "step": 5649 }, { "epoch": 0.051001985918035744, "grad_norm": 3.715955972671509, "learning_rate": 4.745035204910634e-05, "loss": 4.3221, "step": 5650 }, { "epoch": 0.05101101281819823, "grad_norm": 2.041243553161621, "learning_rate": 4.744990070409822e-05, "loss": 4.8206, "step": 5651 }, { "epoch": 0.051020039718360714, "grad_norm": 2.68709659576416, "learning_rate": 4.744944935909009e-05, "loss": 4.8317, "step": 5652 }, { "epoch": 0.051029066618523196, "grad_norm": 4.137860298156738, "learning_rate": 4.744899801408197e-05, "loss": 4.7769, "step": 5653 }, { "epoch": 0.051038093518685684, "grad_norm": 3.5792417526245117, "learning_rate": 4.744854666907384e-05, "loss": 4.491, "step": 5654 }, { "epoch": 0.051047120418848166, "grad_norm": 2.930617094039917, "learning_rate": 4.7448095324065716e-05, "loss": 5.033, "step": 5655 }, { "epoch": 0.051056147319010654, "grad_norm": 3.4660263061523438, "learning_rate": 4.7447643979057595e-05, "loss": 4.2456, "step": 5656 }, { "epoch": 0.051065174219173136, "grad_norm": 2.690627336502075, "learning_rate": 4.744719263404947e-05, "loss": 4.6959, "step": 5657 }, { "epoch": 0.05107420111933562, "grad_norm": 3.258397102355957, "learning_rate": 4.7446741289041346e-05, "loss": 4.7818, "step": 5658 }, { "epoch": 0.051083228019498106, "grad_norm": 3.3170087337493896, "learning_rate": 4.7446289944033226e-05, "loss": 4.705, "step": 5659 }, { "epoch": 0.05109225491966059, "grad_norm": 4.245102405548096, "learning_rate": 4.74458385990251e-05, "loss": 5.0468, "step": 5660 }, { "epoch": 0.051101281819823076, "grad_norm": 2.9131693840026855, "learning_rate": 4.744538725401697e-05, "loss": 5.2282, "step": 5661 }, { "epoch": 0.05111030871998556, "grad_norm": 4.907872200012207, "learning_rate": 4.744493590900885e-05, "loss": 5.3346, "step": 5662 }, { "epoch": 0.05111933562014804, "grad_norm": 3.520360231399536, "learning_rate": 4.744448456400072e-05, "loss": 4.4732, "step": 5663 }, { "epoch": 0.05112836252031053, "grad_norm": 3.3787307739257812, "learning_rate": 4.74440332189926e-05, "loss": 5.3625, "step": 5664 }, { "epoch": 0.05113738942047301, "grad_norm": 3.1141762733459473, "learning_rate": 4.7443581873984474e-05, "loss": 5.0592, "step": 5665 }, { "epoch": 0.05114641632063549, "grad_norm": 2.9018185138702393, "learning_rate": 4.744313052897635e-05, "loss": 4.4595, "step": 5666 }, { "epoch": 0.05115544322079798, "grad_norm": 2.4435336589813232, "learning_rate": 4.7442679183968226e-05, "loss": 4.6595, "step": 5667 }, { "epoch": 0.05116447012096046, "grad_norm": 2.7649433612823486, "learning_rate": 4.74422278389601e-05, "loss": 4.7043, "step": 5668 }, { "epoch": 0.05117349702112295, "grad_norm": 3.403268814086914, "learning_rate": 4.744177649395198e-05, "loss": 4.8841, "step": 5669 }, { "epoch": 0.05118252392128543, "grad_norm": 3.6697561740875244, "learning_rate": 4.744132514894386e-05, "loss": 4.3047, "step": 5670 }, { "epoch": 0.05119155082144791, "grad_norm": 3.185189962387085, "learning_rate": 4.744087380393573e-05, "loss": 5.039, "step": 5671 }, { "epoch": 0.0512005777216104, "grad_norm": 3.166325330734253, "learning_rate": 4.744042245892761e-05, "loss": 5.3077, "step": 5672 }, { "epoch": 0.05120960462177288, "grad_norm": 3.1793408393859863, "learning_rate": 4.743997111391949e-05, "loss": 5.4135, "step": 5673 }, { "epoch": 0.05121863152193537, "grad_norm": 3.4917919635772705, "learning_rate": 4.743951976891135e-05, "loss": 4.5101, "step": 5674 }, { "epoch": 0.05122765842209785, "grad_norm": 3.870628595352173, "learning_rate": 4.743906842390323e-05, "loss": 5.0852, "step": 5675 }, { "epoch": 0.051236685322260334, "grad_norm": 3.274912118911743, "learning_rate": 4.743861707889511e-05, "loss": 4.5546, "step": 5676 }, { "epoch": 0.05124571222242282, "grad_norm": 4.2594122886657715, "learning_rate": 4.7438165733886984e-05, "loss": 5.0474, "step": 5677 }, { "epoch": 0.051254739122585304, "grad_norm": 3.5274295806884766, "learning_rate": 4.7437714388878864e-05, "loss": 4.7499, "step": 5678 }, { "epoch": 0.051263766022747785, "grad_norm": 8.806198120117188, "learning_rate": 4.7437263043870736e-05, "loss": 3.9378, "step": 5679 }, { "epoch": 0.051272792922910274, "grad_norm": 4.137459754943848, "learning_rate": 4.7436811698862615e-05, "loss": 4.1117, "step": 5680 }, { "epoch": 0.051281819823072755, "grad_norm": 4.527613162994385, "learning_rate": 4.743636035385449e-05, "loss": 4.9097, "step": 5681 }, { "epoch": 0.051290846723235244, "grad_norm": 3.8770675659179688, "learning_rate": 4.743590900884636e-05, "loss": 4.8308, "step": 5682 }, { "epoch": 0.051299873623397725, "grad_norm": 3.1901230812072754, "learning_rate": 4.743545766383824e-05, "loss": 4.8135, "step": 5683 }, { "epoch": 0.05130890052356021, "grad_norm": 3.546469211578369, "learning_rate": 4.743500631883012e-05, "loss": 5.1977, "step": 5684 }, { "epoch": 0.051317927423722695, "grad_norm": 2.598146438598633, "learning_rate": 4.743455497382199e-05, "loss": 4.5364, "step": 5685 }, { "epoch": 0.05132695432388518, "grad_norm": 3.8206472396850586, "learning_rate": 4.743410362881387e-05, "loss": 4.5112, "step": 5686 }, { "epoch": 0.051335981224047665, "grad_norm": 3.5504977703094482, "learning_rate": 4.743365228380574e-05, "loss": 4.0076, "step": 5687 }, { "epoch": 0.05134500812421015, "grad_norm": 3.257936716079712, "learning_rate": 4.7433200938797615e-05, "loss": 4.5576, "step": 5688 }, { "epoch": 0.05135403502437263, "grad_norm": 2.3589367866516113, "learning_rate": 4.7432749593789495e-05, "loss": 4.4945, "step": 5689 }, { "epoch": 0.05136306192453512, "grad_norm": 3.2660677433013916, "learning_rate": 4.7432298248781374e-05, "loss": 4.4065, "step": 5690 }, { "epoch": 0.0513720888246976, "grad_norm": 4.748043060302734, "learning_rate": 4.7431846903773246e-05, "loss": 4.9441, "step": 5691 }, { "epoch": 0.05138111572486008, "grad_norm": 3.1016311645507812, "learning_rate": 4.7431395558765125e-05, "loss": 5.0382, "step": 5692 }, { "epoch": 0.05139014262502257, "grad_norm": 2.9423818588256836, "learning_rate": 4.7430944213757e-05, "loss": 4.8754, "step": 5693 }, { "epoch": 0.05139916952518505, "grad_norm": 3.972670078277588, "learning_rate": 4.743049286874887e-05, "loss": 4.6582, "step": 5694 }, { "epoch": 0.05140819642534754, "grad_norm": 3.9877758026123047, "learning_rate": 4.743004152374075e-05, "loss": 5.3185, "step": 5695 }, { "epoch": 0.05141722332551002, "grad_norm": 10.34790325164795, "learning_rate": 4.742959017873262e-05, "loss": 5.6218, "step": 5696 }, { "epoch": 0.0514262502256725, "grad_norm": 3.1104350090026855, "learning_rate": 4.74291388337245e-05, "loss": 4.7519, "step": 5697 }, { "epoch": 0.05143527712583499, "grad_norm": 3.7338595390319824, "learning_rate": 4.742868748871638e-05, "loss": 4.34, "step": 5698 }, { "epoch": 0.05144430402599747, "grad_norm": 2.7300925254821777, "learning_rate": 4.742823614370825e-05, "loss": 4.2612, "step": 5699 }, { "epoch": 0.05145333092615996, "grad_norm": 3.8887369632720947, "learning_rate": 4.742778479870013e-05, "loss": 4.8653, "step": 5700 }, { "epoch": 0.05146235782632244, "grad_norm": 3.876734733581543, "learning_rate": 4.7427333453692005e-05, "loss": 4.8396, "step": 5701 }, { "epoch": 0.05147138472648492, "grad_norm": 3.2355995178222656, "learning_rate": 4.742688210868388e-05, "loss": 4.527, "step": 5702 }, { "epoch": 0.05148041162664741, "grad_norm": 3.2718374729156494, "learning_rate": 4.7426430763675756e-05, "loss": 4.847, "step": 5703 }, { "epoch": 0.05148943852680989, "grad_norm": 4.855353355407715, "learning_rate": 4.742597941866763e-05, "loss": 4.7152, "step": 5704 }, { "epoch": 0.051498465426972374, "grad_norm": 3.490678071975708, "learning_rate": 4.742552807365951e-05, "loss": 4.9383, "step": 5705 }, { "epoch": 0.05150749232713486, "grad_norm": 2.5053579807281494, "learning_rate": 4.742507672865139e-05, "loss": 4.6288, "step": 5706 }, { "epoch": 0.051516519227297344, "grad_norm": 2.7493350505828857, "learning_rate": 4.742462538364326e-05, "loss": 4.7134, "step": 5707 }, { "epoch": 0.05152554612745983, "grad_norm": 2.7806015014648438, "learning_rate": 4.742417403863513e-05, "loss": 4.0881, "step": 5708 }, { "epoch": 0.051534573027622314, "grad_norm": 3.3209218978881836, "learning_rate": 4.742372269362701e-05, "loss": 4.7219, "step": 5709 }, { "epoch": 0.051543599927784796, "grad_norm": 3.4245259761810303, "learning_rate": 4.7423271348618884e-05, "loss": 5.0793, "step": 5710 }, { "epoch": 0.051552626827947284, "grad_norm": 3.852639675140381, "learning_rate": 4.742282000361076e-05, "loss": 4.3634, "step": 5711 }, { "epoch": 0.051561653728109766, "grad_norm": 4.611656665802002, "learning_rate": 4.742236865860264e-05, "loss": 4.4542, "step": 5712 }, { "epoch": 0.051570680628272254, "grad_norm": 2.606557607650757, "learning_rate": 4.7421917313594515e-05, "loss": 4.8572, "step": 5713 }, { "epoch": 0.051579707528434736, "grad_norm": 3.376267433166504, "learning_rate": 4.742146596858639e-05, "loss": 4.4898, "step": 5714 }, { "epoch": 0.05158873442859722, "grad_norm": 3.3888204097747803, "learning_rate": 4.742101462357827e-05, "loss": 4.5602, "step": 5715 }, { "epoch": 0.051597761328759706, "grad_norm": 3.296640396118164, "learning_rate": 4.742056327857014e-05, "loss": 4.7858, "step": 5716 }, { "epoch": 0.05160678822892219, "grad_norm": 3.0737903118133545, "learning_rate": 4.742011193356202e-05, "loss": 4.5878, "step": 5717 }, { "epoch": 0.05161581512908467, "grad_norm": 5.7004313468933105, "learning_rate": 4.741966058855389e-05, "loss": 4.9, "step": 5718 }, { "epoch": 0.05162484202924716, "grad_norm": 4.12216854095459, "learning_rate": 4.741920924354577e-05, "loss": 5.0626, "step": 5719 }, { "epoch": 0.05163386892940964, "grad_norm": 3.7362589836120605, "learning_rate": 4.741875789853765e-05, "loss": 4.2904, "step": 5720 }, { "epoch": 0.05164289582957213, "grad_norm": 3.9042186737060547, "learning_rate": 4.7418306553529515e-05, "loss": 4.9246, "step": 5721 }, { "epoch": 0.05165192272973461, "grad_norm": 3.2804017066955566, "learning_rate": 4.7417855208521394e-05, "loss": 4.9948, "step": 5722 }, { "epoch": 0.05166094962989709, "grad_norm": 3.712752342224121, "learning_rate": 4.7417403863513274e-05, "loss": 4.4856, "step": 5723 }, { "epoch": 0.05166997653005958, "grad_norm": 3.2310028076171875, "learning_rate": 4.7416952518505146e-05, "loss": 4.9998, "step": 5724 }, { "epoch": 0.05167900343022206, "grad_norm": 3.7422568798065186, "learning_rate": 4.7416501173497025e-05, "loss": 4.4929, "step": 5725 }, { "epoch": 0.05168803033038455, "grad_norm": 3.2749850749969482, "learning_rate": 4.74160498284889e-05, "loss": 4.8992, "step": 5726 }, { "epoch": 0.05169705723054703, "grad_norm": 3.3836731910705566, "learning_rate": 4.741559848348078e-05, "loss": 4.3482, "step": 5727 }, { "epoch": 0.05170608413070951, "grad_norm": 3.614001512527466, "learning_rate": 4.741514713847265e-05, "loss": 4.8102, "step": 5728 }, { "epoch": 0.051715111030872, "grad_norm": 3.303567409515381, "learning_rate": 4.741469579346453e-05, "loss": 4.2628, "step": 5729 }, { "epoch": 0.05172413793103448, "grad_norm": 3.5157203674316406, "learning_rate": 4.74142444484564e-05, "loss": 4.794, "step": 5730 }, { "epoch": 0.051733164831196964, "grad_norm": 3.33744740486145, "learning_rate": 4.741379310344828e-05, "loss": 4.9685, "step": 5731 }, { "epoch": 0.05174219173135945, "grad_norm": 3.377634048461914, "learning_rate": 4.741334175844015e-05, "loss": 4.7473, "step": 5732 }, { "epoch": 0.051751218631521934, "grad_norm": 3.8874928951263428, "learning_rate": 4.741289041343203e-05, "loss": 4.7919, "step": 5733 }, { "epoch": 0.05176024553168442, "grad_norm": 4.281409740447998, "learning_rate": 4.7412439068423905e-05, "loss": 4.6683, "step": 5734 }, { "epoch": 0.051769272431846904, "grad_norm": 4.502560138702393, "learning_rate": 4.741198772341578e-05, "loss": 4.6171, "step": 5735 }, { "epoch": 0.051778299332009385, "grad_norm": 4.1363348960876465, "learning_rate": 4.7411536378407656e-05, "loss": 4.9873, "step": 5736 }, { "epoch": 0.051787326232171874, "grad_norm": 3.7606868743896484, "learning_rate": 4.7411085033399535e-05, "loss": 4.4209, "step": 5737 }, { "epoch": 0.051796353132334355, "grad_norm": 3.3716800212860107, "learning_rate": 4.741063368839141e-05, "loss": 3.6451, "step": 5738 }, { "epoch": 0.051805380032496844, "grad_norm": 3.507615089416504, "learning_rate": 4.741018234338329e-05, "loss": 4.5332, "step": 5739 }, { "epoch": 0.051814406932659325, "grad_norm": 4.863586902618408, "learning_rate": 4.740973099837516e-05, "loss": 4.5676, "step": 5740 }, { "epoch": 0.05182343383282181, "grad_norm": 5.469146251678467, "learning_rate": 4.740927965336703e-05, "loss": 4.3373, "step": 5741 }, { "epoch": 0.051832460732984295, "grad_norm": 4.5702643394470215, "learning_rate": 4.740882830835891e-05, "loss": 5.0069, "step": 5742 }, { "epoch": 0.05184148763314678, "grad_norm": 3.797409772872925, "learning_rate": 4.7408376963350784e-05, "loss": 4.7117, "step": 5743 }, { "epoch": 0.05185051453330926, "grad_norm": 2.9002532958984375, "learning_rate": 4.740792561834266e-05, "loss": 5.1316, "step": 5744 }, { "epoch": 0.05185954143347175, "grad_norm": 3.0694539546966553, "learning_rate": 4.740747427333454e-05, "loss": 4.7516, "step": 5745 }, { "epoch": 0.05186856833363423, "grad_norm": 3.9948933124542236, "learning_rate": 4.7407022928326415e-05, "loss": 4.5038, "step": 5746 }, { "epoch": 0.05187759523379672, "grad_norm": 3.076662063598633, "learning_rate": 4.7406571583318294e-05, "loss": 5.0687, "step": 5747 }, { "epoch": 0.0518866221339592, "grad_norm": 3.040865659713745, "learning_rate": 4.7406120238310166e-05, "loss": 4.5753, "step": 5748 }, { "epoch": 0.05189564903412168, "grad_norm": 3.0455262660980225, "learning_rate": 4.740566889330204e-05, "loss": 4.8611, "step": 5749 }, { "epoch": 0.05190467593428417, "grad_norm": 3.377321720123291, "learning_rate": 4.740521754829392e-05, "loss": 4.8953, "step": 5750 }, { "epoch": 0.05191370283444665, "grad_norm": 3.119812488555908, "learning_rate": 4.74047662032858e-05, "loss": 4.9994, "step": 5751 }, { "epoch": 0.05192272973460914, "grad_norm": 5.293556213378906, "learning_rate": 4.740431485827767e-05, "loss": 3.736, "step": 5752 }, { "epoch": 0.05193175663477162, "grad_norm": 2.6273040771484375, "learning_rate": 4.740386351326955e-05, "loss": 4.3225, "step": 5753 }, { "epoch": 0.0519407835349341, "grad_norm": 3.627422332763672, "learning_rate": 4.740341216826142e-05, "loss": 4.8678, "step": 5754 }, { "epoch": 0.05194981043509659, "grad_norm": 4.773269176483154, "learning_rate": 4.7402960823253294e-05, "loss": 5.1313, "step": 5755 }, { "epoch": 0.05195883733525907, "grad_norm": 3.318767547607422, "learning_rate": 4.740250947824517e-05, "loss": 4.8871, "step": 5756 }, { "epoch": 0.05196786423542155, "grad_norm": 3.246304988861084, "learning_rate": 4.7402058133237046e-05, "loss": 5.0702, "step": 5757 }, { "epoch": 0.05197689113558404, "grad_norm": 4.4508209228515625, "learning_rate": 4.7401606788228925e-05, "loss": 5.1392, "step": 5758 }, { "epoch": 0.05198591803574652, "grad_norm": 3.2327842712402344, "learning_rate": 4.7401155443220804e-05, "loss": 4.7371, "step": 5759 }, { "epoch": 0.05199494493590901, "grad_norm": 5.331404685974121, "learning_rate": 4.740070409821268e-05, "loss": 5.1506, "step": 5760 }, { "epoch": 0.05200397183607149, "grad_norm": 3.7324490547180176, "learning_rate": 4.740025275320455e-05, "loss": 4.7958, "step": 5761 }, { "epoch": 0.052012998736233974, "grad_norm": 2.89022159576416, "learning_rate": 4.739980140819643e-05, "loss": 4.7606, "step": 5762 }, { "epoch": 0.05202202563639646, "grad_norm": 4.227753162384033, "learning_rate": 4.73993500631883e-05, "loss": 4.1566, "step": 5763 }, { "epoch": 0.052031052536558944, "grad_norm": 4.002079010009766, "learning_rate": 4.739889871818018e-05, "loss": 4.7153, "step": 5764 }, { "epoch": 0.05204007943672143, "grad_norm": 3.0148563385009766, "learning_rate": 4.739844737317205e-05, "loss": 4.8121, "step": 5765 }, { "epoch": 0.052049106336883914, "grad_norm": 3.660972833633423, "learning_rate": 4.739799602816393e-05, "loss": 4.4102, "step": 5766 }, { "epoch": 0.052058133237046396, "grad_norm": 3.95715594291687, "learning_rate": 4.739754468315581e-05, "loss": 5.1961, "step": 5767 }, { "epoch": 0.052067160137208884, "grad_norm": 4.169416904449463, "learning_rate": 4.739709333814768e-05, "loss": 4.5769, "step": 5768 }, { "epoch": 0.052076187037371366, "grad_norm": 4.3844194412231445, "learning_rate": 4.7396641993139556e-05, "loss": 4.4071, "step": 5769 }, { "epoch": 0.05208521393753385, "grad_norm": 3.927030086517334, "learning_rate": 4.7396190648131435e-05, "loss": 4.4857, "step": 5770 }, { "epoch": 0.052094240837696336, "grad_norm": 4.0903754234313965, "learning_rate": 4.739573930312331e-05, "loss": 4.6703, "step": 5771 }, { "epoch": 0.05210326773785882, "grad_norm": 4.045041561126709, "learning_rate": 4.739528795811519e-05, "loss": 4.5437, "step": 5772 }, { "epoch": 0.052112294638021306, "grad_norm": 3.276824712753296, "learning_rate": 4.7394836613107066e-05, "loss": 4.581, "step": 5773 }, { "epoch": 0.05212132153818379, "grad_norm": 3.678518533706665, "learning_rate": 4.739438526809894e-05, "loss": 5.1394, "step": 5774 }, { "epoch": 0.05213034843834627, "grad_norm": 3.041217803955078, "learning_rate": 4.739393392309081e-05, "loss": 4.9128, "step": 5775 }, { "epoch": 0.05213937533850876, "grad_norm": 4.831362724304199, "learning_rate": 4.739348257808269e-05, "loss": 4.4083, "step": 5776 }, { "epoch": 0.05214840223867124, "grad_norm": 3.870908260345459, "learning_rate": 4.739303123307456e-05, "loss": 5.1249, "step": 5777 }, { "epoch": 0.05215742913883373, "grad_norm": 3.327221155166626, "learning_rate": 4.739257988806644e-05, "loss": 4.9295, "step": 5778 }, { "epoch": 0.05216645603899621, "grad_norm": 2.6160178184509277, "learning_rate": 4.7392128543058315e-05, "loss": 5.25, "step": 5779 }, { "epoch": 0.05217548293915869, "grad_norm": 3.8228604793548584, "learning_rate": 4.7391677198050194e-05, "loss": 4.682, "step": 5780 }, { "epoch": 0.05218450983932118, "grad_norm": 3.2035751342773438, "learning_rate": 4.7391225853042066e-05, "loss": 5.1365, "step": 5781 }, { "epoch": 0.05219353673948366, "grad_norm": 2.8055944442749023, "learning_rate": 4.739077450803394e-05, "loss": 4.9381, "step": 5782 }, { "epoch": 0.05220256363964614, "grad_norm": 3.804544687271118, "learning_rate": 4.739032316302582e-05, "loss": 5.2358, "step": 5783 }, { "epoch": 0.05221159053980863, "grad_norm": 3.4819400310516357, "learning_rate": 4.73898718180177e-05, "loss": 5.0143, "step": 5784 }, { "epoch": 0.05222061743997111, "grad_norm": 3.140237808227539, "learning_rate": 4.738942047300957e-05, "loss": 4.6381, "step": 5785 }, { "epoch": 0.0522296443401336, "grad_norm": 4.109978675842285, "learning_rate": 4.738896912800145e-05, "loss": 4.7302, "step": 5786 }, { "epoch": 0.05223867124029608, "grad_norm": 2.5095086097717285, "learning_rate": 4.738851778299332e-05, "loss": 4.4961, "step": 5787 }, { "epoch": 0.052247698140458564, "grad_norm": 3.264057159423828, "learning_rate": 4.7388066437985194e-05, "loss": 4.8981, "step": 5788 }, { "epoch": 0.05225672504062105, "grad_norm": 11.998217582702637, "learning_rate": 4.738761509297707e-05, "loss": 4.3868, "step": 5789 }, { "epoch": 0.052265751940783534, "grad_norm": 3.25610089302063, "learning_rate": 4.738716374796895e-05, "loss": 4.7018, "step": 5790 }, { "epoch": 0.05227477884094602, "grad_norm": 3.458261489868164, "learning_rate": 4.7386712402960825e-05, "loss": 4.162, "step": 5791 }, { "epoch": 0.052283805741108504, "grad_norm": 3.8527204990386963, "learning_rate": 4.7386261057952704e-05, "loss": 5.0037, "step": 5792 }, { "epoch": 0.052292832641270985, "grad_norm": 3.5062124729156494, "learning_rate": 4.7385809712944576e-05, "loss": 4.5135, "step": 5793 }, { "epoch": 0.052301859541433474, "grad_norm": 4.7043776512146, "learning_rate": 4.7385358367936456e-05, "loss": 4.7076, "step": 5794 }, { "epoch": 0.052310886441595955, "grad_norm": 4.148157596588135, "learning_rate": 4.738490702292833e-05, "loss": 4.955, "step": 5795 }, { "epoch": 0.052319913341758444, "grad_norm": 4.804696559906006, "learning_rate": 4.73844556779202e-05, "loss": 4.7064, "step": 5796 }, { "epoch": 0.052328940241920925, "grad_norm": 3.040548324584961, "learning_rate": 4.738400433291208e-05, "loss": 4.7782, "step": 5797 }, { "epoch": 0.05233796714208341, "grad_norm": 3.9844400882720947, "learning_rate": 4.738355298790396e-05, "loss": 5.0093, "step": 5798 }, { "epoch": 0.052346994042245895, "grad_norm": 3.189008951187134, "learning_rate": 4.738310164289583e-05, "loss": 4.7933, "step": 5799 }, { "epoch": 0.05235602094240838, "grad_norm": 3.8276517391204834, "learning_rate": 4.738265029788771e-05, "loss": 3.735, "step": 5800 }, { "epoch": 0.05236504784257086, "grad_norm": 2.990601062774658, "learning_rate": 4.738219895287958e-05, "loss": 4.173, "step": 5801 }, { "epoch": 0.05237407474273335, "grad_norm": 3.225602388381958, "learning_rate": 4.7381747607871456e-05, "loss": 4.4551, "step": 5802 }, { "epoch": 0.05238310164289583, "grad_norm": 3.010338544845581, "learning_rate": 4.7381296262863335e-05, "loss": 4.5639, "step": 5803 }, { "epoch": 0.05239212854305832, "grad_norm": 3.2943246364593506, "learning_rate": 4.738084491785521e-05, "loss": 4.6338, "step": 5804 }, { "epoch": 0.0524011554432208, "grad_norm": 4.33919095993042, "learning_rate": 4.738039357284709e-05, "loss": 4.3372, "step": 5805 }, { "epoch": 0.05241018234338328, "grad_norm": 3.5038018226623535, "learning_rate": 4.7379942227838966e-05, "loss": 5.1002, "step": 5806 }, { "epoch": 0.05241920924354577, "grad_norm": 3.4562578201293945, "learning_rate": 4.737949088283084e-05, "loss": 4.5001, "step": 5807 }, { "epoch": 0.05242823614370825, "grad_norm": 3.384486436843872, "learning_rate": 4.737903953782271e-05, "loss": 4.4361, "step": 5808 }, { "epoch": 0.05243726304387074, "grad_norm": 3.330496072769165, "learning_rate": 4.737858819281459e-05, "loss": 4.3046, "step": 5809 }, { "epoch": 0.05244628994403322, "grad_norm": 4.26151704788208, "learning_rate": 4.737813684780646e-05, "loss": 4.6451, "step": 5810 }, { "epoch": 0.0524553168441957, "grad_norm": 3.8663578033447266, "learning_rate": 4.737768550279834e-05, "loss": 4.4262, "step": 5811 }, { "epoch": 0.05246434374435819, "grad_norm": 3.717512845993042, "learning_rate": 4.737723415779022e-05, "loss": 4.7762, "step": 5812 }, { "epoch": 0.05247337064452067, "grad_norm": 2.6513471603393555, "learning_rate": 4.7376782812782094e-05, "loss": 4.3486, "step": 5813 }, { "epoch": 0.05248239754468315, "grad_norm": 4.030007839202881, "learning_rate": 4.737633146777397e-05, "loss": 4.8419, "step": 5814 }, { "epoch": 0.05249142444484564, "grad_norm": 3.8176350593566895, "learning_rate": 4.7375880122765845e-05, "loss": 4.3752, "step": 5815 }, { "epoch": 0.05250045134500812, "grad_norm": 4.216625213623047, "learning_rate": 4.737542877775772e-05, "loss": 4.6149, "step": 5816 }, { "epoch": 0.05250947824517061, "grad_norm": 2.6846511363983154, "learning_rate": 4.73749774327496e-05, "loss": 4.7175, "step": 5817 }, { "epoch": 0.05251850514533309, "grad_norm": 3.644289970397949, "learning_rate": 4.737452608774147e-05, "loss": 4.7644, "step": 5818 }, { "epoch": 0.052527532045495574, "grad_norm": 3.8696203231811523, "learning_rate": 4.737407474273335e-05, "loss": 4.4168, "step": 5819 }, { "epoch": 0.05253655894565806, "grad_norm": 3.6751742362976074, "learning_rate": 4.737362339772523e-05, "loss": 5.2936, "step": 5820 }, { "epoch": 0.052545585845820544, "grad_norm": 3.641103744506836, "learning_rate": 4.73731720527171e-05, "loss": 4.2968, "step": 5821 }, { "epoch": 0.05255461274598303, "grad_norm": 3.39176869392395, "learning_rate": 4.737272070770897e-05, "loss": 3.6231, "step": 5822 }, { "epoch": 0.052563639646145514, "grad_norm": 4.872692584991455, "learning_rate": 4.737226936270085e-05, "loss": 4.1524, "step": 5823 }, { "epoch": 0.052572666546307996, "grad_norm": 3.9451253414154053, "learning_rate": 4.7371818017692725e-05, "loss": 5.4311, "step": 5824 }, { "epoch": 0.052581693446470484, "grad_norm": 2.908843994140625, "learning_rate": 4.7371366672684604e-05, "loss": 4.797, "step": 5825 }, { "epoch": 0.052590720346632966, "grad_norm": 4.089914321899414, "learning_rate": 4.7370915327676476e-05, "loss": 4.3341, "step": 5826 }, { "epoch": 0.05259974724679545, "grad_norm": 3.3509788513183594, "learning_rate": 4.7370463982668355e-05, "loss": 4.7448, "step": 5827 }, { "epoch": 0.052608774146957936, "grad_norm": 4.474333763122559, "learning_rate": 4.737001263766023e-05, "loss": 5.0201, "step": 5828 }, { "epoch": 0.05261780104712042, "grad_norm": 3.19724178314209, "learning_rate": 4.73695612926521e-05, "loss": 4.9615, "step": 5829 }, { "epoch": 0.052626827947282906, "grad_norm": 3.716594934463501, "learning_rate": 4.736910994764398e-05, "loss": 4.7995, "step": 5830 }, { "epoch": 0.05263585484744539, "grad_norm": 4.005665302276611, "learning_rate": 4.736865860263586e-05, "loss": 4.3333, "step": 5831 }, { "epoch": 0.05264488174760787, "grad_norm": 2.536726713180542, "learning_rate": 4.736820725762773e-05, "loss": 5.1049, "step": 5832 }, { "epoch": 0.05265390864777036, "grad_norm": 3.569535255432129, "learning_rate": 4.736775591261961e-05, "loss": 4.6619, "step": 5833 }, { "epoch": 0.05266293554793284, "grad_norm": 3.9235339164733887, "learning_rate": 4.736730456761149e-05, "loss": 5.2226, "step": 5834 }, { "epoch": 0.05267196244809533, "grad_norm": 3.246830940246582, "learning_rate": 4.7366853222603356e-05, "loss": 4.9594, "step": 5835 }, { "epoch": 0.05268098934825781, "grad_norm": 4.253094673156738, "learning_rate": 4.7366401877595235e-05, "loss": 4.6551, "step": 5836 }, { "epoch": 0.05269001624842029, "grad_norm": 3.5675129890441895, "learning_rate": 4.7365950532587114e-05, "loss": 4.815, "step": 5837 }, { "epoch": 0.05269904314858278, "grad_norm": 3.316141366958618, "learning_rate": 4.7365499187578986e-05, "loss": 5.3745, "step": 5838 }, { "epoch": 0.05270807004874526, "grad_norm": 3.2170181274414062, "learning_rate": 4.7365047842570866e-05, "loss": 4.3309, "step": 5839 }, { "epoch": 0.05271709694890774, "grad_norm": 2.9026100635528564, "learning_rate": 4.736459649756274e-05, "loss": 4.3852, "step": 5840 }, { "epoch": 0.05272612384907023, "grad_norm": 3.6777679920196533, "learning_rate": 4.736414515255462e-05, "loss": 4.2187, "step": 5841 }, { "epoch": 0.05273515074923271, "grad_norm": 3.83535099029541, "learning_rate": 4.736369380754649e-05, "loss": 4.2428, "step": 5842 }, { "epoch": 0.0527441776493952, "grad_norm": 3.2595458030700684, "learning_rate": 4.736324246253836e-05, "loss": 4.2258, "step": 5843 }, { "epoch": 0.05275320454955768, "grad_norm": 4.03408145904541, "learning_rate": 4.736279111753024e-05, "loss": 4.0794, "step": 5844 }, { "epoch": 0.052762231449720164, "grad_norm": 4.982542514801025, "learning_rate": 4.736233977252212e-05, "loss": 4.2864, "step": 5845 }, { "epoch": 0.05277125834988265, "grad_norm": 3.8984293937683105, "learning_rate": 4.736188842751399e-05, "loss": 4.6999, "step": 5846 }, { "epoch": 0.052780285250045134, "grad_norm": 4.702534198760986, "learning_rate": 4.736143708250587e-05, "loss": 3.6296, "step": 5847 }, { "epoch": 0.05278931215020762, "grad_norm": 5.3626179695129395, "learning_rate": 4.7360985737497745e-05, "loss": 4.1234, "step": 5848 }, { "epoch": 0.052798339050370104, "grad_norm": 4.79099178314209, "learning_rate": 4.736053439248962e-05, "loss": 4.8953, "step": 5849 }, { "epoch": 0.052807365950532585, "grad_norm": 3.383175849914551, "learning_rate": 4.73600830474815e-05, "loss": 4.3202, "step": 5850 }, { "epoch": 0.052816392850695074, "grad_norm": 4.1095805168151855, "learning_rate": 4.7359631702473376e-05, "loss": 5.4505, "step": 5851 }, { "epoch": 0.052825419750857555, "grad_norm": 5.067244052886963, "learning_rate": 4.735918035746525e-05, "loss": 3.7708, "step": 5852 }, { "epoch": 0.05283444665102004, "grad_norm": 2.5610814094543457, "learning_rate": 4.735872901245713e-05, "loss": 4.9535, "step": 5853 }, { "epoch": 0.052843473551182525, "grad_norm": 3.6191041469573975, "learning_rate": 4.7358277667449e-05, "loss": 5.0941, "step": 5854 }, { "epoch": 0.05285250045134501, "grad_norm": 4.23582124710083, "learning_rate": 4.735782632244087e-05, "loss": 3.7152, "step": 5855 }, { "epoch": 0.052861527351507495, "grad_norm": 2.841221570968628, "learning_rate": 4.735737497743275e-05, "loss": 4.6082, "step": 5856 }, { "epoch": 0.05287055425166998, "grad_norm": 3.622562885284424, "learning_rate": 4.7356923632424624e-05, "loss": 4.2336, "step": 5857 }, { "epoch": 0.05287958115183246, "grad_norm": 3.336071491241455, "learning_rate": 4.7356472287416504e-05, "loss": 4.7158, "step": 5858 }, { "epoch": 0.05288860805199495, "grad_norm": 4.76399040222168, "learning_rate": 4.735602094240838e-05, "loss": 5.4856, "step": 5859 }, { "epoch": 0.05289763495215743, "grad_norm": 3.1061244010925293, "learning_rate": 4.7355569597400255e-05, "loss": 4.5257, "step": 5860 }, { "epoch": 0.05290666185231992, "grad_norm": 3.668139696121216, "learning_rate": 4.7355118252392134e-05, "loss": 5.007, "step": 5861 }, { "epoch": 0.0529156887524824, "grad_norm": 4.452606201171875, "learning_rate": 4.735466690738401e-05, "loss": 4.7271, "step": 5862 }, { "epoch": 0.05292471565264488, "grad_norm": 2.656446695327759, "learning_rate": 4.735421556237588e-05, "loss": 4.6193, "step": 5863 }, { "epoch": 0.05293374255280737, "grad_norm": 3.8065531253814697, "learning_rate": 4.735376421736776e-05, "loss": 4.9082, "step": 5864 }, { "epoch": 0.05294276945296985, "grad_norm": 3.783872127532959, "learning_rate": 4.735331287235963e-05, "loss": 4.5683, "step": 5865 }, { "epoch": 0.05295179635313233, "grad_norm": 4.765957832336426, "learning_rate": 4.735286152735151e-05, "loss": 4.9923, "step": 5866 }, { "epoch": 0.05296082325329482, "grad_norm": 3.0385801792144775, "learning_rate": 4.735241018234339e-05, "loss": 4.7399, "step": 5867 }, { "epoch": 0.0529698501534573, "grad_norm": 3.52425479888916, "learning_rate": 4.735195883733526e-05, "loss": 4.1236, "step": 5868 }, { "epoch": 0.05297887705361979, "grad_norm": 3.661271810531616, "learning_rate": 4.7351507492327135e-05, "loss": 4.7334, "step": 5869 }, { "epoch": 0.05298790395378227, "grad_norm": 2.7444865703582764, "learning_rate": 4.7351056147319014e-05, "loss": 4.8487, "step": 5870 }, { "epoch": 0.05299693085394475, "grad_norm": 2.965125799179077, "learning_rate": 4.7350604802310886e-05, "loss": 5.0992, "step": 5871 }, { "epoch": 0.05300595775410724, "grad_norm": 2.6764004230499268, "learning_rate": 4.7350153457302765e-05, "loss": 4.6727, "step": 5872 }, { "epoch": 0.05301498465426972, "grad_norm": 4.349026679992676, "learning_rate": 4.7349702112294645e-05, "loss": 4.749, "step": 5873 }, { "epoch": 0.05302401155443221, "grad_norm": 3.515230178833008, "learning_rate": 4.734925076728652e-05, "loss": 5.2273, "step": 5874 }, { "epoch": 0.05303303845459469, "grad_norm": 3.6980979442596436, "learning_rate": 4.734879942227839e-05, "loss": 5.2708, "step": 5875 }, { "epoch": 0.053042065354757174, "grad_norm": 2.8928005695343018, "learning_rate": 4.734834807727027e-05, "loss": 5.5494, "step": 5876 }, { "epoch": 0.05305109225491966, "grad_norm": 3.1782777309417725, "learning_rate": 4.734789673226214e-05, "loss": 4.1643, "step": 5877 }, { "epoch": 0.053060119155082144, "grad_norm": 4.668824195861816, "learning_rate": 4.734744538725402e-05, "loss": 4.8723, "step": 5878 }, { "epoch": 0.053069146055244626, "grad_norm": 3.6712679862976074, "learning_rate": 4.734699404224589e-05, "loss": 4.3875, "step": 5879 }, { "epoch": 0.053078172955407114, "grad_norm": 2.5084896087646484, "learning_rate": 4.734654269723777e-05, "loss": 4.7432, "step": 5880 }, { "epoch": 0.053087199855569596, "grad_norm": 2.843266487121582, "learning_rate": 4.734609135222965e-05, "loss": 3.9902, "step": 5881 }, { "epoch": 0.053096226755732084, "grad_norm": 2.291668653488159, "learning_rate": 4.734564000722152e-05, "loss": 5.0845, "step": 5882 }, { "epoch": 0.053105253655894566, "grad_norm": 4.128401279449463, "learning_rate": 4.7345188662213396e-05, "loss": 4.2797, "step": 5883 }, { "epoch": 0.05311428055605705, "grad_norm": 3.3272624015808105, "learning_rate": 4.7344737317205276e-05, "loss": 4.813, "step": 5884 }, { "epoch": 0.053123307456219536, "grad_norm": 4.280406475067139, "learning_rate": 4.734428597219715e-05, "loss": 4.3522, "step": 5885 }, { "epoch": 0.05313233435638202, "grad_norm": 3.5278573036193848, "learning_rate": 4.734383462718903e-05, "loss": 4.8732, "step": 5886 }, { "epoch": 0.053141361256544506, "grad_norm": 3.3306920528411865, "learning_rate": 4.73433832821809e-05, "loss": 3.9765, "step": 5887 }, { "epoch": 0.05315038815670699, "grad_norm": 3.3882269859313965, "learning_rate": 4.734293193717278e-05, "loss": 4.1645, "step": 5888 }, { "epoch": 0.05315941505686947, "grad_norm": 7.928360462188721, "learning_rate": 4.734248059216465e-05, "loss": 5.4522, "step": 5889 }, { "epoch": 0.05316844195703196, "grad_norm": 3.5473690032958984, "learning_rate": 4.734202924715653e-05, "loss": 4.0566, "step": 5890 }, { "epoch": 0.05317746885719444, "grad_norm": 3.001668930053711, "learning_rate": 4.73415779021484e-05, "loss": 4.9933, "step": 5891 }, { "epoch": 0.05318649575735692, "grad_norm": 4.775453567504883, "learning_rate": 4.734112655714028e-05, "loss": 4.483, "step": 5892 }, { "epoch": 0.05319552265751941, "grad_norm": 3.2269723415374756, "learning_rate": 4.7340675212132155e-05, "loss": 4.1356, "step": 5893 }, { "epoch": 0.05320454955768189, "grad_norm": 3.3302226066589355, "learning_rate": 4.7340223867124034e-05, "loss": 4.174, "step": 5894 }, { "epoch": 0.05321357645784438, "grad_norm": 4.505125522613525, "learning_rate": 4.733977252211591e-05, "loss": 5.163, "step": 5895 }, { "epoch": 0.05322260335800686, "grad_norm": 4.092109203338623, "learning_rate": 4.733932117710778e-05, "loss": 4.5343, "step": 5896 }, { "epoch": 0.05323163025816934, "grad_norm": 3.21462345123291, "learning_rate": 4.733886983209966e-05, "loss": 4.8705, "step": 5897 }, { "epoch": 0.05324065715833183, "grad_norm": 3.4045772552490234, "learning_rate": 4.733841848709154e-05, "loss": 4.3094, "step": 5898 }, { "epoch": 0.05324968405849431, "grad_norm": 3.853644609451294, "learning_rate": 4.733796714208341e-05, "loss": 4.764, "step": 5899 }, { "epoch": 0.0532587109586568, "grad_norm": 3.613879442214966, "learning_rate": 4.733751579707529e-05, "loss": 4.8125, "step": 5900 }, { "epoch": 0.05326773785881928, "grad_norm": 4.589334011077881, "learning_rate": 4.733706445206716e-05, "loss": 3.2154, "step": 5901 }, { "epoch": 0.053276764758981764, "grad_norm": 5.281139373779297, "learning_rate": 4.7336613107059034e-05, "loss": 4.8348, "step": 5902 }, { "epoch": 0.05328579165914425, "grad_norm": 4.314511299133301, "learning_rate": 4.7336161762050913e-05, "loss": 5.086, "step": 5903 }, { "epoch": 0.053294818559306734, "grad_norm": 4.818948745727539, "learning_rate": 4.7335710417042786e-05, "loss": 5.1228, "step": 5904 }, { "epoch": 0.053303845459469215, "grad_norm": 2.911592483520508, "learning_rate": 4.7335259072034665e-05, "loss": 4.3519, "step": 5905 }, { "epoch": 0.053312872359631704, "grad_norm": 3.148514747619629, "learning_rate": 4.7334807727026544e-05, "loss": 4.49, "step": 5906 }, { "epoch": 0.053321899259794185, "grad_norm": 3.149502992630005, "learning_rate": 4.733435638201842e-05, "loss": 5.0591, "step": 5907 }, { "epoch": 0.053330926159956674, "grad_norm": 3.7653486728668213, "learning_rate": 4.7333905037010296e-05, "loss": 4.8906, "step": 5908 }, { "epoch": 0.053339953060119155, "grad_norm": 2.9780077934265137, "learning_rate": 4.733345369200217e-05, "loss": 4.9809, "step": 5909 }, { "epoch": 0.05334897996028164, "grad_norm": 3.2674808502197266, "learning_rate": 4.733300234699404e-05, "loss": 4.6527, "step": 5910 }, { "epoch": 0.053358006860444125, "grad_norm": 3.804725408554077, "learning_rate": 4.733255100198592e-05, "loss": 3.6273, "step": 5911 }, { "epoch": 0.05336703376060661, "grad_norm": 6.854763031005859, "learning_rate": 4.73320996569778e-05, "loss": 5.1121, "step": 5912 }, { "epoch": 0.053376060660769095, "grad_norm": 4.272799491882324, "learning_rate": 4.733164831196967e-05, "loss": 3.5075, "step": 5913 }, { "epoch": 0.05338508756093158, "grad_norm": 2.7007036209106445, "learning_rate": 4.733119696696155e-05, "loss": 5.189, "step": 5914 }, { "epoch": 0.05339411446109406, "grad_norm": 4.096935749053955, "learning_rate": 4.7330745621953424e-05, "loss": 3.8498, "step": 5915 }, { "epoch": 0.05340314136125655, "grad_norm": 3.7724456787109375, "learning_rate": 4.7330294276945296e-05, "loss": 5.1557, "step": 5916 }, { "epoch": 0.05341216826141903, "grad_norm": 3.601390838623047, "learning_rate": 4.7329842931937175e-05, "loss": 4.9629, "step": 5917 }, { "epoch": 0.05342119516158151, "grad_norm": 4.662482738494873, "learning_rate": 4.732939158692905e-05, "loss": 4.4826, "step": 5918 }, { "epoch": 0.053430222061744, "grad_norm": 2.808262825012207, "learning_rate": 4.732894024192093e-05, "loss": 4.5885, "step": 5919 }, { "epoch": 0.05343924896190648, "grad_norm": 3.8350841999053955, "learning_rate": 4.7328488896912806e-05, "loss": 4.7101, "step": 5920 }, { "epoch": 0.05344827586206897, "grad_norm": 2.991462230682373, "learning_rate": 4.732803755190468e-05, "loss": 4.8536, "step": 5921 }, { "epoch": 0.05345730276223145, "grad_norm": 4.02279806137085, "learning_rate": 4.732758620689655e-05, "loss": 4.6423, "step": 5922 }, { "epoch": 0.05346632966239393, "grad_norm": 4.948504447937012, "learning_rate": 4.732713486188843e-05, "loss": 4.837, "step": 5923 }, { "epoch": 0.05347535656255642, "grad_norm": 3.218116521835327, "learning_rate": 4.73266835168803e-05, "loss": 4.5338, "step": 5924 }, { "epoch": 0.0534843834627189, "grad_norm": 3.9334871768951416, "learning_rate": 4.732623217187218e-05, "loss": 4.4861, "step": 5925 }, { "epoch": 0.05349341036288139, "grad_norm": 4.29694128036499, "learning_rate": 4.7325780826864055e-05, "loss": 3.5969, "step": 5926 }, { "epoch": 0.05350243726304387, "grad_norm": 3.4099721908569336, "learning_rate": 4.7325329481855934e-05, "loss": 4.7262, "step": 5927 }, { "epoch": 0.05351146416320635, "grad_norm": 3.0541985034942627, "learning_rate": 4.732487813684781e-05, "loss": 4.0523, "step": 5928 }, { "epoch": 0.05352049106336884, "grad_norm": 3.027698040008545, "learning_rate": 4.732442679183968e-05, "loss": 4.7142, "step": 5929 }, { "epoch": 0.05352951796353132, "grad_norm": 3.2731211185455322, "learning_rate": 4.732397544683156e-05, "loss": 4.3072, "step": 5930 }, { "epoch": 0.053538544863693804, "grad_norm": 3.939873456954956, "learning_rate": 4.732352410182344e-05, "loss": 4.9798, "step": 5931 }, { "epoch": 0.05354757176385629, "grad_norm": 5.2634172439575195, "learning_rate": 4.732307275681531e-05, "loss": 3.6799, "step": 5932 }, { "epoch": 0.053556598664018774, "grad_norm": 2.806960344314575, "learning_rate": 4.732262141180719e-05, "loss": 5.4202, "step": 5933 }, { "epoch": 0.05356562556418126, "grad_norm": 7.462500095367432, "learning_rate": 4.732217006679907e-05, "loss": 5.6423, "step": 5934 }, { "epoch": 0.053574652464343744, "grad_norm": 3.2199461460113525, "learning_rate": 4.732171872179094e-05, "loss": 4.3981, "step": 5935 }, { "epoch": 0.053583679364506226, "grad_norm": 3.1774373054504395, "learning_rate": 4.732126737678281e-05, "loss": 5.0133, "step": 5936 }, { "epoch": 0.053592706264668714, "grad_norm": 3.3676300048828125, "learning_rate": 4.732081603177469e-05, "loss": 4.6431, "step": 5937 }, { "epoch": 0.053601733164831196, "grad_norm": 3.2770676612854004, "learning_rate": 4.7320364686766565e-05, "loss": 4.4205, "step": 5938 }, { "epoch": 0.053610760064993684, "grad_norm": 3.3848350048065186, "learning_rate": 4.7319913341758444e-05, "loss": 4.9413, "step": 5939 }, { "epoch": 0.053619786965156166, "grad_norm": 4.035752296447754, "learning_rate": 4.731946199675032e-05, "loss": 4.6611, "step": 5940 }, { "epoch": 0.05362881386531865, "grad_norm": 3.180201530456543, "learning_rate": 4.7319010651742196e-05, "loss": 4.4155, "step": 5941 }, { "epoch": 0.053637840765481136, "grad_norm": 3.738398551940918, "learning_rate": 4.731855930673407e-05, "loss": 4.3271, "step": 5942 }, { "epoch": 0.05364686766564362, "grad_norm": 4.108572483062744, "learning_rate": 4.731810796172594e-05, "loss": 5.0399, "step": 5943 }, { "epoch": 0.0536558945658061, "grad_norm": 3.983797550201416, "learning_rate": 4.731765661671782e-05, "loss": 5.3307, "step": 5944 }, { "epoch": 0.05366492146596859, "grad_norm": 2.6571428775787354, "learning_rate": 4.73172052717097e-05, "loss": 5.4144, "step": 5945 }, { "epoch": 0.05367394836613107, "grad_norm": 7.028806209564209, "learning_rate": 4.731675392670157e-05, "loss": 4.812, "step": 5946 }, { "epoch": 0.05368297526629356, "grad_norm": 3.3182225227355957, "learning_rate": 4.731630258169345e-05, "loss": 4.4016, "step": 5947 }, { "epoch": 0.05369200216645604, "grad_norm": 3.168154239654541, "learning_rate": 4.7315851236685323e-05, "loss": 4.4745, "step": 5948 }, { "epoch": 0.05370102906661852, "grad_norm": 3.007617235183716, "learning_rate": 4.7315399891677196e-05, "loss": 5.0329, "step": 5949 }, { "epoch": 0.05371005596678101, "grad_norm": 2.911010503768921, "learning_rate": 4.7314948546669075e-05, "loss": 4.585, "step": 5950 }, { "epoch": 0.05371908286694349, "grad_norm": 4.153873920440674, "learning_rate": 4.7314497201660954e-05, "loss": 3.8263, "step": 5951 }, { "epoch": 0.05372810976710598, "grad_norm": 3.517833709716797, "learning_rate": 4.731404585665283e-05, "loss": 4.2428, "step": 5952 }, { "epoch": 0.05373713666726846, "grad_norm": 2.8690242767333984, "learning_rate": 4.7313594511644706e-05, "loss": 4.8934, "step": 5953 }, { "epoch": 0.05374616356743094, "grad_norm": 2.7216556072235107, "learning_rate": 4.731314316663658e-05, "loss": 4.6977, "step": 5954 }, { "epoch": 0.05375519046759343, "grad_norm": 2.4763123989105225, "learning_rate": 4.731269182162846e-05, "loss": 4.5266, "step": 5955 }, { "epoch": 0.05376421736775591, "grad_norm": 3.654582977294922, "learning_rate": 4.731224047662033e-05, "loss": 4.9136, "step": 5956 }, { "epoch": 0.053773244267918394, "grad_norm": 3.418917655944824, "learning_rate": 4.73117891316122e-05, "loss": 5.0264, "step": 5957 }, { "epoch": 0.05378227116808088, "grad_norm": 2.9174158573150635, "learning_rate": 4.731133778660408e-05, "loss": 4.7094, "step": 5958 }, { "epoch": 0.053791298068243364, "grad_norm": 3.3303046226501465, "learning_rate": 4.731088644159596e-05, "loss": 4.1379, "step": 5959 }, { "epoch": 0.05380032496840585, "grad_norm": 3.485036849975586, "learning_rate": 4.7310435096587834e-05, "loss": 4.5885, "step": 5960 }, { "epoch": 0.053809351868568334, "grad_norm": 3.8367178440093994, "learning_rate": 4.730998375157971e-05, "loss": 5.2479, "step": 5961 }, { "epoch": 0.053818378768730815, "grad_norm": 3.9543964862823486, "learning_rate": 4.7309532406571585e-05, "loss": 5.2617, "step": 5962 }, { "epoch": 0.053827405668893304, "grad_norm": 6.885345458984375, "learning_rate": 4.730908106156346e-05, "loss": 4.2771, "step": 5963 }, { "epoch": 0.053836432569055785, "grad_norm": 4.170760154724121, "learning_rate": 4.730862971655534e-05, "loss": 4.4978, "step": 5964 }, { "epoch": 0.053845459469218274, "grad_norm": 2.872943162918091, "learning_rate": 4.730817837154721e-05, "loss": 5.1231, "step": 5965 }, { "epoch": 0.053854486369380755, "grad_norm": 3.7878191471099854, "learning_rate": 4.730772702653909e-05, "loss": 4.909, "step": 5966 }, { "epoch": 0.05386351326954324, "grad_norm": 4.055604457855225, "learning_rate": 4.730727568153097e-05, "loss": 4.0408, "step": 5967 }, { "epoch": 0.053872540169705725, "grad_norm": 3.812819242477417, "learning_rate": 4.730682433652284e-05, "loss": 4.7313, "step": 5968 }, { "epoch": 0.05388156706986821, "grad_norm": 3.245337963104248, "learning_rate": 4.730637299151471e-05, "loss": 5.2146, "step": 5969 }, { "epoch": 0.05389059397003069, "grad_norm": 3.9604153633117676, "learning_rate": 4.730592164650659e-05, "loss": 3.9141, "step": 5970 }, { "epoch": 0.05389962087019318, "grad_norm": 3.4840309619903564, "learning_rate": 4.7305470301498465e-05, "loss": 4.1781, "step": 5971 }, { "epoch": 0.05390864777035566, "grad_norm": 3.1285173892974854, "learning_rate": 4.7305018956490344e-05, "loss": 5.0612, "step": 5972 }, { "epoch": 0.05391767467051815, "grad_norm": 5.531972408294678, "learning_rate": 4.730456761148222e-05, "loss": 4.6666, "step": 5973 }, { "epoch": 0.05392670157068063, "grad_norm": 2.9028427600860596, "learning_rate": 4.7304116266474096e-05, "loss": 4.7682, "step": 5974 }, { "epoch": 0.05393572847084311, "grad_norm": 3.4814908504486084, "learning_rate": 4.7303664921465975e-05, "loss": 4.8238, "step": 5975 }, { "epoch": 0.0539447553710056, "grad_norm": 5.196049213409424, "learning_rate": 4.730321357645785e-05, "loss": 5.0069, "step": 5976 }, { "epoch": 0.05395378227116808, "grad_norm": 4.6310224533081055, "learning_rate": 4.730276223144972e-05, "loss": 5.2346, "step": 5977 }, { "epoch": 0.05396280917133057, "grad_norm": 4.276399612426758, "learning_rate": 4.73023108864416e-05, "loss": 4.907, "step": 5978 }, { "epoch": 0.05397183607149305, "grad_norm": 4.313297748565674, "learning_rate": 4.730185954143347e-05, "loss": 4.2681, "step": 5979 }, { "epoch": 0.05398086297165553, "grad_norm": 2.9634599685668945, "learning_rate": 4.730140819642535e-05, "loss": 4.1844, "step": 5980 }, { "epoch": 0.05398988987181802, "grad_norm": 2.4920156002044678, "learning_rate": 4.730095685141723e-05, "loss": 4.3403, "step": 5981 }, { "epoch": 0.0539989167719805, "grad_norm": 2.984196901321411, "learning_rate": 4.73005055064091e-05, "loss": 4.2112, "step": 5982 }, { "epoch": 0.05400794367214298, "grad_norm": 6.299097537994385, "learning_rate": 4.7300054161400975e-05, "loss": 4.6557, "step": 5983 }, { "epoch": 0.05401697057230547, "grad_norm": 3.814028263092041, "learning_rate": 4.7299602816392854e-05, "loss": 3.8649, "step": 5984 }, { "epoch": 0.05402599747246795, "grad_norm": 4.944756507873535, "learning_rate": 4.729915147138473e-05, "loss": 4.6513, "step": 5985 }, { "epoch": 0.05403502437263044, "grad_norm": 3.3232080936431885, "learning_rate": 4.7298700126376606e-05, "loss": 4.7259, "step": 5986 }, { "epoch": 0.05404405127279292, "grad_norm": 3.5127031803131104, "learning_rate": 4.729824878136848e-05, "loss": 4.5567, "step": 5987 }, { "epoch": 0.054053078172955404, "grad_norm": 4.4722161293029785, "learning_rate": 4.729779743636036e-05, "loss": 3.853, "step": 5988 }, { "epoch": 0.05406210507311789, "grad_norm": 3.743612051010132, "learning_rate": 4.729734609135223e-05, "loss": 4.9231, "step": 5989 }, { "epoch": 0.054071131973280374, "grad_norm": 3.609246253967285, "learning_rate": 4.72968947463441e-05, "loss": 3.9778, "step": 5990 }, { "epoch": 0.05408015887344286, "grad_norm": 3.4927761554718018, "learning_rate": 4.729644340133598e-05, "loss": 4.3596, "step": 5991 }, { "epoch": 0.054089185773605344, "grad_norm": 3.6923625469207764, "learning_rate": 4.729599205632786e-05, "loss": 4.4438, "step": 5992 }, { "epoch": 0.054098212673767826, "grad_norm": 4.612232685089111, "learning_rate": 4.7295540711319733e-05, "loss": 5.3897, "step": 5993 }, { "epoch": 0.054107239573930314, "grad_norm": 5.422158718109131, "learning_rate": 4.729508936631161e-05, "loss": 4.1451, "step": 5994 }, { "epoch": 0.054116266474092796, "grad_norm": 2.7122952938079834, "learning_rate": 4.729463802130349e-05, "loss": 5.2479, "step": 5995 }, { "epoch": 0.05412529337425528, "grad_norm": 4.295980930328369, "learning_rate": 4.729418667629536e-05, "loss": 4.5473, "step": 5996 }, { "epoch": 0.054134320274417766, "grad_norm": 2.9396064281463623, "learning_rate": 4.729373533128724e-05, "loss": 5.0244, "step": 5997 }, { "epoch": 0.05414334717458025, "grad_norm": 5.964011192321777, "learning_rate": 4.7293283986279116e-05, "loss": 5.0396, "step": 5998 }, { "epoch": 0.054152374074742736, "grad_norm": 3.9639410972595215, "learning_rate": 4.729283264127099e-05, "loss": 4.3633, "step": 5999 }, { "epoch": 0.05416140097490522, "grad_norm": 2.9059970378875732, "learning_rate": 4.729238129626287e-05, "loss": 4.871, "step": 6000 }, { "epoch": 0.0541704278750677, "grad_norm": 3.6348729133605957, "learning_rate": 4.729192995125474e-05, "loss": 5.4127, "step": 6001 }, { "epoch": 0.05417945477523019, "grad_norm": 4.8474297523498535, "learning_rate": 4.729147860624662e-05, "loss": 3.0903, "step": 6002 }, { "epoch": 0.05418848167539267, "grad_norm": 3.710805892944336, "learning_rate": 4.729102726123849e-05, "loss": 4.3942, "step": 6003 }, { "epoch": 0.05419750857555516, "grad_norm": 4.014278411865234, "learning_rate": 4.7290575916230364e-05, "loss": 4.5478, "step": 6004 }, { "epoch": 0.05420653547571764, "grad_norm": 3.2344515323638916, "learning_rate": 4.7290124571222244e-05, "loss": 5.0321, "step": 6005 }, { "epoch": 0.05421556237588012, "grad_norm": 3.638416290283203, "learning_rate": 4.728967322621412e-05, "loss": 5.0216, "step": 6006 }, { "epoch": 0.05422458927604261, "grad_norm": 3.980314016342163, "learning_rate": 4.7289221881205995e-05, "loss": 4.8347, "step": 6007 }, { "epoch": 0.05423361617620509, "grad_norm": 3.4927241802215576, "learning_rate": 4.7288770536197875e-05, "loss": 4.5232, "step": 6008 }, { "epoch": 0.05424264307636757, "grad_norm": 3.1619582176208496, "learning_rate": 4.728831919118975e-05, "loss": 4.5025, "step": 6009 }, { "epoch": 0.05425166997653006, "grad_norm": 3.0956220626831055, "learning_rate": 4.728786784618162e-05, "loss": 4.6297, "step": 6010 }, { "epoch": 0.05426069687669254, "grad_norm": 2.662940502166748, "learning_rate": 4.72874165011735e-05, "loss": 4.8624, "step": 6011 }, { "epoch": 0.05426972377685503, "grad_norm": 3.891547679901123, "learning_rate": 4.728696515616538e-05, "loss": 4.8502, "step": 6012 }, { "epoch": 0.05427875067701751, "grad_norm": 3.7971770763397217, "learning_rate": 4.728651381115725e-05, "loss": 5.1986, "step": 6013 }, { "epoch": 0.054287777577179994, "grad_norm": 3.8930301666259766, "learning_rate": 4.728606246614913e-05, "loss": 4.3373, "step": 6014 }, { "epoch": 0.05429680447734248, "grad_norm": 3.427396535873413, "learning_rate": 4.7285611121141e-05, "loss": 3.8911, "step": 6015 }, { "epoch": 0.054305831377504964, "grad_norm": 3.584163188934326, "learning_rate": 4.7285159776132875e-05, "loss": 4.9712, "step": 6016 }, { "epoch": 0.05431485827766745, "grad_norm": 3.9850008487701416, "learning_rate": 4.7284708431124754e-05, "loss": 4.8182, "step": 6017 }, { "epoch": 0.054323885177829934, "grad_norm": 3.7500827312469482, "learning_rate": 4.7284257086116626e-05, "loss": 4.2614, "step": 6018 }, { "epoch": 0.054332912077992415, "grad_norm": 2.73582124710083, "learning_rate": 4.7283805741108506e-05, "loss": 4.6225, "step": 6019 }, { "epoch": 0.054341938978154904, "grad_norm": 3.210749387741089, "learning_rate": 4.7283354396100385e-05, "loss": 4.7705, "step": 6020 }, { "epoch": 0.054350965878317385, "grad_norm": 3.5645484924316406, "learning_rate": 4.728290305109226e-05, "loss": 5.0285, "step": 6021 }, { "epoch": 0.05435999277847987, "grad_norm": 3.9532089233398438, "learning_rate": 4.7282451706084137e-05, "loss": 4.9889, "step": 6022 }, { "epoch": 0.054369019678642355, "grad_norm": 3.7946853637695312, "learning_rate": 4.728200036107601e-05, "loss": 4.7024, "step": 6023 }, { "epoch": 0.05437804657880484, "grad_norm": 3.134411573410034, "learning_rate": 4.728154901606788e-05, "loss": 4.6449, "step": 6024 }, { "epoch": 0.054387073478967325, "grad_norm": 2.8790924549102783, "learning_rate": 4.728109767105976e-05, "loss": 4.9377, "step": 6025 }, { "epoch": 0.05439610037912981, "grad_norm": 3.7872800827026367, "learning_rate": 4.728064632605163e-05, "loss": 4.1788, "step": 6026 }, { "epoch": 0.05440512727929229, "grad_norm": 3.369556427001953, "learning_rate": 4.728019498104351e-05, "loss": 4.8413, "step": 6027 }, { "epoch": 0.05441415417945478, "grad_norm": 5.4930243492126465, "learning_rate": 4.727974363603539e-05, "loss": 5.1139, "step": 6028 }, { "epoch": 0.05442318107961726, "grad_norm": 3.6170525550842285, "learning_rate": 4.7279292291027264e-05, "loss": 4.6298, "step": 6029 }, { "epoch": 0.05443220797977975, "grad_norm": 3.6538617610931396, "learning_rate": 4.727884094601914e-05, "loss": 4.9956, "step": 6030 }, { "epoch": 0.05444123487994223, "grad_norm": 4.331655979156494, "learning_rate": 4.7278389601011016e-05, "loss": 4.7247, "step": 6031 }, { "epoch": 0.05445026178010471, "grad_norm": 3.062182664871216, "learning_rate": 4.727793825600289e-05, "loss": 5.0297, "step": 6032 }, { "epoch": 0.0544592886802672, "grad_norm": 2.8457632064819336, "learning_rate": 4.727748691099477e-05, "loss": 4.9248, "step": 6033 }, { "epoch": 0.05446831558042968, "grad_norm": 3.9812586307525635, "learning_rate": 4.727703556598665e-05, "loss": 3.8725, "step": 6034 }, { "epoch": 0.05447734248059216, "grad_norm": 3.7143633365631104, "learning_rate": 4.727658422097852e-05, "loss": 4.7227, "step": 6035 }, { "epoch": 0.05448636938075465, "grad_norm": 3.1558241844177246, "learning_rate": 4.727613287597039e-05, "loss": 4.5556, "step": 6036 }, { "epoch": 0.05449539628091713, "grad_norm": 3.6974265575408936, "learning_rate": 4.727568153096227e-05, "loss": 5.0893, "step": 6037 }, { "epoch": 0.05450442318107962, "grad_norm": 4.207555770874023, "learning_rate": 4.7275230185954143e-05, "loss": 4.4642, "step": 6038 }, { "epoch": 0.0545134500812421, "grad_norm": 3.572953224182129, "learning_rate": 4.727477884094602e-05, "loss": 4.7703, "step": 6039 }, { "epoch": 0.05452247698140458, "grad_norm": 4.169456958770752, "learning_rate": 4.7274327495937895e-05, "loss": 5.1478, "step": 6040 }, { "epoch": 0.05453150388156707, "grad_norm": 3.055760622024536, "learning_rate": 4.7273876150929774e-05, "loss": 4.5957, "step": 6041 }, { "epoch": 0.05454053078172955, "grad_norm": 2.745572090148926, "learning_rate": 4.7273424805921654e-05, "loss": 4.7947, "step": 6042 }, { "epoch": 0.05454955768189204, "grad_norm": 3.3633203506469727, "learning_rate": 4.727297346091352e-05, "loss": 4.8282, "step": 6043 }, { "epoch": 0.05455858458205452, "grad_norm": 3.2359721660614014, "learning_rate": 4.72725221159054e-05, "loss": 4.3242, "step": 6044 }, { "epoch": 0.054567611482217004, "grad_norm": 5.2045183181762695, "learning_rate": 4.727207077089728e-05, "loss": 4.1472, "step": 6045 }, { "epoch": 0.05457663838237949, "grad_norm": 3.365570545196533, "learning_rate": 4.727161942588915e-05, "loss": 4.8737, "step": 6046 }, { "epoch": 0.054585665282541974, "grad_norm": 3.3393681049346924, "learning_rate": 4.727116808088103e-05, "loss": 3.9247, "step": 6047 }, { "epoch": 0.054594692182704456, "grad_norm": 2.4362809658050537, "learning_rate": 4.72707167358729e-05, "loss": 5.065, "step": 6048 }, { "epoch": 0.054603719082866944, "grad_norm": 5.868264198303223, "learning_rate": 4.727026539086478e-05, "loss": 5.1037, "step": 6049 }, { "epoch": 0.054612745983029426, "grad_norm": 3.212029218673706, "learning_rate": 4.7269814045856654e-05, "loss": 4.3467, "step": 6050 }, { "epoch": 0.054621772883191914, "grad_norm": 4.430450439453125, "learning_rate": 4.726936270084853e-05, "loss": 3.8661, "step": 6051 }, { "epoch": 0.054630799783354396, "grad_norm": 4.24215841293335, "learning_rate": 4.7268911355840405e-05, "loss": 4.6786, "step": 6052 }, { "epoch": 0.05463982668351688, "grad_norm": 5.224486827850342, "learning_rate": 4.7268460010832285e-05, "loss": 5.35, "step": 6053 }, { "epoch": 0.054648853583679366, "grad_norm": 2.7449121475219727, "learning_rate": 4.726800866582416e-05, "loss": 4.9616, "step": 6054 }, { "epoch": 0.05465788048384185, "grad_norm": 5.737008571624756, "learning_rate": 4.7267557320816036e-05, "loss": 4.7688, "step": 6055 }, { "epoch": 0.054666907384004336, "grad_norm": 2.8301455974578857, "learning_rate": 4.726710597580791e-05, "loss": 4.555, "step": 6056 }, { "epoch": 0.05467593428416682, "grad_norm": 3.547492504119873, "learning_rate": 4.726665463079978e-05, "loss": 4.8957, "step": 6057 }, { "epoch": 0.0546849611843293, "grad_norm": 3.6772637367248535, "learning_rate": 4.726620328579166e-05, "loss": 4.7721, "step": 6058 }, { "epoch": 0.05469398808449179, "grad_norm": 3.8631255626678467, "learning_rate": 4.726575194078354e-05, "loss": 4.6524, "step": 6059 }, { "epoch": 0.05470301498465427, "grad_norm": 4.29941987991333, "learning_rate": 4.726530059577541e-05, "loss": 5.5604, "step": 6060 }, { "epoch": 0.05471204188481675, "grad_norm": 2.591251850128174, "learning_rate": 4.726484925076729e-05, "loss": 4.7973, "step": 6061 }, { "epoch": 0.05472106878497924, "grad_norm": 4.093677043914795, "learning_rate": 4.7264397905759164e-05, "loss": 4.7104, "step": 6062 }, { "epoch": 0.05473009568514172, "grad_norm": 3.06119704246521, "learning_rate": 4.7263946560751036e-05, "loss": 4.724, "step": 6063 }, { "epoch": 0.05473912258530421, "grad_norm": 2.7407608032226562, "learning_rate": 4.7263495215742916e-05, "loss": 4.9409, "step": 6064 }, { "epoch": 0.05474814948546669, "grad_norm": 3.490800380706787, "learning_rate": 4.726304387073479e-05, "loss": 4.3717, "step": 6065 }, { "epoch": 0.05475717638562917, "grad_norm": 3.8028438091278076, "learning_rate": 4.726259252572667e-05, "loss": 4.6941, "step": 6066 }, { "epoch": 0.05476620328579166, "grad_norm": 3.956493377685547, "learning_rate": 4.7262141180718547e-05, "loss": 4.3941, "step": 6067 }, { "epoch": 0.05477523018595414, "grad_norm": 3.110586166381836, "learning_rate": 4.726168983571042e-05, "loss": 4.2945, "step": 6068 }, { "epoch": 0.05478425708611663, "grad_norm": 4.149788856506348, "learning_rate": 4.72612384907023e-05, "loss": 5.2365, "step": 6069 }, { "epoch": 0.05479328398627911, "grad_norm": 4.044622421264648, "learning_rate": 4.726078714569417e-05, "loss": 4.6156, "step": 6070 }, { "epoch": 0.054802310886441594, "grad_norm": 3.4613943099975586, "learning_rate": 4.726033580068604e-05, "loss": 5.2095, "step": 6071 }, { "epoch": 0.05481133778660408, "grad_norm": 3.8782386779785156, "learning_rate": 4.725988445567792e-05, "loss": 4.6571, "step": 6072 }, { "epoch": 0.054820364686766564, "grad_norm": 3.137953281402588, "learning_rate": 4.72594331106698e-05, "loss": 4.5251, "step": 6073 }, { "epoch": 0.054829391586929045, "grad_norm": 3.212353229522705, "learning_rate": 4.7258981765661674e-05, "loss": 5.0541, "step": 6074 }, { "epoch": 0.054838418487091534, "grad_norm": 2.800755262374878, "learning_rate": 4.7258530420653553e-05, "loss": 3.5908, "step": 6075 }, { "epoch": 0.054847445387254015, "grad_norm": 4.003439903259277, "learning_rate": 4.7258079075645426e-05, "loss": 4.664, "step": 6076 }, { "epoch": 0.054856472287416504, "grad_norm": 2.396484375, "learning_rate": 4.72576277306373e-05, "loss": 5.1973, "step": 6077 }, { "epoch": 0.054865499187578985, "grad_norm": 3.1651089191436768, "learning_rate": 4.725717638562918e-05, "loss": 4.7682, "step": 6078 }, { "epoch": 0.05487452608774147, "grad_norm": 3.7247464656829834, "learning_rate": 4.725672504062105e-05, "loss": 4.8726, "step": 6079 }, { "epoch": 0.054883552987903955, "grad_norm": 3.3959901332855225, "learning_rate": 4.725627369561293e-05, "loss": 4.8771, "step": 6080 }, { "epoch": 0.05489257988806644, "grad_norm": 3.590242862701416, "learning_rate": 4.725582235060481e-05, "loss": 4.6639, "step": 6081 }, { "epoch": 0.054901606788228925, "grad_norm": 3.9309141635894775, "learning_rate": 4.725537100559668e-05, "loss": 5.162, "step": 6082 }, { "epoch": 0.05491063368839141, "grad_norm": 2.935804605484009, "learning_rate": 4.7254919660588553e-05, "loss": 4.9062, "step": 6083 }, { "epoch": 0.05491966058855389, "grad_norm": 3.604069471359253, "learning_rate": 4.725446831558043e-05, "loss": 3.9636, "step": 6084 }, { "epoch": 0.05492868748871638, "grad_norm": 3.78283953666687, "learning_rate": 4.7254016970572305e-05, "loss": 4.3438, "step": 6085 }, { "epoch": 0.05493771438887886, "grad_norm": 3.27306866645813, "learning_rate": 4.7253565625564184e-05, "loss": 3.6348, "step": 6086 }, { "epoch": 0.05494674128904134, "grad_norm": 3.2675774097442627, "learning_rate": 4.725311428055606e-05, "loss": 4.8131, "step": 6087 }, { "epoch": 0.05495576818920383, "grad_norm": 3.5877139568328857, "learning_rate": 4.7252662935547936e-05, "loss": 4.9164, "step": 6088 }, { "epoch": 0.05496479508936631, "grad_norm": 3.312955379486084, "learning_rate": 4.7252211590539815e-05, "loss": 5.0661, "step": 6089 }, { "epoch": 0.0549738219895288, "grad_norm": 6.8062944412231445, "learning_rate": 4.725176024553168e-05, "loss": 4.8931, "step": 6090 }, { "epoch": 0.05498284888969128, "grad_norm": 2.8573782444000244, "learning_rate": 4.725130890052356e-05, "loss": 4.9498, "step": 6091 }, { "epoch": 0.05499187578985376, "grad_norm": 2.9611599445343018, "learning_rate": 4.725085755551544e-05, "loss": 3.974, "step": 6092 }, { "epoch": 0.05500090269001625, "grad_norm": 3.643566846847534, "learning_rate": 4.725040621050731e-05, "loss": 4.9182, "step": 6093 }, { "epoch": 0.05500992959017873, "grad_norm": 3.5815961360931396, "learning_rate": 4.724995486549919e-05, "loss": 4.795, "step": 6094 }, { "epoch": 0.05501895649034122, "grad_norm": 3.6657774448394775, "learning_rate": 4.724950352049107e-05, "loss": 4.8938, "step": 6095 }, { "epoch": 0.0550279833905037, "grad_norm": 2.913487672805786, "learning_rate": 4.724905217548294e-05, "loss": 4.4269, "step": 6096 }, { "epoch": 0.05503701029066618, "grad_norm": 3.6759581565856934, "learning_rate": 4.7248600830474815e-05, "loss": 5.0862, "step": 6097 }, { "epoch": 0.05504603719082867, "grad_norm": 2.6736702919006348, "learning_rate": 4.7248149485466695e-05, "loss": 4.8475, "step": 6098 }, { "epoch": 0.05505506409099115, "grad_norm": 4.929628372192383, "learning_rate": 4.724769814045857e-05, "loss": 4.5643, "step": 6099 }, { "epoch": 0.055064090991153634, "grad_norm": 2.9200782775878906, "learning_rate": 4.7247246795450446e-05, "loss": 4.4297, "step": 6100 }, { "epoch": 0.05507311789131612, "grad_norm": 3.313648223876953, "learning_rate": 4.724679545044232e-05, "loss": 4.5278, "step": 6101 }, { "epoch": 0.055082144791478604, "grad_norm": 3.2823264598846436, "learning_rate": 4.72463441054342e-05, "loss": 4.806, "step": 6102 }, { "epoch": 0.05509117169164109, "grad_norm": 3.446861505508423, "learning_rate": 4.724589276042607e-05, "loss": 5.0562, "step": 6103 }, { "epoch": 0.055100198591803574, "grad_norm": 4.8419294357299805, "learning_rate": 4.724544141541794e-05, "loss": 4.1639, "step": 6104 }, { "epoch": 0.055109225491966056, "grad_norm": 3.501958131790161, "learning_rate": 4.724499007040982e-05, "loss": 5.241, "step": 6105 }, { "epoch": 0.055118252392128544, "grad_norm": 3.5515310764312744, "learning_rate": 4.72445387254017e-05, "loss": 4.1595, "step": 6106 }, { "epoch": 0.055127279292291026, "grad_norm": 4.202483177185059, "learning_rate": 4.7244087380393574e-05, "loss": 4.8304, "step": 6107 }, { "epoch": 0.055136306192453514, "grad_norm": 3.284696578979492, "learning_rate": 4.724363603538545e-05, "loss": 4.8554, "step": 6108 }, { "epoch": 0.055145333092615996, "grad_norm": 3.9499142169952393, "learning_rate": 4.7243184690377326e-05, "loss": 4.6712, "step": 6109 }, { "epoch": 0.05515435999277848, "grad_norm": 4.540812015533447, "learning_rate": 4.72427333453692e-05, "loss": 4.4704, "step": 6110 }, { "epoch": 0.055163386892940966, "grad_norm": 4.373757839202881, "learning_rate": 4.724228200036108e-05, "loss": 4.6371, "step": 6111 }, { "epoch": 0.05517241379310345, "grad_norm": 4.252433776855469, "learning_rate": 4.7241830655352957e-05, "loss": 3.4662, "step": 6112 }, { "epoch": 0.05518144069326593, "grad_norm": 4.066745281219482, "learning_rate": 4.724137931034483e-05, "loss": 4.4687, "step": 6113 }, { "epoch": 0.05519046759342842, "grad_norm": 2.705814838409424, "learning_rate": 4.724092796533671e-05, "loss": 5.0163, "step": 6114 }, { "epoch": 0.0551994944935909, "grad_norm": 3.9825375080108643, "learning_rate": 4.724047662032858e-05, "loss": 4.4351, "step": 6115 }, { "epoch": 0.05520852139375339, "grad_norm": 4.090001583099365, "learning_rate": 4.724002527532046e-05, "loss": 4.5123, "step": 6116 }, { "epoch": 0.05521754829391587, "grad_norm": 3.2042407989501953, "learning_rate": 4.723957393031233e-05, "loss": 4.7331, "step": 6117 }, { "epoch": 0.05522657519407835, "grad_norm": 3.479675054550171, "learning_rate": 4.7239122585304205e-05, "loss": 4.4962, "step": 6118 }, { "epoch": 0.05523560209424084, "grad_norm": 4.271698951721191, "learning_rate": 4.7238671240296084e-05, "loss": 4.6859, "step": 6119 }, { "epoch": 0.05524462899440332, "grad_norm": 3.6579205989837646, "learning_rate": 4.7238219895287963e-05, "loss": 3.4925, "step": 6120 }, { "epoch": 0.05525365589456581, "grad_norm": 4.235774517059326, "learning_rate": 4.7237768550279836e-05, "loss": 5.0494, "step": 6121 }, { "epoch": 0.05526268279472829, "grad_norm": 2.884336471557617, "learning_rate": 4.7237317205271715e-05, "loss": 4.4596, "step": 6122 }, { "epoch": 0.05527170969489077, "grad_norm": 3.526580333709717, "learning_rate": 4.723686586026359e-05, "loss": 4.8823, "step": 6123 }, { "epoch": 0.05528073659505326, "grad_norm": 4.5164923667907715, "learning_rate": 4.723641451525546e-05, "loss": 3.2486, "step": 6124 }, { "epoch": 0.05528976349521574, "grad_norm": 3.6026651859283447, "learning_rate": 4.723596317024734e-05, "loss": 4.9162, "step": 6125 }, { "epoch": 0.055298790395378224, "grad_norm": 2.7201571464538574, "learning_rate": 4.723551182523921e-05, "loss": 5.2755, "step": 6126 }, { "epoch": 0.05530781729554071, "grad_norm": 3.1810247898101807, "learning_rate": 4.723506048023109e-05, "loss": 4.2409, "step": 6127 }, { "epoch": 0.055316844195703194, "grad_norm": 3.819366931915283, "learning_rate": 4.723460913522297e-05, "loss": 4.1382, "step": 6128 }, { "epoch": 0.05532587109586568, "grad_norm": 3.4000911712646484, "learning_rate": 4.723415779021484e-05, "loss": 4.5812, "step": 6129 }, { "epoch": 0.055334897996028164, "grad_norm": 3.6086130142211914, "learning_rate": 4.7233706445206715e-05, "loss": 4.9901, "step": 6130 }, { "epoch": 0.055343924896190645, "grad_norm": 5.699713706970215, "learning_rate": 4.7233255100198594e-05, "loss": 4.8706, "step": 6131 }, { "epoch": 0.055352951796353134, "grad_norm": 3.033979892730713, "learning_rate": 4.723280375519047e-05, "loss": 4.0292, "step": 6132 }, { "epoch": 0.055361978696515615, "grad_norm": 3.075996160507202, "learning_rate": 4.7232352410182346e-05, "loss": 5.0889, "step": 6133 }, { "epoch": 0.055371005596678104, "grad_norm": 3.6570377349853516, "learning_rate": 4.7231901065174225e-05, "loss": 4.558, "step": 6134 }, { "epoch": 0.055380032496840585, "grad_norm": 3.832706928253174, "learning_rate": 4.72314497201661e-05, "loss": 4.5925, "step": 6135 }, { "epoch": 0.05538905939700307, "grad_norm": 3.52089262008667, "learning_rate": 4.723099837515798e-05, "loss": 4.7321, "step": 6136 }, { "epoch": 0.055398086297165555, "grad_norm": 3.5836880207061768, "learning_rate": 4.723054703014985e-05, "loss": 4.1306, "step": 6137 }, { "epoch": 0.05540711319732804, "grad_norm": 2.821826219558716, "learning_rate": 4.723009568514172e-05, "loss": 5.1689, "step": 6138 }, { "epoch": 0.055416140097490525, "grad_norm": 4.0472798347473145, "learning_rate": 4.72296443401336e-05, "loss": 5.2033, "step": 6139 }, { "epoch": 0.05542516699765301, "grad_norm": 3.2245821952819824, "learning_rate": 4.7229192995125474e-05, "loss": 4.639, "step": 6140 }, { "epoch": 0.05543419389781549, "grad_norm": 4.106179714202881, "learning_rate": 4.722874165011735e-05, "loss": 4.373, "step": 6141 }, { "epoch": 0.05544322079797798, "grad_norm": 5.223020076751709, "learning_rate": 4.722829030510923e-05, "loss": 4.2749, "step": 6142 }, { "epoch": 0.05545224769814046, "grad_norm": 3.3120641708374023, "learning_rate": 4.7227838960101105e-05, "loss": 5.1444, "step": 6143 }, { "epoch": 0.05546127459830294, "grad_norm": 3.3207221031188965, "learning_rate": 4.722738761509298e-05, "loss": 5.0517, "step": 6144 }, { "epoch": 0.05547030149846543, "grad_norm": 2.602496385574341, "learning_rate": 4.7226936270084856e-05, "loss": 4.4347, "step": 6145 }, { "epoch": 0.05547932839862791, "grad_norm": 3.6456563472747803, "learning_rate": 4.722648492507673e-05, "loss": 4.7942, "step": 6146 }, { "epoch": 0.0554883552987904, "grad_norm": 4.184974670410156, "learning_rate": 4.722603358006861e-05, "loss": 4.7513, "step": 6147 }, { "epoch": 0.05549738219895288, "grad_norm": 5.114792823791504, "learning_rate": 4.722558223506048e-05, "loss": 3.703, "step": 6148 }, { "epoch": 0.05550640909911536, "grad_norm": 3.219432830810547, "learning_rate": 4.722513089005236e-05, "loss": 5.2636, "step": 6149 }, { "epoch": 0.05551543599927785, "grad_norm": 3.8142600059509277, "learning_rate": 4.722467954504423e-05, "loss": 3.7924, "step": 6150 }, { "epoch": 0.05552446289944033, "grad_norm": 3.0677456855773926, "learning_rate": 4.7224228200036105e-05, "loss": 5.4337, "step": 6151 }, { "epoch": 0.05553348979960282, "grad_norm": 5.64909553527832, "learning_rate": 4.7223776855027984e-05, "loss": 4.5212, "step": 6152 }, { "epoch": 0.0555425166997653, "grad_norm": 3.637817144393921, "learning_rate": 4.722332551001986e-05, "loss": 4.893, "step": 6153 }, { "epoch": 0.05555154359992778, "grad_norm": 4.185501575469971, "learning_rate": 4.7222874165011736e-05, "loss": 4.0734, "step": 6154 }, { "epoch": 0.05556057050009027, "grad_norm": 3.1615724563598633, "learning_rate": 4.7222422820003615e-05, "loss": 4.506, "step": 6155 }, { "epoch": 0.05556959740025275, "grad_norm": 4.877957344055176, "learning_rate": 4.7221971474995494e-05, "loss": 4.1331, "step": 6156 }, { "epoch": 0.055578624300415234, "grad_norm": 3.9289915561676025, "learning_rate": 4.722152012998736e-05, "loss": 4.2696, "step": 6157 }, { "epoch": 0.05558765120057772, "grad_norm": 4.199972152709961, "learning_rate": 4.722106878497924e-05, "loss": 4.3553, "step": 6158 }, { "epoch": 0.055596678100740204, "grad_norm": 2.8148353099823, "learning_rate": 4.722061743997112e-05, "loss": 4.7759, "step": 6159 }, { "epoch": 0.05560570500090269, "grad_norm": 3.522333860397339, "learning_rate": 4.722016609496299e-05, "loss": 4.3442, "step": 6160 }, { "epoch": 0.055614731901065174, "grad_norm": 3.592142105102539, "learning_rate": 4.721971474995487e-05, "loss": 5.1533, "step": 6161 }, { "epoch": 0.055623758801227656, "grad_norm": 4.304500579833984, "learning_rate": 4.721926340494674e-05, "loss": 4.1586, "step": 6162 }, { "epoch": 0.055632785701390144, "grad_norm": 3.868359088897705, "learning_rate": 4.721881205993862e-05, "loss": 4.7987, "step": 6163 }, { "epoch": 0.055641812601552626, "grad_norm": 3.2388808727264404, "learning_rate": 4.7218360714930494e-05, "loss": 4.426, "step": 6164 }, { "epoch": 0.055650839501715114, "grad_norm": 3.7524240016937256, "learning_rate": 4.721790936992237e-05, "loss": 4.5995, "step": 6165 }, { "epoch": 0.055659866401877596, "grad_norm": 3.4521164894104004, "learning_rate": 4.7217458024914246e-05, "loss": 5.0911, "step": 6166 }, { "epoch": 0.05566889330204008, "grad_norm": 3.5116257667541504, "learning_rate": 4.7217006679906125e-05, "loss": 4.3598, "step": 6167 }, { "epoch": 0.055677920202202566, "grad_norm": 4.985110282897949, "learning_rate": 4.7216555334898e-05, "loss": 3.8137, "step": 6168 }, { "epoch": 0.05568694710236505, "grad_norm": 3.4627327919006348, "learning_rate": 4.721610398988988e-05, "loss": 5.215, "step": 6169 }, { "epoch": 0.05569597400252753, "grad_norm": 4.8835039138793945, "learning_rate": 4.721565264488175e-05, "loss": 4.5947, "step": 6170 }, { "epoch": 0.05570500090269002, "grad_norm": 3.6093032360076904, "learning_rate": 4.721520129987362e-05, "loss": 3.7825, "step": 6171 }, { "epoch": 0.0557140278028525, "grad_norm": 5.799022674560547, "learning_rate": 4.72147499548655e-05, "loss": 4.5616, "step": 6172 }, { "epoch": 0.05572305470301499, "grad_norm": 2.8355321884155273, "learning_rate": 4.721429860985738e-05, "loss": 4.9868, "step": 6173 }, { "epoch": 0.05573208160317747, "grad_norm": 2.604872465133667, "learning_rate": 4.721384726484925e-05, "loss": 5.1663, "step": 6174 }, { "epoch": 0.05574110850333995, "grad_norm": 3.6831393241882324, "learning_rate": 4.721339591984113e-05, "loss": 4.9799, "step": 6175 }, { "epoch": 0.05575013540350244, "grad_norm": 3.4079172611236572, "learning_rate": 4.7212944574833004e-05, "loss": 4.1756, "step": 6176 }, { "epoch": 0.05575916230366492, "grad_norm": 3.81488037109375, "learning_rate": 4.721249322982488e-05, "loss": 4.4715, "step": 6177 }, { "epoch": 0.05576818920382741, "grad_norm": 3.437406301498413, "learning_rate": 4.7212041884816756e-05, "loss": 4.6132, "step": 6178 }, { "epoch": 0.05577721610398989, "grad_norm": 3.4014976024627686, "learning_rate": 4.721159053980863e-05, "loss": 4.8361, "step": 6179 }, { "epoch": 0.05578624300415237, "grad_norm": 2.692537546157837, "learning_rate": 4.721113919480051e-05, "loss": 4.7438, "step": 6180 }, { "epoch": 0.05579526990431486, "grad_norm": 3.0194551944732666, "learning_rate": 4.721068784979239e-05, "loss": 4.7664, "step": 6181 }, { "epoch": 0.05580429680447734, "grad_norm": 3.348266363143921, "learning_rate": 4.721023650478426e-05, "loss": 4.737, "step": 6182 }, { "epoch": 0.055813323704639824, "grad_norm": 3.860067844390869, "learning_rate": 4.720978515977614e-05, "loss": 5.0917, "step": 6183 }, { "epoch": 0.05582235060480231, "grad_norm": 4.663966178894043, "learning_rate": 4.720933381476801e-05, "loss": 5.1262, "step": 6184 }, { "epoch": 0.055831377504964794, "grad_norm": 3.4520740509033203, "learning_rate": 4.7208882469759884e-05, "loss": 4.0008, "step": 6185 }, { "epoch": 0.05584040440512728, "grad_norm": 2.6202733516693115, "learning_rate": 4.720843112475176e-05, "loss": 4.5403, "step": 6186 }, { "epoch": 0.055849431305289764, "grad_norm": 3.937584400177002, "learning_rate": 4.7207979779743635e-05, "loss": 4.8278, "step": 6187 }, { "epoch": 0.055858458205452245, "grad_norm": 4.176036834716797, "learning_rate": 4.7207528434735515e-05, "loss": 4.5074, "step": 6188 }, { "epoch": 0.055867485105614734, "grad_norm": 3.7395436763763428, "learning_rate": 4.7207077089727394e-05, "loss": 4.996, "step": 6189 }, { "epoch": 0.055876512005777215, "grad_norm": 3.01954984664917, "learning_rate": 4.7206625744719266e-05, "loss": 3.6239, "step": 6190 }, { "epoch": 0.055885538905939704, "grad_norm": 4.604175567626953, "learning_rate": 4.720617439971114e-05, "loss": 3.9537, "step": 6191 }, { "epoch": 0.055894565806102185, "grad_norm": 3.9751288890838623, "learning_rate": 4.720572305470302e-05, "loss": 4.0187, "step": 6192 }, { "epoch": 0.05590359270626467, "grad_norm": 2.9967620372772217, "learning_rate": 4.720527170969489e-05, "loss": 3.9287, "step": 6193 }, { "epoch": 0.055912619606427155, "grad_norm": 3.617676258087158, "learning_rate": 4.720482036468677e-05, "loss": 4.6075, "step": 6194 }, { "epoch": 0.05592164650658964, "grad_norm": 3.0503318309783936, "learning_rate": 4.720436901967865e-05, "loss": 4.5213, "step": 6195 }, { "epoch": 0.05593067340675212, "grad_norm": 3.817427158355713, "learning_rate": 4.720391767467052e-05, "loss": 4.6218, "step": 6196 }, { "epoch": 0.05593970030691461, "grad_norm": 3.6150107383728027, "learning_rate": 4.7203466329662394e-05, "loss": 4.4769, "step": 6197 }, { "epoch": 0.05594872720707709, "grad_norm": 4.0913920402526855, "learning_rate": 4.720301498465427e-05, "loss": 5.2784, "step": 6198 }, { "epoch": 0.05595775410723958, "grad_norm": 5.259833812713623, "learning_rate": 4.7202563639646146e-05, "loss": 4.2496, "step": 6199 }, { "epoch": 0.05596678100740206, "grad_norm": 4.008776664733887, "learning_rate": 4.7202112294638025e-05, "loss": 4.189, "step": 6200 }, { "epoch": 0.05597580790756454, "grad_norm": 3.2292861938476562, "learning_rate": 4.72016609496299e-05, "loss": 5.0291, "step": 6201 }, { "epoch": 0.05598483480772703, "grad_norm": 2.9246394634246826, "learning_rate": 4.7201209604621777e-05, "loss": 4.3507, "step": 6202 }, { "epoch": 0.05599386170788951, "grad_norm": 4.181618690490723, "learning_rate": 4.7200758259613656e-05, "loss": 4.7729, "step": 6203 }, { "epoch": 0.056002888608052, "grad_norm": 5.264355659484863, "learning_rate": 4.720030691460552e-05, "loss": 4.474, "step": 6204 }, { "epoch": 0.05601191550821448, "grad_norm": 2.899179458618164, "learning_rate": 4.71998555695974e-05, "loss": 5.1163, "step": 6205 }, { "epoch": 0.05602094240837696, "grad_norm": 2.2407357692718506, "learning_rate": 4.719940422458928e-05, "loss": 4.7952, "step": 6206 }, { "epoch": 0.05602996930853945, "grad_norm": 3.446411609649658, "learning_rate": 4.719895287958115e-05, "loss": 4.9199, "step": 6207 }, { "epoch": 0.05603899620870193, "grad_norm": 3.526940107345581, "learning_rate": 4.719850153457303e-05, "loss": 4.9177, "step": 6208 }, { "epoch": 0.05604802310886441, "grad_norm": 3.886094808578491, "learning_rate": 4.7198050189564904e-05, "loss": 4.5561, "step": 6209 }, { "epoch": 0.0560570500090269, "grad_norm": 2.6116559505462646, "learning_rate": 4.7197598844556783e-05, "loss": 4.6253, "step": 6210 }, { "epoch": 0.05606607690918938, "grad_norm": 3.7141382694244385, "learning_rate": 4.7197147499548656e-05, "loss": 5.1227, "step": 6211 }, { "epoch": 0.05607510380935187, "grad_norm": 3.5179965496063232, "learning_rate": 4.7196696154540535e-05, "loss": 3.7885, "step": 6212 }, { "epoch": 0.05608413070951435, "grad_norm": 3.792290210723877, "learning_rate": 4.719624480953241e-05, "loss": 3.7182, "step": 6213 }, { "epoch": 0.056093157609676834, "grad_norm": 3.430088520050049, "learning_rate": 4.719579346452429e-05, "loss": 4.8987, "step": 6214 }, { "epoch": 0.05610218450983932, "grad_norm": 4.256494522094727, "learning_rate": 4.719534211951616e-05, "loss": 4.8599, "step": 6215 }, { "epoch": 0.056111211410001804, "grad_norm": 5.986686706542969, "learning_rate": 4.719489077450804e-05, "loss": 4.5969, "step": 6216 }, { "epoch": 0.05612023831016429, "grad_norm": 2.3893229961395264, "learning_rate": 4.719443942949991e-05, "loss": 4.5587, "step": 6217 }, { "epoch": 0.056129265210326774, "grad_norm": 6.523247241973877, "learning_rate": 4.7193988084491783e-05, "loss": 5.0593, "step": 6218 }, { "epoch": 0.056138292110489256, "grad_norm": 2.775435209274292, "learning_rate": 4.719353673948366e-05, "loss": 5.2373, "step": 6219 }, { "epoch": 0.056147319010651744, "grad_norm": 3.1100707054138184, "learning_rate": 4.719308539447554e-05, "loss": 4.9059, "step": 6220 }, { "epoch": 0.056156345910814226, "grad_norm": 3.271153688430786, "learning_rate": 4.7192634049467414e-05, "loss": 4.7554, "step": 6221 }, { "epoch": 0.05616537281097671, "grad_norm": 2.7505548000335693, "learning_rate": 4.7192182704459294e-05, "loss": 4.05, "step": 6222 }, { "epoch": 0.056174399711139196, "grad_norm": 2.4128353595733643, "learning_rate": 4.7191731359451166e-05, "loss": 4.6542, "step": 6223 }, { "epoch": 0.05618342661130168, "grad_norm": 3.4739491939544678, "learning_rate": 4.719128001444304e-05, "loss": 4.5866, "step": 6224 }, { "epoch": 0.056192453511464166, "grad_norm": 3.4983561038970947, "learning_rate": 4.719082866943492e-05, "loss": 4.2425, "step": 6225 }, { "epoch": 0.05620148041162665, "grad_norm": 4.789651393890381, "learning_rate": 4.719037732442679e-05, "loss": 5.253, "step": 6226 }, { "epoch": 0.05621050731178913, "grad_norm": 3.119616985321045, "learning_rate": 4.718992597941867e-05, "loss": 4.7634, "step": 6227 }, { "epoch": 0.05621953421195162, "grad_norm": 3.4532742500305176, "learning_rate": 4.718947463441055e-05, "loss": 4.7134, "step": 6228 }, { "epoch": 0.0562285611121141, "grad_norm": 2.9045584201812744, "learning_rate": 4.718902328940242e-05, "loss": 4.7769, "step": 6229 }, { "epoch": 0.05623758801227659, "grad_norm": 2.6757142543792725, "learning_rate": 4.71885719443943e-05, "loss": 4.8683, "step": 6230 }, { "epoch": 0.05624661491243907, "grad_norm": 4.642531871795654, "learning_rate": 4.718812059938617e-05, "loss": 4.5433, "step": 6231 }, { "epoch": 0.05625564181260155, "grad_norm": 3.5891361236572266, "learning_rate": 4.7187669254378045e-05, "loss": 5.1993, "step": 6232 }, { "epoch": 0.05626466871276404, "grad_norm": 2.237863540649414, "learning_rate": 4.7187217909369925e-05, "loss": 5.0733, "step": 6233 }, { "epoch": 0.05627369561292652, "grad_norm": 4.543962001800537, "learning_rate": 4.7186766564361804e-05, "loss": 4.251, "step": 6234 }, { "epoch": 0.056282722513089, "grad_norm": 3.833000421524048, "learning_rate": 4.7186315219353676e-05, "loss": 4.4928, "step": 6235 }, { "epoch": 0.05629174941325149, "grad_norm": 3.2602853775024414, "learning_rate": 4.7185863874345556e-05, "loss": 4.7569, "step": 6236 }, { "epoch": 0.05630077631341397, "grad_norm": 3.278153896331787, "learning_rate": 4.718541252933743e-05, "loss": 4.5901, "step": 6237 }, { "epoch": 0.05630980321357646, "grad_norm": 3.226351499557495, "learning_rate": 4.71849611843293e-05, "loss": 5.2875, "step": 6238 }, { "epoch": 0.05631883011373894, "grad_norm": 3.1008810997009277, "learning_rate": 4.718450983932118e-05, "loss": 4.731, "step": 6239 }, { "epoch": 0.056327857013901424, "grad_norm": 4.060003757476807, "learning_rate": 4.718405849431305e-05, "loss": 4.6491, "step": 6240 }, { "epoch": 0.05633688391406391, "grad_norm": 4.125527858734131, "learning_rate": 4.718360714930493e-05, "loss": 4.6779, "step": 6241 }, { "epoch": 0.056345910814226394, "grad_norm": 4.731974124908447, "learning_rate": 4.718315580429681e-05, "loss": 4.2617, "step": 6242 }, { "epoch": 0.05635493771438888, "grad_norm": 3.06571888923645, "learning_rate": 4.718270445928868e-05, "loss": 4.686, "step": 6243 }, { "epoch": 0.056363964614551364, "grad_norm": 4.39947509765625, "learning_rate": 4.7182253114280556e-05, "loss": 4.677, "step": 6244 }, { "epoch": 0.056372991514713845, "grad_norm": 4.526091575622559, "learning_rate": 4.7181801769272435e-05, "loss": 4.8663, "step": 6245 }, { "epoch": 0.056382018414876334, "grad_norm": 2.874542474746704, "learning_rate": 4.718135042426431e-05, "loss": 4.9153, "step": 6246 }, { "epoch": 0.056391045315038815, "grad_norm": 2.590252637863159, "learning_rate": 4.7180899079256187e-05, "loss": 4.5101, "step": 6247 }, { "epoch": 0.0564000722152013, "grad_norm": 4.16680383682251, "learning_rate": 4.718044773424806e-05, "loss": 4.3725, "step": 6248 }, { "epoch": 0.056409099115363785, "grad_norm": 3.567594051361084, "learning_rate": 4.717999638923994e-05, "loss": 4.9622, "step": 6249 }, { "epoch": 0.05641812601552627, "grad_norm": 3.8062071800231934, "learning_rate": 4.717954504423182e-05, "loss": 4.1406, "step": 6250 }, { "epoch": 0.056427152915688755, "grad_norm": 3.6948726177215576, "learning_rate": 4.717909369922368e-05, "loss": 5.0132, "step": 6251 }, { "epoch": 0.05643617981585124, "grad_norm": 5.173949241638184, "learning_rate": 4.717864235421556e-05, "loss": 4.7496, "step": 6252 }, { "epoch": 0.05644520671601372, "grad_norm": 3.9633967876434326, "learning_rate": 4.717819100920744e-05, "loss": 4.8238, "step": 6253 }, { "epoch": 0.05645423361617621, "grad_norm": 3.3513731956481934, "learning_rate": 4.7177739664199314e-05, "loss": 4.6997, "step": 6254 }, { "epoch": 0.05646326051633869, "grad_norm": 9.178075790405273, "learning_rate": 4.7177288319191193e-05, "loss": 3.8556, "step": 6255 }, { "epoch": 0.05647228741650118, "grad_norm": 3.7177529335021973, "learning_rate": 4.717683697418307e-05, "loss": 4.5304, "step": 6256 }, { "epoch": 0.05648131431666366, "grad_norm": 2.7397098541259766, "learning_rate": 4.7176385629174945e-05, "loss": 5.2009, "step": 6257 }, { "epoch": 0.05649034121682614, "grad_norm": 3.8261959552764893, "learning_rate": 4.717593428416682e-05, "loss": 4.3935, "step": 6258 }, { "epoch": 0.05649936811698863, "grad_norm": 3.026420831680298, "learning_rate": 4.71754829391587e-05, "loss": 5.1153, "step": 6259 }, { "epoch": 0.05650839501715111, "grad_norm": 3.1360740661621094, "learning_rate": 4.717503159415057e-05, "loss": 4.5041, "step": 6260 }, { "epoch": 0.05651742191731359, "grad_norm": 3.5450363159179688, "learning_rate": 4.717458024914245e-05, "loss": 4.8467, "step": 6261 }, { "epoch": 0.05652644881747608, "grad_norm": 3.1923606395721436, "learning_rate": 4.717412890413432e-05, "loss": 4.4277, "step": 6262 }, { "epoch": 0.05653547571763856, "grad_norm": 2.4140591621398926, "learning_rate": 4.71736775591262e-05, "loss": 4.9693, "step": 6263 }, { "epoch": 0.05654450261780105, "grad_norm": 3.745128870010376, "learning_rate": 4.717322621411807e-05, "loss": 4.6841, "step": 6264 }, { "epoch": 0.05655352951796353, "grad_norm": 4.012075901031494, "learning_rate": 4.7172774869109945e-05, "loss": 4.7396, "step": 6265 }, { "epoch": 0.05656255641812601, "grad_norm": 4.434304237365723, "learning_rate": 4.7172323524101824e-05, "loss": 4.7525, "step": 6266 }, { "epoch": 0.0565715833182885, "grad_norm": 2.659534454345703, "learning_rate": 4.7171872179093704e-05, "loss": 5.0404, "step": 6267 }, { "epoch": 0.05658061021845098, "grad_norm": 2.8560972213745117, "learning_rate": 4.7171420834085576e-05, "loss": 4.6326, "step": 6268 }, { "epoch": 0.05658963711861347, "grad_norm": 3.2751963138580322, "learning_rate": 4.7170969489077455e-05, "loss": 4.7781, "step": 6269 }, { "epoch": 0.05659866401877595, "grad_norm": 3.4168875217437744, "learning_rate": 4.717051814406933e-05, "loss": 4.8136, "step": 6270 }, { "epoch": 0.056607690918938434, "grad_norm": 5.0536980628967285, "learning_rate": 4.71700667990612e-05, "loss": 4.0599, "step": 6271 }, { "epoch": 0.05661671781910092, "grad_norm": 3.3898873329162598, "learning_rate": 4.716961545405308e-05, "loss": 4.8206, "step": 6272 }, { "epoch": 0.056625744719263404, "grad_norm": 4.161417007446289, "learning_rate": 4.716916410904496e-05, "loss": 4.8182, "step": 6273 }, { "epoch": 0.056634771619425886, "grad_norm": 2.7210004329681396, "learning_rate": 4.716871276403683e-05, "loss": 4.7769, "step": 6274 }, { "epoch": 0.056643798519588374, "grad_norm": 3.5857884883880615, "learning_rate": 4.716826141902871e-05, "loss": 4.9318, "step": 6275 }, { "epoch": 0.056652825419750856, "grad_norm": 4.852811813354492, "learning_rate": 4.716781007402058e-05, "loss": 4.4656, "step": 6276 }, { "epoch": 0.056661852319913344, "grad_norm": 4.317056655883789, "learning_rate": 4.716735872901246e-05, "loss": 4.8451, "step": 6277 }, { "epoch": 0.056670879220075826, "grad_norm": 5.740682601928711, "learning_rate": 4.7166907384004335e-05, "loss": 4.9992, "step": 6278 }, { "epoch": 0.05667990612023831, "grad_norm": 2.4093825817108154, "learning_rate": 4.716645603899621e-05, "loss": 4.8052, "step": 6279 }, { "epoch": 0.056688933020400796, "grad_norm": 2.995878219604492, "learning_rate": 4.7166004693988086e-05, "loss": 4.687, "step": 6280 }, { "epoch": 0.05669795992056328, "grad_norm": 3.519474744796753, "learning_rate": 4.7165553348979966e-05, "loss": 5.3828, "step": 6281 }, { "epoch": 0.056706986820725766, "grad_norm": 7.202792167663574, "learning_rate": 4.716510200397184e-05, "loss": 4.6257, "step": 6282 }, { "epoch": 0.05671601372088825, "grad_norm": 3.960545063018799, "learning_rate": 4.716465065896372e-05, "loss": 5.048, "step": 6283 }, { "epoch": 0.05672504062105073, "grad_norm": 3.2881102561950684, "learning_rate": 4.716419931395559e-05, "loss": 4.4214, "step": 6284 }, { "epoch": 0.05673406752121322, "grad_norm": 4.138964653015137, "learning_rate": 4.716374796894746e-05, "loss": 4.9206, "step": 6285 }, { "epoch": 0.0567430944213757, "grad_norm": 2.9202868938446045, "learning_rate": 4.716329662393934e-05, "loss": 5.45, "step": 6286 }, { "epoch": 0.05675212132153818, "grad_norm": 3.539827823638916, "learning_rate": 4.7162845278931214e-05, "loss": 5.3046, "step": 6287 }, { "epoch": 0.05676114822170067, "grad_norm": 3.47406005859375, "learning_rate": 4.716239393392309e-05, "loss": 4.4875, "step": 6288 }, { "epoch": 0.05677017512186315, "grad_norm": 4.464563846588135, "learning_rate": 4.716194258891497e-05, "loss": 5.1356, "step": 6289 }, { "epoch": 0.05677920202202564, "grad_norm": 2.7887086868286133, "learning_rate": 4.7161491243906845e-05, "loss": 4.6827, "step": 6290 }, { "epoch": 0.05678822892218812, "grad_norm": 4.531978130340576, "learning_rate": 4.716103989889872e-05, "loss": 4.8851, "step": 6291 }, { "epoch": 0.0567972558223506, "grad_norm": 3.693305253982544, "learning_rate": 4.7160588553890597e-05, "loss": 3.592, "step": 6292 }, { "epoch": 0.05680628272251309, "grad_norm": 3.3706719875335693, "learning_rate": 4.716013720888247e-05, "loss": 4.076, "step": 6293 }, { "epoch": 0.05681530962267557, "grad_norm": 4.476597785949707, "learning_rate": 4.715968586387435e-05, "loss": 4.5016, "step": 6294 }, { "epoch": 0.05682433652283806, "grad_norm": 3.9716899394989014, "learning_rate": 4.715923451886623e-05, "loss": 3.471, "step": 6295 }, { "epoch": 0.05683336342300054, "grad_norm": 5.264523983001709, "learning_rate": 4.71587831738581e-05, "loss": 4.7169, "step": 6296 }, { "epoch": 0.056842390323163024, "grad_norm": 2.9821865558624268, "learning_rate": 4.715833182884998e-05, "loss": 5.1301, "step": 6297 }, { "epoch": 0.05685141722332551, "grad_norm": 3.8128724098205566, "learning_rate": 4.715788048384185e-05, "loss": 4.108, "step": 6298 }, { "epoch": 0.056860444123487994, "grad_norm": 3.3624062538146973, "learning_rate": 4.7157429138833724e-05, "loss": 5.0366, "step": 6299 }, { "epoch": 0.056869471023650475, "grad_norm": 3.998509407043457, "learning_rate": 4.71569777938256e-05, "loss": 4.2304, "step": 6300 }, { "epoch": 0.056878497923812964, "grad_norm": 3.6072685718536377, "learning_rate": 4.7156526448817476e-05, "loss": 4.4688, "step": 6301 }, { "epoch": 0.056887524823975445, "grad_norm": 4.129689693450928, "learning_rate": 4.7156075103809355e-05, "loss": 4.6438, "step": 6302 }, { "epoch": 0.056896551724137934, "grad_norm": 3.5261080265045166, "learning_rate": 4.7155623758801234e-05, "loss": 4.6098, "step": 6303 }, { "epoch": 0.056905578624300415, "grad_norm": 3.0728485584259033, "learning_rate": 4.715517241379311e-05, "loss": 5.1872, "step": 6304 }, { "epoch": 0.0569146055244629, "grad_norm": 3.7849817276000977, "learning_rate": 4.715472106878498e-05, "loss": 5.3096, "step": 6305 }, { "epoch": 0.056923632424625385, "grad_norm": 3.651799440383911, "learning_rate": 4.715426972377686e-05, "loss": 4.435, "step": 6306 }, { "epoch": 0.05693265932478787, "grad_norm": 3.338006019592285, "learning_rate": 4.715381837876873e-05, "loss": 4.206, "step": 6307 }, { "epoch": 0.056941686224950355, "grad_norm": 3.1152594089508057, "learning_rate": 4.715336703376061e-05, "loss": 4.8087, "step": 6308 }, { "epoch": 0.05695071312511284, "grad_norm": 3.4513843059539795, "learning_rate": 4.715291568875248e-05, "loss": 4.6656, "step": 6309 }, { "epoch": 0.05695974002527532, "grad_norm": 3.983344554901123, "learning_rate": 4.715246434374436e-05, "loss": 4.669, "step": 6310 }, { "epoch": 0.05696876692543781, "grad_norm": 3.402801275253296, "learning_rate": 4.7152012998736234e-05, "loss": 5.2992, "step": 6311 }, { "epoch": 0.05697779382560029, "grad_norm": 3.1010098457336426, "learning_rate": 4.715156165372811e-05, "loss": 5.0357, "step": 6312 }, { "epoch": 0.05698682072576277, "grad_norm": 4.645699501037598, "learning_rate": 4.7151110308719986e-05, "loss": 4.6262, "step": 6313 }, { "epoch": 0.05699584762592526, "grad_norm": 3.12093448638916, "learning_rate": 4.7150658963711865e-05, "loss": 4.8842, "step": 6314 }, { "epoch": 0.05700487452608774, "grad_norm": 2.8998758792877197, "learning_rate": 4.715020761870374e-05, "loss": 5.0168, "step": 6315 }, { "epoch": 0.05701390142625023, "grad_norm": 3.9995055198669434, "learning_rate": 4.714975627369562e-05, "loss": 4.835, "step": 6316 }, { "epoch": 0.05702292832641271, "grad_norm": 3.3709516525268555, "learning_rate": 4.7149304928687496e-05, "loss": 4.4408, "step": 6317 }, { "epoch": 0.05703195522657519, "grad_norm": 3.323092222213745, "learning_rate": 4.714885358367936e-05, "loss": 4.0949, "step": 6318 }, { "epoch": 0.05704098212673768, "grad_norm": 3.975764274597168, "learning_rate": 4.714840223867124e-05, "loss": 4.4805, "step": 6319 }, { "epoch": 0.05705000902690016, "grad_norm": 3.2736618518829346, "learning_rate": 4.714795089366312e-05, "loss": 4.4164, "step": 6320 }, { "epoch": 0.05705903592706265, "grad_norm": 3.2413442134857178, "learning_rate": 4.714749954865499e-05, "loss": 4.7116, "step": 6321 }, { "epoch": 0.05706806282722513, "grad_norm": 3.157644271850586, "learning_rate": 4.714704820364687e-05, "loss": 3.9587, "step": 6322 }, { "epoch": 0.05707708972738761, "grad_norm": 2.7486720085144043, "learning_rate": 4.7146596858638745e-05, "loss": 4.6463, "step": 6323 }, { "epoch": 0.0570861166275501, "grad_norm": 3.662583351135254, "learning_rate": 4.7146145513630624e-05, "loss": 4.5272, "step": 6324 }, { "epoch": 0.05709514352771258, "grad_norm": 3.6576972007751465, "learning_rate": 4.7145694168622496e-05, "loss": 4.748, "step": 6325 }, { "epoch": 0.057104170427875064, "grad_norm": 3.460855007171631, "learning_rate": 4.714524282361437e-05, "loss": 4.5317, "step": 6326 }, { "epoch": 0.05711319732803755, "grad_norm": 3.7257399559020996, "learning_rate": 4.714479147860625e-05, "loss": 4.521, "step": 6327 }, { "epoch": 0.057122224228200034, "grad_norm": 4.330859661102295, "learning_rate": 4.714434013359813e-05, "loss": 4.8937, "step": 6328 }, { "epoch": 0.05713125112836252, "grad_norm": 4.02699089050293, "learning_rate": 4.714388878859e-05, "loss": 4.6746, "step": 6329 }, { "epoch": 0.057140278028525004, "grad_norm": 2.4987306594848633, "learning_rate": 4.714343744358188e-05, "loss": 4.7717, "step": 6330 }, { "epoch": 0.057149304928687486, "grad_norm": 3.9684767723083496, "learning_rate": 4.714298609857375e-05, "loss": 4.8122, "step": 6331 }, { "epoch": 0.057158331828849974, "grad_norm": 4.186667442321777, "learning_rate": 4.7142534753565624e-05, "loss": 4.7948, "step": 6332 }, { "epoch": 0.057167358729012456, "grad_norm": 3.5759291648864746, "learning_rate": 4.71420834085575e-05, "loss": 4.4534, "step": 6333 }, { "epoch": 0.057176385629174944, "grad_norm": 3.0765774250030518, "learning_rate": 4.714163206354938e-05, "loss": 4.4396, "step": 6334 }, { "epoch": 0.057185412529337426, "grad_norm": 4.316167831420898, "learning_rate": 4.7141180718541255e-05, "loss": 4.4673, "step": 6335 }, { "epoch": 0.05719443942949991, "grad_norm": 2.58408784866333, "learning_rate": 4.7140729373533134e-05, "loss": 4.9962, "step": 6336 }, { "epoch": 0.057203466329662396, "grad_norm": 4.8675856590271, "learning_rate": 4.7140278028525007e-05, "loss": 4.3742, "step": 6337 }, { "epoch": 0.05721249322982488, "grad_norm": 4.302143573760986, "learning_rate": 4.713982668351688e-05, "loss": 4.7554, "step": 6338 }, { "epoch": 0.05722152012998736, "grad_norm": 3.777733564376831, "learning_rate": 4.713937533850876e-05, "loss": 4.6556, "step": 6339 }, { "epoch": 0.05723054703014985, "grad_norm": 3.2539713382720947, "learning_rate": 4.713892399350063e-05, "loss": 4.8573, "step": 6340 }, { "epoch": 0.05723957393031233, "grad_norm": 3.8643550872802734, "learning_rate": 4.713847264849251e-05, "loss": 5.0271, "step": 6341 }, { "epoch": 0.05724860083047482, "grad_norm": 3.833827257156372, "learning_rate": 4.713802130348439e-05, "loss": 4.899, "step": 6342 }, { "epoch": 0.0572576277306373, "grad_norm": 2.279327869415283, "learning_rate": 4.713756995847626e-05, "loss": 4.4815, "step": 6343 }, { "epoch": 0.05726665463079978, "grad_norm": 4.051000595092773, "learning_rate": 4.713711861346814e-05, "loss": 4.8961, "step": 6344 }, { "epoch": 0.05727568153096227, "grad_norm": 2.6795883178710938, "learning_rate": 4.713666726846001e-05, "loss": 4.8207, "step": 6345 }, { "epoch": 0.05728470843112475, "grad_norm": 4.239276885986328, "learning_rate": 4.7136215923451886e-05, "loss": 4.5383, "step": 6346 }, { "epoch": 0.05729373533128724, "grad_norm": 3.90470027923584, "learning_rate": 4.7135764578443765e-05, "loss": 4.9054, "step": 6347 }, { "epoch": 0.05730276223144972, "grad_norm": 3.113978624343872, "learning_rate": 4.713531323343564e-05, "loss": 5.2291, "step": 6348 }, { "epoch": 0.0573117891316122, "grad_norm": 3.3741109371185303, "learning_rate": 4.713486188842752e-05, "loss": 5.2919, "step": 6349 }, { "epoch": 0.05732081603177469, "grad_norm": 3.3065459728240967, "learning_rate": 4.7134410543419396e-05, "loss": 4.6595, "step": 6350 }, { "epoch": 0.05732984293193717, "grad_norm": 4.951126575469971, "learning_rate": 4.713395919841127e-05, "loss": 4.0229, "step": 6351 }, { "epoch": 0.057338869832099654, "grad_norm": 2.9660143852233887, "learning_rate": 4.713350785340314e-05, "loss": 4.8693, "step": 6352 }, { "epoch": 0.05734789673226214, "grad_norm": 2.8556325435638428, "learning_rate": 4.713305650839502e-05, "loss": 4.1685, "step": 6353 }, { "epoch": 0.057356923632424624, "grad_norm": 3.843149423599243, "learning_rate": 4.713260516338689e-05, "loss": 4.1819, "step": 6354 }, { "epoch": 0.05736595053258711, "grad_norm": 2.886930227279663, "learning_rate": 4.713215381837877e-05, "loss": 4.5103, "step": 6355 }, { "epoch": 0.057374977432749594, "grad_norm": 6.29197883605957, "learning_rate": 4.713170247337065e-05, "loss": 4.6007, "step": 6356 }, { "epoch": 0.057384004332912075, "grad_norm": 4.549216270446777, "learning_rate": 4.7131251128362524e-05, "loss": 4.421, "step": 6357 }, { "epoch": 0.057393031233074564, "grad_norm": 3.5124151706695557, "learning_rate": 4.7130799783354396e-05, "loss": 4.1096, "step": 6358 }, { "epoch": 0.057402058133237045, "grad_norm": 3.795133113861084, "learning_rate": 4.7130348438346275e-05, "loss": 4.6403, "step": 6359 }, { "epoch": 0.057411085033399534, "grad_norm": 3.203681707382202, "learning_rate": 4.712989709333815e-05, "loss": 4.3211, "step": 6360 }, { "epoch": 0.057420111933562015, "grad_norm": 4.706249713897705, "learning_rate": 4.712944574833003e-05, "loss": 4.6021, "step": 6361 }, { "epoch": 0.0574291388337245, "grad_norm": 4.038034915924072, "learning_rate": 4.71289944033219e-05, "loss": 4.5065, "step": 6362 }, { "epoch": 0.057438165733886985, "grad_norm": 2.8319833278656006, "learning_rate": 4.712854305831378e-05, "loss": 5.3316, "step": 6363 }, { "epoch": 0.05744719263404947, "grad_norm": 3.4883368015289307, "learning_rate": 4.712809171330566e-05, "loss": 4.155, "step": 6364 }, { "epoch": 0.05745621953421195, "grad_norm": 2.3310463428497314, "learning_rate": 4.7127640368297524e-05, "loss": 4.5313, "step": 6365 }, { "epoch": 0.05746524643437444, "grad_norm": 2.420294761657715, "learning_rate": 4.71271890232894e-05, "loss": 5.135, "step": 6366 }, { "epoch": 0.05747427333453692, "grad_norm": 3.703866720199585, "learning_rate": 4.712673767828128e-05, "loss": 4.5963, "step": 6367 }, { "epoch": 0.05748330023469941, "grad_norm": 3.173525333404541, "learning_rate": 4.7126286333273155e-05, "loss": 5.403, "step": 6368 }, { "epoch": 0.05749232713486189, "grad_norm": 3.5494472980499268, "learning_rate": 4.7125834988265034e-05, "loss": 4.278, "step": 6369 }, { "epoch": 0.05750135403502437, "grad_norm": 3.0481584072113037, "learning_rate": 4.7125383643256906e-05, "loss": 4.6451, "step": 6370 }, { "epoch": 0.05751038093518686, "grad_norm": 3.160554885864258, "learning_rate": 4.7124932298248786e-05, "loss": 4.3525, "step": 6371 }, { "epoch": 0.05751940783534934, "grad_norm": 3.1851770877838135, "learning_rate": 4.712448095324066e-05, "loss": 4.4767, "step": 6372 }, { "epoch": 0.05752843473551183, "grad_norm": 2.875558853149414, "learning_rate": 4.712402960823254e-05, "loss": 4.8259, "step": 6373 }, { "epoch": 0.05753746163567431, "grad_norm": 3.698575496673584, "learning_rate": 4.712357826322441e-05, "loss": 4.7078, "step": 6374 }, { "epoch": 0.05754648853583679, "grad_norm": 5.083904266357422, "learning_rate": 4.712312691821629e-05, "loss": 4.9597, "step": 6375 }, { "epoch": 0.05755551543599928, "grad_norm": 2.9286534786224365, "learning_rate": 4.712267557320816e-05, "loss": 4.5382, "step": 6376 }, { "epoch": 0.05756454233616176, "grad_norm": 2.9020421504974365, "learning_rate": 4.712222422820004e-05, "loss": 4.9167, "step": 6377 }, { "epoch": 0.05757356923632424, "grad_norm": 4.178226470947266, "learning_rate": 4.712177288319191e-05, "loss": 4.7901, "step": 6378 }, { "epoch": 0.05758259613648673, "grad_norm": 3.588573455810547, "learning_rate": 4.7121321538183786e-05, "loss": 5.183, "step": 6379 }, { "epoch": 0.05759162303664921, "grad_norm": 4.269277095794678, "learning_rate": 4.7120870193175665e-05, "loss": 5.2434, "step": 6380 }, { "epoch": 0.0576006499368117, "grad_norm": 3.627678871154785, "learning_rate": 4.7120418848167544e-05, "loss": 4.7343, "step": 6381 }, { "epoch": 0.05760967683697418, "grad_norm": 3.300657272338867, "learning_rate": 4.7119967503159417e-05, "loss": 4.432, "step": 6382 }, { "epoch": 0.057618703737136664, "grad_norm": 4.675553798675537, "learning_rate": 4.7119516158151296e-05, "loss": 4.7373, "step": 6383 }, { "epoch": 0.05762773063729915, "grad_norm": 3.168470621109009, "learning_rate": 4.711906481314317e-05, "loss": 4.7864, "step": 6384 }, { "epoch": 0.057636757537461634, "grad_norm": 3.98557186126709, "learning_rate": 4.711861346813504e-05, "loss": 4.6826, "step": 6385 }, { "epoch": 0.05764578443762412, "grad_norm": 4.286021709442139, "learning_rate": 4.711816212312692e-05, "loss": 4.7292, "step": 6386 }, { "epoch": 0.057654811337786604, "grad_norm": 2.5943338871002197, "learning_rate": 4.711771077811879e-05, "loss": 5.1405, "step": 6387 }, { "epoch": 0.057663838237949086, "grad_norm": 2.9568307399749756, "learning_rate": 4.711725943311067e-05, "loss": 5.0239, "step": 6388 }, { "epoch": 0.057672865138111575, "grad_norm": 3.5010504722595215, "learning_rate": 4.711680808810255e-05, "loss": 5.0331, "step": 6389 }, { "epoch": 0.057681892038274056, "grad_norm": 3.8756964206695557, "learning_rate": 4.711635674309442e-05, "loss": 4.5419, "step": 6390 }, { "epoch": 0.05769091893843654, "grad_norm": 3.302154779434204, "learning_rate": 4.71159053980863e-05, "loss": 4.3441, "step": 6391 }, { "epoch": 0.057699945838599026, "grad_norm": 4.249909400939941, "learning_rate": 4.7115454053078175e-05, "loss": 5.1441, "step": 6392 }, { "epoch": 0.05770897273876151, "grad_norm": 3.994034767150879, "learning_rate": 4.711500270807005e-05, "loss": 4.5855, "step": 6393 }, { "epoch": 0.057717999638923996, "grad_norm": 3.242525339126587, "learning_rate": 4.711455136306193e-05, "loss": 4.5413, "step": 6394 }, { "epoch": 0.05772702653908648, "grad_norm": 2.9685192108154297, "learning_rate": 4.7114100018053806e-05, "loss": 4.5175, "step": 6395 }, { "epoch": 0.05773605343924896, "grad_norm": 3.3289906978607178, "learning_rate": 4.711364867304568e-05, "loss": 4.3528, "step": 6396 }, { "epoch": 0.05774508033941145, "grad_norm": 5.273959159851074, "learning_rate": 4.711319732803756e-05, "loss": 4.4916, "step": 6397 }, { "epoch": 0.05775410723957393, "grad_norm": 3.4910354614257812, "learning_rate": 4.711274598302943e-05, "loss": 4.5172, "step": 6398 }, { "epoch": 0.05776313413973642, "grad_norm": 3.3733227252960205, "learning_rate": 4.71122946380213e-05, "loss": 4.5912, "step": 6399 }, { "epoch": 0.0577721610398989, "grad_norm": 3.5590550899505615, "learning_rate": 4.711184329301318e-05, "loss": 5.1689, "step": 6400 }, { "epoch": 0.05778118794006138, "grad_norm": 3.7992348670959473, "learning_rate": 4.7111391948005054e-05, "loss": 4.7712, "step": 6401 }, { "epoch": 0.05779021484022387, "grad_norm": 3.1381747722625732, "learning_rate": 4.7110940602996934e-05, "loss": 4.4283, "step": 6402 }, { "epoch": 0.05779924174038635, "grad_norm": 4.020051956176758, "learning_rate": 4.711048925798881e-05, "loss": 5.2394, "step": 6403 }, { "epoch": 0.05780826864054883, "grad_norm": 4.015910625457764, "learning_rate": 4.7110037912980685e-05, "loss": 5.0634, "step": 6404 }, { "epoch": 0.05781729554071132, "grad_norm": 2.6987226009368896, "learning_rate": 4.710958656797256e-05, "loss": 5.1484, "step": 6405 }, { "epoch": 0.0578263224408738, "grad_norm": 3.5593199729919434, "learning_rate": 4.710913522296444e-05, "loss": 4.551, "step": 6406 }, { "epoch": 0.05783534934103629, "grad_norm": 4.029634952545166, "learning_rate": 4.710868387795631e-05, "loss": 5.0207, "step": 6407 }, { "epoch": 0.05784437624119877, "grad_norm": 3.2671611309051514, "learning_rate": 4.710823253294819e-05, "loss": 4.6433, "step": 6408 }, { "epoch": 0.057853403141361254, "grad_norm": 6.469050407409668, "learning_rate": 4.710778118794006e-05, "loss": 3.9645, "step": 6409 }, { "epoch": 0.05786243004152374, "grad_norm": 3.4559831619262695, "learning_rate": 4.710732984293194e-05, "loss": 4.3335, "step": 6410 }, { "epoch": 0.057871456941686224, "grad_norm": 2.7821786403656006, "learning_rate": 4.710687849792382e-05, "loss": 5.0031, "step": 6411 }, { "epoch": 0.05788048384184871, "grad_norm": 4.1350932121276855, "learning_rate": 4.7106427152915685e-05, "loss": 3.9644, "step": 6412 }, { "epoch": 0.057889510742011194, "grad_norm": 4.141672611236572, "learning_rate": 4.7105975807907565e-05, "loss": 4.777, "step": 6413 }, { "epoch": 0.057898537642173675, "grad_norm": 4.318477153778076, "learning_rate": 4.7105524462899444e-05, "loss": 5.3687, "step": 6414 }, { "epoch": 0.057907564542336164, "grad_norm": 3.2833762168884277, "learning_rate": 4.7105073117891316e-05, "loss": 4.2727, "step": 6415 }, { "epoch": 0.057916591442498645, "grad_norm": 3.621267080307007, "learning_rate": 4.7104621772883196e-05, "loss": 4.545, "step": 6416 }, { "epoch": 0.05792561834266113, "grad_norm": 5.12749719619751, "learning_rate": 4.7104170427875075e-05, "loss": 5.3664, "step": 6417 }, { "epoch": 0.057934645242823615, "grad_norm": 2.6906135082244873, "learning_rate": 4.710371908286695e-05, "loss": 5.5691, "step": 6418 }, { "epoch": 0.0579436721429861, "grad_norm": 6.030412673950195, "learning_rate": 4.710326773785882e-05, "loss": 4.8368, "step": 6419 }, { "epoch": 0.057952699043148585, "grad_norm": 5.422453880310059, "learning_rate": 4.71028163928507e-05, "loss": 3.887, "step": 6420 }, { "epoch": 0.05796172594331107, "grad_norm": 3.59533953666687, "learning_rate": 4.710236504784257e-05, "loss": 4.8291, "step": 6421 }, { "epoch": 0.05797075284347355, "grad_norm": 3.7998106479644775, "learning_rate": 4.710191370283445e-05, "loss": 4.9699, "step": 6422 }, { "epoch": 0.05797977974363604, "grad_norm": 3.93979549407959, "learning_rate": 4.710146235782632e-05, "loss": 4.664, "step": 6423 }, { "epoch": 0.05798880664379852, "grad_norm": 3.3175814151763916, "learning_rate": 4.71010110128182e-05, "loss": 5.1419, "step": 6424 }, { "epoch": 0.05799783354396101, "grad_norm": 4.981836318969727, "learning_rate": 4.7100559667810075e-05, "loss": 4.2337, "step": 6425 }, { "epoch": 0.05800686044412349, "grad_norm": 3.730112075805664, "learning_rate": 4.710010832280195e-05, "loss": 4.4294, "step": 6426 }, { "epoch": 0.05801588734428597, "grad_norm": 2.7335922718048096, "learning_rate": 4.7099656977793827e-05, "loss": 4.7808, "step": 6427 }, { "epoch": 0.05802491424444846, "grad_norm": 4.285132884979248, "learning_rate": 4.7099205632785706e-05, "loss": 4.701, "step": 6428 }, { "epoch": 0.05803394114461094, "grad_norm": 3.039649486541748, "learning_rate": 4.709875428777758e-05, "loss": 4.7909, "step": 6429 }, { "epoch": 0.05804296804477342, "grad_norm": 3.346266269683838, "learning_rate": 4.709830294276946e-05, "loss": 4.6705, "step": 6430 }, { "epoch": 0.05805199494493591, "grad_norm": 4.190643310546875, "learning_rate": 4.709785159776133e-05, "loss": 4.8282, "step": 6431 }, { "epoch": 0.05806102184509839, "grad_norm": 3.1078078746795654, "learning_rate": 4.70974002527532e-05, "loss": 4.6482, "step": 6432 }, { "epoch": 0.05807004874526088, "grad_norm": 3.4460933208465576, "learning_rate": 4.709694890774508e-05, "loss": 4.533, "step": 6433 }, { "epoch": 0.05807907564542336, "grad_norm": 4.517588138580322, "learning_rate": 4.709649756273696e-05, "loss": 3.7283, "step": 6434 }, { "epoch": 0.05808810254558584, "grad_norm": 2.766638994216919, "learning_rate": 4.709604621772883e-05, "loss": 5.0657, "step": 6435 }, { "epoch": 0.05809712944574833, "grad_norm": 3.5161280632019043, "learning_rate": 4.709559487272071e-05, "loss": 5.0789, "step": 6436 }, { "epoch": 0.05810615634591081, "grad_norm": 4.398107051849365, "learning_rate": 4.7095143527712585e-05, "loss": 4.0882, "step": 6437 }, { "epoch": 0.0581151832460733, "grad_norm": 3.7402148246765137, "learning_rate": 4.7094692182704464e-05, "loss": 4.3355, "step": 6438 }, { "epoch": 0.05812421014623578, "grad_norm": 3.8466851711273193, "learning_rate": 4.709424083769634e-05, "loss": 4.7012, "step": 6439 }, { "epoch": 0.058133237046398265, "grad_norm": 3.077263832092285, "learning_rate": 4.709378949268821e-05, "loss": 5.2304, "step": 6440 }, { "epoch": 0.05814226394656075, "grad_norm": 3.8317854404449463, "learning_rate": 4.709333814768009e-05, "loss": 4.7283, "step": 6441 }, { "epoch": 0.058151290846723235, "grad_norm": 4.281594276428223, "learning_rate": 4.709288680267197e-05, "loss": 5.1105, "step": 6442 }, { "epoch": 0.058160317746885716, "grad_norm": 3.5791127681732178, "learning_rate": 4.709243545766384e-05, "loss": 4.6929, "step": 6443 }, { "epoch": 0.058169344647048205, "grad_norm": 3.214526891708374, "learning_rate": 4.709198411265572e-05, "loss": 5.0726, "step": 6444 }, { "epoch": 0.058178371547210686, "grad_norm": 4.388269424438477, "learning_rate": 4.709153276764759e-05, "loss": 4.1197, "step": 6445 }, { "epoch": 0.058187398447373175, "grad_norm": 3.141505241394043, "learning_rate": 4.7091081422639464e-05, "loss": 5.3153, "step": 6446 }, { "epoch": 0.058196425347535656, "grad_norm": 3.896786689758301, "learning_rate": 4.7090630077631344e-05, "loss": 4.4479, "step": 6447 }, { "epoch": 0.05820545224769814, "grad_norm": 3.446129322052002, "learning_rate": 4.7090178732623216e-05, "loss": 4.7859, "step": 6448 }, { "epoch": 0.058214479147860626, "grad_norm": 5.172661781311035, "learning_rate": 4.7089727387615095e-05, "loss": 4.6785, "step": 6449 }, { "epoch": 0.05822350604802311, "grad_norm": 2.9769155979156494, "learning_rate": 4.7089276042606975e-05, "loss": 5.0378, "step": 6450 }, { "epoch": 0.058232532948185596, "grad_norm": 2.2838082313537598, "learning_rate": 4.708882469759885e-05, "loss": 5.0733, "step": 6451 }, { "epoch": 0.05824155984834808, "grad_norm": 3.748356819152832, "learning_rate": 4.708837335259072e-05, "loss": 4.4537, "step": 6452 }, { "epoch": 0.05825058674851056, "grad_norm": 4.0148725509643555, "learning_rate": 4.70879220075826e-05, "loss": 5.1549, "step": 6453 }, { "epoch": 0.05825961364867305, "grad_norm": 6.280846118927002, "learning_rate": 4.708747066257447e-05, "loss": 4.8629, "step": 6454 }, { "epoch": 0.05826864054883553, "grad_norm": 4.482416152954102, "learning_rate": 4.708701931756635e-05, "loss": 4.6748, "step": 6455 }, { "epoch": 0.05827766744899801, "grad_norm": 3.183548927307129, "learning_rate": 4.708656797255823e-05, "loss": 4.7121, "step": 6456 }, { "epoch": 0.0582866943491605, "grad_norm": 3.7478415966033936, "learning_rate": 4.70861166275501e-05, "loss": 4.4177, "step": 6457 }, { "epoch": 0.05829572124932298, "grad_norm": 3.7222561836242676, "learning_rate": 4.708566528254198e-05, "loss": 4.8154, "step": 6458 }, { "epoch": 0.05830474814948547, "grad_norm": 3.663501501083374, "learning_rate": 4.7085213937533854e-05, "loss": 4.1787, "step": 6459 }, { "epoch": 0.05831377504964795, "grad_norm": 3.3614561557769775, "learning_rate": 4.7084762592525726e-05, "loss": 4.6272, "step": 6460 }, { "epoch": 0.05832280194981043, "grad_norm": 3.012223720550537, "learning_rate": 4.7084311247517606e-05, "loss": 5.1616, "step": 6461 }, { "epoch": 0.05833182884997292, "grad_norm": 3.4466326236724854, "learning_rate": 4.708385990250948e-05, "loss": 4.7705, "step": 6462 }, { "epoch": 0.0583408557501354, "grad_norm": 3.0097620487213135, "learning_rate": 4.708340855750136e-05, "loss": 4.3523, "step": 6463 }, { "epoch": 0.05834988265029789, "grad_norm": 3.8803675174713135, "learning_rate": 4.7082957212493236e-05, "loss": 4.6158, "step": 6464 }, { "epoch": 0.05835890955046037, "grad_norm": 3.007166624069214, "learning_rate": 4.708250586748511e-05, "loss": 4.955, "step": 6465 }, { "epoch": 0.058367936450622854, "grad_norm": 3.768725633621216, "learning_rate": 4.708205452247698e-05, "loss": 5.4122, "step": 6466 }, { "epoch": 0.05837696335078534, "grad_norm": 3.990399122238159, "learning_rate": 4.708160317746886e-05, "loss": 4.9762, "step": 6467 }, { "epoch": 0.058385990250947824, "grad_norm": 3.560946226119995, "learning_rate": 4.708115183246073e-05, "loss": 4.301, "step": 6468 }, { "epoch": 0.058395017151110305, "grad_norm": 4.471059799194336, "learning_rate": 4.708070048745261e-05, "loss": 3.4531, "step": 6469 }, { "epoch": 0.058404044051272794, "grad_norm": 3.0925633907318115, "learning_rate": 4.7080249142444485e-05, "loss": 4.559, "step": 6470 }, { "epoch": 0.058413070951435275, "grad_norm": 3.4093081951141357, "learning_rate": 4.7079797797436364e-05, "loss": 4.8458, "step": 6471 }, { "epoch": 0.058422097851597764, "grad_norm": 4.91413688659668, "learning_rate": 4.7079346452428237e-05, "loss": 5.0674, "step": 6472 }, { "epoch": 0.058431124751760245, "grad_norm": 7.131696701049805, "learning_rate": 4.707889510742011e-05, "loss": 4.708, "step": 6473 }, { "epoch": 0.05844015165192273, "grad_norm": 3.9285197257995605, "learning_rate": 4.707844376241199e-05, "loss": 4.9218, "step": 6474 }, { "epoch": 0.058449178552085215, "grad_norm": 4.046018123626709, "learning_rate": 4.707799241740387e-05, "loss": 4.9219, "step": 6475 }, { "epoch": 0.0584582054522477, "grad_norm": 3.937344789505005, "learning_rate": 4.707754107239574e-05, "loss": 4.9185, "step": 6476 }, { "epoch": 0.058467232352410185, "grad_norm": 2.825216293334961, "learning_rate": 4.707708972738762e-05, "loss": 4.8219, "step": 6477 }, { "epoch": 0.05847625925257267, "grad_norm": 4.753947734832764, "learning_rate": 4.70766383823795e-05, "loss": 4.26, "step": 6478 }, { "epoch": 0.05848528615273515, "grad_norm": 6.928279876708984, "learning_rate": 4.7076187037371364e-05, "loss": 4.5336, "step": 6479 }, { "epoch": 0.05849431305289764, "grad_norm": 3.2829110622406006, "learning_rate": 4.707573569236324e-05, "loss": 3.7116, "step": 6480 }, { "epoch": 0.05850333995306012, "grad_norm": 3.6837522983551025, "learning_rate": 4.707528434735512e-05, "loss": 4.9063, "step": 6481 }, { "epoch": 0.0585123668532226, "grad_norm": 3.5728707313537598, "learning_rate": 4.7074833002346995e-05, "loss": 5.1741, "step": 6482 }, { "epoch": 0.05852139375338509, "grad_norm": 3.221684694290161, "learning_rate": 4.7074381657338874e-05, "loss": 4.8187, "step": 6483 }, { "epoch": 0.05853042065354757, "grad_norm": 3.663402795791626, "learning_rate": 4.707393031233075e-05, "loss": 4.7893, "step": 6484 }, { "epoch": 0.05853944755371006, "grad_norm": 2.9615581035614014, "learning_rate": 4.7073478967322626e-05, "loss": 5.1923, "step": 6485 }, { "epoch": 0.05854847445387254, "grad_norm": 2.5254287719726562, "learning_rate": 4.70730276223145e-05, "loss": 4.8981, "step": 6486 }, { "epoch": 0.05855750135403502, "grad_norm": 3.497598171234131, "learning_rate": 4.707257627730637e-05, "loss": 4.671, "step": 6487 }, { "epoch": 0.05856652825419751, "grad_norm": 3.1413912773132324, "learning_rate": 4.707212493229825e-05, "loss": 4.1312, "step": 6488 }, { "epoch": 0.05857555515435999, "grad_norm": 3.3642220497131348, "learning_rate": 4.707167358729013e-05, "loss": 4.7997, "step": 6489 }, { "epoch": 0.05858458205452248, "grad_norm": 2.8432703018188477, "learning_rate": 4.7071222242282e-05, "loss": 5.0555, "step": 6490 }, { "epoch": 0.05859360895468496, "grad_norm": 3.4262099266052246, "learning_rate": 4.707077089727388e-05, "loss": 5.1289, "step": 6491 }, { "epoch": 0.05860263585484744, "grad_norm": 3.3504929542541504, "learning_rate": 4.7070319552265754e-05, "loss": 4.8155, "step": 6492 }, { "epoch": 0.05861166275500993, "grad_norm": 4.635591506958008, "learning_rate": 4.7069868207257626e-05, "loss": 3.9924, "step": 6493 }, { "epoch": 0.05862068965517241, "grad_norm": 6.249157905578613, "learning_rate": 4.7069416862249505e-05, "loss": 4.6973, "step": 6494 }, { "epoch": 0.0586297165553349, "grad_norm": 3.101529836654663, "learning_rate": 4.7068965517241385e-05, "loss": 4.5314, "step": 6495 }, { "epoch": 0.05863874345549738, "grad_norm": 3.077880620956421, "learning_rate": 4.706851417223326e-05, "loss": 4.5794, "step": 6496 }, { "epoch": 0.058647770355659865, "grad_norm": 3.75472354888916, "learning_rate": 4.7068062827225136e-05, "loss": 4.6047, "step": 6497 }, { "epoch": 0.05865679725582235, "grad_norm": 3.348395347595215, "learning_rate": 4.706761148221701e-05, "loss": 4.9504, "step": 6498 }, { "epoch": 0.058665824155984835, "grad_norm": 3.0467312335968018, "learning_rate": 4.706716013720888e-05, "loss": 5.054, "step": 6499 }, { "epoch": 0.058674851056147316, "grad_norm": 3.568157196044922, "learning_rate": 4.706670879220076e-05, "loss": 4.2968, "step": 6500 }, { "epoch": 0.058683877956309805, "grad_norm": 3.3298697471618652, "learning_rate": 4.706625744719263e-05, "loss": 4.5974, "step": 6501 }, { "epoch": 0.058692904856472286, "grad_norm": 3.2103934288024902, "learning_rate": 4.706580610218451e-05, "loss": 5.0738, "step": 6502 }, { "epoch": 0.058701931756634775, "grad_norm": 4.430436134338379, "learning_rate": 4.706535475717639e-05, "loss": 5.2485, "step": 6503 }, { "epoch": 0.058710958656797256, "grad_norm": 2.6119120121002197, "learning_rate": 4.7064903412168264e-05, "loss": 4.4227, "step": 6504 }, { "epoch": 0.05871998555695974, "grad_norm": 2.8253934383392334, "learning_rate": 4.706445206716014e-05, "loss": 4.5561, "step": 6505 }, { "epoch": 0.058729012457122226, "grad_norm": 4.338256359100342, "learning_rate": 4.7064000722152016e-05, "loss": 3.9397, "step": 6506 }, { "epoch": 0.05873803935728471, "grad_norm": 3.0145206451416016, "learning_rate": 4.706354937714389e-05, "loss": 4.5771, "step": 6507 }, { "epoch": 0.058747066257447196, "grad_norm": 4.228024959564209, "learning_rate": 4.706309803213577e-05, "loss": 5.3297, "step": 6508 }, { "epoch": 0.05875609315760968, "grad_norm": 4.399771213531494, "learning_rate": 4.706264668712764e-05, "loss": 4.8036, "step": 6509 }, { "epoch": 0.05876512005777216, "grad_norm": 3.1941635608673096, "learning_rate": 4.706219534211952e-05, "loss": 5.2975, "step": 6510 }, { "epoch": 0.05877414695793465, "grad_norm": 2.876450300216675, "learning_rate": 4.70617439971114e-05, "loss": 4.411, "step": 6511 }, { "epoch": 0.05878317385809713, "grad_norm": 2.264091968536377, "learning_rate": 4.706129265210327e-05, "loss": 4.9703, "step": 6512 }, { "epoch": 0.05879220075825961, "grad_norm": 3.001321792602539, "learning_rate": 4.706084130709514e-05, "loss": 4.7944, "step": 6513 }, { "epoch": 0.0588012276584221, "grad_norm": 3.334681510925293, "learning_rate": 4.706038996208702e-05, "loss": 5.0977, "step": 6514 }, { "epoch": 0.05881025455858458, "grad_norm": 4.059973239898682, "learning_rate": 4.7059938617078895e-05, "loss": 4.8857, "step": 6515 }, { "epoch": 0.05881928145874707, "grad_norm": 3.508077621459961, "learning_rate": 4.7059487272070774e-05, "loss": 4.6091, "step": 6516 }, { "epoch": 0.05882830835890955, "grad_norm": 3.787297487258911, "learning_rate": 4.705903592706265e-05, "loss": 4.5139, "step": 6517 }, { "epoch": 0.05883733525907203, "grad_norm": 4.233112335205078, "learning_rate": 4.7058584582054526e-05, "loss": 4.9429, "step": 6518 }, { "epoch": 0.05884636215923452, "grad_norm": 4.276304244995117, "learning_rate": 4.70581332370464e-05, "loss": 4.3961, "step": 6519 }, { "epoch": 0.058855389059397, "grad_norm": 4.253217697143555, "learning_rate": 4.705768189203828e-05, "loss": 4.5781, "step": 6520 }, { "epoch": 0.05886441595955949, "grad_norm": 3.3942995071411133, "learning_rate": 4.705723054703015e-05, "loss": 4.6082, "step": 6521 }, { "epoch": 0.05887344285972197, "grad_norm": 4.132242679595947, "learning_rate": 4.705677920202203e-05, "loss": 5.079, "step": 6522 }, { "epoch": 0.058882469759884454, "grad_norm": 4.193413734436035, "learning_rate": 4.70563278570139e-05, "loss": 5.5498, "step": 6523 }, { "epoch": 0.05889149666004694, "grad_norm": 3.612151622772217, "learning_rate": 4.705587651200578e-05, "loss": 4.4613, "step": 6524 }, { "epoch": 0.058900523560209424, "grad_norm": 5.19556999206543, "learning_rate": 4.705542516699766e-05, "loss": 4.1199, "step": 6525 }, { "epoch": 0.058909550460371905, "grad_norm": 3.389911413192749, "learning_rate": 4.7054973821989526e-05, "loss": 4.4857, "step": 6526 }, { "epoch": 0.058918577360534394, "grad_norm": 4.228262424468994, "learning_rate": 4.7054522476981405e-05, "loss": 5.005, "step": 6527 }, { "epoch": 0.058927604260696875, "grad_norm": 3.491006374359131, "learning_rate": 4.7054071131973284e-05, "loss": 4.7792, "step": 6528 }, { "epoch": 0.058936631160859364, "grad_norm": 3.0533812046051025, "learning_rate": 4.705361978696516e-05, "loss": 5.1195, "step": 6529 }, { "epoch": 0.058945658061021845, "grad_norm": 3.7608869075775146, "learning_rate": 4.7053168441957036e-05, "loss": 4.2951, "step": 6530 }, { "epoch": 0.05895468496118433, "grad_norm": 3.2819244861602783, "learning_rate": 4.705271709694891e-05, "loss": 4.613, "step": 6531 }, { "epoch": 0.058963711861346815, "grad_norm": 3.1773321628570557, "learning_rate": 4.705226575194079e-05, "loss": 4.3581, "step": 6532 }, { "epoch": 0.0589727387615093, "grad_norm": 3.6264424324035645, "learning_rate": 4.705181440693266e-05, "loss": 5.137, "step": 6533 }, { "epoch": 0.058981765661671785, "grad_norm": 3.7532241344451904, "learning_rate": 4.705136306192454e-05, "loss": 4.2296, "step": 6534 }, { "epoch": 0.05899079256183427, "grad_norm": 2.7104036808013916, "learning_rate": 4.705091171691641e-05, "loss": 3.7898, "step": 6535 }, { "epoch": 0.05899981946199675, "grad_norm": 4.756623268127441, "learning_rate": 4.705046037190829e-05, "loss": 5.5679, "step": 6536 }, { "epoch": 0.05900884636215924, "grad_norm": 3.454115867614746, "learning_rate": 4.7050009026900164e-05, "loss": 4.8699, "step": 6537 }, { "epoch": 0.05901787326232172, "grad_norm": 3.300593614578247, "learning_rate": 4.704955768189204e-05, "loss": 4.6703, "step": 6538 }, { "epoch": 0.0590269001624842, "grad_norm": 3.93403697013855, "learning_rate": 4.7049106336883915e-05, "loss": 4.7379, "step": 6539 }, { "epoch": 0.05903592706264669, "grad_norm": 3.251718044281006, "learning_rate": 4.704865499187579e-05, "loss": 4.1781, "step": 6540 }, { "epoch": 0.05904495396280917, "grad_norm": 4.861167907714844, "learning_rate": 4.704820364686767e-05, "loss": 4.0911, "step": 6541 }, { "epoch": 0.05905398086297166, "grad_norm": 3.416719913482666, "learning_rate": 4.7047752301859546e-05, "loss": 4.6346, "step": 6542 }, { "epoch": 0.05906300776313414, "grad_norm": 2.9872071743011475, "learning_rate": 4.704730095685142e-05, "loss": 4.9208, "step": 6543 }, { "epoch": 0.05907203466329662, "grad_norm": 2.9588005542755127, "learning_rate": 4.70468496118433e-05, "loss": 4.8511, "step": 6544 }, { "epoch": 0.05908106156345911, "grad_norm": 2.9845049381256104, "learning_rate": 4.704639826683517e-05, "loss": 4.9524, "step": 6545 }, { "epoch": 0.05909008846362159, "grad_norm": 2.8687546253204346, "learning_rate": 4.704594692182704e-05, "loss": 4.9415, "step": 6546 }, { "epoch": 0.05909911536378408, "grad_norm": 3.4795072078704834, "learning_rate": 4.704549557681892e-05, "loss": 4.7319, "step": 6547 }, { "epoch": 0.05910814226394656, "grad_norm": 3.2101197242736816, "learning_rate": 4.7045044231810795e-05, "loss": 5.0071, "step": 6548 }, { "epoch": 0.05911716916410904, "grad_norm": 3.992906093597412, "learning_rate": 4.7044592886802674e-05, "loss": 4.7738, "step": 6549 }, { "epoch": 0.05912619606427153, "grad_norm": 3.902296304702759, "learning_rate": 4.704414154179455e-05, "loss": 4.8325, "step": 6550 }, { "epoch": 0.05913522296443401, "grad_norm": 3.2591805458068848, "learning_rate": 4.7043690196786426e-05, "loss": 4.533, "step": 6551 }, { "epoch": 0.059144249864596495, "grad_norm": 3.235778570175171, "learning_rate": 4.7043238851778305e-05, "loss": 4.9308, "step": 6552 }, { "epoch": 0.05915327676475898, "grad_norm": 3.006363868713379, "learning_rate": 4.704278750677018e-05, "loss": 4.9605, "step": 6553 }, { "epoch": 0.059162303664921465, "grad_norm": 3.25514817237854, "learning_rate": 4.704233616176205e-05, "loss": 3.9874, "step": 6554 }, { "epoch": 0.05917133056508395, "grad_norm": 2.8009960651397705, "learning_rate": 4.704188481675393e-05, "loss": 5.1202, "step": 6555 }, { "epoch": 0.059180357465246435, "grad_norm": 4.62996244430542, "learning_rate": 4.704143347174581e-05, "loss": 4.0384, "step": 6556 }, { "epoch": 0.059189384365408916, "grad_norm": 6.034083843231201, "learning_rate": 4.704098212673768e-05, "loss": 3.1878, "step": 6557 }, { "epoch": 0.059198411265571405, "grad_norm": 3.161693572998047, "learning_rate": 4.704053078172956e-05, "loss": 4.3267, "step": 6558 }, { "epoch": 0.059207438165733886, "grad_norm": 3.8204736709594727, "learning_rate": 4.704007943672143e-05, "loss": 4.1598, "step": 6559 }, { "epoch": 0.059216465065896375, "grad_norm": 4.62429141998291, "learning_rate": 4.7039628091713305e-05, "loss": 4.1364, "step": 6560 }, { "epoch": 0.059225491966058856, "grad_norm": 5.569972515106201, "learning_rate": 4.7039176746705184e-05, "loss": 4.6601, "step": 6561 }, { "epoch": 0.05923451886622134, "grad_norm": 4.25709867477417, "learning_rate": 4.7038725401697057e-05, "loss": 4.8973, "step": 6562 }, { "epoch": 0.059243545766383826, "grad_norm": 2.7749316692352295, "learning_rate": 4.7038274056688936e-05, "loss": 5.0454, "step": 6563 }, { "epoch": 0.05925257266654631, "grad_norm": 3.534383535385132, "learning_rate": 4.7037822711680815e-05, "loss": 4.9759, "step": 6564 }, { "epoch": 0.05926159956670879, "grad_norm": 2.4657721519470215, "learning_rate": 4.703737136667269e-05, "loss": 4.547, "step": 6565 }, { "epoch": 0.05927062646687128, "grad_norm": 3.3643527030944824, "learning_rate": 4.703692002166456e-05, "loss": 4.7505, "step": 6566 }, { "epoch": 0.05927965336703376, "grad_norm": 3.008723497390747, "learning_rate": 4.703646867665644e-05, "loss": 4.4193, "step": 6567 }, { "epoch": 0.05928868026719625, "grad_norm": 3.8634064197540283, "learning_rate": 4.703601733164831e-05, "loss": 4.114, "step": 6568 }, { "epoch": 0.05929770716735873, "grad_norm": 3.5549428462982178, "learning_rate": 4.703556598664019e-05, "loss": 3.4017, "step": 6569 }, { "epoch": 0.05930673406752121, "grad_norm": 3.6553428173065186, "learning_rate": 4.703511464163206e-05, "loss": 4.5501, "step": 6570 }, { "epoch": 0.0593157609676837, "grad_norm": 8.8362455368042, "learning_rate": 4.703466329662394e-05, "loss": 4.7037, "step": 6571 }, { "epoch": 0.05932478786784618, "grad_norm": 3.887830972671509, "learning_rate": 4.703421195161582e-05, "loss": 3.976, "step": 6572 }, { "epoch": 0.05933381476800867, "grad_norm": 4.455444812774658, "learning_rate": 4.703376060660769e-05, "loss": 4.7183, "step": 6573 }, { "epoch": 0.05934284166817115, "grad_norm": 3.8675360679626465, "learning_rate": 4.703330926159957e-05, "loss": 3.9411, "step": 6574 }, { "epoch": 0.05935186856833363, "grad_norm": 3.8500823974609375, "learning_rate": 4.7032857916591446e-05, "loss": 4.6685, "step": 6575 }, { "epoch": 0.05936089546849612, "grad_norm": 3.25907564163208, "learning_rate": 4.703240657158332e-05, "loss": 4.1571, "step": 6576 }, { "epoch": 0.0593699223686586, "grad_norm": 3.2689266204833984, "learning_rate": 4.70319552265752e-05, "loss": 4.0166, "step": 6577 }, { "epoch": 0.059378949268821084, "grad_norm": 4.135339260101318, "learning_rate": 4.703150388156708e-05, "loss": 4.6067, "step": 6578 }, { "epoch": 0.05938797616898357, "grad_norm": 4.098046779632568, "learning_rate": 4.703105253655895e-05, "loss": 4.3068, "step": 6579 }, { "epoch": 0.059397003069146054, "grad_norm": 4.0612897872924805, "learning_rate": 4.703060119155082e-05, "loss": 4.6856, "step": 6580 }, { "epoch": 0.05940602996930854, "grad_norm": 2.6508772373199463, "learning_rate": 4.70301498465427e-05, "loss": 4.8493, "step": 6581 }, { "epoch": 0.059415056869471024, "grad_norm": 2.630941867828369, "learning_rate": 4.7029698501534574e-05, "loss": 4.9358, "step": 6582 }, { "epoch": 0.059424083769633505, "grad_norm": 4.634462833404541, "learning_rate": 4.702924715652645e-05, "loss": 4.9391, "step": 6583 }, { "epoch": 0.059433110669795994, "grad_norm": 3.1530303955078125, "learning_rate": 4.7028795811518325e-05, "loss": 5.0036, "step": 6584 }, { "epoch": 0.059442137569958475, "grad_norm": 3.247039794921875, "learning_rate": 4.7028344466510205e-05, "loss": 4.9759, "step": 6585 }, { "epoch": 0.059451164470120964, "grad_norm": 3.9777212142944336, "learning_rate": 4.702789312150208e-05, "loss": 4.5915, "step": 6586 }, { "epoch": 0.059460191370283445, "grad_norm": 3.7012202739715576, "learning_rate": 4.702744177649395e-05, "loss": 4.3167, "step": 6587 }, { "epoch": 0.05946921827044593, "grad_norm": 3.325385808944702, "learning_rate": 4.702699043148583e-05, "loss": 5.0733, "step": 6588 }, { "epoch": 0.059478245170608415, "grad_norm": 3.0901641845703125, "learning_rate": 4.702653908647771e-05, "loss": 4.7525, "step": 6589 }, { "epoch": 0.0594872720707709, "grad_norm": 3.5475831031799316, "learning_rate": 4.702608774146958e-05, "loss": 4.933, "step": 6590 }, { "epoch": 0.05949629897093338, "grad_norm": 3.914700746536255, "learning_rate": 4.702563639646146e-05, "loss": 4.5544, "step": 6591 }, { "epoch": 0.05950532587109587, "grad_norm": 3.237689971923828, "learning_rate": 4.702518505145333e-05, "loss": 5.2862, "step": 6592 }, { "epoch": 0.05951435277125835, "grad_norm": 3.900207281112671, "learning_rate": 4.7024733706445205e-05, "loss": 4.8491, "step": 6593 }, { "epoch": 0.05952337967142084, "grad_norm": 3.992142915725708, "learning_rate": 4.7024282361437084e-05, "loss": 4.3617, "step": 6594 }, { "epoch": 0.05953240657158332, "grad_norm": 3.5270631313323975, "learning_rate": 4.702383101642896e-05, "loss": 4.5497, "step": 6595 }, { "epoch": 0.0595414334717458, "grad_norm": 4.3769354820251465, "learning_rate": 4.7023379671420836e-05, "loss": 4.3138, "step": 6596 }, { "epoch": 0.05955046037190829, "grad_norm": 2.9466636180877686, "learning_rate": 4.7022928326412715e-05, "loss": 4.7, "step": 6597 }, { "epoch": 0.05955948727207077, "grad_norm": 3.893280267715454, "learning_rate": 4.702247698140459e-05, "loss": 4.6721, "step": 6598 }, { "epoch": 0.05956851417223326, "grad_norm": 3.4964969158172607, "learning_rate": 4.7022025636396466e-05, "loss": 4.7022, "step": 6599 }, { "epoch": 0.05957754107239574, "grad_norm": 4.312137126922607, "learning_rate": 4.702157429138834e-05, "loss": 3.7523, "step": 6600 }, { "epoch": 0.05958656797255822, "grad_norm": 4.201840400695801, "learning_rate": 4.702112294638021e-05, "loss": 4.3097, "step": 6601 }, { "epoch": 0.05959559487272071, "grad_norm": 2.738265037536621, "learning_rate": 4.702067160137209e-05, "loss": 5.1299, "step": 6602 }, { "epoch": 0.05960462177288319, "grad_norm": 4.297179698944092, "learning_rate": 4.702022025636397e-05, "loss": 4.3762, "step": 6603 }, { "epoch": 0.05961364867304567, "grad_norm": 2.935858726501465, "learning_rate": 4.701976891135584e-05, "loss": 4.829, "step": 6604 }, { "epoch": 0.05962267557320816, "grad_norm": 3.508833885192871, "learning_rate": 4.701931756634772e-05, "loss": 4.7399, "step": 6605 }, { "epoch": 0.05963170247337064, "grad_norm": 4.68440055847168, "learning_rate": 4.7018866221339594e-05, "loss": 4.5816, "step": 6606 }, { "epoch": 0.05964072937353313, "grad_norm": 3.974325180053711, "learning_rate": 4.7018414876331467e-05, "loss": 5.2779, "step": 6607 }, { "epoch": 0.05964975627369561, "grad_norm": 3.3377323150634766, "learning_rate": 4.7017963531323346e-05, "loss": 4.8207, "step": 6608 }, { "epoch": 0.059658783173858095, "grad_norm": 3.224130868911743, "learning_rate": 4.701751218631522e-05, "loss": 4.9321, "step": 6609 }, { "epoch": 0.05966781007402058, "grad_norm": 3.7462995052337646, "learning_rate": 4.70170608413071e-05, "loss": 4.8255, "step": 6610 }, { "epoch": 0.059676836974183065, "grad_norm": 3.74760103225708, "learning_rate": 4.701660949629898e-05, "loss": 4.7468, "step": 6611 }, { "epoch": 0.05968586387434555, "grad_norm": 3.6333703994750977, "learning_rate": 4.701615815129085e-05, "loss": 4.9439, "step": 6612 }, { "epoch": 0.059694890774508035, "grad_norm": 2.9573071002960205, "learning_rate": 4.701570680628272e-05, "loss": 4.6462, "step": 6613 }, { "epoch": 0.059703917674670516, "grad_norm": 3.2333943843841553, "learning_rate": 4.70152554612746e-05, "loss": 4.5493, "step": 6614 }, { "epoch": 0.059712944574833005, "grad_norm": 2.8104774951934814, "learning_rate": 4.701480411626647e-05, "loss": 3.9503, "step": 6615 }, { "epoch": 0.059721971474995486, "grad_norm": 3.376915216445923, "learning_rate": 4.701435277125835e-05, "loss": 4.2442, "step": 6616 }, { "epoch": 0.05973099837515797, "grad_norm": 3.6977806091308594, "learning_rate": 4.701390142625023e-05, "loss": 4.8443, "step": 6617 }, { "epoch": 0.059740025275320456, "grad_norm": 3.733571767807007, "learning_rate": 4.7013450081242104e-05, "loss": 3.7376, "step": 6618 }, { "epoch": 0.05974905217548294, "grad_norm": 4.927757740020752, "learning_rate": 4.7012998736233984e-05, "loss": 3.54, "step": 6619 }, { "epoch": 0.059758079075645426, "grad_norm": 2.875370502471924, "learning_rate": 4.7012547391225856e-05, "loss": 4.7512, "step": 6620 }, { "epoch": 0.05976710597580791, "grad_norm": 3.396104335784912, "learning_rate": 4.701209604621773e-05, "loss": 4.3034, "step": 6621 }, { "epoch": 0.05977613287597039, "grad_norm": 4.911276817321777, "learning_rate": 4.701164470120961e-05, "loss": 4.8627, "step": 6622 }, { "epoch": 0.05978515977613288, "grad_norm": 3.0999972820281982, "learning_rate": 4.701119335620148e-05, "loss": 5.5161, "step": 6623 }, { "epoch": 0.05979418667629536, "grad_norm": 3.9064202308654785, "learning_rate": 4.701074201119336e-05, "loss": 5.3645, "step": 6624 }, { "epoch": 0.05980321357645785, "grad_norm": 3.775470495223999, "learning_rate": 4.701029066618524e-05, "loss": 5.2218, "step": 6625 }, { "epoch": 0.05981224047662033, "grad_norm": 3.4595282077789307, "learning_rate": 4.700983932117711e-05, "loss": 4.4149, "step": 6626 }, { "epoch": 0.05982126737678281, "grad_norm": 3.396811008453369, "learning_rate": 4.7009387976168984e-05, "loss": 4.6951, "step": 6627 }, { "epoch": 0.0598302942769453, "grad_norm": 3.925663709640503, "learning_rate": 4.700893663116086e-05, "loss": 4.5048, "step": 6628 }, { "epoch": 0.05983932117710778, "grad_norm": 3.8184704780578613, "learning_rate": 4.7008485286152735e-05, "loss": 4.2601, "step": 6629 }, { "epoch": 0.05984834807727026, "grad_norm": 4.630029678344727, "learning_rate": 4.7008033941144615e-05, "loss": 4.8231, "step": 6630 }, { "epoch": 0.05985737497743275, "grad_norm": 3.451819658279419, "learning_rate": 4.700758259613649e-05, "loss": 4.7501, "step": 6631 }, { "epoch": 0.05986640187759523, "grad_norm": 3.8553378582000732, "learning_rate": 4.7007131251128366e-05, "loss": 4.693, "step": 6632 }, { "epoch": 0.05987542877775772, "grad_norm": 4.276189804077148, "learning_rate": 4.700667990612024e-05, "loss": 5.0561, "step": 6633 }, { "epoch": 0.0598844556779202, "grad_norm": 3.8061962127685547, "learning_rate": 4.700622856111211e-05, "loss": 5.0417, "step": 6634 }, { "epoch": 0.059893482578082684, "grad_norm": 2.88122820854187, "learning_rate": 4.700577721610399e-05, "loss": 5.0453, "step": 6635 }, { "epoch": 0.05990250947824517, "grad_norm": 4.130718231201172, "learning_rate": 4.700532587109587e-05, "loss": 4.3533, "step": 6636 }, { "epoch": 0.059911536378407654, "grad_norm": 3.313340663909912, "learning_rate": 4.700487452608774e-05, "loss": 5.016, "step": 6637 }, { "epoch": 0.05992056327857014, "grad_norm": 3.1600348949432373, "learning_rate": 4.700442318107962e-05, "loss": 4.9367, "step": 6638 }, { "epoch": 0.059929590178732624, "grad_norm": 3.730065107345581, "learning_rate": 4.70039718360715e-05, "loss": 4.7212, "step": 6639 }, { "epoch": 0.059938617078895105, "grad_norm": 3.462066888809204, "learning_rate": 4.7003520491063366e-05, "loss": 4.4499, "step": 6640 }, { "epoch": 0.059947643979057594, "grad_norm": 3.5277822017669678, "learning_rate": 4.7003069146055246e-05, "loss": 4.3601, "step": 6641 }, { "epoch": 0.059956670879220075, "grad_norm": 3.5614571571350098, "learning_rate": 4.7002617801047125e-05, "loss": 4.836, "step": 6642 }, { "epoch": 0.05996569777938256, "grad_norm": 4.359102249145508, "learning_rate": 4.7002166456039e-05, "loss": 5.1286, "step": 6643 }, { "epoch": 0.059974724679545045, "grad_norm": 3.761045217514038, "learning_rate": 4.7001715111030876e-05, "loss": 4.6347, "step": 6644 }, { "epoch": 0.05998375157970753, "grad_norm": 3.2871131896972656, "learning_rate": 4.700126376602275e-05, "loss": 4.4747, "step": 6645 }, { "epoch": 0.059992778479870015, "grad_norm": 4.177505016326904, "learning_rate": 4.700081242101463e-05, "loss": 4.6937, "step": 6646 }, { "epoch": 0.0600018053800325, "grad_norm": 4.3289666175842285, "learning_rate": 4.70003610760065e-05, "loss": 4.615, "step": 6647 }, { "epoch": 0.06001083228019498, "grad_norm": 3.6117873191833496, "learning_rate": 4.699990973099837e-05, "loss": 4.7103, "step": 6648 }, { "epoch": 0.06001985918035747, "grad_norm": 3.4072232246398926, "learning_rate": 4.699945838599025e-05, "loss": 4.9363, "step": 6649 }, { "epoch": 0.06002888608051995, "grad_norm": 3.517713785171509, "learning_rate": 4.699900704098213e-05, "loss": 5.0052, "step": 6650 }, { "epoch": 0.06003791298068244, "grad_norm": 4.881925106048584, "learning_rate": 4.6998555695974004e-05, "loss": 4.7129, "step": 6651 }, { "epoch": 0.06004693988084492, "grad_norm": 3.8023135662078857, "learning_rate": 4.699810435096588e-05, "loss": 4.6161, "step": 6652 }, { "epoch": 0.0600559667810074, "grad_norm": 7.538517951965332, "learning_rate": 4.6997653005957756e-05, "loss": 4.8611, "step": 6653 }, { "epoch": 0.06006499368116989, "grad_norm": 4.833907127380371, "learning_rate": 4.699720166094963e-05, "loss": 4.6103, "step": 6654 }, { "epoch": 0.06007402058133237, "grad_norm": 6.136104106903076, "learning_rate": 4.699675031594151e-05, "loss": 4.8952, "step": 6655 }, { "epoch": 0.06008304748149485, "grad_norm": 3.512786626815796, "learning_rate": 4.699629897093339e-05, "loss": 4.7352, "step": 6656 }, { "epoch": 0.06009207438165734, "grad_norm": 3.971506118774414, "learning_rate": 4.699584762592526e-05, "loss": 3.934, "step": 6657 }, { "epoch": 0.06010110128181982, "grad_norm": 3.6230359077453613, "learning_rate": 4.699539628091714e-05, "loss": 3.8038, "step": 6658 }, { "epoch": 0.06011012818198231, "grad_norm": 3.3393516540527344, "learning_rate": 4.699494493590901e-05, "loss": 4.2087, "step": 6659 }, { "epoch": 0.06011915508214479, "grad_norm": 3.909679889678955, "learning_rate": 4.699449359090088e-05, "loss": 4.7491, "step": 6660 }, { "epoch": 0.06012818198230727, "grad_norm": 3.5932278633117676, "learning_rate": 4.699404224589276e-05, "loss": 4.7779, "step": 6661 }, { "epoch": 0.06013720888246976, "grad_norm": 4.132503986358643, "learning_rate": 4.6993590900884635e-05, "loss": 4.7543, "step": 6662 }, { "epoch": 0.06014623578263224, "grad_norm": 4.380534648895264, "learning_rate": 4.6993139555876514e-05, "loss": 4.6386, "step": 6663 }, { "epoch": 0.06015526268279473, "grad_norm": 3.348998546600342, "learning_rate": 4.6992688210868394e-05, "loss": 4.5263, "step": 6664 }, { "epoch": 0.06016428958295721, "grad_norm": 2.951011896133423, "learning_rate": 4.6992236865860266e-05, "loss": 5.3523, "step": 6665 }, { "epoch": 0.060173316483119695, "grad_norm": 2.852586269378662, "learning_rate": 4.6991785520852145e-05, "loss": 4.4756, "step": 6666 }, { "epoch": 0.06018234338328218, "grad_norm": 3.311051368713379, "learning_rate": 4.699133417584402e-05, "loss": 4.4118, "step": 6667 }, { "epoch": 0.060191370283444665, "grad_norm": 2.617917537689209, "learning_rate": 4.699088283083589e-05, "loss": 5.246, "step": 6668 }, { "epoch": 0.060200397183607146, "grad_norm": 3.0390491485595703, "learning_rate": 4.699043148582777e-05, "loss": 4.7675, "step": 6669 }, { "epoch": 0.060209424083769635, "grad_norm": 3.899905204772949, "learning_rate": 4.698998014081964e-05, "loss": 4.2824, "step": 6670 }, { "epoch": 0.060218450983932116, "grad_norm": 3.281874418258667, "learning_rate": 4.698952879581152e-05, "loss": 4.6062, "step": 6671 }, { "epoch": 0.060227477884094605, "grad_norm": 4.351020336151123, "learning_rate": 4.69890774508034e-05, "loss": 4.872, "step": 6672 }, { "epoch": 0.060236504784257086, "grad_norm": 4.0447678565979, "learning_rate": 4.698862610579527e-05, "loss": 4.2911, "step": 6673 }, { "epoch": 0.06024553168441957, "grad_norm": 2.7678518295288086, "learning_rate": 4.6988174760787145e-05, "loss": 4.6941, "step": 6674 }, { "epoch": 0.060254558584582056, "grad_norm": 2.8946712017059326, "learning_rate": 4.6987723415779025e-05, "loss": 4.4864, "step": 6675 }, { "epoch": 0.06026358548474454, "grad_norm": 3.491353750228882, "learning_rate": 4.69872720707709e-05, "loss": 4.3988, "step": 6676 }, { "epoch": 0.060272612384907026, "grad_norm": 3.633157968521118, "learning_rate": 4.6986820725762776e-05, "loss": 5.1512, "step": 6677 }, { "epoch": 0.06028163928506951, "grad_norm": 3.084839344024658, "learning_rate": 4.6986369380754655e-05, "loss": 5.0188, "step": 6678 }, { "epoch": 0.06029066618523199, "grad_norm": 3.364701509475708, "learning_rate": 4.698591803574653e-05, "loss": 4.4217, "step": 6679 }, { "epoch": 0.06029969308539448, "grad_norm": 3.901160955429077, "learning_rate": 4.69854666907384e-05, "loss": 4.3737, "step": 6680 }, { "epoch": 0.06030871998555696, "grad_norm": 3.9353439807891846, "learning_rate": 4.698501534573028e-05, "loss": 4.3925, "step": 6681 }, { "epoch": 0.06031774688571944, "grad_norm": 3.714836835861206, "learning_rate": 4.698456400072215e-05, "loss": 4.334, "step": 6682 }, { "epoch": 0.06032677378588193, "grad_norm": 3.596566915512085, "learning_rate": 4.698411265571403e-05, "loss": 4.5623, "step": 6683 }, { "epoch": 0.06033580068604441, "grad_norm": 3.878495454788208, "learning_rate": 4.6983661310705904e-05, "loss": 4.573, "step": 6684 }, { "epoch": 0.0603448275862069, "grad_norm": 3.5782217979431152, "learning_rate": 4.698320996569778e-05, "loss": 4.6936, "step": 6685 }, { "epoch": 0.06035385448636938, "grad_norm": 3.5528368949890137, "learning_rate": 4.698275862068966e-05, "loss": 4.8589, "step": 6686 }, { "epoch": 0.06036288138653186, "grad_norm": 3.0999374389648438, "learning_rate": 4.698230727568153e-05, "loss": 4.1213, "step": 6687 }, { "epoch": 0.06037190828669435, "grad_norm": 3.222841262817383, "learning_rate": 4.698185593067341e-05, "loss": 4.584, "step": 6688 }, { "epoch": 0.06038093518685683, "grad_norm": 2.4214015007019043, "learning_rate": 4.6981404585665286e-05, "loss": 4.8746, "step": 6689 }, { "epoch": 0.06038996208701932, "grad_norm": 3.4425337314605713, "learning_rate": 4.698095324065716e-05, "loss": 5.0262, "step": 6690 }, { "epoch": 0.0603989889871818, "grad_norm": 3.1698060035705566, "learning_rate": 4.698050189564904e-05, "loss": 4.8943, "step": 6691 }, { "epoch": 0.060408015887344284, "grad_norm": 2.926945686340332, "learning_rate": 4.698005055064091e-05, "loss": 4.9888, "step": 6692 }, { "epoch": 0.06041704278750677, "grad_norm": 2.6721620559692383, "learning_rate": 4.697959920563279e-05, "loss": 4.8852, "step": 6693 }, { "epoch": 0.060426069687669254, "grad_norm": 2.9492006301879883, "learning_rate": 4.697914786062466e-05, "loss": 3.6342, "step": 6694 }, { "epoch": 0.060435096587831735, "grad_norm": 3.6778085231781006, "learning_rate": 4.697869651561654e-05, "loss": 4.4573, "step": 6695 }, { "epoch": 0.060444123487994224, "grad_norm": 3.390319585800171, "learning_rate": 4.6978245170608414e-05, "loss": 4.7475, "step": 6696 }, { "epoch": 0.060453150388156705, "grad_norm": 3.4254298210144043, "learning_rate": 4.697779382560029e-05, "loss": 4.5635, "step": 6697 }, { "epoch": 0.060462177288319194, "grad_norm": 3.5867090225219727, "learning_rate": 4.6977342480592166e-05, "loss": 4.0576, "step": 6698 }, { "epoch": 0.060471204188481675, "grad_norm": 3.1631650924682617, "learning_rate": 4.6976891135584045e-05, "loss": 4.7047, "step": 6699 }, { "epoch": 0.06048023108864416, "grad_norm": 3.082411050796509, "learning_rate": 4.697643979057592e-05, "loss": 5.704, "step": 6700 }, { "epoch": 0.060489257988806645, "grad_norm": 3.110832452774048, "learning_rate": 4.697598844556779e-05, "loss": 4.6827, "step": 6701 }, { "epoch": 0.06049828488896913, "grad_norm": 3.0582821369171143, "learning_rate": 4.697553710055967e-05, "loss": 4.779, "step": 6702 }, { "epoch": 0.060507311789131615, "grad_norm": 3.9845640659332275, "learning_rate": 4.697508575555155e-05, "loss": 4.6838, "step": 6703 }, { "epoch": 0.0605163386892941, "grad_norm": 2.907881021499634, "learning_rate": 4.697463441054342e-05, "loss": 5.1407, "step": 6704 }, { "epoch": 0.06052536558945658, "grad_norm": 2.3470699787139893, "learning_rate": 4.69741830655353e-05, "loss": 4.7571, "step": 6705 }, { "epoch": 0.06053439248961907, "grad_norm": 6.748230934143066, "learning_rate": 4.697373172052717e-05, "loss": 5.2348, "step": 6706 }, { "epoch": 0.06054341938978155, "grad_norm": 3.263119697570801, "learning_rate": 4.6973280375519045e-05, "loss": 4.3836, "step": 6707 }, { "epoch": 0.06055244628994403, "grad_norm": 3.6569273471832275, "learning_rate": 4.6972829030510924e-05, "loss": 4.6541, "step": 6708 }, { "epoch": 0.06056147319010652, "grad_norm": 5.563722610473633, "learning_rate": 4.69723776855028e-05, "loss": 4.6129, "step": 6709 }, { "epoch": 0.060570500090269, "grad_norm": 3.3922183513641357, "learning_rate": 4.6971926340494676e-05, "loss": 4.4943, "step": 6710 }, { "epoch": 0.06057952699043149, "grad_norm": 2.6296207904815674, "learning_rate": 4.6971474995486555e-05, "loss": 4.7249, "step": 6711 }, { "epoch": 0.06058855389059397, "grad_norm": 4.563413619995117, "learning_rate": 4.697102365047843e-05, "loss": 4.2712, "step": 6712 }, { "epoch": 0.06059758079075645, "grad_norm": 2.9299368858337402, "learning_rate": 4.697057230547031e-05, "loss": 5.1256, "step": 6713 }, { "epoch": 0.06060660769091894, "grad_norm": 3.3787407875061035, "learning_rate": 4.697012096046218e-05, "loss": 4.3725, "step": 6714 }, { "epoch": 0.06061563459108142, "grad_norm": 3.670048236846924, "learning_rate": 4.696966961545405e-05, "loss": 4.7307, "step": 6715 }, { "epoch": 0.06062466149124391, "grad_norm": 5.405137538909912, "learning_rate": 4.696921827044593e-05, "loss": 4.7812, "step": 6716 }, { "epoch": 0.06063368839140639, "grad_norm": 4.940077781677246, "learning_rate": 4.696876692543781e-05, "loss": 4.6175, "step": 6717 }, { "epoch": 0.06064271529156887, "grad_norm": 3.9963743686676025, "learning_rate": 4.696831558042968e-05, "loss": 4.6343, "step": 6718 }, { "epoch": 0.06065174219173136, "grad_norm": 6.533657550811768, "learning_rate": 4.696786423542156e-05, "loss": 5.1143, "step": 6719 }, { "epoch": 0.06066076909189384, "grad_norm": 2.7104997634887695, "learning_rate": 4.6967412890413434e-05, "loss": 5.0737, "step": 6720 }, { "epoch": 0.060669795992056325, "grad_norm": 5.557425022125244, "learning_rate": 4.696696154540531e-05, "loss": 4.6017, "step": 6721 }, { "epoch": 0.06067882289221881, "grad_norm": 3.742349863052368, "learning_rate": 4.6966510200397186e-05, "loss": 4.3946, "step": 6722 }, { "epoch": 0.060687849792381295, "grad_norm": 3.5630886554718018, "learning_rate": 4.696605885538906e-05, "loss": 4.3673, "step": 6723 }, { "epoch": 0.06069687669254378, "grad_norm": 5.24250602722168, "learning_rate": 4.696560751038094e-05, "loss": 4.7505, "step": 6724 }, { "epoch": 0.060705903592706265, "grad_norm": 4.531944274902344, "learning_rate": 4.696515616537282e-05, "loss": 3.5244, "step": 6725 }, { "epoch": 0.060714930492868746, "grad_norm": 3.3970179557800293, "learning_rate": 4.696470482036469e-05, "loss": 5.332, "step": 6726 }, { "epoch": 0.060723957393031235, "grad_norm": 4.695334434509277, "learning_rate": 4.696425347535656e-05, "loss": 5.1854, "step": 6727 }, { "epoch": 0.060732984293193716, "grad_norm": 3.3724188804626465, "learning_rate": 4.696380213034844e-05, "loss": 5.19, "step": 6728 }, { "epoch": 0.060742011193356205, "grad_norm": 2.86102294921875, "learning_rate": 4.6963350785340314e-05, "loss": 4.9242, "step": 6729 }, { "epoch": 0.060751038093518686, "grad_norm": 3.5658791065216064, "learning_rate": 4.696289944033219e-05, "loss": 4.9277, "step": 6730 }, { "epoch": 0.06076006499368117, "grad_norm": 3.1909494400024414, "learning_rate": 4.6962448095324065e-05, "loss": 4.7714, "step": 6731 }, { "epoch": 0.060769091893843656, "grad_norm": 3.438453197479248, "learning_rate": 4.6961996750315945e-05, "loss": 3.7875, "step": 6732 }, { "epoch": 0.06077811879400614, "grad_norm": 3.59157395362854, "learning_rate": 4.6961545405307824e-05, "loss": 4.5957, "step": 6733 }, { "epoch": 0.06078714569416862, "grad_norm": 3.844794750213623, "learning_rate": 4.696109406029969e-05, "loss": 5.2577, "step": 6734 }, { "epoch": 0.06079617259433111, "grad_norm": 4.9194769859313965, "learning_rate": 4.696064271529157e-05, "loss": 4.5972, "step": 6735 }, { "epoch": 0.06080519949449359, "grad_norm": 2.686781883239746, "learning_rate": 4.696019137028345e-05, "loss": 4.8649, "step": 6736 }, { "epoch": 0.06081422639465608, "grad_norm": 3.750300645828247, "learning_rate": 4.695974002527532e-05, "loss": 3.4529, "step": 6737 }, { "epoch": 0.06082325329481856, "grad_norm": 4.140266418457031, "learning_rate": 4.69592886802672e-05, "loss": 4.5547, "step": 6738 }, { "epoch": 0.06083228019498104, "grad_norm": 3.4766547679901123, "learning_rate": 4.695883733525908e-05, "loss": 4.714, "step": 6739 }, { "epoch": 0.06084130709514353, "grad_norm": 4.6088128089904785, "learning_rate": 4.695838599025095e-05, "loss": 5.3401, "step": 6740 }, { "epoch": 0.06085033399530601, "grad_norm": 2.625774621963501, "learning_rate": 4.6957934645242824e-05, "loss": 5.2707, "step": 6741 }, { "epoch": 0.0608593608954685, "grad_norm": 4.092885494232178, "learning_rate": 4.69574833002347e-05, "loss": 4.6542, "step": 6742 }, { "epoch": 0.06086838779563098, "grad_norm": 3.429698944091797, "learning_rate": 4.6957031955226576e-05, "loss": 4.6424, "step": 6743 }, { "epoch": 0.06087741469579346, "grad_norm": 3.5770394802093506, "learning_rate": 4.6956580610218455e-05, "loss": 4.6276, "step": 6744 }, { "epoch": 0.06088644159595595, "grad_norm": 3.1236894130706787, "learning_rate": 4.695612926521033e-05, "loss": 5.3945, "step": 6745 }, { "epoch": 0.06089546849611843, "grad_norm": 4.73247766494751, "learning_rate": 4.695567792020221e-05, "loss": 4.8043, "step": 6746 }, { "epoch": 0.060904495396280914, "grad_norm": 2.4791884422302246, "learning_rate": 4.695522657519408e-05, "loss": 5.1243, "step": 6747 }, { "epoch": 0.0609135222964434, "grad_norm": 3.1676533222198486, "learning_rate": 4.695477523018595e-05, "loss": 5.189, "step": 6748 }, { "epoch": 0.060922549196605884, "grad_norm": 3.480325698852539, "learning_rate": 4.695432388517783e-05, "loss": 4.9999, "step": 6749 }, { "epoch": 0.06093157609676837, "grad_norm": 3.0157928466796875, "learning_rate": 4.695387254016971e-05, "loss": 4.4891, "step": 6750 }, { "epoch": 0.060940602996930854, "grad_norm": 3.155704975128174, "learning_rate": 4.695342119516158e-05, "loss": 5.1052, "step": 6751 }, { "epoch": 0.060949629897093335, "grad_norm": 4.086731433868408, "learning_rate": 4.695296985015346e-05, "loss": 4.1925, "step": 6752 }, { "epoch": 0.060958656797255824, "grad_norm": 4.711173057556152, "learning_rate": 4.6952518505145334e-05, "loss": 4.3303, "step": 6753 }, { "epoch": 0.060967683697418305, "grad_norm": 2.4319570064544678, "learning_rate": 4.695206716013721e-05, "loss": 4.6936, "step": 6754 }, { "epoch": 0.060976710597580794, "grad_norm": 2.6704416275024414, "learning_rate": 4.6951615815129086e-05, "loss": 5.0671, "step": 6755 }, { "epoch": 0.060985737497743275, "grad_norm": 3.9773950576782227, "learning_rate": 4.6951164470120965e-05, "loss": 3.8134, "step": 6756 }, { "epoch": 0.06099476439790576, "grad_norm": 3.1982104778289795, "learning_rate": 4.695071312511284e-05, "loss": 4.1219, "step": 6757 }, { "epoch": 0.061003791298068245, "grad_norm": 5.469961643218994, "learning_rate": 4.695026178010472e-05, "loss": 4.0959, "step": 6758 }, { "epoch": 0.06101281819823073, "grad_norm": 3.9842000007629395, "learning_rate": 4.694981043509659e-05, "loss": 4.897, "step": 6759 }, { "epoch": 0.06102184509839321, "grad_norm": 4.066778182983398, "learning_rate": 4.694935909008847e-05, "loss": 4.5914, "step": 6760 }, { "epoch": 0.0610308719985557, "grad_norm": 2.766974687576294, "learning_rate": 4.694890774508034e-05, "loss": 4.983, "step": 6761 }, { "epoch": 0.06103989889871818, "grad_norm": 3.487025260925293, "learning_rate": 4.6948456400072214e-05, "loss": 4.0306, "step": 6762 }, { "epoch": 0.06104892579888067, "grad_norm": 3.7847092151641846, "learning_rate": 4.694800505506409e-05, "loss": 4.5531, "step": 6763 }, { "epoch": 0.06105795269904315, "grad_norm": 3.443115472793579, "learning_rate": 4.694755371005597e-05, "loss": 4.7657, "step": 6764 }, { "epoch": 0.06106697959920563, "grad_norm": 3.2825517654418945, "learning_rate": 4.6947102365047844e-05, "loss": 5.1135, "step": 6765 }, { "epoch": 0.06107600649936812, "grad_norm": 2.9250645637512207, "learning_rate": 4.6946651020039724e-05, "loss": 4.9226, "step": 6766 }, { "epoch": 0.0610850333995306, "grad_norm": 4.083537578582764, "learning_rate": 4.6946199675031596e-05, "loss": 4.4096, "step": 6767 }, { "epoch": 0.06109406029969309, "grad_norm": 4.081089973449707, "learning_rate": 4.694574833002347e-05, "loss": 4.6035, "step": 6768 }, { "epoch": 0.06110308719985557, "grad_norm": 3.102266550064087, "learning_rate": 4.694529698501535e-05, "loss": 4.5924, "step": 6769 }, { "epoch": 0.06111211410001805, "grad_norm": 3.8634817600250244, "learning_rate": 4.694484564000722e-05, "loss": 4.5474, "step": 6770 }, { "epoch": 0.06112114100018054, "grad_norm": 2.823061943054199, "learning_rate": 4.69443942949991e-05, "loss": 4.8002, "step": 6771 }, { "epoch": 0.06113016790034302, "grad_norm": 3.655071973800659, "learning_rate": 4.694394294999098e-05, "loss": 4.227, "step": 6772 }, { "epoch": 0.0611391948005055, "grad_norm": 2.542250394821167, "learning_rate": 4.694349160498285e-05, "loss": 4.9498, "step": 6773 }, { "epoch": 0.06114822170066799, "grad_norm": 3.916388750076294, "learning_rate": 4.6943040259974724e-05, "loss": 4.5688, "step": 6774 }, { "epoch": 0.06115724860083047, "grad_norm": 4.4543633460998535, "learning_rate": 4.69425889149666e-05, "loss": 5.376, "step": 6775 }, { "epoch": 0.06116627550099296, "grad_norm": 2.869489908218384, "learning_rate": 4.6942137569958475e-05, "loss": 4.935, "step": 6776 }, { "epoch": 0.06117530240115544, "grad_norm": 2.928711175918579, "learning_rate": 4.6941686224950355e-05, "loss": 4.4935, "step": 6777 }, { "epoch": 0.061184329301317925, "grad_norm": 4.097861289978027, "learning_rate": 4.6941234879942234e-05, "loss": 4.7537, "step": 6778 }, { "epoch": 0.06119335620148041, "grad_norm": 3.8122801780700684, "learning_rate": 4.6940783534934106e-05, "loss": 4.826, "step": 6779 }, { "epoch": 0.061202383101642895, "grad_norm": 3.2508277893066406, "learning_rate": 4.6940332189925986e-05, "loss": 4.5158, "step": 6780 }, { "epoch": 0.06121141000180538, "grad_norm": 3.3721227645874023, "learning_rate": 4.693988084491786e-05, "loss": 4.055, "step": 6781 }, { "epoch": 0.061220436901967865, "grad_norm": 3.6351866722106934, "learning_rate": 4.693942949990973e-05, "loss": 4.4954, "step": 6782 }, { "epoch": 0.061229463802130346, "grad_norm": 3.4459421634674072, "learning_rate": 4.693897815490161e-05, "loss": 4.4948, "step": 6783 }, { "epoch": 0.061238490702292835, "grad_norm": 2.760138750076294, "learning_rate": 4.693852680989348e-05, "loss": 4.8418, "step": 6784 }, { "epoch": 0.061247517602455316, "grad_norm": 3.5073862075805664, "learning_rate": 4.693807546488536e-05, "loss": 4.3861, "step": 6785 }, { "epoch": 0.0612565445026178, "grad_norm": 3.598010301589966, "learning_rate": 4.693762411987724e-05, "loss": 4.4213, "step": 6786 }, { "epoch": 0.061265571402780286, "grad_norm": 3.4983022212982178, "learning_rate": 4.693717277486911e-05, "loss": 4.6271, "step": 6787 }, { "epoch": 0.06127459830294277, "grad_norm": 3.5032830238342285, "learning_rate": 4.6936721429860986e-05, "loss": 4.9068, "step": 6788 }, { "epoch": 0.061283625203105256, "grad_norm": 4.972364902496338, "learning_rate": 4.6936270084852865e-05, "loss": 4.403, "step": 6789 }, { "epoch": 0.06129265210326774, "grad_norm": 3.8401551246643066, "learning_rate": 4.693581873984474e-05, "loss": 4.7132, "step": 6790 }, { "epoch": 0.06130167900343022, "grad_norm": 3.609673500061035, "learning_rate": 4.693536739483662e-05, "loss": 4.8403, "step": 6791 }, { "epoch": 0.06131070590359271, "grad_norm": 4.461460590362549, "learning_rate": 4.693491604982849e-05, "loss": 4.5248, "step": 6792 }, { "epoch": 0.06131973280375519, "grad_norm": 4.251713752746582, "learning_rate": 4.693446470482037e-05, "loss": 4.9708, "step": 6793 }, { "epoch": 0.06132875970391768, "grad_norm": 2.526789665222168, "learning_rate": 4.693401335981224e-05, "loss": 5.0094, "step": 6794 }, { "epoch": 0.06133778660408016, "grad_norm": 4.64372444152832, "learning_rate": 4.693356201480411e-05, "loss": 4.9307, "step": 6795 }, { "epoch": 0.06134681350424264, "grad_norm": 3.618823766708374, "learning_rate": 4.693311066979599e-05, "loss": 3.8086, "step": 6796 }, { "epoch": 0.06135584040440513, "grad_norm": 3.638753890991211, "learning_rate": 4.693265932478787e-05, "loss": 5.172, "step": 6797 }, { "epoch": 0.06136486730456761, "grad_norm": 3.8965635299682617, "learning_rate": 4.6932207979779744e-05, "loss": 5.0194, "step": 6798 }, { "epoch": 0.06137389420473009, "grad_norm": 4.6606621742248535, "learning_rate": 4.6931756634771623e-05, "loss": 4.608, "step": 6799 }, { "epoch": 0.06138292110489258, "grad_norm": 2.9582009315490723, "learning_rate": 4.69313052897635e-05, "loss": 5.1906, "step": 6800 }, { "epoch": 0.06139194800505506, "grad_norm": 3.1266446113586426, "learning_rate": 4.693085394475537e-05, "loss": 4.3559, "step": 6801 }, { "epoch": 0.06140097490521755, "grad_norm": 3.3738276958465576, "learning_rate": 4.693040259974725e-05, "loss": 4.8379, "step": 6802 }, { "epoch": 0.06141000180538003, "grad_norm": 3.543168544769287, "learning_rate": 4.692995125473913e-05, "loss": 4.3345, "step": 6803 }, { "epoch": 0.061419028705542514, "grad_norm": 3.822425603866577, "learning_rate": 4.6929499909731e-05, "loss": 4.457, "step": 6804 }, { "epoch": 0.061428055605705, "grad_norm": 2.803858757019043, "learning_rate": 4.692904856472288e-05, "loss": 4.9145, "step": 6805 }, { "epoch": 0.061437082505867484, "grad_norm": 3.0258400440216064, "learning_rate": 4.692859721971475e-05, "loss": 4.7509, "step": 6806 }, { "epoch": 0.06144610940602997, "grad_norm": 6.323482990264893, "learning_rate": 4.692814587470663e-05, "loss": 4.5051, "step": 6807 }, { "epoch": 0.061455136306192454, "grad_norm": 4.426280975341797, "learning_rate": 4.69276945296985e-05, "loss": 4.2852, "step": 6808 }, { "epoch": 0.061464163206354935, "grad_norm": 3.7374048233032227, "learning_rate": 4.6927243184690375e-05, "loss": 5.2544, "step": 6809 }, { "epoch": 0.061473190106517424, "grad_norm": 3.45654034614563, "learning_rate": 4.6926791839682254e-05, "loss": 4.4969, "step": 6810 }, { "epoch": 0.061482217006679905, "grad_norm": 2.827939033508301, "learning_rate": 4.6926340494674134e-05, "loss": 4.8571, "step": 6811 }, { "epoch": 0.06149124390684239, "grad_norm": 4.462057113647461, "learning_rate": 4.6925889149666006e-05, "loss": 4.9585, "step": 6812 }, { "epoch": 0.061500270807004875, "grad_norm": 2.465303659439087, "learning_rate": 4.6925437804657885e-05, "loss": 4.4557, "step": 6813 }, { "epoch": 0.06150929770716736, "grad_norm": 3.7786476612091064, "learning_rate": 4.692498645964976e-05, "loss": 4.6842, "step": 6814 }, { "epoch": 0.061518324607329845, "grad_norm": 3.186659574508667, "learning_rate": 4.692453511464163e-05, "loss": 4.3989, "step": 6815 }, { "epoch": 0.06152735150749233, "grad_norm": 3.437548875808716, "learning_rate": 4.692408376963351e-05, "loss": 5.1208, "step": 6816 }, { "epoch": 0.06153637840765481, "grad_norm": 3.587411880493164, "learning_rate": 4.692363242462539e-05, "loss": 4.5417, "step": 6817 }, { "epoch": 0.0615454053078173, "grad_norm": 3.022906541824341, "learning_rate": 4.692318107961726e-05, "loss": 4.4592, "step": 6818 }, { "epoch": 0.06155443220797978, "grad_norm": 3.0215532779693604, "learning_rate": 4.692272973460914e-05, "loss": 4.5083, "step": 6819 }, { "epoch": 0.06156345910814227, "grad_norm": 4.67625617980957, "learning_rate": 4.692227838960101e-05, "loss": 3.4764, "step": 6820 }, { "epoch": 0.06157248600830475, "grad_norm": 4.090109348297119, "learning_rate": 4.6921827044592885e-05, "loss": 4.8651, "step": 6821 }, { "epoch": 0.06158151290846723, "grad_norm": 3.8170361518859863, "learning_rate": 4.6921375699584765e-05, "loss": 5.0133, "step": 6822 }, { "epoch": 0.06159053980862972, "grad_norm": 3.399414300918579, "learning_rate": 4.692092435457664e-05, "loss": 5.239, "step": 6823 }, { "epoch": 0.0615995667087922, "grad_norm": 2.985508680343628, "learning_rate": 4.6920473009568516e-05, "loss": 4.4786, "step": 6824 }, { "epoch": 0.06160859360895468, "grad_norm": 3.7167341709136963, "learning_rate": 4.6920021664560396e-05, "loss": 4.8411, "step": 6825 }, { "epoch": 0.06161762050911717, "grad_norm": 3.010728120803833, "learning_rate": 4.691957031955227e-05, "loss": 4.9269, "step": 6826 }, { "epoch": 0.06162664740927965, "grad_norm": 4.174037933349609, "learning_rate": 4.691911897454415e-05, "loss": 3.3319, "step": 6827 }, { "epoch": 0.06163567430944214, "grad_norm": 3.9510324001312256, "learning_rate": 4.691866762953602e-05, "loss": 4.054, "step": 6828 }, { "epoch": 0.06164470120960462, "grad_norm": 3.1888606548309326, "learning_rate": 4.691821628452789e-05, "loss": 4.2956, "step": 6829 }, { "epoch": 0.0616537281097671, "grad_norm": 4.230090618133545, "learning_rate": 4.691776493951977e-05, "loss": 4.6741, "step": 6830 }, { "epoch": 0.06166275500992959, "grad_norm": 3.775552272796631, "learning_rate": 4.6917313594511644e-05, "loss": 4.3445, "step": 6831 }, { "epoch": 0.06167178191009207, "grad_norm": 3.492523670196533, "learning_rate": 4.691686224950352e-05, "loss": 4.7398, "step": 6832 }, { "epoch": 0.06168080881025456, "grad_norm": 4.879342079162598, "learning_rate": 4.69164109044954e-05, "loss": 3.6954, "step": 6833 }, { "epoch": 0.06168983571041704, "grad_norm": 3.4285600185394287, "learning_rate": 4.6915959559487275e-05, "loss": 4.7439, "step": 6834 }, { "epoch": 0.061698862610579525, "grad_norm": 3.6227266788482666, "learning_rate": 4.691550821447915e-05, "loss": 4.8033, "step": 6835 }, { "epoch": 0.06170788951074201, "grad_norm": 3.2446181774139404, "learning_rate": 4.691505686947103e-05, "loss": 4.5749, "step": 6836 }, { "epoch": 0.061716916410904495, "grad_norm": 2.1848580837249756, "learning_rate": 4.69146055244629e-05, "loss": 4.8819, "step": 6837 }, { "epoch": 0.06172594331106698, "grad_norm": 4.539522171020508, "learning_rate": 4.691415417945478e-05, "loss": 3.878, "step": 6838 }, { "epoch": 0.061734970211229465, "grad_norm": 4.419719696044922, "learning_rate": 4.691370283444666e-05, "loss": 5.0758, "step": 6839 }, { "epoch": 0.061743997111391946, "grad_norm": 3.800002098083496, "learning_rate": 4.691325148943853e-05, "loss": 4.819, "step": 6840 }, { "epoch": 0.061753024011554435, "grad_norm": 3.2295291423797607, "learning_rate": 4.69128001444304e-05, "loss": 3.919, "step": 6841 }, { "epoch": 0.061762050911716916, "grad_norm": 5.1325554847717285, "learning_rate": 4.691234879942228e-05, "loss": 4.5747, "step": 6842 }, { "epoch": 0.0617710778118794, "grad_norm": 2.965731620788574, "learning_rate": 4.6911897454414154e-05, "loss": 4.8167, "step": 6843 }, { "epoch": 0.061780104712041886, "grad_norm": 3.2677972316741943, "learning_rate": 4.6911446109406033e-05, "loss": 5.4213, "step": 6844 }, { "epoch": 0.06178913161220437, "grad_norm": 3.1744723320007324, "learning_rate": 4.6910994764397906e-05, "loss": 4.6336, "step": 6845 }, { "epoch": 0.061798158512366856, "grad_norm": 3.4638075828552246, "learning_rate": 4.6910543419389785e-05, "loss": 4.52, "step": 6846 }, { "epoch": 0.06180718541252934, "grad_norm": 3.04221773147583, "learning_rate": 4.6910092074381664e-05, "loss": 4.3647, "step": 6847 }, { "epoch": 0.06181621231269182, "grad_norm": 5.242077827453613, "learning_rate": 4.690964072937353e-05, "loss": 3.3844, "step": 6848 }, { "epoch": 0.06182523921285431, "grad_norm": 4.195073127746582, "learning_rate": 4.690918938436541e-05, "loss": 4.31, "step": 6849 }, { "epoch": 0.06183426611301679, "grad_norm": 3.0607175827026367, "learning_rate": 4.690873803935729e-05, "loss": 5.0351, "step": 6850 }, { "epoch": 0.06184329301317928, "grad_norm": 3.881013870239258, "learning_rate": 4.690828669434916e-05, "loss": 4.7177, "step": 6851 }, { "epoch": 0.06185231991334176, "grad_norm": 3.950795888900757, "learning_rate": 4.690783534934104e-05, "loss": 4.7068, "step": 6852 }, { "epoch": 0.06186134681350424, "grad_norm": 3.1778182983398438, "learning_rate": 4.690738400433291e-05, "loss": 4.6356, "step": 6853 }, { "epoch": 0.06187037371366673, "grad_norm": 3.0362956523895264, "learning_rate": 4.690693265932479e-05, "loss": 4.2584, "step": 6854 }, { "epoch": 0.06187940061382921, "grad_norm": 2.8640291690826416, "learning_rate": 4.6906481314316664e-05, "loss": 3.8862, "step": 6855 }, { "epoch": 0.06188842751399169, "grad_norm": 4.389331817626953, "learning_rate": 4.690602996930854e-05, "loss": 4.8809, "step": 6856 }, { "epoch": 0.06189745441415418, "grad_norm": 4.48484468460083, "learning_rate": 4.6905578624300416e-05, "loss": 5.154, "step": 6857 }, { "epoch": 0.06190648131431666, "grad_norm": 4.169193744659424, "learning_rate": 4.6905127279292295e-05, "loss": 4.6056, "step": 6858 }, { "epoch": 0.06191550821447915, "grad_norm": 2.910235643386841, "learning_rate": 4.690467593428417e-05, "loss": 4.218, "step": 6859 }, { "epoch": 0.06192453511464163, "grad_norm": 3.507169723510742, "learning_rate": 4.690422458927605e-05, "loss": 4.5928, "step": 6860 }, { "epoch": 0.061933562014804114, "grad_norm": 3.7893810272216797, "learning_rate": 4.6903773244267926e-05, "loss": 4.383, "step": 6861 }, { "epoch": 0.0619425889149666, "grad_norm": 3.0574400424957275, "learning_rate": 4.690332189925979e-05, "loss": 5.2116, "step": 6862 }, { "epoch": 0.061951615815129084, "grad_norm": 4.6319193840026855, "learning_rate": 4.690287055425167e-05, "loss": 4.5553, "step": 6863 }, { "epoch": 0.06196064271529157, "grad_norm": 2.5757575035095215, "learning_rate": 4.690241920924355e-05, "loss": 4.8414, "step": 6864 }, { "epoch": 0.061969669615454054, "grad_norm": 3.1346964836120605, "learning_rate": 4.690196786423542e-05, "loss": 5.3286, "step": 6865 }, { "epoch": 0.061978696515616535, "grad_norm": 3.5494630336761475, "learning_rate": 4.69015165192273e-05, "loss": 4.1526, "step": 6866 }, { "epoch": 0.061987723415779024, "grad_norm": 3.3170573711395264, "learning_rate": 4.6901065174219175e-05, "loss": 4.9338, "step": 6867 }, { "epoch": 0.061996750315941505, "grad_norm": 3.695282459259033, "learning_rate": 4.690061382921105e-05, "loss": 4.4449, "step": 6868 }, { "epoch": 0.06200577721610399, "grad_norm": 4.642669200897217, "learning_rate": 4.6900162484202926e-05, "loss": 4.6368, "step": 6869 }, { "epoch": 0.062014804116266475, "grad_norm": 4.14042329788208, "learning_rate": 4.68997111391948e-05, "loss": 5.0199, "step": 6870 }, { "epoch": 0.06202383101642896, "grad_norm": 3.5866923332214355, "learning_rate": 4.689925979418668e-05, "loss": 5.238, "step": 6871 }, { "epoch": 0.062032857916591445, "grad_norm": 3.6746721267700195, "learning_rate": 4.689880844917856e-05, "loss": 4.6148, "step": 6872 }, { "epoch": 0.06204188481675393, "grad_norm": 4.251859188079834, "learning_rate": 4.689835710417043e-05, "loss": 3.2577, "step": 6873 }, { "epoch": 0.06205091171691641, "grad_norm": 3.3956615924835205, "learning_rate": 4.689790575916231e-05, "loss": 4.6453, "step": 6874 }, { "epoch": 0.0620599386170789, "grad_norm": 3.0356297492980957, "learning_rate": 4.689745441415418e-05, "loss": 4.5736, "step": 6875 }, { "epoch": 0.06206896551724138, "grad_norm": 3.0762665271759033, "learning_rate": 4.6897003069146054e-05, "loss": 4.9982, "step": 6876 }, { "epoch": 0.06207799241740387, "grad_norm": 3.3060522079467773, "learning_rate": 4.689655172413793e-05, "loss": 4.7152, "step": 6877 }, { "epoch": 0.06208701931756635, "grad_norm": 3.4944355487823486, "learning_rate": 4.689610037912981e-05, "loss": 4.876, "step": 6878 }, { "epoch": 0.06209604621772883, "grad_norm": 6.067193508148193, "learning_rate": 4.6895649034121685e-05, "loss": 5.2754, "step": 6879 }, { "epoch": 0.06210507311789132, "grad_norm": 3.863104820251465, "learning_rate": 4.6895197689113564e-05, "loss": 4.339, "step": 6880 }, { "epoch": 0.0621141000180538, "grad_norm": 4.136638641357422, "learning_rate": 4.689474634410544e-05, "loss": 4.8993, "step": 6881 }, { "epoch": 0.06212312691821628, "grad_norm": 3.401439666748047, "learning_rate": 4.689429499909731e-05, "loss": 5.0451, "step": 6882 }, { "epoch": 0.06213215381837877, "grad_norm": 2.991854429244995, "learning_rate": 4.689384365408919e-05, "loss": 4.4401, "step": 6883 }, { "epoch": 0.06214118071854125, "grad_norm": 4.359419345855713, "learning_rate": 4.689339230908106e-05, "loss": 4.0021, "step": 6884 }, { "epoch": 0.06215020761870374, "grad_norm": 3.2860677242279053, "learning_rate": 4.689294096407294e-05, "loss": 4.3308, "step": 6885 }, { "epoch": 0.06215923451886622, "grad_norm": 3.107642412185669, "learning_rate": 4.689248961906482e-05, "loss": 4.6516, "step": 6886 }, { "epoch": 0.0621682614190287, "grad_norm": 4.091547012329102, "learning_rate": 4.689203827405669e-05, "loss": 4.3547, "step": 6887 }, { "epoch": 0.06217728831919119, "grad_norm": 3.4950501918792725, "learning_rate": 4.6891586929048564e-05, "loss": 4.9113, "step": 6888 }, { "epoch": 0.06218631521935367, "grad_norm": 4.1770830154418945, "learning_rate": 4.6891135584040443e-05, "loss": 4.4359, "step": 6889 }, { "epoch": 0.06219534211951616, "grad_norm": 3.2172982692718506, "learning_rate": 4.6890684239032316e-05, "loss": 4.849, "step": 6890 }, { "epoch": 0.06220436901967864, "grad_norm": 3.470777750015259, "learning_rate": 4.6890232894024195e-05, "loss": 4.5443, "step": 6891 }, { "epoch": 0.062213395919841125, "grad_norm": 3.2426888942718506, "learning_rate": 4.688978154901607e-05, "loss": 4.7352, "step": 6892 }, { "epoch": 0.06222242282000361, "grad_norm": 4.2189106941223145, "learning_rate": 4.688933020400795e-05, "loss": 5.4642, "step": 6893 }, { "epoch": 0.062231449720166095, "grad_norm": 3.556638717651367, "learning_rate": 4.6888878858999826e-05, "loss": 4.6194, "step": 6894 }, { "epoch": 0.062240476620328576, "grad_norm": 2.9775915145874023, "learning_rate": 4.688842751399169e-05, "loss": 5.1724, "step": 6895 }, { "epoch": 0.062249503520491065, "grad_norm": 3.8273327350616455, "learning_rate": 4.688797616898357e-05, "loss": 4.7523, "step": 6896 }, { "epoch": 0.062258530420653546, "grad_norm": 3.3361899852752686, "learning_rate": 4.688752482397545e-05, "loss": 4.701, "step": 6897 }, { "epoch": 0.062267557320816035, "grad_norm": 3.5036566257476807, "learning_rate": 4.688707347896732e-05, "loss": 5.2385, "step": 6898 }, { "epoch": 0.062276584220978516, "grad_norm": 3.887620687484741, "learning_rate": 4.68866221339592e-05, "loss": 4.5427, "step": 6899 }, { "epoch": 0.062285611121141, "grad_norm": 4.206681728363037, "learning_rate": 4.688617078895108e-05, "loss": 4.7352, "step": 6900 }, { "epoch": 0.062294638021303486, "grad_norm": 4.160096168518066, "learning_rate": 4.6885719443942954e-05, "loss": 5.4746, "step": 6901 }, { "epoch": 0.06230366492146597, "grad_norm": 3.8422353267669678, "learning_rate": 4.6885268098934826e-05, "loss": 4.8108, "step": 6902 }, { "epoch": 0.062312691821628456, "grad_norm": 3.985771894454956, "learning_rate": 4.6884816753926705e-05, "loss": 4.8158, "step": 6903 }, { "epoch": 0.06232171872179094, "grad_norm": 3.5890231132507324, "learning_rate": 4.688436540891858e-05, "loss": 4.9358, "step": 6904 }, { "epoch": 0.06233074562195342, "grad_norm": 4.620809555053711, "learning_rate": 4.688391406391046e-05, "loss": 4.8713, "step": 6905 }, { "epoch": 0.06233977252211591, "grad_norm": 2.741858720779419, "learning_rate": 4.688346271890233e-05, "loss": 4.4886, "step": 6906 }, { "epoch": 0.06234879942227839, "grad_norm": 3.518129825592041, "learning_rate": 4.688301137389421e-05, "loss": 4.2222, "step": 6907 }, { "epoch": 0.06235782632244087, "grad_norm": 4.166593074798584, "learning_rate": 4.688256002888608e-05, "loss": 4.5259, "step": 6908 }, { "epoch": 0.06236685322260336, "grad_norm": 3.4215145111083984, "learning_rate": 4.6882108683877954e-05, "loss": 4.4561, "step": 6909 }, { "epoch": 0.06237588012276584, "grad_norm": 3.3346166610717773, "learning_rate": 4.688165733886983e-05, "loss": 4.6921, "step": 6910 }, { "epoch": 0.06238490702292833, "grad_norm": 3.965230703353882, "learning_rate": 4.688120599386171e-05, "loss": 4.9677, "step": 6911 }, { "epoch": 0.06239393392309081, "grad_norm": 3.1700963973999023, "learning_rate": 4.6880754648853585e-05, "loss": 4.8366, "step": 6912 }, { "epoch": 0.06240296082325329, "grad_norm": 3.17930269241333, "learning_rate": 4.6880303303845464e-05, "loss": 5.3399, "step": 6913 }, { "epoch": 0.06241198772341578, "grad_norm": 3.9842336177825928, "learning_rate": 4.6879851958837336e-05, "loss": 5.1281, "step": 6914 }, { "epoch": 0.06242101462357826, "grad_norm": 4.680858612060547, "learning_rate": 4.687940061382921e-05, "loss": 4.5221, "step": 6915 }, { "epoch": 0.06243004152374075, "grad_norm": 3.734769821166992, "learning_rate": 4.687894926882109e-05, "loss": 4.446, "step": 6916 }, { "epoch": 0.06243906842390323, "grad_norm": 3.258756637573242, "learning_rate": 4.687849792381297e-05, "loss": 5.2214, "step": 6917 }, { "epoch": 0.062448095324065714, "grad_norm": 3.644209623336792, "learning_rate": 4.687804657880484e-05, "loss": 4.4515, "step": 6918 }, { "epoch": 0.0624571222242282, "grad_norm": 3.6343507766723633, "learning_rate": 4.687759523379672e-05, "loss": 4.7541, "step": 6919 }, { "epoch": 0.062466149124390684, "grad_norm": 4.433508396148682, "learning_rate": 4.687714388878859e-05, "loss": 3.7675, "step": 6920 }, { "epoch": 0.062475176024553165, "grad_norm": 4.062785625457764, "learning_rate": 4.687669254378047e-05, "loss": 4.5656, "step": 6921 }, { "epoch": 0.062484202924715654, "grad_norm": 3.790107250213623, "learning_rate": 4.687624119877234e-05, "loss": 4.4941, "step": 6922 }, { "epoch": 0.062493229824878135, "grad_norm": 3.288579225540161, "learning_rate": 4.6875789853764216e-05, "loss": 5.0906, "step": 6923 }, { "epoch": 0.06250225672504062, "grad_norm": 3.044581651687622, "learning_rate": 4.6875338508756095e-05, "loss": 4.8226, "step": 6924 }, { "epoch": 0.0625112836252031, "grad_norm": 3.277047872543335, "learning_rate": 4.6874887163747974e-05, "loss": 4.5668, "step": 6925 }, { "epoch": 0.0625203105253656, "grad_norm": 2.4854705333709717, "learning_rate": 4.687443581873985e-05, "loss": 4.4511, "step": 6926 }, { "epoch": 0.06252933742552808, "grad_norm": 3.1375389099121094, "learning_rate": 4.6873984473731726e-05, "loss": 4.8139, "step": 6927 }, { "epoch": 0.06253836432569056, "grad_norm": 3.150805711746216, "learning_rate": 4.68735331287236e-05, "loss": 5.0884, "step": 6928 }, { "epoch": 0.06254739122585304, "grad_norm": 4.251956939697266, "learning_rate": 4.687308178371547e-05, "loss": 4.7243, "step": 6929 }, { "epoch": 0.06255641812601552, "grad_norm": 3.320108652114868, "learning_rate": 4.687263043870735e-05, "loss": 4.8846, "step": 6930 }, { "epoch": 0.06256544502617802, "grad_norm": 3.0209672451019287, "learning_rate": 4.687217909369922e-05, "loss": 5.2671, "step": 6931 }, { "epoch": 0.0625744719263405, "grad_norm": 3.1125965118408203, "learning_rate": 4.68717277486911e-05, "loss": 4.8189, "step": 6932 }, { "epoch": 0.06258349882650298, "grad_norm": 8.108853340148926, "learning_rate": 4.687127640368298e-05, "loss": 5.0222, "step": 6933 }, { "epoch": 0.06259252572666546, "grad_norm": 4.556732654571533, "learning_rate": 4.6870825058674853e-05, "loss": 5.0551, "step": 6934 }, { "epoch": 0.06260155262682794, "grad_norm": 7.465206146240234, "learning_rate": 4.6870373713666726e-05, "loss": 4.1917, "step": 6935 }, { "epoch": 0.06261057952699044, "grad_norm": 3.1060311794281006, "learning_rate": 4.6869922368658605e-05, "loss": 5.1767, "step": 6936 }, { "epoch": 0.06261960642715292, "grad_norm": 2.212986469268799, "learning_rate": 4.686947102365048e-05, "loss": 5.278, "step": 6937 }, { "epoch": 0.0626286333273154, "grad_norm": 4.672826290130615, "learning_rate": 4.686901967864236e-05, "loss": 5.0574, "step": 6938 }, { "epoch": 0.06263766022747788, "grad_norm": 3.8608953952789307, "learning_rate": 4.6868568333634236e-05, "loss": 4.565, "step": 6939 }, { "epoch": 0.06264668712764036, "grad_norm": 3.5942490100860596, "learning_rate": 4.686811698862611e-05, "loss": 4.5051, "step": 6940 }, { "epoch": 0.06265571402780286, "grad_norm": 3.3287808895111084, "learning_rate": 4.686766564361799e-05, "loss": 4.5688, "step": 6941 }, { "epoch": 0.06266474092796534, "grad_norm": 3.7792649269104004, "learning_rate": 4.686721429860986e-05, "loss": 4.5448, "step": 6942 }, { "epoch": 0.06267376782812782, "grad_norm": 3.505699396133423, "learning_rate": 4.686676295360173e-05, "loss": 4.6021, "step": 6943 }, { "epoch": 0.0626827947282903, "grad_norm": 4.002392292022705, "learning_rate": 4.686631160859361e-05, "loss": 4.9597, "step": 6944 }, { "epoch": 0.06269182162845278, "grad_norm": 3.887861490249634, "learning_rate": 4.6865860263585484e-05, "loss": 4.7756, "step": 6945 }, { "epoch": 0.06270084852861528, "grad_norm": 3.729240655899048, "learning_rate": 4.6865408918577364e-05, "loss": 4.5322, "step": 6946 }, { "epoch": 0.06270987542877776, "grad_norm": 3.461310863494873, "learning_rate": 4.686495757356924e-05, "loss": 4.5363, "step": 6947 }, { "epoch": 0.06271890232894024, "grad_norm": 3.8402416706085205, "learning_rate": 4.6864506228561115e-05, "loss": 4.5152, "step": 6948 }, { "epoch": 0.06272792922910272, "grad_norm": 3.8679416179656982, "learning_rate": 4.686405488355299e-05, "loss": 4.4114, "step": 6949 }, { "epoch": 0.0627369561292652, "grad_norm": 3.5062973499298096, "learning_rate": 4.686360353854487e-05, "loss": 4.5083, "step": 6950 }, { "epoch": 0.06274598302942769, "grad_norm": 3.223860502243042, "learning_rate": 4.686315219353674e-05, "loss": 5.003, "step": 6951 }, { "epoch": 0.06275500992959018, "grad_norm": 3.9280426502227783, "learning_rate": 4.686270084852862e-05, "loss": 4.8159, "step": 6952 }, { "epoch": 0.06276403682975266, "grad_norm": 3.729253053665161, "learning_rate": 4.686224950352049e-05, "loss": 4.1397, "step": 6953 }, { "epoch": 0.06277306372991515, "grad_norm": 3.3329734802246094, "learning_rate": 4.686179815851237e-05, "loss": 4.4862, "step": 6954 }, { "epoch": 0.06278209063007763, "grad_norm": 3.297262191772461, "learning_rate": 4.686134681350424e-05, "loss": 4.852, "step": 6955 }, { "epoch": 0.06279111753024011, "grad_norm": 2.8509838581085205, "learning_rate": 4.6860895468496115e-05, "loss": 4.9015, "step": 6956 }, { "epoch": 0.0628001444304026, "grad_norm": 4.579135417938232, "learning_rate": 4.6860444123487995e-05, "loss": 4.7025, "step": 6957 }, { "epoch": 0.06280917133056509, "grad_norm": 4.713012218475342, "learning_rate": 4.6859992778479874e-05, "loss": 5.3753, "step": 6958 }, { "epoch": 0.06281819823072757, "grad_norm": 3.98848819732666, "learning_rate": 4.6859541433471746e-05, "loss": 4.3437, "step": 6959 }, { "epoch": 0.06282722513089005, "grad_norm": 3.0139007568359375, "learning_rate": 4.6859090088463626e-05, "loss": 4.4881, "step": 6960 }, { "epoch": 0.06283625203105253, "grad_norm": 2.85874080657959, "learning_rate": 4.6858638743455505e-05, "loss": 5.1371, "step": 6961 }, { "epoch": 0.06284527893121503, "grad_norm": 2.5394070148468018, "learning_rate": 4.685818739844737e-05, "loss": 4.4591, "step": 6962 }, { "epoch": 0.06285430583137751, "grad_norm": 3.9147660732269287, "learning_rate": 4.685773605343925e-05, "loss": 4.2996, "step": 6963 }, { "epoch": 0.06286333273153999, "grad_norm": 2.9852614402770996, "learning_rate": 4.685728470843113e-05, "loss": 5.1768, "step": 6964 }, { "epoch": 0.06287235963170247, "grad_norm": 3.7539937496185303, "learning_rate": 4.6856833363423e-05, "loss": 5.2817, "step": 6965 }, { "epoch": 0.06288138653186495, "grad_norm": 3.311048984527588, "learning_rate": 4.685638201841488e-05, "loss": 4.9271, "step": 6966 }, { "epoch": 0.06289041343202745, "grad_norm": 3.216820001602173, "learning_rate": 4.685593067340675e-05, "loss": 4.8231, "step": 6967 }, { "epoch": 0.06289944033218993, "grad_norm": 4.087973594665527, "learning_rate": 4.685547932839863e-05, "loss": 4.8366, "step": 6968 }, { "epoch": 0.06290846723235241, "grad_norm": 4.860591888427734, "learning_rate": 4.6855027983390505e-05, "loss": 4.4617, "step": 6969 }, { "epoch": 0.06291749413251489, "grad_norm": 3.532550573348999, "learning_rate": 4.685457663838238e-05, "loss": 4.3433, "step": 6970 }, { "epoch": 0.06292652103267737, "grad_norm": 3.479543924331665, "learning_rate": 4.685412529337426e-05, "loss": 4.7039, "step": 6971 }, { "epoch": 0.06293554793283987, "grad_norm": 3.5776524543762207, "learning_rate": 4.6853673948366136e-05, "loss": 4.4978, "step": 6972 }, { "epoch": 0.06294457483300235, "grad_norm": 3.6564791202545166, "learning_rate": 4.685322260335801e-05, "loss": 4.766, "step": 6973 }, { "epoch": 0.06295360173316483, "grad_norm": 3.9632370471954346, "learning_rate": 4.685277125834989e-05, "loss": 4.7265, "step": 6974 }, { "epoch": 0.06296262863332731, "grad_norm": 3.390533685684204, "learning_rate": 4.685231991334176e-05, "loss": 4.5167, "step": 6975 }, { "epoch": 0.0629716555334898, "grad_norm": 3.090771198272705, "learning_rate": 4.685186856833363e-05, "loss": 4.9339, "step": 6976 }, { "epoch": 0.06298068243365228, "grad_norm": 4.323919296264648, "learning_rate": 4.685141722332551e-05, "loss": 4.3379, "step": 6977 }, { "epoch": 0.06298970933381477, "grad_norm": 3.0865588188171387, "learning_rate": 4.685096587831739e-05, "loss": 4.729, "step": 6978 }, { "epoch": 0.06299873623397725, "grad_norm": 3.414363145828247, "learning_rate": 4.6850514533309263e-05, "loss": 5.1573, "step": 6979 }, { "epoch": 0.06300776313413974, "grad_norm": 3.3089632987976074, "learning_rate": 4.685006318830114e-05, "loss": 4.7929, "step": 6980 }, { "epoch": 0.06301679003430222, "grad_norm": 3.2940096855163574, "learning_rate": 4.6849611843293015e-05, "loss": 4.9126, "step": 6981 }, { "epoch": 0.0630258169344647, "grad_norm": 4.432494640350342, "learning_rate": 4.684916049828489e-05, "loss": 5.3139, "step": 6982 }, { "epoch": 0.0630348438346272, "grad_norm": 3.8694889545440674, "learning_rate": 4.684870915327677e-05, "loss": 4.9971, "step": 6983 }, { "epoch": 0.06304387073478968, "grad_norm": 3.2457919120788574, "learning_rate": 4.684825780826864e-05, "loss": 4.6628, "step": 6984 }, { "epoch": 0.06305289763495216, "grad_norm": 3.569544553756714, "learning_rate": 4.684780646326052e-05, "loss": 4.5845, "step": 6985 }, { "epoch": 0.06306192453511464, "grad_norm": 2.106092691421509, "learning_rate": 4.68473551182524e-05, "loss": 5.1865, "step": 6986 }, { "epoch": 0.06307095143527712, "grad_norm": 3.1625654697418213, "learning_rate": 4.684690377324427e-05, "loss": 4.8397, "step": 6987 }, { "epoch": 0.06307997833543962, "grad_norm": 4.725551128387451, "learning_rate": 4.684645242823615e-05, "loss": 3.1118, "step": 6988 }, { "epoch": 0.0630890052356021, "grad_norm": 3.3974051475524902, "learning_rate": 4.684600108322802e-05, "loss": 4.6632, "step": 6989 }, { "epoch": 0.06309803213576458, "grad_norm": 3.6378917694091797, "learning_rate": 4.6845549738219894e-05, "loss": 5.3369, "step": 6990 }, { "epoch": 0.06310705903592706, "grad_norm": 3.0946147441864014, "learning_rate": 4.6845098393211774e-05, "loss": 4.4228, "step": 6991 }, { "epoch": 0.06311608593608954, "grad_norm": 2.741288423538208, "learning_rate": 4.6844647048203646e-05, "loss": 4.8452, "step": 6992 }, { "epoch": 0.06312511283625204, "grad_norm": 3.3439102172851562, "learning_rate": 4.6844195703195525e-05, "loss": 4.9756, "step": 6993 }, { "epoch": 0.06313413973641452, "grad_norm": 3.6129684448242188, "learning_rate": 4.6843744358187405e-05, "loss": 4.7822, "step": 6994 }, { "epoch": 0.063143166636577, "grad_norm": 3.4464633464813232, "learning_rate": 4.684329301317928e-05, "loss": 4.7951, "step": 6995 }, { "epoch": 0.06315219353673948, "grad_norm": 3.3581924438476562, "learning_rate": 4.684284166817115e-05, "loss": 3.9939, "step": 6996 }, { "epoch": 0.06316122043690196, "grad_norm": 3.6704626083374023, "learning_rate": 4.684239032316303e-05, "loss": 4.0928, "step": 6997 }, { "epoch": 0.06317024733706446, "grad_norm": 2.93603253364563, "learning_rate": 4.68419389781549e-05, "loss": 4.2568, "step": 6998 }, { "epoch": 0.06317927423722694, "grad_norm": 3.8456201553344727, "learning_rate": 4.684148763314678e-05, "loss": 4.6643, "step": 6999 }, { "epoch": 0.06318830113738942, "grad_norm": 3.228607654571533, "learning_rate": 4.684103628813866e-05, "loss": 4.4025, "step": 7000 }, { "epoch": 0.0631973280375519, "grad_norm": 3.833224296569824, "learning_rate": 4.684058494313053e-05, "loss": 4.5571, "step": 7001 }, { "epoch": 0.06320635493771438, "grad_norm": 3.902143955230713, "learning_rate": 4.6840133598122405e-05, "loss": 4.5716, "step": 7002 }, { "epoch": 0.06321538183787687, "grad_norm": 3.3849246501922607, "learning_rate": 4.6839682253114284e-05, "loss": 5.2027, "step": 7003 }, { "epoch": 0.06322440873803936, "grad_norm": 3.7638776302337646, "learning_rate": 4.6839230908106156e-05, "loss": 5.4036, "step": 7004 }, { "epoch": 0.06323343563820184, "grad_norm": 4.092964172363281, "learning_rate": 4.6838779563098036e-05, "loss": 3.6951, "step": 7005 }, { "epoch": 0.06324246253836432, "grad_norm": 3.541717290878296, "learning_rate": 4.683832821808991e-05, "loss": 4.0712, "step": 7006 }, { "epoch": 0.0632514894385268, "grad_norm": 4.194136142730713, "learning_rate": 4.683787687308179e-05, "loss": 4.5661, "step": 7007 }, { "epoch": 0.06326051633868929, "grad_norm": 5.428067207336426, "learning_rate": 4.6837425528073667e-05, "loss": 4.8787, "step": 7008 }, { "epoch": 0.06326954323885178, "grad_norm": 3.152499198913574, "learning_rate": 4.683697418306553e-05, "loss": 4.4652, "step": 7009 }, { "epoch": 0.06327857013901426, "grad_norm": 3.9483096599578857, "learning_rate": 4.683652283805741e-05, "loss": 3.4683, "step": 7010 }, { "epoch": 0.06328759703917675, "grad_norm": 4.197038173675537, "learning_rate": 4.683607149304929e-05, "loss": 4.4085, "step": 7011 }, { "epoch": 0.06329662393933923, "grad_norm": 3.5734429359436035, "learning_rate": 4.683562014804116e-05, "loss": 4.349, "step": 7012 }, { "epoch": 0.06330565083950171, "grad_norm": 3.3319997787475586, "learning_rate": 4.683516880303304e-05, "loss": 4.0253, "step": 7013 }, { "epoch": 0.0633146777396642, "grad_norm": 3.5570895671844482, "learning_rate": 4.6834717458024915e-05, "loss": 4.9698, "step": 7014 }, { "epoch": 0.06332370463982669, "grad_norm": 3.3961546421051025, "learning_rate": 4.6834266113016794e-05, "loss": 4.6865, "step": 7015 }, { "epoch": 0.06333273153998917, "grad_norm": 2.3973515033721924, "learning_rate": 4.683381476800867e-05, "loss": 4.9822, "step": 7016 }, { "epoch": 0.06334175844015165, "grad_norm": 3.3565614223480225, "learning_rate": 4.683336342300054e-05, "loss": 4.621, "step": 7017 }, { "epoch": 0.06335078534031413, "grad_norm": 3.1804776191711426, "learning_rate": 4.683291207799242e-05, "loss": 4.2486, "step": 7018 }, { "epoch": 0.06335981224047663, "grad_norm": 3.032827615737915, "learning_rate": 4.68324607329843e-05, "loss": 4.6137, "step": 7019 }, { "epoch": 0.06336883914063911, "grad_norm": 3.723503589630127, "learning_rate": 4.683200938797617e-05, "loss": 4.8369, "step": 7020 }, { "epoch": 0.06337786604080159, "grad_norm": 2.9843904972076416, "learning_rate": 4.683155804296805e-05, "loss": 4.3943, "step": 7021 }, { "epoch": 0.06338689294096407, "grad_norm": 2.5751235485076904, "learning_rate": 4.683110669795993e-05, "loss": 4.7371, "step": 7022 }, { "epoch": 0.06339591984112655, "grad_norm": 4.451903820037842, "learning_rate": 4.6830655352951794e-05, "loss": 3.6097, "step": 7023 }, { "epoch": 0.06340494674128905, "grad_norm": 3.03914213180542, "learning_rate": 4.6830204007943673e-05, "loss": 4.7877, "step": 7024 }, { "epoch": 0.06341397364145153, "grad_norm": 2.584151268005371, "learning_rate": 4.682975266293555e-05, "loss": 4.8172, "step": 7025 }, { "epoch": 0.06342300054161401, "grad_norm": 2.987582206726074, "learning_rate": 4.6829301317927425e-05, "loss": 5.0042, "step": 7026 }, { "epoch": 0.06343202744177649, "grad_norm": 5.010430812835693, "learning_rate": 4.6828849972919304e-05, "loss": 4.6789, "step": 7027 }, { "epoch": 0.06344105434193897, "grad_norm": 3.079700469970703, "learning_rate": 4.682839862791118e-05, "loss": 4.6216, "step": 7028 }, { "epoch": 0.06345008124210147, "grad_norm": 4.0656843185424805, "learning_rate": 4.682794728290305e-05, "loss": 4.0285, "step": 7029 }, { "epoch": 0.06345910814226395, "grad_norm": 3.7147481441497803, "learning_rate": 4.682749593789493e-05, "loss": 4.61, "step": 7030 }, { "epoch": 0.06346813504242643, "grad_norm": 3.439406394958496, "learning_rate": 4.68270445928868e-05, "loss": 4.449, "step": 7031 }, { "epoch": 0.06347716194258891, "grad_norm": 3.70395565032959, "learning_rate": 4.682659324787868e-05, "loss": 4.5579, "step": 7032 }, { "epoch": 0.0634861888427514, "grad_norm": 3.738011598587036, "learning_rate": 4.682614190287056e-05, "loss": 4.7245, "step": 7033 }, { "epoch": 0.06349521574291388, "grad_norm": 3.644300937652588, "learning_rate": 4.682569055786243e-05, "loss": 4.1625, "step": 7034 }, { "epoch": 0.06350424264307637, "grad_norm": 2.8983607292175293, "learning_rate": 4.682523921285431e-05, "loss": 4.927, "step": 7035 }, { "epoch": 0.06351326954323885, "grad_norm": 3.8593366146087646, "learning_rate": 4.6824787867846184e-05, "loss": 5.1493, "step": 7036 }, { "epoch": 0.06352229644340134, "grad_norm": 2.766049385070801, "learning_rate": 4.6824336522838056e-05, "loss": 4.9488, "step": 7037 }, { "epoch": 0.06353132334356382, "grad_norm": 6.521785736083984, "learning_rate": 4.6823885177829935e-05, "loss": 3.8395, "step": 7038 }, { "epoch": 0.0635403502437263, "grad_norm": 4.131359100341797, "learning_rate": 4.6823433832821815e-05, "loss": 4.1022, "step": 7039 }, { "epoch": 0.0635493771438888, "grad_norm": 4.096398830413818, "learning_rate": 4.682298248781369e-05, "loss": 4.7321, "step": 7040 }, { "epoch": 0.06355840404405128, "grad_norm": 4.111522674560547, "learning_rate": 4.6822531142805566e-05, "loss": 4.5905, "step": 7041 }, { "epoch": 0.06356743094421376, "grad_norm": 3.3394527435302734, "learning_rate": 4.682207979779744e-05, "loss": 4.4478, "step": 7042 }, { "epoch": 0.06357645784437624, "grad_norm": 3.5160598754882812, "learning_rate": 4.682162845278931e-05, "loss": 4.6594, "step": 7043 }, { "epoch": 0.06358548474453872, "grad_norm": 2.3931374549865723, "learning_rate": 4.682117710778119e-05, "loss": 5.3143, "step": 7044 }, { "epoch": 0.06359451164470122, "grad_norm": 3.2984585762023926, "learning_rate": 4.682072576277306e-05, "loss": 5.0229, "step": 7045 }, { "epoch": 0.0636035385448637, "grad_norm": 3.505396842956543, "learning_rate": 4.682027441776494e-05, "loss": 4.8496, "step": 7046 }, { "epoch": 0.06361256544502618, "grad_norm": 3.985382318496704, "learning_rate": 4.681982307275682e-05, "loss": 4.853, "step": 7047 }, { "epoch": 0.06362159234518866, "grad_norm": 3.348701238632202, "learning_rate": 4.6819371727748694e-05, "loss": 4.4533, "step": 7048 }, { "epoch": 0.06363061924535114, "grad_norm": 3.866962432861328, "learning_rate": 4.6818920382740566e-05, "loss": 4.3959, "step": 7049 }, { "epoch": 0.06363964614551364, "grad_norm": 3.1576972007751465, "learning_rate": 4.6818469037732446e-05, "loss": 4.4961, "step": 7050 }, { "epoch": 0.06364867304567612, "grad_norm": 4.290769100189209, "learning_rate": 4.681801769272432e-05, "loss": 5.1598, "step": 7051 }, { "epoch": 0.0636576999458386, "grad_norm": 3.0916123390197754, "learning_rate": 4.68175663477162e-05, "loss": 4.8231, "step": 7052 }, { "epoch": 0.06366672684600108, "grad_norm": 3.3448996543884277, "learning_rate": 4.681711500270807e-05, "loss": 4.9098, "step": 7053 }, { "epoch": 0.06367575374616356, "grad_norm": 2.4858851432800293, "learning_rate": 4.681666365769995e-05, "loss": 4.5183, "step": 7054 }, { "epoch": 0.06368478064632606, "grad_norm": 2.839124917984009, "learning_rate": 4.681621231269183e-05, "loss": 4.9516, "step": 7055 }, { "epoch": 0.06369380754648854, "grad_norm": 3.189720630645752, "learning_rate": 4.6815760967683694e-05, "loss": 5.304, "step": 7056 }, { "epoch": 0.06370283444665102, "grad_norm": 3.908611297607422, "learning_rate": 4.681530962267557e-05, "loss": 4.7028, "step": 7057 }, { "epoch": 0.0637118613468135, "grad_norm": 3.0532491207122803, "learning_rate": 4.681485827766745e-05, "loss": 4.7488, "step": 7058 }, { "epoch": 0.06372088824697598, "grad_norm": 4.514667987823486, "learning_rate": 4.6814406932659325e-05, "loss": 5.0899, "step": 7059 }, { "epoch": 0.06372991514713847, "grad_norm": 4.3267598152160645, "learning_rate": 4.6813955587651204e-05, "loss": 5.3485, "step": 7060 }, { "epoch": 0.06373894204730096, "grad_norm": 2.5067946910858154, "learning_rate": 4.6813504242643083e-05, "loss": 4.4934, "step": 7061 }, { "epoch": 0.06374796894746344, "grad_norm": 4.780242919921875, "learning_rate": 4.6813052897634956e-05, "loss": 4.8306, "step": 7062 }, { "epoch": 0.06375699584762592, "grad_norm": 2.899937152862549, "learning_rate": 4.681260155262683e-05, "loss": 4.4981, "step": 7063 }, { "epoch": 0.0637660227477884, "grad_norm": 3.616680145263672, "learning_rate": 4.681215020761871e-05, "loss": 5.1965, "step": 7064 }, { "epoch": 0.06377504964795089, "grad_norm": 4.998465538024902, "learning_rate": 4.681169886261058e-05, "loss": 3.7878, "step": 7065 }, { "epoch": 0.06378407654811338, "grad_norm": 3.013414144515991, "learning_rate": 4.681124751760246e-05, "loss": 4.6826, "step": 7066 }, { "epoch": 0.06379310344827586, "grad_norm": 3.630406618118286, "learning_rate": 4.681079617259433e-05, "loss": 4.9855, "step": 7067 }, { "epoch": 0.06380213034843835, "grad_norm": 3.2101669311523438, "learning_rate": 4.681034482758621e-05, "loss": 4.8434, "step": 7068 }, { "epoch": 0.06381115724860083, "grad_norm": 3.2469849586486816, "learning_rate": 4.6809893482578083e-05, "loss": 4.656, "step": 7069 }, { "epoch": 0.06382018414876331, "grad_norm": 3.3221850395202637, "learning_rate": 4.6809442137569956e-05, "loss": 4.3353, "step": 7070 }, { "epoch": 0.0638292110489258, "grad_norm": 3.302231788635254, "learning_rate": 4.6808990792561835e-05, "loss": 4.6016, "step": 7071 }, { "epoch": 0.06383823794908829, "grad_norm": 3.6780035495758057, "learning_rate": 4.6808539447553714e-05, "loss": 5.1297, "step": 7072 }, { "epoch": 0.06384726484925077, "grad_norm": 3.9727730751037598, "learning_rate": 4.680808810254559e-05, "loss": 3.2682, "step": 7073 }, { "epoch": 0.06385629174941325, "grad_norm": 3.7296552658081055, "learning_rate": 4.6807636757537466e-05, "loss": 4.3569, "step": 7074 }, { "epoch": 0.06386531864957573, "grad_norm": 3.681774854660034, "learning_rate": 4.680718541252934e-05, "loss": 4.9564, "step": 7075 }, { "epoch": 0.06387434554973823, "grad_norm": 2.713925361633301, "learning_rate": 4.680673406752121e-05, "loss": 4.8392, "step": 7076 }, { "epoch": 0.06388337244990071, "grad_norm": 3.442030191421509, "learning_rate": 4.680628272251309e-05, "loss": 4.4992, "step": 7077 }, { "epoch": 0.06389239935006319, "grad_norm": 4.010642051696777, "learning_rate": 4.680583137750497e-05, "loss": 4.4921, "step": 7078 }, { "epoch": 0.06390142625022567, "grad_norm": 3.2538695335388184, "learning_rate": 4.680538003249684e-05, "loss": 4.8505, "step": 7079 }, { "epoch": 0.06391045315038815, "grad_norm": 3.6369025707244873, "learning_rate": 4.680492868748872e-05, "loss": 4.4336, "step": 7080 }, { "epoch": 0.06391948005055065, "grad_norm": 3.796349048614502, "learning_rate": 4.6804477342480594e-05, "loss": 4.2728, "step": 7081 }, { "epoch": 0.06392850695071313, "grad_norm": 3.4423327445983887, "learning_rate": 4.680402599747247e-05, "loss": 4.8254, "step": 7082 }, { "epoch": 0.06393753385087561, "grad_norm": 3.444413900375366, "learning_rate": 4.6803574652464345e-05, "loss": 4.414, "step": 7083 }, { "epoch": 0.06394656075103809, "grad_norm": 3.5183122158050537, "learning_rate": 4.680312330745622e-05, "loss": 5.5808, "step": 7084 }, { "epoch": 0.06395558765120057, "grad_norm": 2.696781873703003, "learning_rate": 4.68026719624481e-05, "loss": 4.8346, "step": 7085 }, { "epoch": 0.06396461455136306, "grad_norm": 2.8942296504974365, "learning_rate": 4.6802220617439976e-05, "loss": 4.6495, "step": 7086 }, { "epoch": 0.06397364145152555, "grad_norm": 3.384556293487549, "learning_rate": 4.680176927243185e-05, "loss": 4.6717, "step": 7087 }, { "epoch": 0.06398266835168803, "grad_norm": 3.2840709686279297, "learning_rate": 4.680131792742373e-05, "loss": 5.43, "step": 7088 }, { "epoch": 0.06399169525185051, "grad_norm": 3.2945077419281006, "learning_rate": 4.68008665824156e-05, "loss": 4.2827, "step": 7089 }, { "epoch": 0.064000722152013, "grad_norm": 3.3453924655914307, "learning_rate": 4.680041523740747e-05, "loss": 5.0733, "step": 7090 }, { "epoch": 0.06400974905217548, "grad_norm": 3.238759994506836, "learning_rate": 4.679996389239935e-05, "loss": 3.9309, "step": 7091 }, { "epoch": 0.06401877595233797, "grad_norm": 3.2115023136138916, "learning_rate": 4.6799512547391225e-05, "loss": 4.2475, "step": 7092 }, { "epoch": 0.06402780285250045, "grad_norm": 3.235637664794922, "learning_rate": 4.6799061202383104e-05, "loss": 4.8101, "step": 7093 }, { "epoch": 0.06403682975266294, "grad_norm": 3.686413288116455, "learning_rate": 4.679860985737498e-05, "loss": 4.5456, "step": 7094 }, { "epoch": 0.06404585665282542, "grad_norm": 3.8886256217956543, "learning_rate": 4.6798158512366856e-05, "loss": 4.7636, "step": 7095 }, { "epoch": 0.0640548835529879, "grad_norm": 3.0184831619262695, "learning_rate": 4.679770716735873e-05, "loss": 4.7916, "step": 7096 }, { "epoch": 0.0640639104531504, "grad_norm": 3.3987338542938232, "learning_rate": 4.679725582235061e-05, "loss": 4.7733, "step": 7097 }, { "epoch": 0.06407293735331288, "grad_norm": 3.2009284496307373, "learning_rate": 4.679680447734248e-05, "loss": 4.017, "step": 7098 }, { "epoch": 0.06408196425347536, "grad_norm": 4.054732322692871, "learning_rate": 4.679635313233436e-05, "loss": 4.5273, "step": 7099 }, { "epoch": 0.06409099115363784, "grad_norm": 3.5214903354644775, "learning_rate": 4.679590178732624e-05, "loss": 5.2859, "step": 7100 }, { "epoch": 0.06410001805380032, "grad_norm": 4.049001693725586, "learning_rate": 4.679545044231811e-05, "loss": 4.9744, "step": 7101 }, { "epoch": 0.06410904495396282, "grad_norm": 2.796804904937744, "learning_rate": 4.679499909730999e-05, "loss": 4.4498, "step": 7102 }, { "epoch": 0.0641180718541253, "grad_norm": 5.299804210662842, "learning_rate": 4.679454775230186e-05, "loss": 4.8432, "step": 7103 }, { "epoch": 0.06412709875428778, "grad_norm": 3.6776328086853027, "learning_rate": 4.6794096407293735e-05, "loss": 5.7281, "step": 7104 }, { "epoch": 0.06413612565445026, "grad_norm": 3.5174946784973145, "learning_rate": 4.6793645062285614e-05, "loss": 5.1466, "step": 7105 }, { "epoch": 0.06414515255461274, "grad_norm": 3.421264410018921, "learning_rate": 4.6793193717277487e-05, "loss": 4.7921, "step": 7106 }, { "epoch": 0.06415417945477524, "grad_norm": 4.007692813873291, "learning_rate": 4.6792742372269366e-05, "loss": 4.9129, "step": 7107 }, { "epoch": 0.06416320635493772, "grad_norm": 3.7908153533935547, "learning_rate": 4.6792291027261245e-05, "loss": 4.3821, "step": 7108 }, { "epoch": 0.0641722332551002, "grad_norm": 3.519692897796631, "learning_rate": 4.679183968225312e-05, "loss": 4.49, "step": 7109 }, { "epoch": 0.06418126015526268, "grad_norm": 4.346581935882568, "learning_rate": 4.679138833724499e-05, "loss": 4.5934, "step": 7110 }, { "epoch": 0.06419028705542516, "grad_norm": 3.03493070602417, "learning_rate": 4.679093699223687e-05, "loss": 4.6168, "step": 7111 }, { "epoch": 0.06419931395558764, "grad_norm": 3.5855960845947266, "learning_rate": 4.679048564722874e-05, "loss": 4.224, "step": 7112 }, { "epoch": 0.06420834085575014, "grad_norm": 6.112953186035156, "learning_rate": 4.679003430222062e-05, "loss": 4.677, "step": 7113 }, { "epoch": 0.06421736775591262, "grad_norm": 5.640876770019531, "learning_rate": 4.6789582957212493e-05, "loss": 4.9251, "step": 7114 }, { "epoch": 0.0642263946560751, "grad_norm": 3.803809642791748, "learning_rate": 4.678913161220437e-05, "loss": 3.3088, "step": 7115 }, { "epoch": 0.06423542155623758, "grad_norm": 3.6662604808807373, "learning_rate": 4.6788680267196245e-05, "loss": 4.9112, "step": 7116 }, { "epoch": 0.06424444845640007, "grad_norm": 3.4746718406677246, "learning_rate": 4.678822892218812e-05, "loss": 4.5479, "step": 7117 }, { "epoch": 0.06425347535656256, "grad_norm": 3.591418981552124, "learning_rate": 4.678777757718e-05, "loss": 5.485, "step": 7118 }, { "epoch": 0.06426250225672504, "grad_norm": 3.337080478668213, "learning_rate": 4.6787326232171876e-05, "loss": 4.6732, "step": 7119 }, { "epoch": 0.06427152915688752, "grad_norm": 2.790210008621216, "learning_rate": 4.678687488716375e-05, "loss": 4.445, "step": 7120 }, { "epoch": 0.06428055605705, "grad_norm": 3.875788927078247, "learning_rate": 4.678642354215563e-05, "loss": 4.6458, "step": 7121 }, { "epoch": 0.06428958295721249, "grad_norm": 3.8841946125030518, "learning_rate": 4.678597219714751e-05, "loss": 4.2563, "step": 7122 }, { "epoch": 0.06429860985737498, "grad_norm": 4.312118053436279, "learning_rate": 4.678552085213937e-05, "loss": 4.3287, "step": 7123 }, { "epoch": 0.06430763675753746, "grad_norm": 4.061550140380859, "learning_rate": 4.678506950713125e-05, "loss": 4.8147, "step": 7124 }, { "epoch": 0.06431666365769995, "grad_norm": 3.159391403198242, "learning_rate": 4.678461816212313e-05, "loss": 4.8254, "step": 7125 }, { "epoch": 0.06432569055786243, "grad_norm": 3.7845115661621094, "learning_rate": 4.6784166817115004e-05, "loss": 3.9826, "step": 7126 }, { "epoch": 0.06433471745802491, "grad_norm": 4.06912088394165, "learning_rate": 4.678371547210688e-05, "loss": 5.1486, "step": 7127 }, { "epoch": 0.0643437443581874, "grad_norm": 3.630479574203491, "learning_rate": 4.6783264127098755e-05, "loss": 5.0805, "step": 7128 }, { "epoch": 0.06435277125834989, "grad_norm": 3.9546475410461426, "learning_rate": 4.6782812782090635e-05, "loss": 4.6498, "step": 7129 }, { "epoch": 0.06436179815851237, "grad_norm": 4.319364547729492, "learning_rate": 4.678236143708251e-05, "loss": 4.4301, "step": 7130 }, { "epoch": 0.06437082505867485, "grad_norm": 3.3444314002990723, "learning_rate": 4.678191009207438e-05, "loss": 4.5124, "step": 7131 }, { "epoch": 0.06437985195883733, "grad_norm": 3.2135329246520996, "learning_rate": 4.678145874706626e-05, "loss": 5.1744, "step": 7132 }, { "epoch": 0.06438887885899983, "grad_norm": 2.7894296646118164, "learning_rate": 4.678100740205814e-05, "loss": 4.8244, "step": 7133 }, { "epoch": 0.06439790575916231, "grad_norm": 3.8114266395568848, "learning_rate": 4.678055605705001e-05, "loss": 4.3114, "step": 7134 }, { "epoch": 0.06440693265932479, "grad_norm": 4.193461894989014, "learning_rate": 4.678010471204189e-05, "loss": 4.0963, "step": 7135 }, { "epoch": 0.06441595955948727, "grad_norm": 2.724790573120117, "learning_rate": 4.677965336703376e-05, "loss": 4.6867, "step": 7136 }, { "epoch": 0.06442498645964975, "grad_norm": 3.276548147201538, "learning_rate": 4.6779202022025635e-05, "loss": 4.4314, "step": 7137 }, { "epoch": 0.06443401335981223, "grad_norm": 3.148749351501465, "learning_rate": 4.6778750677017514e-05, "loss": 4.6237, "step": 7138 }, { "epoch": 0.06444304025997473, "grad_norm": 4.044842720031738, "learning_rate": 4.677829933200939e-05, "loss": 5.6767, "step": 7139 }, { "epoch": 0.06445206716013721, "grad_norm": 4.11122989654541, "learning_rate": 4.6777847987001266e-05, "loss": 4.3663, "step": 7140 }, { "epoch": 0.06446109406029969, "grad_norm": 3.0428590774536133, "learning_rate": 4.6777396641993145e-05, "loss": 4.8887, "step": 7141 }, { "epoch": 0.06447012096046217, "grad_norm": 4.152416229248047, "learning_rate": 4.677694529698502e-05, "loss": 5.1676, "step": 7142 }, { "epoch": 0.06447914786062466, "grad_norm": 3.9775757789611816, "learning_rate": 4.677649395197689e-05, "loss": 4.5411, "step": 7143 }, { "epoch": 0.06448817476078715, "grad_norm": 4.1225385665893555, "learning_rate": 4.677604260696877e-05, "loss": 5.0564, "step": 7144 }, { "epoch": 0.06449720166094963, "grad_norm": 4.070682048797607, "learning_rate": 4.677559126196064e-05, "loss": 5.1235, "step": 7145 }, { "epoch": 0.06450622856111211, "grad_norm": 3.4491074085235596, "learning_rate": 4.677513991695252e-05, "loss": 5.2782, "step": 7146 }, { "epoch": 0.0645152554612746, "grad_norm": 4.519309997558594, "learning_rate": 4.67746885719444e-05, "loss": 4.4837, "step": 7147 }, { "epoch": 0.06452428236143708, "grad_norm": 3.8207340240478516, "learning_rate": 4.677423722693627e-05, "loss": 5.4104, "step": 7148 }, { "epoch": 0.06453330926159957, "grad_norm": 2.781797170639038, "learning_rate": 4.677378588192815e-05, "loss": 4.3857, "step": 7149 }, { "epoch": 0.06454233616176205, "grad_norm": 4.37318754196167, "learning_rate": 4.6773334536920024e-05, "loss": 3.9281, "step": 7150 }, { "epoch": 0.06455136306192454, "grad_norm": 4.315132141113281, "learning_rate": 4.6772883191911897e-05, "loss": 4.3423, "step": 7151 }, { "epoch": 0.06456038996208702, "grad_norm": 4.6629486083984375, "learning_rate": 4.6772431846903776e-05, "loss": 5.634, "step": 7152 }, { "epoch": 0.0645694168622495, "grad_norm": 3.380305767059326, "learning_rate": 4.677198050189565e-05, "loss": 4.2364, "step": 7153 }, { "epoch": 0.064578443762412, "grad_norm": 4.543047904968262, "learning_rate": 4.677152915688753e-05, "loss": 5.301, "step": 7154 }, { "epoch": 0.06458747066257448, "grad_norm": 3.302558183670044, "learning_rate": 4.677107781187941e-05, "loss": 4.7428, "step": 7155 }, { "epoch": 0.06459649756273696, "grad_norm": 3.8478949069976807, "learning_rate": 4.677062646687128e-05, "loss": 5.6313, "step": 7156 }, { "epoch": 0.06460552446289944, "grad_norm": 2.421454429626465, "learning_rate": 4.677017512186315e-05, "loss": 4.6479, "step": 7157 }, { "epoch": 0.06461455136306192, "grad_norm": 3.0650177001953125, "learning_rate": 4.676972377685503e-05, "loss": 4.4981, "step": 7158 }, { "epoch": 0.06462357826322442, "grad_norm": 2.973402261734009, "learning_rate": 4.6769272431846903e-05, "loss": 4.4339, "step": 7159 }, { "epoch": 0.0646326051633869, "grad_norm": 2.4311764240264893, "learning_rate": 4.676882108683878e-05, "loss": 4.841, "step": 7160 }, { "epoch": 0.06464163206354938, "grad_norm": 4.274649620056152, "learning_rate": 4.676836974183066e-05, "loss": 3.7626, "step": 7161 }, { "epoch": 0.06465065896371186, "grad_norm": 3.5152366161346436, "learning_rate": 4.6767918396822534e-05, "loss": 5.0548, "step": 7162 }, { "epoch": 0.06465968586387434, "grad_norm": 4.534857749938965, "learning_rate": 4.676746705181441e-05, "loss": 4.3141, "step": 7163 }, { "epoch": 0.06466871276403682, "grad_norm": 2.8144354820251465, "learning_rate": 4.6767015706806286e-05, "loss": 4.5609, "step": 7164 }, { "epoch": 0.06467773966419932, "grad_norm": 4.381448268890381, "learning_rate": 4.676656436179816e-05, "loss": 4.1458, "step": 7165 }, { "epoch": 0.0646867665643618, "grad_norm": 4.282710552215576, "learning_rate": 4.676611301679004e-05, "loss": 4.5352, "step": 7166 }, { "epoch": 0.06469579346452428, "grad_norm": 4.353124141693115, "learning_rate": 4.676566167178191e-05, "loss": 4.451, "step": 7167 }, { "epoch": 0.06470482036468676, "grad_norm": 2.8326170444488525, "learning_rate": 4.676521032677379e-05, "loss": 4.609, "step": 7168 }, { "epoch": 0.06471384726484924, "grad_norm": 2.8816630840301514, "learning_rate": 4.676475898176567e-05, "loss": 5.143, "step": 7169 }, { "epoch": 0.06472287416501174, "grad_norm": 3.39595890045166, "learning_rate": 4.6764307636757534e-05, "loss": 4.6258, "step": 7170 }, { "epoch": 0.06473190106517422, "grad_norm": 3.790661573410034, "learning_rate": 4.6763856291749414e-05, "loss": 5.0153, "step": 7171 }, { "epoch": 0.0647409279653367, "grad_norm": 3.257749080657959, "learning_rate": 4.676340494674129e-05, "loss": 4.7624, "step": 7172 }, { "epoch": 0.06474995486549918, "grad_norm": 4.792238712310791, "learning_rate": 4.6762953601733165e-05, "loss": 4.3553, "step": 7173 }, { "epoch": 0.06475898176566167, "grad_norm": 3.654580593109131, "learning_rate": 4.6762502256725045e-05, "loss": 4.5928, "step": 7174 }, { "epoch": 0.06476800866582416, "grad_norm": 3.1408045291900635, "learning_rate": 4.676205091171692e-05, "loss": 4.2235, "step": 7175 }, { "epoch": 0.06477703556598664, "grad_norm": 5.185642719268799, "learning_rate": 4.6761599566708796e-05, "loss": 4.8174, "step": 7176 }, { "epoch": 0.06478606246614912, "grad_norm": 3.7629711627960205, "learning_rate": 4.676114822170067e-05, "loss": 4.7186, "step": 7177 }, { "epoch": 0.0647950893663116, "grad_norm": 3.85353946685791, "learning_rate": 4.676069687669254e-05, "loss": 4.9312, "step": 7178 }, { "epoch": 0.06480411626647409, "grad_norm": 3.5096938610076904, "learning_rate": 4.676024553168442e-05, "loss": 4.6498, "step": 7179 }, { "epoch": 0.06481314316663658, "grad_norm": 2.6704628467559814, "learning_rate": 4.67597941866763e-05, "loss": 5.2222, "step": 7180 }, { "epoch": 0.06482217006679906, "grad_norm": 3.9454762935638428, "learning_rate": 4.675934284166817e-05, "loss": 4.7688, "step": 7181 }, { "epoch": 0.06483119696696155, "grad_norm": 3.417121171951294, "learning_rate": 4.675889149666005e-05, "loss": 3.8908, "step": 7182 }, { "epoch": 0.06484022386712403, "grad_norm": 3.4270544052124023, "learning_rate": 4.675844015165193e-05, "loss": 4.5661, "step": 7183 }, { "epoch": 0.06484925076728651, "grad_norm": 3.673567056655884, "learning_rate": 4.6757988806643796e-05, "loss": 4.4585, "step": 7184 }, { "epoch": 0.064858277667449, "grad_norm": 3.977586507797241, "learning_rate": 4.6757537461635676e-05, "loss": 3.888, "step": 7185 }, { "epoch": 0.06486730456761149, "grad_norm": 4.207317352294922, "learning_rate": 4.6757086116627555e-05, "loss": 4.6815, "step": 7186 }, { "epoch": 0.06487633146777397, "grad_norm": 3.7470486164093018, "learning_rate": 4.675663477161943e-05, "loss": 4.9787, "step": 7187 }, { "epoch": 0.06488535836793645, "grad_norm": 3.778327226638794, "learning_rate": 4.6756183426611307e-05, "loss": 5.5558, "step": 7188 }, { "epoch": 0.06489438526809893, "grad_norm": 3.367694139480591, "learning_rate": 4.675573208160318e-05, "loss": 4.9497, "step": 7189 }, { "epoch": 0.06490341216826141, "grad_norm": 2.5688247680664062, "learning_rate": 4.675528073659505e-05, "loss": 5.3099, "step": 7190 }, { "epoch": 0.06491243906842391, "grad_norm": 3.8373308181762695, "learning_rate": 4.675482939158693e-05, "loss": 4.4633, "step": 7191 }, { "epoch": 0.06492146596858639, "grad_norm": 4.7811598777771, "learning_rate": 4.67543780465788e-05, "loss": 4.1546, "step": 7192 }, { "epoch": 0.06493049286874887, "grad_norm": 3.534262180328369, "learning_rate": 4.675392670157068e-05, "loss": 5.1391, "step": 7193 }, { "epoch": 0.06493951976891135, "grad_norm": 2.839252471923828, "learning_rate": 4.675347535656256e-05, "loss": 4.852, "step": 7194 }, { "epoch": 0.06494854666907383, "grad_norm": 3.9539623260498047, "learning_rate": 4.6753024011554434e-05, "loss": 4.4802, "step": 7195 }, { "epoch": 0.06495757356923633, "grad_norm": 4.336818218231201, "learning_rate": 4.675257266654631e-05, "loss": 4.7238, "step": 7196 }, { "epoch": 0.06496660046939881, "grad_norm": 3.450955390930176, "learning_rate": 4.6752121321538186e-05, "loss": 4.346, "step": 7197 }, { "epoch": 0.06497562736956129, "grad_norm": 2.9352831840515137, "learning_rate": 4.675166997653006e-05, "loss": 4.4331, "step": 7198 }, { "epoch": 0.06498465426972377, "grad_norm": 3.4426779747009277, "learning_rate": 4.675121863152194e-05, "loss": 4.5978, "step": 7199 }, { "epoch": 0.06499368116988626, "grad_norm": 3.5935745239257812, "learning_rate": 4.675076728651382e-05, "loss": 4.6647, "step": 7200 }, { "epoch": 0.06500270807004875, "grad_norm": 3.039365291595459, "learning_rate": 4.675031594150569e-05, "loss": 4.8645, "step": 7201 }, { "epoch": 0.06501173497021123, "grad_norm": 3.1943793296813965, "learning_rate": 4.674986459649757e-05, "loss": 4.0382, "step": 7202 }, { "epoch": 0.06502076187037371, "grad_norm": 4.357545375823975, "learning_rate": 4.674941325148944e-05, "loss": 4.5909, "step": 7203 }, { "epoch": 0.0650297887705362, "grad_norm": 3.5981314182281494, "learning_rate": 4.6748961906481313e-05, "loss": 3.8565, "step": 7204 }, { "epoch": 0.06503881567069868, "grad_norm": 3.5887036323547363, "learning_rate": 4.674851056147319e-05, "loss": 5.4167, "step": 7205 }, { "epoch": 0.06504784257086117, "grad_norm": 3.2993593215942383, "learning_rate": 4.6748059216465065e-05, "loss": 4.7446, "step": 7206 }, { "epoch": 0.06505686947102365, "grad_norm": 6.728585720062256, "learning_rate": 4.6747607871456944e-05, "loss": 4.7004, "step": 7207 }, { "epoch": 0.06506589637118614, "grad_norm": 4.292425155639648, "learning_rate": 4.6747156526448824e-05, "loss": 4.0391, "step": 7208 }, { "epoch": 0.06507492327134862, "grad_norm": 2.7971041202545166, "learning_rate": 4.6746705181440696e-05, "loss": 5.0373, "step": 7209 }, { "epoch": 0.0650839501715111, "grad_norm": 3.867292881011963, "learning_rate": 4.674625383643257e-05, "loss": 5.2936, "step": 7210 }, { "epoch": 0.0650929770716736, "grad_norm": 4.041898727416992, "learning_rate": 4.674580249142445e-05, "loss": 4.3508, "step": 7211 }, { "epoch": 0.06510200397183608, "grad_norm": 3.5521888732910156, "learning_rate": 4.674535114641632e-05, "loss": 4.6031, "step": 7212 }, { "epoch": 0.06511103087199856, "grad_norm": 3.1940317153930664, "learning_rate": 4.67448998014082e-05, "loss": 4.738, "step": 7213 }, { "epoch": 0.06512005777216104, "grad_norm": 3.476600408554077, "learning_rate": 4.674444845640007e-05, "loss": 4.2814, "step": 7214 }, { "epoch": 0.06512908467232352, "grad_norm": 6.677481174468994, "learning_rate": 4.674399711139195e-05, "loss": 4.8053, "step": 7215 }, { "epoch": 0.065138111572486, "grad_norm": 3.08152174949646, "learning_rate": 4.674354576638383e-05, "loss": 5.2727, "step": 7216 }, { "epoch": 0.0651471384726485, "grad_norm": 3.3846371173858643, "learning_rate": 4.6743094421375696e-05, "loss": 4.8424, "step": 7217 }, { "epoch": 0.06515616537281098, "grad_norm": 5.627836227416992, "learning_rate": 4.6742643076367575e-05, "loss": 4.3467, "step": 7218 }, { "epoch": 0.06516519227297346, "grad_norm": 3.8375372886657715, "learning_rate": 4.6742191731359455e-05, "loss": 4.7265, "step": 7219 }, { "epoch": 0.06517421917313594, "grad_norm": 4.49609899520874, "learning_rate": 4.674174038635133e-05, "loss": 4.2585, "step": 7220 }, { "epoch": 0.06518324607329842, "grad_norm": 5.7685322761535645, "learning_rate": 4.6741289041343206e-05, "loss": 4.4262, "step": 7221 }, { "epoch": 0.06519227297346092, "grad_norm": 4.209146976470947, "learning_rate": 4.6740837696335086e-05, "loss": 4.7658, "step": 7222 }, { "epoch": 0.0652012998736234, "grad_norm": 3.7373204231262207, "learning_rate": 4.674038635132696e-05, "loss": 4.0381, "step": 7223 }, { "epoch": 0.06521032677378588, "grad_norm": 4.319340229034424, "learning_rate": 4.673993500631883e-05, "loss": 5.0758, "step": 7224 }, { "epoch": 0.06521935367394836, "grad_norm": 4.499049186706543, "learning_rate": 4.673948366131071e-05, "loss": 4.9558, "step": 7225 }, { "epoch": 0.06522838057411084, "grad_norm": 3.2711751461029053, "learning_rate": 4.673903231630258e-05, "loss": 4.5519, "step": 7226 }, { "epoch": 0.06523740747427334, "grad_norm": 4.380200386047363, "learning_rate": 4.673858097129446e-05, "loss": 5.2226, "step": 7227 }, { "epoch": 0.06524643437443582, "grad_norm": 4.134674549102783, "learning_rate": 4.6738129626286334e-05, "loss": 4.3404, "step": 7228 }, { "epoch": 0.0652554612745983, "grad_norm": 3.7416679859161377, "learning_rate": 4.673767828127821e-05, "loss": 4.7205, "step": 7229 }, { "epoch": 0.06526448817476078, "grad_norm": 2.3816215991973877, "learning_rate": 4.673722693627009e-05, "loss": 4.8009, "step": 7230 }, { "epoch": 0.06527351507492327, "grad_norm": 3.210402727127075, "learning_rate": 4.673677559126196e-05, "loss": 4.3439, "step": 7231 }, { "epoch": 0.06528254197508576, "grad_norm": 3.6412031650543213, "learning_rate": 4.673632424625384e-05, "loss": 4.5061, "step": 7232 }, { "epoch": 0.06529156887524824, "grad_norm": 3.7803070545196533, "learning_rate": 4.6735872901245717e-05, "loss": 3.5908, "step": 7233 }, { "epoch": 0.06530059577541072, "grad_norm": 2.6741764545440674, "learning_rate": 4.673542155623759e-05, "loss": 4.9889, "step": 7234 }, { "epoch": 0.0653096226755732, "grad_norm": 3.2748358249664307, "learning_rate": 4.673497021122947e-05, "loss": 4.7019, "step": 7235 }, { "epoch": 0.06531864957573569, "grad_norm": 5.113977432250977, "learning_rate": 4.673451886622134e-05, "loss": 4.5757, "step": 7236 }, { "epoch": 0.06532767647589818, "grad_norm": 3.687612533569336, "learning_rate": 4.673406752121321e-05, "loss": 4.5803, "step": 7237 }, { "epoch": 0.06533670337606066, "grad_norm": 4.180397033691406, "learning_rate": 4.673361617620509e-05, "loss": 4.6076, "step": 7238 }, { "epoch": 0.06534573027622315, "grad_norm": 3.4573681354522705, "learning_rate": 4.673316483119697e-05, "loss": 4.5031, "step": 7239 }, { "epoch": 0.06535475717638563, "grad_norm": 4.648081302642822, "learning_rate": 4.6732713486188844e-05, "loss": 4.5131, "step": 7240 }, { "epoch": 0.06536378407654811, "grad_norm": 3.1557414531707764, "learning_rate": 4.673226214118072e-05, "loss": 5.2022, "step": 7241 }, { "epoch": 0.06537281097671059, "grad_norm": 3.5703468322753906, "learning_rate": 4.6731810796172596e-05, "loss": 4.3481, "step": 7242 }, { "epoch": 0.06538183787687309, "grad_norm": 3.326023817062378, "learning_rate": 4.6731359451164475e-05, "loss": 4.7237, "step": 7243 }, { "epoch": 0.06539086477703557, "grad_norm": 6.307315826416016, "learning_rate": 4.673090810615635e-05, "loss": 3.3526, "step": 7244 }, { "epoch": 0.06539989167719805, "grad_norm": 3.157796859741211, "learning_rate": 4.673045676114822e-05, "loss": 5.1301, "step": 7245 }, { "epoch": 0.06540891857736053, "grad_norm": 3.0984644889831543, "learning_rate": 4.67300054161401e-05, "loss": 4.4738, "step": 7246 }, { "epoch": 0.06541794547752301, "grad_norm": 3.0793256759643555, "learning_rate": 4.672955407113198e-05, "loss": 4.129, "step": 7247 }, { "epoch": 0.06542697237768551, "grad_norm": 3.33970046043396, "learning_rate": 4.672910272612385e-05, "loss": 4.2339, "step": 7248 }, { "epoch": 0.06543599927784799, "grad_norm": 4.302135467529297, "learning_rate": 4.672865138111573e-05, "loss": 4.8665, "step": 7249 }, { "epoch": 0.06544502617801047, "grad_norm": 3.4159677028656006, "learning_rate": 4.67282000361076e-05, "loss": 4.6314, "step": 7250 }, { "epoch": 0.06545405307817295, "grad_norm": 3.9666247367858887, "learning_rate": 4.6727748691099475e-05, "loss": 4.3805, "step": 7251 }, { "epoch": 0.06546307997833543, "grad_norm": 3.373180866241455, "learning_rate": 4.6727297346091354e-05, "loss": 4.664, "step": 7252 }, { "epoch": 0.06547210687849793, "grad_norm": 3.532845973968506, "learning_rate": 4.672684600108323e-05, "loss": 4.641, "step": 7253 }, { "epoch": 0.06548113377866041, "grad_norm": 3.8012537956237793, "learning_rate": 4.6726394656075106e-05, "loss": 4.6331, "step": 7254 }, { "epoch": 0.06549016067882289, "grad_norm": 3.333562135696411, "learning_rate": 4.6725943311066985e-05, "loss": 5.2087, "step": 7255 }, { "epoch": 0.06549918757898537, "grad_norm": 4.1398115158081055, "learning_rate": 4.672549196605886e-05, "loss": 3.8961, "step": 7256 }, { "epoch": 0.06550821447914786, "grad_norm": 4.3188700675964355, "learning_rate": 4.672504062105073e-05, "loss": 4.6504, "step": 7257 }, { "epoch": 0.06551724137931035, "grad_norm": 3.3146963119506836, "learning_rate": 4.672458927604261e-05, "loss": 4.8292, "step": 7258 }, { "epoch": 0.06552626827947283, "grad_norm": 3.4933886528015137, "learning_rate": 4.672413793103448e-05, "loss": 4.4451, "step": 7259 }, { "epoch": 0.06553529517963531, "grad_norm": 3.127575159072876, "learning_rate": 4.672368658602636e-05, "loss": 4.8986, "step": 7260 }, { "epoch": 0.0655443220797978, "grad_norm": 3.57073974609375, "learning_rate": 4.672323524101824e-05, "loss": 4.3415, "step": 7261 }, { "epoch": 0.06555334897996028, "grad_norm": 3.5964438915252686, "learning_rate": 4.672278389601011e-05, "loss": 4.1731, "step": 7262 }, { "epoch": 0.06556237588012277, "grad_norm": 4.041140556335449, "learning_rate": 4.672233255100199e-05, "loss": 3.9633, "step": 7263 }, { "epoch": 0.06557140278028525, "grad_norm": 3.590664863586426, "learning_rate": 4.6721881205993865e-05, "loss": 4.7375, "step": 7264 }, { "epoch": 0.06558042968044774, "grad_norm": 3.7701218128204346, "learning_rate": 4.672142986098574e-05, "loss": 4.5661, "step": 7265 }, { "epoch": 0.06558945658061022, "grad_norm": 3.2059288024902344, "learning_rate": 4.6720978515977616e-05, "loss": 3.9566, "step": 7266 }, { "epoch": 0.0655984834807727, "grad_norm": 5.1885666847229, "learning_rate": 4.672052717096949e-05, "loss": 4.6545, "step": 7267 }, { "epoch": 0.06560751038093518, "grad_norm": 4.052884101867676, "learning_rate": 4.672007582596137e-05, "loss": 4.5173, "step": 7268 }, { "epoch": 0.06561653728109768, "grad_norm": 3.596318006515503, "learning_rate": 4.671962448095325e-05, "loss": 4.3367, "step": 7269 }, { "epoch": 0.06562556418126016, "grad_norm": 3.5122671127319336, "learning_rate": 4.671917313594512e-05, "loss": 4.5739, "step": 7270 }, { "epoch": 0.06563459108142264, "grad_norm": 3.5331594944000244, "learning_rate": 4.671872179093699e-05, "loss": 5.7715, "step": 7271 }, { "epoch": 0.06564361798158512, "grad_norm": 4.473720073699951, "learning_rate": 4.671827044592887e-05, "loss": 5.1483, "step": 7272 }, { "epoch": 0.0656526448817476, "grad_norm": 3.810987949371338, "learning_rate": 4.6717819100920744e-05, "loss": 3.5204, "step": 7273 }, { "epoch": 0.0656616717819101, "grad_norm": 3.2105190753936768, "learning_rate": 4.671736775591262e-05, "loss": 4.414, "step": 7274 }, { "epoch": 0.06567069868207258, "grad_norm": 3.5620248317718506, "learning_rate": 4.6716916410904496e-05, "loss": 4.9917, "step": 7275 }, { "epoch": 0.06567972558223506, "grad_norm": 2.6674106121063232, "learning_rate": 4.6716465065896375e-05, "loss": 5.0627, "step": 7276 }, { "epoch": 0.06568875248239754, "grad_norm": 2.951796531677246, "learning_rate": 4.671601372088825e-05, "loss": 4.915, "step": 7277 }, { "epoch": 0.06569777938256002, "grad_norm": 3.573784589767456, "learning_rate": 4.671556237588012e-05, "loss": 4.4539, "step": 7278 }, { "epoch": 0.06570680628272252, "grad_norm": 6.432697296142578, "learning_rate": 4.6715111030872e-05, "loss": 4.1176, "step": 7279 }, { "epoch": 0.065715833182885, "grad_norm": 3.722686290740967, "learning_rate": 4.671465968586388e-05, "loss": 4.4191, "step": 7280 }, { "epoch": 0.06572486008304748, "grad_norm": 4.371450901031494, "learning_rate": 4.671420834085575e-05, "loss": 4.831, "step": 7281 }, { "epoch": 0.06573388698320996, "grad_norm": 4.768301486968994, "learning_rate": 4.671375699584763e-05, "loss": 4.2206, "step": 7282 }, { "epoch": 0.06574291388337244, "grad_norm": 3.9793388843536377, "learning_rate": 4.671330565083951e-05, "loss": 4.7262, "step": 7283 }, { "epoch": 0.06575194078353494, "grad_norm": 2.9058735370635986, "learning_rate": 4.6712854305831375e-05, "loss": 4.9318, "step": 7284 }, { "epoch": 0.06576096768369742, "grad_norm": 3.2955338954925537, "learning_rate": 4.6712402960823254e-05, "loss": 4.196, "step": 7285 }, { "epoch": 0.0657699945838599, "grad_norm": 3.2718887329101562, "learning_rate": 4.671195161581513e-05, "loss": 4.3943, "step": 7286 }, { "epoch": 0.06577902148402238, "grad_norm": 4.173015594482422, "learning_rate": 4.6711500270807006e-05, "loss": 4.675, "step": 7287 }, { "epoch": 0.06578804838418487, "grad_norm": 6.910688400268555, "learning_rate": 4.6711048925798885e-05, "loss": 4.016, "step": 7288 }, { "epoch": 0.06579707528434736, "grad_norm": 3.9680378437042236, "learning_rate": 4.671059758079076e-05, "loss": 4.9446, "step": 7289 }, { "epoch": 0.06580610218450984, "grad_norm": 3.3819360733032227, "learning_rate": 4.671014623578264e-05, "loss": 4.5067, "step": 7290 }, { "epoch": 0.06581512908467232, "grad_norm": 3.884910821914673, "learning_rate": 4.670969489077451e-05, "loss": 4.4018, "step": 7291 }, { "epoch": 0.0658241559848348, "grad_norm": 2.938577890396118, "learning_rate": 4.670924354576638e-05, "loss": 4.8057, "step": 7292 }, { "epoch": 0.06583318288499729, "grad_norm": 3.141239643096924, "learning_rate": 4.670879220075826e-05, "loss": 4.7049, "step": 7293 }, { "epoch": 0.06584220978515977, "grad_norm": 6.00602388381958, "learning_rate": 4.670834085575014e-05, "loss": 4.1956, "step": 7294 }, { "epoch": 0.06585123668532226, "grad_norm": 4.195224761962891, "learning_rate": 4.670788951074201e-05, "loss": 4.3714, "step": 7295 }, { "epoch": 0.06586026358548475, "grad_norm": 3.553236484527588, "learning_rate": 4.670743816573389e-05, "loss": 4.5135, "step": 7296 }, { "epoch": 0.06586929048564723, "grad_norm": 2.6551198959350586, "learning_rate": 4.6706986820725764e-05, "loss": 4.8907, "step": 7297 }, { "epoch": 0.06587831738580971, "grad_norm": 3.898707389831543, "learning_rate": 4.670653547571764e-05, "loss": 5.1952, "step": 7298 }, { "epoch": 0.06588734428597219, "grad_norm": 2.8265271186828613, "learning_rate": 4.6706084130709516e-05, "loss": 5.3409, "step": 7299 }, { "epoch": 0.06589637118613469, "grad_norm": 5.069133281707764, "learning_rate": 4.6705632785701395e-05, "loss": 4.0508, "step": 7300 }, { "epoch": 0.06590539808629717, "grad_norm": 3.3921642303466797, "learning_rate": 4.670518144069327e-05, "loss": 4.6124, "step": 7301 }, { "epoch": 0.06591442498645965, "grad_norm": 3.512275218963623, "learning_rate": 4.670473009568515e-05, "loss": 4.4937, "step": 7302 }, { "epoch": 0.06592345188662213, "grad_norm": 3.7475640773773193, "learning_rate": 4.670427875067702e-05, "loss": 4.4376, "step": 7303 }, { "epoch": 0.06593247878678461, "grad_norm": 3.175933599472046, "learning_rate": 4.670382740566889e-05, "loss": 4.7388, "step": 7304 }, { "epoch": 0.06594150568694711, "grad_norm": 3.1719048023223877, "learning_rate": 4.670337606066077e-05, "loss": 4.0818, "step": 7305 }, { "epoch": 0.06595053258710959, "grad_norm": 3.9833624362945557, "learning_rate": 4.6702924715652644e-05, "loss": 4.9922, "step": 7306 }, { "epoch": 0.06595955948727207, "grad_norm": 3.913773775100708, "learning_rate": 4.670247337064452e-05, "loss": 4.6925, "step": 7307 }, { "epoch": 0.06596858638743455, "grad_norm": 3.4446797370910645, "learning_rate": 4.67020220256364e-05, "loss": 4.8376, "step": 7308 }, { "epoch": 0.06597761328759703, "grad_norm": 4.871633052825928, "learning_rate": 4.6701570680628275e-05, "loss": 3.9039, "step": 7309 }, { "epoch": 0.06598664018775953, "grad_norm": 3.671469211578369, "learning_rate": 4.6701119335620154e-05, "loss": 4.6405, "step": 7310 }, { "epoch": 0.06599566708792201, "grad_norm": 3.670114278793335, "learning_rate": 4.6700667990612026e-05, "loss": 4.8727, "step": 7311 }, { "epoch": 0.06600469398808449, "grad_norm": 3.1913440227508545, "learning_rate": 4.67002166456039e-05, "loss": 4.5277, "step": 7312 }, { "epoch": 0.06601372088824697, "grad_norm": 3.554476499557495, "learning_rate": 4.669976530059578e-05, "loss": 4.8546, "step": 7313 }, { "epoch": 0.06602274778840946, "grad_norm": 3.746915102005005, "learning_rate": 4.669931395558765e-05, "loss": 4.9676, "step": 7314 }, { "epoch": 0.06603177468857195, "grad_norm": 3.6876838207244873, "learning_rate": 4.669886261057953e-05, "loss": 5.2116, "step": 7315 }, { "epoch": 0.06604080158873443, "grad_norm": 5.653449535369873, "learning_rate": 4.669841126557141e-05, "loss": 5.1264, "step": 7316 }, { "epoch": 0.06604982848889691, "grad_norm": 3.240682601928711, "learning_rate": 4.669795992056328e-05, "loss": 4.9956, "step": 7317 }, { "epoch": 0.0660588553890594, "grad_norm": 3.9867751598358154, "learning_rate": 4.6697508575555154e-05, "loss": 3.6673, "step": 7318 }, { "epoch": 0.06606788228922188, "grad_norm": 3.7656760215759277, "learning_rate": 4.669705723054703e-05, "loss": 5.0114, "step": 7319 }, { "epoch": 0.06607690918938436, "grad_norm": 3.8269755840301514, "learning_rate": 4.6696605885538906e-05, "loss": 4.2, "step": 7320 }, { "epoch": 0.06608593608954685, "grad_norm": 3.527554750442505, "learning_rate": 4.6696154540530785e-05, "loss": 4.7376, "step": 7321 }, { "epoch": 0.06609496298970934, "grad_norm": 5.27197265625, "learning_rate": 4.6695703195522664e-05, "loss": 4.7263, "step": 7322 }, { "epoch": 0.06610398988987182, "grad_norm": 3.391829490661621, "learning_rate": 4.6695251850514537e-05, "loss": 4.9953, "step": 7323 }, { "epoch": 0.0661130167900343, "grad_norm": 4.154562473297119, "learning_rate": 4.669480050550641e-05, "loss": 4.4624, "step": 7324 }, { "epoch": 0.06612204369019678, "grad_norm": 3.25657057762146, "learning_rate": 4.669434916049829e-05, "loss": 5.0353, "step": 7325 }, { "epoch": 0.06613107059035928, "grad_norm": 3.977414846420288, "learning_rate": 4.669389781549016e-05, "loss": 4.8491, "step": 7326 }, { "epoch": 0.06614009749052176, "grad_norm": 7.0864152908325195, "learning_rate": 4.669344647048204e-05, "loss": 5.2407, "step": 7327 }, { "epoch": 0.06614912439068424, "grad_norm": 3.4155683517456055, "learning_rate": 4.669299512547391e-05, "loss": 3.9739, "step": 7328 }, { "epoch": 0.06615815129084672, "grad_norm": 3.5353667736053467, "learning_rate": 4.669254378046579e-05, "loss": 5.2474, "step": 7329 }, { "epoch": 0.0661671781910092, "grad_norm": 4.7584075927734375, "learning_rate": 4.669209243545767e-05, "loss": 3.8821, "step": 7330 }, { "epoch": 0.0661762050911717, "grad_norm": 3.088505983352661, "learning_rate": 4.6691641090449537e-05, "loss": 4.7609, "step": 7331 }, { "epoch": 0.06618523199133418, "grad_norm": 3.8194620609283447, "learning_rate": 4.6691189745441416e-05, "loss": 4.5149, "step": 7332 }, { "epoch": 0.06619425889149666, "grad_norm": 3.7749297618865967, "learning_rate": 4.6690738400433295e-05, "loss": 4.5341, "step": 7333 }, { "epoch": 0.06620328579165914, "grad_norm": 3.0625438690185547, "learning_rate": 4.669028705542517e-05, "loss": 4.824, "step": 7334 }, { "epoch": 0.06621231269182162, "grad_norm": 4.61579704284668, "learning_rate": 4.668983571041705e-05, "loss": 5.2181, "step": 7335 }, { "epoch": 0.06622133959198412, "grad_norm": 3.8303353786468506, "learning_rate": 4.668938436540892e-05, "loss": 4.2854, "step": 7336 }, { "epoch": 0.0662303664921466, "grad_norm": 3.7371935844421387, "learning_rate": 4.66889330204008e-05, "loss": 4.5593, "step": 7337 }, { "epoch": 0.06623939339230908, "grad_norm": 3.8577916622161865, "learning_rate": 4.668848167539267e-05, "loss": 4.8359, "step": 7338 }, { "epoch": 0.06624842029247156, "grad_norm": 2.561638593673706, "learning_rate": 4.6688030330384543e-05, "loss": 5.3059, "step": 7339 }, { "epoch": 0.06625744719263404, "grad_norm": 3.1469690799713135, "learning_rate": 4.668757898537642e-05, "loss": 5.2175, "step": 7340 }, { "epoch": 0.06626647409279654, "grad_norm": 4.53573751449585, "learning_rate": 4.66871276403683e-05, "loss": 3.8394, "step": 7341 }, { "epoch": 0.06627550099295902, "grad_norm": 3.230356454849243, "learning_rate": 4.6686676295360174e-05, "loss": 4.8245, "step": 7342 }, { "epoch": 0.0662845278931215, "grad_norm": 3.594715118408203, "learning_rate": 4.6686224950352054e-05, "loss": 4.6515, "step": 7343 }, { "epoch": 0.06629355479328398, "grad_norm": 4.583065986633301, "learning_rate": 4.668577360534393e-05, "loss": 4.8918, "step": 7344 }, { "epoch": 0.06630258169344647, "grad_norm": 2.9132542610168457, "learning_rate": 4.66853222603358e-05, "loss": 4.8338, "step": 7345 }, { "epoch": 0.06631160859360895, "grad_norm": 2.9658873081207275, "learning_rate": 4.668487091532768e-05, "loss": 4.8841, "step": 7346 }, { "epoch": 0.06632063549377144, "grad_norm": 4.177502155303955, "learning_rate": 4.668441957031956e-05, "loss": 4.5067, "step": 7347 }, { "epoch": 0.06632966239393392, "grad_norm": 3.489905834197998, "learning_rate": 4.668396822531143e-05, "loss": 4.8852, "step": 7348 }, { "epoch": 0.0663386892940964, "grad_norm": 5.0370917320251465, "learning_rate": 4.668351688030331e-05, "loss": 4.8309, "step": 7349 }, { "epoch": 0.06634771619425889, "grad_norm": 4.5802435874938965, "learning_rate": 4.668306553529518e-05, "loss": 4.7311, "step": 7350 }, { "epoch": 0.06635674309442137, "grad_norm": 2.9905169010162354, "learning_rate": 4.6682614190287054e-05, "loss": 4.6911, "step": 7351 }, { "epoch": 0.06636576999458386, "grad_norm": 3.8762333393096924, "learning_rate": 4.668216284527893e-05, "loss": 5.0597, "step": 7352 }, { "epoch": 0.06637479689474635, "grad_norm": 3.0775580406188965, "learning_rate": 4.6681711500270805e-05, "loss": 4.7371, "step": 7353 }, { "epoch": 0.06638382379490883, "grad_norm": 4.6911163330078125, "learning_rate": 4.6681260155262685e-05, "loss": 5.1545, "step": 7354 }, { "epoch": 0.06639285069507131, "grad_norm": 3.8423399925231934, "learning_rate": 4.6680808810254564e-05, "loss": 4.464, "step": 7355 }, { "epoch": 0.06640187759523379, "grad_norm": 3.67838191986084, "learning_rate": 4.6680357465246436e-05, "loss": 4.3056, "step": 7356 }, { "epoch": 0.06641090449539629, "grad_norm": 3.593843936920166, "learning_rate": 4.6679906120238316e-05, "loss": 4.4278, "step": 7357 }, { "epoch": 0.06641993139555877, "grad_norm": 2.7377700805664062, "learning_rate": 4.667945477523019e-05, "loss": 4.8378, "step": 7358 }, { "epoch": 0.06642895829572125, "grad_norm": 4.059517860412598, "learning_rate": 4.667900343022206e-05, "loss": 4.7641, "step": 7359 }, { "epoch": 0.06643798519588373, "grad_norm": 3.318969488143921, "learning_rate": 4.667855208521394e-05, "loss": 5.1352, "step": 7360 }, { "epoch": 0.06644701209604621, "grad_norm": 2.624145984649658, "learning_rate": 4.667810074020582e-05, "loss": 4.6818, "step": 7361 }, { "epoch": 0.06645603899620871, "grad_norm": 3.6028411388397217, "learning_rate": 4.667764939519769e-05, "loss": 4.5667, "step": 7362 }, { "epoch": 0.06646506589637119, "grad_norm": 4.970089912414551, "learning_rate": 4.667719805018957e-05, "loss": 3.7474, "step": 7363 }, { "epoch": 0.06647409279653367, "grad_norm": 3.5147206783294678, "learning_rate": 4.667674670518144e-05, "loss": 4.5731, "step": 7364 }, { "epoch": 0.06648311969669615, "grad_norm": 3.293118715286255, "learning_rate": 4.6676295360173316e-05, "loss": 4.4235, "step": 7365 }, { "epoch": 0.06649214659685863, "grad_norm": 3.1965272426605225, "learning_rate": 4.6675844015165195e-05, "loss": 4.8208, "step": 7366 }, { "epoch": 0.06650117349702113, "grad_norm": 4.645400524139404, "learning_rate": 4.667539267015707e-05, "loss": 5.2368, "step": 7367 }, { "epoch": 0.06651020039718361, "grad_norm": 3.750589370727539, "learning_rate": 4.6674941325148947e-05, "loss": 4.3078, "step": 7368 }, { "epoch": 0.06651922729734609, "grad_norm": 3.641045093536377, "learning_rate": 4.6674489980140826e-05, "loss": 4.8278, "step": 7369 }, { "epoch": 0.06652825419750857, "grad_norm": 3.871427297592163, "learning_rate": 4.66740386351327e-05, "loss": 4.2915, "step": 7370 }, { "epoch": 0.06653728109767106, "grad_norm": 4.439202308654785, "learning_rate": 4.667358729012457e-05, "loss": 4.3124, "step": 7371 }, { "epoch": 0.06654630799783355, "grad_norm": 5.851142406463623, "learning_rate": 4.667313594511645e-05, "loss": 3.9404, "step": 7372 }, { "epoch": 0.06655533489799603, "grad_norm": 3.319091320037842, "learning_rate": 4.667268460010832e-05, "loss": 4.7023, "step": 7373 }, { "epoch": 0.06656436179815851, "grad_norm": 3.7914698123931885, "learning_rate": 4.66722332551002e-05, "loss": 4.2963, "step": 7374 }, { "epoch": 0.066573388698321, "grad_norm": 3.9418563842773438, "learning_rate": 4.6671781910092074e-05, "loss": 5.2852, "step": 7375 }, { "epoch": 0.06658241559848348, "grad_norm": 3.941542863845825, "learning_rate": 4.667133056508395e-05, "loss": 5.0269, "step": 7376 }, { "epoch": 0.06659144249864596, "grad_norm": 3.899106979370117, "learning_rate": 4.667087922007583e-05, "loss": 4.804, "step": 7377 }, { "epoch": 0.06660046939880845, "grad_norm": 3.6181721687316895, "learning_rate": 4.66704278750677e-05, "loss": 5.0341, "step": 7378 }, { "epoch": 0.06660949629897094, "grad_norm": 3.3074493408203125, "learning_rate": 4.666997653005958e-05, "loss": 4.9379, "step": 7379 }, { "epoch": 0.06661852319913342, "grad_norm": 3.5475313663482666, "learning_rate": 4.666952518505146e-05, "loss": 4.5637, "step": 7380 }, { "epoch": 0.0666275500992959, "grad_norm": 4.243101119995117, "learning_rate": 4.666907384004333e-05, "loss": 4.6189, "step": 7381 }, { "epoch": 0.06663657699945838, "grad_norm": 4.04753303527832, "learning_rate": 4.666862249503521e-05, "loss": 4.337, "step": 7382 }, { "epoch": 0.06664560389962088, "grad_norm": 3.6710762977600098, "learning_rate": 4.666817115002709e-05, "loss": 5.1811, "step": 7383 }, { "epoch": 0.06665463079978336, "grad_norm": 3.390346050262451, "learning_rate": 4.666771980501896e-05, "loss": 4.9707, "step": 7384 }, { "epoch": 0.06666365769994584, "grad_norm": 4.034848213195801, "learning_rate": 4.666726846001083e-05, "loss": 4.5664, "step": 7385 }, { "epoch": 0.06667268460010832, "grad_norm": 3.018874168395996, "learning_rate": 4.666681711500271e-05, "loss": 4.7038, "step": 7386 }, { "epoch": 0.0666817115002708, "grad_norm": 2.5760202407836914, "learning_rate": 4.6666365769994584e-05, "loss": 4.8188, "step": 7387 }, { "epoch": 0.0666907384004333, "grad_norm": 3.9002695083618164, "learning_rate": 4.6665914424986464e-05, "loss": 5.4957, "step": 7388 }, { "epoch": 0.06669976530059578, "grad_norm": 4.891604423522949, "learning_rate": 4.6665463079978336e-05, "loss": 5.1974, "step": 7389 }, { "epoch": 0.06670879220075826, "grad_norm": 3.1753649711608887, "learning_rate": 4.6665011734970215e-05, "loss": 4.5748, "step": 7390 }, { "epoch": 0.06671781910092074, "grad_norm": 3.653787136077881, "learning_rate": 4.6664560389962095e-05, "loss": 4.4815, "step": 7391 }, { "epoch": 0.06672684600108322, "grad_norm": 4.736242294311523, "learning_rate": 4.666410904495396e-05, "loss": 4.2299, "step": 7392 }, { "epoch": 0.06673587290124572, "grad_norm": 2.881544351577759, "learning_rate": 4.666365769994584e-05, "loss": 4.2242, "step": 7393 }, { "epoch": 0.0667448998014082, "grad_norm": 3.5678796768188477, "learning_rate": 4.666320635493772e-05, "loss": 4.8633, "step": 7394 }, { "epoch": 0.06675392670157068, "grad_norm": 4.472013473510742, "learning_rate": 4.666275500992959e-05, "loss": 3.7476, "step": 7395 }, { "epoch": 0.06676295360173316, "grad_norm": 3.755150079727173, "learning_rate": 4.666230366492147e-05, "loss": 4.0645, "step": 7396 }, { "epoch": 0.06677198050189564, "grad_norm": 3.051609754562378, "learning_rate": 4.666185231991334e-05, "loss": 4.8093, "step": 7397 }, { "epoch": 0.06678100740205814, "grad_norm": 5.030146598815918, "learning_rate": 4.6661400974905215e-05, "loss": 4.9003, "step": 7398 }, { "epoch": 0.06679003430222062, "grad_norm": 2.7502596378326416, "learning_rate": 4.6660949629897095e-05, "loss": 5.0527, "step": 7399 }, { "epoch": 0.0667990612023831, "grad_norm": 2.7148427963256836, "learning_rate": 4.6660498284888974e-05, "loss": 4.6843, "step": 7400 }, { "epoch": 0.06680808810254558, "grad_norm": 3.4205989837646484, "learning_rate": 4.6660046939880846e-05, "loss": 4.2243, "step": 7401 }, { "epoch": 0.06681711500270807, "grad_norm": 2.7396085262298584, "learning_rate": 4.6659595594872726e-05, "loss": 4.7478, "step": 7402 }, { "epoch": 0.06682614190287055, "grad_norm": 3.1774191856384277, "learning_rate": 4.66591442498646e-05, "loss": 5.4772, "step": 7403 }, { "epoch": 0.06683516880303304, "grad_norm": 2.939692974090576, "learning_rate": 4.665869290485648e-05, "loss": 4.5083, "step": 7404 }, { "epoch": 0.06684419570319552, "grad_norm": 4.509622573852539, "learning_rate": 4.665824155984835e-05, "loss": 3.8058, "step": 7405 }, { "epoch": 0.066853222603358, "grad_norm": 4.098762512207031, "learning_rate": 4.665779021484022e-05, "loss": 4.261, "step": 7406 }, { "epoch": 0.06686224950352049, "grad_norm": 3.9924161434173584, "learning_rate": 4.66573388698321e-05, "loss": 5.2762, "step": 7407 }, { "epoch": 0.06687127640368297, "grad_norm": 2.339094638824463, "learning_rate": 4.665688752482398e-05, "loss": 4.4626, "step": 7408 }, { "epoch": 0.06688030330384546, "grad_norm": 2.712385416030884, "learning_rate": 4.665643617981585e-05, "loss": 4.1221, "step": 7409 }, { "epoch": 0.06688933020400795, "grad_norm": 3.673612356185913, "learning_rate": 4.665598483480773e-05, "loss": 4.4662, "step": 7410 }, { "epoch": 0.06689835710417043, "grad_norm": 3.820929527282715, "learning_rate": 4.6655533489799605e-05, "loss": 4.9508, "step": 7411 }, { "epoch": 0.06690738400433291, "grad_norm": 3.817647695541382, "learning_rate": 4.665508214479148e-05, "loss": 4.4886, "step": 7412 }, { "epoch": 0.06691641090449539, "grad_norm": 3.7573435306549072, "learning_rate": 4.6654630799783357e-05, "loss": 4.5011, "step": 7413 }, { "epoch": 0.06692543780465789, "grad_norm": 2.469740867614746, "learning_rate": 4.665417945477523e-05, "loss": 4.9281, "step": 7414 }, { "epoch": 0.06693446470482037, "grad_norm": 4.014022350311279, "learning_rate": 4.665372810976711e-05, "loss": 4.515, "step": 7415 }, { "epoch": 0.06694349160498285, "grad_norm": 3.2026753425598145, "learning_rate": 4.665327676475899e-05, "loss": 4.6923, "step": 7416 }, { "epoch": 0.06695251850514533, "grad_norm": 2.861842632293701, "learning_rate": 4.665282541975086e-05, "loss": 4.5969, "step": 7417 }, { "epoch": 0.06696154540530781, "grad_norm": 3.3181309700012207, "learning_rate": 4.665237407474273e-05, "loss": 5.3119, "step": 7418 }, { "epoch": 0.06697057230547031, "grad_norm": 5.628209590911865, "learning_rate": 4.665192272973461e-05, "loss": 4.7615, "step": 7419 }, { "epoch": 0.06697959920563279, "grad_norm": 3.791247606277466, "learning_rate": 4.6651471384726484e-05, "loss": 4.6296, "step": 7420 }, { "epoch": 0.06698862610579527, "grad_norm": 2.4110889434814453, "learning_rate": 4.665102003971836e-05, "loss": 4.5782, "step": 7421 }, { "epoch": 0.06699765300595775, "grad_norm": 3.7053701877593994, "learning_rate": 4.665056869471024e-05, "loss": 5.2917, "step": 7422 }, { "epoch": 0.06700667990612023, "grad_norm": 3.183018922805786, "learning_rate": 4.6650117349702115e-05, "loss": 5.0155, "step": 7423 }, { "epoch": 0.06701570680628273, "grad_norm": 4.653709411621094, "learning_rate": 4.6649666004693994e-05, "loss": 4.5716, "step": 7424 }, { "epoch": 0.06702473370644521, "grad_norm": 3.5124030113220215, "learning_rate": 4.664921465968587e-05, "loss": 4.4601, "step": 7425 }, { "epoch": 0.06703376060660769, "grad_norm": 4.0398993492126465, "learning_rate": 4.664876331467774e-05, "loss": 4.7458, "step": 7426 }, { "epoch": 0.06704278750677017, "grad_norm": 3.3738670349121094, "learning_rate": 4.664831196966962e-05, "loss": 4.6659, "step": 7427 }, { "epoch": 0.06705181440693266, "grad_norm": 3.6787045001983643, "learning_rate": 4.664786062466149e-05, "loss": 4.2098, "step": 7428 }, { "epoch": 0.06706084130709514, "grad_norm": 3.5089001655578613, "learning_rate": 4.664740927965337e-05, "loss": 4.4616, "step": 7429 }, { "epoch": 0.06706986820725763, "grad_norm": 3.294485092163086, "learning_rate": 4.664695793464525e-05, "loss": 4.8376, "step": 7430 }, { "epoch": 0.06707889510742011, "grad_norm": 3.9527668952941895, "learning_rate": 4.664650658963712e-05, "loss": 4.3635, "step": 7431 }, { "epoch": 0.0670879220075826, "grad_norm": 3.5919933319091797, "learning_rate": 4.6646055244628994e-05, "loss": 4.1625, "step": 7432 }, { "epoch": 0.06709694890774508, "grad_norm": 4.679066181182861, "learning_rate": 4.6645603899620874e-05, "loss": 4.7601, "step": 7433 }, { "epoch": 0.06710597580790756, "grad_norm": 3.3902194499969482, "learning_rate": 4.6645152554612746e-05, "loss": 4.9994, "step": 7434 }, { "epoch": 0.06711500270807005, "grad_norm": 3.649254083633423, "learning_rate": 4.6644701209604625e-05, "loss": 4.4637, "step": 7435 }, { "epoch": 0.06712402960823254, "grad_norm": 4.626873970031738, "learning_rate": 4.66442498645965e-05, "loss": 4.8909, "step": 7436 }, { "epoch": 0.06713305650839502, "grad_norm": 3.4310333728790283, "learning_rate": 4.664379851958838e-05, "loss": 4.7101, "step": 7437 }, { "epoch": 0.0671420834085575, "grad_norm": 3.7138330936431885, "learning_rate": 4.664334717458025e-05, "loss": 4.4257, "step": 7438 }, { "epoch": 0.06715111030871998, "grad_norm": 4.901238918304443, "learning_rate": 4.664289582957212e-05, "loss": 4.9263, "step": 7439 }, { "epoch": 0.06716013720888248, "grad_norm": 5.608785629272461, "learning_rate": 4.6642444484564e-05, "loss": 5.1751, "step": 7440 }, { "epoch": 0.06716916410904496, "grad_norm": 4.226010799407959, "learning_rate": 4.664199313955588e-05, "loss": 4.6717, "step": 7441 }, { "epoch": 0.06717819100920744, "grad_norm": 3.2731103897094727, "learning_rate": 4.664154179454775e-05, "loss": 4.7074, "step": 7442 }, { "epoch": 0.06718721790936992, "grad_norm": 2.9253361225128174, "learning_rate": 4.664109044953963e-05, "loss": 5.178, "step": 7443 }, { "epoch": 0.0671962448095324, "grad_norm": 4.080262184143066, "learning_rate": 4.664063910453151e-05, "loss": 4.5728, "step": 7444 }, { "epoch": 0.0672052717096949, "grad_norm": 3.6819939613342285, "learning_rate": 4.664018775952338e-05, "loss": 4.5915, "step": 7445 }, { "epoch": 0.06721429860985738, "grad_norm": 4.076644420623779, "learning_rate": 4.6639736414515256e-05, "loss": 4.1381, "step": 7446 }, { "epoch": 0.06722332551001986, "grad_norm": 3.4561867713928223, "learning_rate": 4.6639285069507136e-05, "loss": 5.0644, "step": 7447 }, { "epoch": 0.06723235241018234, "grad_norm": 3.8556010723114014, "learning_rate": 4.663883372449901e-05, "loss": 4.4665, "step": 7448 }, { "epoch": 0.06724137931034482, "grad_norm": 3.8505637645721436, "learning_rate": 4.663838237949089e-05, "loss": 4.8583, "step": 7449 }, { "epoch": 0.06725040621050732, "grad_norm": 3.683403491973877, "learning_rate": 4.663793103448276e-05, "loss": 4.6704, "step": 7450 }, { "epoch": 0.0672594331106698, "grad_norm": 3.479522466659546, "learning_rate": 4.663747968947464e-05, "loss": 4.7697, "step": 7451 }, { "epoch": 0.06726846001083228, "grad_norm": 3.333137273788452, "learning_rate": 4.663702834446651e-05, "loss": 4.4208, "step": 7452 }, { "epoch": 0.06727748691099476, "grad_norm": 4.030351638793945, "learning_rate": 4.6636576999458384e-05, "loss": 4.1426, "step": 7453 }, { "epoch": 0.06728651381115724, "grad_norm": 4.166952133178711, "learning_rate": 4.663612565445026e-05, "loss": 4.8698, "step": 7454 }, { "epoch": 0.06729554071131973, "grad_norm": 3.481732130050659, "learning_rate": 4.663567430944214e-05, "loss": 5.1768, "step": 7455 }, { "epoch": 0.06730456761148222, "grad_norm": 3.812450647354126, "learning_rate": 4.6635222964434015e-05, "loss": 4.7425, "step": 7456 }, { "epoch": 0.0673135945116447, "grad_norm": 3.9003000259399414, "learning_rate": 4.6634771619425894e-05, "loss": 4.9688, "step": 7457 }, { "epoch": 0.06732262141180718, "grad_norm": 3.0878894329071045, "learning_rate": 4.6634320274417766e-05, "loss": 5.115, "step": 7458 }, { "epoch": 0.06733164831196967, "grad_norm": 3.533963203430176, "learning_rate": 4.663386892940964e-05, "loss": 4.8893, "step": 7459 }, { "epoch": 0.06734067521213215, "grad_norm": 3.745898485183716, "learning_rate": 4.663341758440152e-05, "loss": 4.7697, "step": 7460 }, { "epoch": 0.06734970211229464, "grad_norm": 4.724710941314697, "learning_rate": 4.66329662393934e-05, "loss": 4.3632, "step": 7461 }, { "epoch": 0.06735872901245712, "grad_norm": 3.446186065673828, "learning_rate": 4.663251489438527e-05, "loss": 4.614, "step": 7462 }, { "epoch": 0.0673677559126196, "grad_norm": 2.939695358276367, "learning_rate": 4.663206354937715e-05, "loss": 4.596, "step": 7463 }, { "epoch": 0.06737678281278209, "grad_norm": 4.386322498321533, "learning_rate": 4.663161220436902e-05, "loss": 4.7573, "step": 7464 }, { "epoch": 0.06738580971294457, "grad_norm": 3.2776200771331787, "learning_rate": 4.6631160859360894e-05, "loss": 5.1311, "step": 7465 }, { "epoch": 0.06739483661310706, "grad_norm": 3.7876033782958984, "learning_rate": 4.663070951435277e-05, "loss": 4.9492, "step": 7466 }, { "epoch": 0.06740386351326955, "grad_norm": 2.883434295654297, "learning_rate": 4.6630258169344646e-05, "loss": 4.6622, "step": 7467 }, { "epoch": 0.06741289041343203, "grad_norm": 3.6707050800323486, "learning_rate": 4.6629806824336525e-05, "loss": 4.603, "step": 7468 }, { "epoch": 0.06742191731359451, "grad_norm": 2.944802761077881, "learning_rate": 4.6629355479328404e-05, "loss": 4.4764, "step": 7469 }, { "epoch": 0.06743094421375699, "grad_norm": 3.2092058658599854, "learning_rate": 4.662890413432028e-05, "loss": 3.6753, "step": 7470 }, { "epoch": 0.06743997111391949, "grad_norm": 4.304442882537842, "learning_rate": 4.6628452789312156e-05, "loss": 5.0695, "step": 7471 }, { "epoch": 0.06744899801408197, "grad_norm": 4.005500793457031, "learning_rate": 4.662800144430403e-05, "loss": 5.3963, "step": 7472 }, { "epoch": 0.06745802491424445, "grad_norm": 4.996784687042236, "learning_rate": 4.66275500992959e-05, "loss": 4.8255, "step": 7473 }, { "epoch": 0.06746705181440693, "grad_norm": 5.124020099639893, "learning_rate": 4.662709875428778e-05, "loss": 3.3521, "step": 7474 }, { "epoch": 0.06747607871456941, "grad_norm": 3.5014727115631104, "learning_rate": 4.662664740927965e-05, "loss": 3.9501, "step": 7475 }, { "epoch": 0.06748510561473191, "grad_norm": 3.898608922958374, "learning_rate": 4.662619606427153e-05, "loss": 4.3612, "step": 7476 }, { "epoch": 0.06749413251489439, "grad_norm": 3.405350685119629, "learning_rate": 4.662574471926341e-05, "loss": 4.2381, "step": 7477 }, { "epoch": 0.06750315941505687, "grad_norm": 2.586223840713501, "learning_rate": 4.6625293374255284e-05, "loss": 4.384, "step": 7478 }, { "epoch": 0.06751218631521935, "grad_norm": 4.356416702270508, "learning_rate": 4.6624842029247156e-05, "loss": 4.8782, "step": 7479 }, { "epoch": 0.06752121321538183, "grad_norm": 3.723505973815918, "learning_rate": 4.6624390684239035e-05, "loss": 4.6654, "step": 7480 }, { "epoch": 0.06753024011554432, "grad_norm": 3.419954299926758, "learning_rate": 4.662393933923091e-05, "loss": 4.9779, "step": 7481 }, { "epoch": 0.06753926701570681, "grad_norm": 3.215726852416992, "learning_rate": 4.662348799422279e-05, "loss": 4.8263, "step": 7482 }, { "epoch": 0.06754829391586929, "grad_norm": 4.083481311798096, "learning_rate": 4.6623036649214666e-05, "loss": 4.6278, "step": 7483 }, { "epoch": 0.06755732081603177, "grad_norm": 2.7879674434661865, "learning_rate": 4.662258530420654e-05, "loss": 4.9945, "step": 7484 }, { "epoch": 0.06756634771619426, "grad_norm": 4.368890285491943, "learning_rate": 4.662213395919841e-05, "loss": 3.7301, "step": 7485 }, { "epoch": 0.06757537461635674, "grad_norm": 3.7449703216552734, "learning_rate": 4.662168261419029e-05, "loss": 4.1559, "step": 7486 }, { "epoch": 0.06758440151651923, "grad_norm": 4.223385810852051, "learning_rate": 4.662123126918216e-05, "loss": 4.2469, "step": 7487 }, { "epoch": 0.06759342841668171, "grad_norm": 3.117063522338867, "learning_rate": 4.662077992417404e-05, "loss": 4.6306, "step": 7488 }, { "epoch": 0.0676024553168442, "grad_norm": 3.386733055114746, "learning_rate": 4.6620328579165915e-05, "loss": 4.7808, "step": 7489 }, { "epoch": 0.06761148221700668, "grad_norm": 2.8121206760406494, "learning_rate": 4.6619877234157794e-05, "loss": 4.9871, "step": 7490 }, { "epoch": 0.06762050911716916, "grad_norm": 4.020061016082764, "learning_rate": 4.661942588914967e-05, "loss": 4.2685, "step": 7491 }, { "epoch": 0.06762953601733165, "grad_norm": 3.5182621479034424, "learning_rate": 4.661897454414154e-05, "loss": 4.9057, "step": 7492 }, { "epoch": 0.06763856291749414, "grad_norm": 4.211935520172119, "learning_rate": 4.661852319913342e-05, "loss": 4.2593, "step": 7493 }, { "epoch": 0.06764758981765662, "grad_norm": 2.876373529434204, "learning_rate": 4.66180718541253e-05, "loss": 4.8751, "step": 7494 }, { "epoch": 0.0676566167178191, "grad_norm": 4.0738911628723145, "learning_rate": 4.661762050911717e-05, "loss": 4.199, "step": 7495 }, { "epoch": 0.06766564361798158, "grad_norm": 3.7163915634155273, "learning_rate": 4.661716916410905e-05, "loss": 4.1821, "step": 7496 }, { "epoch": 0.06767467051814408, "grad_norm": 4.426020622253418, "learning_rate": 4.661671781910092e-05, "loss": 4.5628, "step": 7497 }, { "epoch": 0.06768369741830656, "grad_norm": 3.007641077041626, "learning_rate": 4.66162664740928e-05, "loss": 4.9889, "step": 7498 }, { "epoch": 0.06769272431846904, "grad_norm": 3.982405185699463, "learning_rate": 4.661581512908467e-05, "loss": 4.2592, "step": 7499 }, { "epoch": 0.06770175121863152, "grad_norm": 2.411383867263794, "learning_rate": 4.6615363784076546e-05, "loss": 4.9749, "step": 7500 }, { "epoch": 0.067710778118794, "grad_norm": 2.2800564765930176, "learning_rate": 4.6614912439068425e-05, "loss": 5.3706, "step": 7501 }, { "epoch": 0.0677198050189565, "grad_norm": 4.724740505218506, "learning_rate": 4.6614461094060304e-05, "loss": 4.7727, "step": 7502 }, { "epoch": 0.06772883191911898, "grad_norm": 4.18139123916626, "learning_rate": 4.6614009749052176e-05, "loss": 5.0836, "step": 7503 }, { "epoch": 0.06773785881928146, "grad_norm": 2.9247992038726807, "learning_rate": 4.6613558404044056e-05, "loss": 4.9903, "step": 7504 }, { "epoch": 0.06774688571944394, "grad_norm": 2.815865993499756, "learning_rate": 4.6613107059035935e-05, "loss": 5.0709, "step": 7505 }, { "epoch": 0.06775591261960642, "grad_norm": 3.776252508163452, "learning_rate": 4.66126557140278e-05, "loss": 4.8455, "step": 7506 }, { "epoch": 0.0677649395197689, "grad_norm": 3.53583025932312, "learning_rate": 4.661220436901968e-05, "loss": 4.9795, "step": 7507 }, { "epoch": 0.0677739664199314, "grad_norm": 3.524467706680298, "learning_rate": 4.661175302401156e-05, "loss": 4.0189, "step": 7508 }, { "epoch": 0.06778299332009388, "grad_norm": 3.544867992401123, "learning_rate": 4.661130167900343e-05, "loss": 4.4748, "step": 7509 }, { "epoch": 0.06779202022025636, "grad_norm": 3.223742961883545, "learning_rate": 4.661085033399531e-05, "loss": 4.8552, "step": 7510 }, { "epoch": 0.06780104712041884, "grad_norm": 3.3340630531311035, "learning_rate": 4.661039898898718e-05, "loss": 4.297, "step": 7511 }, { "epoch": 0.06781007402058133, "grad_norm": 3.8375604152679443, "learning_rate": 4.6609947643979056e-05, "loss": 4.5541, "step": 7512 }, { "epoch": 0.06781910092074382, "grad_norm": 3.9800236225128174, "learning_rate": 4.6609496298970935e-05, "loss": 4.3001, "step": 7513 }, { "epoch": 0.0678281278209063, "grad_norm": 3.512427568435669, "learning_rate": 4.660904495396281e-05, "loss": 4.2261, "step": 7514 }, { "epoch": 0.06783715472106878, "grad_norm": 3.681434392929077, "learning_rate": 4.660859360895469e-05, "loss": 5.0814, "step": 7515 }, { "epoch": 0.06784618162123127, "grad_norm": 4.392979621887207, "learning_rate": 4.6608142263946566e-05, "loss": 5.2095, "step": 7516 }, { "epoch": 0.06785520852139375, "grad_norm": 3.4805819988250732, "learning_rate": 4.660769091893844e-05, "loss": 4.4922, "step": 7517 }, { "epoch": 0.06786423542155624, "grad_norm": 4.131979942321777, "learning_rate": 4.660723957393032e-05, "loss": 5.0889, "step": 7518 }, { "epoch": 0.06787326232171872, "grad_norm": 3.8552043437957764, "learning_rate": 4.660678822892219e-05, "loss": 4.4894, "step": 7519 }, { "epoch": 0.0678822892218812, "grad_norm": 3.532613754272461, "learning_rate": 4.660633688391406e-05, "loss": 5.0015, "step": 7520 }, { "epoch": 0.06789131612204369, "grad_norm": 5.054487705230713, "learning_rate": 4.660588553890594e-05, "loss": 4.5815, "step": 7521 }, { "epoch": 0.06790034302220617, "grad_norm": 4.644036769866943, "learning_rate": 4.660543419389782e-05, "loss": 5.0692, "step": 7522 }, { "epoch": 0.06790936992236866, "grad_norm": 3.2636005878448486, "learning_rate": 4.6604982848889694e-05, "loss": 4.8127, "step": 7523 }, { "epoch": 0.06791839682253115, "grad_norm": 3.9190587997436523, "learning_rate": 4.660453150388157e-05, "loss": 4.6819, "step": 7524 }, { "epoch": 0.06792742372269363, "grad_norm": 3.933187961578369, "learning_rate": 4.6604080158873445e-05, "loss": 4.4938, "step": 7525 }, { "epoch": 0.06793645062285611, "grad_norm": 4.574535846710205, "learning_rate": 4.660362881386532e-05, "loss": 4.564, "step": 7526 }, { "epoch": 0.06794547752301859, "grad_norm": 3.981309652328491, "learning_rate": 4.66031774688572e-05, "loss": 4.6204, "step": 7527 }, { "epoch": 0.06795450442318109, "grad_norm": 3.5272226333618164, "learning_rate": 4.660272612384907e-05, "loss": 5.3834, "step": 7528 }, { "epoch": 0.06796353132334357, "grad_norm": 3.1852009296417236, "learning_rate": 4.660227477884095e-05, "loss": 4.799, "step": 7529 }, { "epoch": 0.06797255822350605, "grad_norm": 3.3304800987243652, "learning_rate": 4.660182343383283e-05, "loss": 4.9838, "step": 7530 }, { "epoch": 0.06798158512366853, "grad_norm": 3.618072509765625, "learning_rate": 4.66013720888247e-05, "loss": 4.3266, "step": 7531 }, { "epoch": 0.06799061202383101, "grad_norm": 3.771425724029541, "learning_rate": 4.660092074381657e-05, "loss": 4.9478, "step": 7532 }, { "epoch": 0.0679996389239935, "grad_norm": 3.552961587905884, "learning_rate": 4.660046939880845e-05, "loss": 4.6699, "step": 7533 }, { "epoch": 0.06800866582415599, "grad_norm": 3.609039306640625, "learning_rate": 4.6600018053800325e-05, "loss": 4.7578, "step": 7534 }, { "epoch": 0.06801769272431847, "grad_norm": 3.304379463195801, "learning_rate": 4.6599566708792204e-05, "loss": 5.3296, "step": 7535 }, { "epoch": 0.06802671962448095, "grad_norm": 4.010908603668213, "learning_rate": 4.6599115363784076e-05, "loss": 4.4526, "step": 7536 }, { "epoch": 0.06803574652464343, "grad_norm": 3.4188480377197266, "learning_rate": 4.6598664018775955e-05, "loss": 4.0769, "step": 7537 }, { "epoch": 0.06804477342480592, "grad_norm": 3.380971670150757, "learning_rate": 4.6598212673767835e-05, "loss": 3.9823, "step": 7538 }, { "epoch": 0.06805380032496841, "grad_norm": 4.72296667098999, "learning_rate": 4.65977613287597e-05, "loss": 5.3597, "step": 7539 }, { "epoch": 0.06806282722513089, "grad_norm": 3.785306692123413, "learning_rate": 4.659730998375158e-05, "loss": 4.695, "step": 7540 }, { "epoch": 0.06807185412529337, "grad_norm": 3.607482433319092, "learning_rate": 4.659685863874346e-05, "loss": 5.0449, "step": 7541 }, { "epoch": 0.06808088102545586, "grad_norm": 4.032645225524902, "learning_rate": 4.659640729373533e-05, "loss": 4.97, "step": 7542 }, { "epoch": 0.06808990792561834, "grad_norm": 5.415939807891846, "learning_rate": 4.659595594872721e-05, "loss": 4.5272, "step": 7543 }, { "epoch": 0.06809893482578083, "grad_norm": 6.083359718322754, "learning_rate": 4.659550460371909e-05, "loss": 4.3302, "step": 7544 }, { "epoch": 0.06810796172594331, "grad_norm": 2.6436781883239746, "learning_rate": 4.659505325871096e-05, "loss": 4.835, "step": 7545 }, { "epoch": 0.0681169886261058, "grad_norm": 8.257261276245117, "learning_rate": 4.6594601913702835e-05, "loss": 4.6133, "step": 7546 }, { "epoch": 0.06812601552626828, "grad_norm": 3.4519364833831787, "learning_rate": 4.6594150568694714e-05, "loss": 4.4923, "step": 7547 }, { "epoch": 0.06813504242643076, "grad_norm": 2.507232427597046, "learning_rate": 4.6593699223686586e-05, "loss": 4.435, "step": 7548 }, { "epoch": 0.06814406932659325, "grad_norm": 2.913181781768799, "learning_rate": 4.6593247878678466e-05, "loss": 5.0611, "step": 7549 }, { "epoch": 0.06815309622675574, "grad_norm": 3.1347708702087402, "learning_rate": 4.659279653367034e-05, "loss": 4.8402, "step": 7550 }, { "epoch": 0.06816212312691822, "grad_norm": 4.047896385192871, "learning_rate": 4.659234518866222e-05, "loss": 5.0283, "step": 7551 }, { "epoch": 0.0681711500270807, "grad_norm": 3.7192399501800537, "learning_rate": 4.65918938436541e-05, "loss": 5.0358, "step": 7552 }, { "epoch": 0.06818017692724318, "grad_norm": 5.087768077850342, "learning_rate": 4.659144249864596e-05, "loss": 4.6751, "step": 7553 }, { "epoch": 0.06818920382740568, "grad_norm": 2.741581678390503, "learning_rate": 4.659099115363784e-05, "loss": 5.107, "step": 7554 }, { "epoch": 0.06819823072756816, "grad_norm": 4.365182399749756, "learning_rate": 4.659053980862972e-05, "loss": 4.9637, "step": 7555 }, { "epoch": 0.06820725762773064, "grad_norm": 3.806419849395752, "learning_rate": 4.659008846362159e-05, "loss": 4.647, "step": 7556 }, { "epoch": 0.06821628452789312, "grad_norm": 3.8180525302886963, "learning_rate": 4.658963711861347e-05, "loss": 4.3855, "step": 7557 }, { "epoch": 0.0682253114280556, "grad_norm": 3.890648126602173, "learning_rate": 4.6589185773605345e-05, "loss": 5.3655, "step": 7558 }, { "epoch": 0.06823433832821808, "grad_norm": 3.1930477619171143, "learning_rate": 4.658873442859722e-05, "loss": 4.6864, "step": 7559 }, { "epoch": 0.06824336522838058, "grad_norm": 3.4954702854156494, "learning_rate": 4.65882830835891e-05, "loss": 4.4659, "step": 7560 }, { "epoch": 0.06825239212854306, "grad_norm": 3.8700759410858154, "learning_rate": 4.6587831738580976e-05, "loss": 4.2793, "step": 7561 }, { "epoch": 0.06826141902870554, "grad_norm": 3.5135273933410645, "learning_rate": 4.658738039357285e-05, "loss": 4.5341, "step": 7562 }, { "epoch": 0.06827044592886802, "grad_norm": 3.6523654460906982, "learning_rate": 4.658692904856473e-05, "loss": 4.6929, "step": 7563 }, { "epoch": 0.0682794728290305, "grad_norm": 3.8457415103912354, "learning_rate": 4.65864777035566e-05, "loss": 4.7215, "step": 7564 }, { "epoch": 0.068288499729193, "grad_norm": 3.4350368976593018, "learning_rate": 4.658602635854848e-05, "loss": 4.5775, "step": 7565 }, { "epoch": 0.06829752662935548, "grad_norm": 2.643397331237793, "learning_rate": 4.658557501354035e-05, "loss": 4.8, "step": 7566 }, { "epoch": 0.06830655352951796, "grad_norm": 2.3796844482421875, "learning_rate": 4.6585123668532224e-05, "loss": 4.5726, "step": 7567 }, { "epoch": 0.06831558042968044, "grad_norm": 3.7821333408355713, "learning_rate": 4.6584672323524104e-05, "loss": 4.6361, "step": 7568 }, { "epoch": 0.06832460732984293, "grad_norm": 3.020379066467285, "learning_rate": 4.658422097851598e-05, "loss": 5.1114, "step": 7569 }, { "epoch": 0.06833363423000542, "grad_norm": 3.68902325630188, "learning_rate": 4.6583769633507855e-05, "loss": 3.8908, "step": 7570 }, { "epoch": 0.0683426611301679, "grad_norm": 4.02341890335083, "learning_rate": 4.6583318288499734e-05, "loss": 4.8369, "step": 7571 }, { "epoch": 0.06835168803033038, "grad_norm": 3.7262654304504395, "learning_rate": 4.658286694349161e-05, "loss": 4.5285, "step": 7572 }, { "epoch": 0.06836071493049287, "grad_norm": 4.315223693847656, "learning_rate": 4.658241559848348e-05, "loss": 4.6939, "step": 7573 }, { "epoch": 0.06836974183065535, "grad_norm": 3.016624689102173, "learning_rate": 4.658196425347536e-05, "loss": 4.4577, "step": 7574 }, { "epoch": 0.06837876873081784, "grad_norm": 2.4877610206604004, "learning_rate": 4.658151290846723e-05, "loss": 4.8324, "step": 7575 }, { "epoch": 0.06838779563098032, "grad_norm": 5.665232181549072, "learning_rate": 4.658106156345911e-05, "loss": 4.4391, "step": 7576 }, { "epoch": 0.0683968225311428, "grad_norm": 2.7843308448791504, "learning_rate": 4.658061021845099e-05, "loss": 4.2259, "step": 7577 }, { "epoch": 0.06840584943130529, "grad_norm": 3.5072124004364014, "learning_rate": 4.658015887344286e-05, "loss": 4.808, "step": 7578 }, { "epoch": 0.06841487633146777, "grad_norm": 4.531610488891602, "learning_rate": 4.6579707528434735e-05, "loss": 4.2737, "step": 7579 }, { "epoch": 0.06842390323163026, "grad_norm": 4.161898612976074, "learning_rate": 4.6579256183426614e-05, "loss": 4.0154, "step": 7580 }, { "epoch": 0.06843293013179275, "grad_norm": 3.3315114974975586, "learning_rate": 4.6578804838418486e-05, "loss": 4.2115, "step": 7581 }, { "epoch": 0.06844195703195523, "grad_norm": 3.330472707748413, "learning_rate": 4.6578353493410365e-05, "loss": 4.7036, "step": 7582 }, { "epoch": 0.06845098393211771, "grad_norm": 3.422657012939453, "learning_rate": 4.6577902148402245e-05, "loss": 5.0673, "step": 7583 }, { "epoch": 0.06846001083228019, "grad_norm": 3.2729175090789795, "learning_rate": 4.657745080339412e-05, "loss": 5.0288, "step": 7584 }, { "epoch": 0.06846903773244267, "grad_norm": 4.734395980834961, "learning_rate": 4.6576999458385996e-05, "loss": 5.4804, "step": 7585 }, { "epoch": 0.06847806463260517, "grad_norm": 3.1456289291381836, "learning_rate": 4.657654811337787e-05, "loss": 5.2643, "step": 7586 }, { "epoch": 0.06848709153276765, "grad_norm": 3.628561496734619, "learning_rate": 4.657609676836974e-05, "loss": 4.7218, "step": 7587 }, { "epoch": 0.06849611843293013, "grad_norm": 3.1609370708465576, "learning_rate": 4.657564542336162e-05, "loss": 5.0387, "step": 7588 }, { "epoch": 0.06850514533309261, "grad_norm": 5.695820331573486, "learning_rate": 4.657519407835349e-05, "loss": 5.1138, "step": 7589 }, { "epoch": 0.0685141722332551, "grad_norm": 3.490633964538574, "learning_rate": 4.657474273334537e-05, "loss": 4.1379, "step": 7590 }, { "epoch": 0.06852319913341759, "grad_norm": 3.657461166381836, "learning_rate": 4.657429138833725e-05, "loss": 4.5873, "step": 7591 }, { "epoch": 0.06853222603358007, "grad_norm": 4.156826019287109, "learning_rate": 4.6573840043329124e-05, "loss": 3.9868, "step": 7592 }, { "epoch": 0.06854125293374255, "grad_norm": 3.7740774154663086, "learning_rate": 4.6573388698320996e-05, "loss": 5.3611, "step": 7593 }, { "epoch": 0.06855027983390503, "grad_norm": 2.7373769283294678, "learning_rate": 4.6572937353312876e-05, "loss": 4.3504, "step": 7594 }, { "epoch": 0.06855930673406752, "grad_norm": 3.447028160095215, "learning_rate": 4.657248600830475e-05, "loss": 4.8784, "step": 7595 }, { "epoch": 0.06856833363423001, "grad_norm": 3.3762013912200928, "learning_rate": 4.657203466329663e-05, "loss": 4.8008, "step": 7596 }, { "epoch": 0.06857736053439249, "grad_norm": 4.002870559692383, "learning_rate": 4.65715833182885e-05, "loss": 4.1905, "step": 7597 }, { "epoch": 0.06858638743455497, "grad_norm": 4.183661460876465, "learning_rate": 4.657113197328038e-05, "loss": 4.5119, "step": 7598 }, { "epoch": 0.06859541433471746, "grad_norm": 3.2360057830810547, "learning_rate": 4.657068062827226e-05, "loss": 4.1178, "step": 7599 }, { "epoch": 0.06860444123487994, "grad_norm": 3.275059938430786, "learning_rate": 4.6570229283264124e-05, "loss": 4.908, "step": 7600 }, { "epoch": 0.06861346813504243, "grad_norm": 3.3582520484924316, "learning_rate": 4.6569777938256e-05, "loss": 3.9764, "step": 7601 }, { "epoch": 0.06862249503520491, "grad_norm": 3.5936098098754883, "learning_rate": 4.656932659324788e-05, "loss": 5.0954, "step": 7602 }, { "epoch": 0.0686315219353674, "grad_norm": 3.3019309043884277, "learning_rate": 4.6568875248239755e-05, "loss": 4.3188, "step": 7603 }, { "epoch": 0.06864054883552988, "grad_norm": 4.539477348327637, "learning_rate": 4.6568423903231634e-05, "loss": 4.2213, "step": 7604 }, { "epoch": 0.06864957573569236, "grad_norm": 3.560323715209961, "learning_rate": 4.6567972558223513e-05, "loss": 4.3325, "step": 7605 }, { "epoch": 0.06865860263585485, "grad_norm": 3.7023706436157227, "learning_rate": 4.656752121321538e-05, "loss": 5.3849, "step": 7606 }, { "epoch": 0.06866762953601734, "grad_norm": 3.4296369552612305, "learning_rate": 4.656706986820726e-05, "loss": 4.4816, "step": 7607 }, { "epoch": 0.06867665643617982, "grad_norm": 3.3323986530303955, "learning_rate": 4.656661852319914e-05, "loss": 5.6102, "step": 7608 }, { "epoch": 0.0686856833363423, "grad_norm": 3.7456345558166504, "learning_rate": 4.656616717819101e-05, "loss": 4.9256, "step": 7609 }, { "epoch": 0.06869471023650478, "grad_norm": 3.995978355407715, "learning_rate": 4.656571583318289e-05, "loss": 4.9493, "step": 7610 }, { "epoch": 0.06870373713666726, "grad_norm": 3.2104337215423584, "learning_rate": 4.656526448817476e-05, "loss": 4.7277, "step": 7611 }, { "epoch": 0.06871276403682976, "grad_norm": 4.457736015319824, "learning_rate": 4.656481314316664e-05, "loss": 5.1865, "step": 7612 }, { "epoch": 0.06872179093699224, "grad_norm": 3.1817104816436768, "learning_rate": 4.6564361798158514e-05, "loss": 4.3683, "step": 7613 }, { "epoch": 0.06873081783715472, "grad_norm": 5.5845723152160645, "learning_rate": 4.6563910453150386e-05, "loss": 5.1826, "step": 7614 }, { "epoch": 0.0687398447373172, "grad_norm": 3.223065137863159, "learning_rate": 4.6563459108142265e-05, "loss": 4.9696, "step": 7615 }, { "epoch": 0.06874887163747968, "grad_norm": 3.7508530616760254, "learning_rate": 4.6563007763134144e-05, "loss": 4.4881, "step": 7616 }, { "epoch": 0.06875789853764218, "grad_norm": 2.904898166656494, "learning_rate": 4.656255641812602e-05, "loss": 4.3854, "step": 7617 }, { "epoch": 0.06876692543780466, "grad_norm": 2.7698638439178467, "learning_rate": 4.6562105073117896e-05, "loss": 5.0593, "step": 7618 }, { "epoch": 0.06877595233796714, "grad_norm": 2.794609785079956, "learning_rate": 4.656165372810977e-05, "loss": 4.966, "step": 7619 }, { "epoch": 0.06878497923812962, "grad_norm": 3.3423407077789307, "learning_rate": 4.656120238310164e-05, "loss": 4.9561, "step": 7620 }, { "epoch": 0.0687940061382921, "grad_norm": 3.7005250453948975, "learning_rate": 4.656075103809352e-05, "loss": 4.6178, "step": 7621 }, { "epoch": 0.0688030330384546, "grad_norm": 4.310872554779053, "learning_rate": 4.65602996930854e-05, "loss": 4.7671, "step": 7622 }, { "epoch": 0.06881205993861708, "grad_norm": 2.562527894973755, "learning_rate": 4.655984834807727e-05, "loss": 4.9869, "step": 7623 }, { "epoch": 0.06882108683877956, "grad_norm": 4.0396528244018555, "learning_rate": 4.655939700306915e-05, "loss": 4.5485, "step": 7624 }, { "epoch": 0.06883011373894204, "grad_norm": 2.7784321308135986, "learning_rate": 4.6558945658061024e-05, "loss": 5.3334, "step": 7625 }, { "epoch": 0.06883914063910453, "grad_norm": 3.84379243850708, "learning_rate": 4.6558494313052896e-05, "loss": 5.0986, "step": 7626 }, { "epoch": 0.06884816753926702, "grad_norm": 3.6271095275878906, "learning_rate": 4.6558042968044775e-05, "loss": 4.27, "step": 7627 }, { "epoch": 0.0688571944394295, "grad_norm": 3.4257986545562744, "learning_rate": 4.655759162303665e-05, "loss": 4.8296, "step": 7628 }, { "epoch": 0.06886622133959198, "grad_norm": 3.6508500576019287, "learning_rate": 4.655714027802853e-05, "loss": 4.6601, "step": 7629 }, { "epoch": 0.06887524823975447, "grad_norm": 3.8421778678894043, "learning_rate": 4.6556688933020406e-05, "loss": 4.6114, "step": 7630 }, { "epoch": 0.06888427513991695, "grad_norm": 4.389159679412842, "learning_rate": 4.655623758801228e-05, "loss": 3.6516, "step": 7631 }, { "epoch": 0.06889330204007944, "grad_norm": 3.2806482315063477, "learning_rate": 4.655578624300416e-05, "loss": 5.1729, "step": 7632 }, { "epoch": 0.06890232894024192, "grad_norm": 3.356548547744751, "learning_rate": 4.655533489799603e-05, "loss": 4.6126, "step": 7633 }, { "epoch": 0.0689113558404044, "grad_norm": 3.1322219371795654, "learning_rate": 4.65548835529879e-05, "loss": 4.7064, "step": 7634 }, { "epoch": 0.06892038274056689, "grad_norm": 2.8827991485595703, "learning_rate": 4.655443220797978e-05, "loss": 4.4758, "step": 7635 }, { "epoch": 0.06892940964072937, "grad_norm": 3.860358715057373, "learning_rate": 4.6553980862971655e-05, "loss": 4.0794, "step": 7636 }, { "epoch": 0.06893843654089185, "grad_norm": 3.7443900108337402, "learning_rate": 4.6553529517963534e-05, "loss": 4.7981, "step": 7637 }, { "epoch": 0.06894746344105435, "grad_norm": 2.7275562286376953, "learning_rate": 4.655307817295541e-05, "loss": 4.5586, "step": 7638 }, { "epoch": 0.06895649034121683, "grad_norm": 3.452206611633301, "learning_rate": 4.6552626827947286e-05, "loss": 4.1408, "step": 7639 }, { "epoch": 0.06896551724137931, "grad_norm": 3.3389968872070312, "learning_rate": 4.655217548293916e-05, "loss": 4.7336, "step": 7640 }, { "epoch": 0.06897454414154179, "grad_norm": 3.361581802368164, "learning_rate": 4.655172413793104e-05, "loss": 4.3572, "step": 7641 }, { "epoch": 0.06898357104170427, "grad_norm": 3.035259246826172, "learning_rate": 4.655127279292291e-05, "loss": 4.4721, "step": 7642 }, { "epoch": 0.06899259794186677, "grad_norm": 2.963804244995117, "learning_rate": 4.655082144791479e-05, "loss": 4.3295, "step": 7643 }, { "epoch": 0.06900162484202925, "grad_norm": 3.6481170654296875, "learning_rate": 4.655037010290667e-05, "loss": 4.3219, "step": 7644 }, { "epoch": 0.06901065174219173, "grad_norm": 3.475221872329712, "learning_rate": 4.654991875789854e-05, "loss": 4.7003, "step": 7645 }, { "epoch": 0.06901967864235421, "grad_norm": 4.0882568359375, "learning_rate": 4.654946741289041e-05, "loss": 4.6265, "step": 7646 }, { "epoch": 0.0690287055425167, "grad_norm": 3.764036178588867, "learning_rate": 4.654901606788229e-05, "loss": 4.6692, "step": 7647 }, { "epoch": 0.06903773244267919, "grad_norm": 2.4122860431671143, "learning_rate": 4.6548564722874165e-05, "loss": 4.7361, "step": 7648 }, { "epoch": 0.06904675934284167, "grad_norm": 2.8847646713256836, "learning_rate": 4.6548113377866044e-05, "loss": 4.0397, "step": 7649 }, { "epoch": 0.06905578624300415, "grad_norm": 3.086378574371338, "learning_rate": 4.654766203285792e-05, "loss": 4.341, "step": 7650 }, { "epoch": 0.06906481314316663, "grad_norm": 2.708672285079956, "learning_rate": 4.6547210687849796e-05, "loss": 5.1282, "step": 7651 }, { "epoch": 0.06907384004332912, "grad_norm": 3.584158182144165, "learning_rate": 4.6546759342841675e-05, "loss": 4.2051, "step": 7652 }, { "epoch": 0.06908286694349161, "grad_norm": 2.9935927391052246, "learning_rate": 4.654630799783354e-05, "loss": 3.1694, "step": 7653 }, { "epoch": 0.06909189384365409, "grad_norm": 3.0899107456207275, "learning_rate": 4.654585665282542e-05, "loss": 4.4519, "step": 7654 }, { "epoch": 0.06910092074381657, "grad_norm": 2.9469430446624756, "learning_rate": 4.65454053078173e-05, "loss": 4.8179, "step": 7655 }, { "epoch": 0.06910994764397906, "grad_norm": 2.6786680221557617, "learning_rate": 4.654495396280917e-05, "loss": 4.8594, "step": 7656 }, { "epoch": 0.06911897454414154, "grad_norm": 3.277883291244507, "learning_rate": 4.654450261780105e-05, "loss": 5.065, "step": 7657 }, { "epoch": 0.06912800144430403, "grad_norm": 4.321727275848389, "learning_rate": 4.6544051272792924e-05, "loss": 5.3296, "step": 7658 }, { "epoch": 0.06913702834446651, "grad_norm": 4.61374568939209, "learning_rate": 4.65435999277848e-05, "loss": 3.5488, "step": 7659 }, { "epoch": 0.069146055244629, "grad_norm": 5.189788818359375, "learning_rate": 4.6543148582776675e-05, "loss": 5.2375, "step": 7660 }, { "epoch": 0.06915508214479148, "grad_norm": 3.8446178436279297, "learning_rate": 4.654269723776855e-05, "loss": 4.513, "step": 7661 }, { "epoch": 0.06916410904495396, "grad_norm": 4.415157318115234, "learning_rate": 4.654224589276043e-05, "loss": 4.6299, "step": 7662 }, { "epoch": 0.06917313594511644, "grad_norm": 3.9616072177886963, "learning_rate": 4.6541794547752306e-05, "loss": 4.5449, "step": 7663 }, { "epoch": 0.06918216284527894, "grad_norm": 4.646270275115967, "learning_rate": 4.654134320274418e-05, "loss": 4.5396, "step": 7664 }, { "epoch": 0.06919118974544142, "grad_norm": 3.572969913482666, "learning_rate": 4.654089185773606e-05, "loss": 4.5762, "step": 7665 }, { "epoch": 0.0692002166456039, "grad_norm": 6.259682655334473, "learning_rate": 4.654044051272794e-05, "loss": 4.8826, "step": 7666 }, { "epoch": 0.06920924354576638, "grad_norm": 3.3732798099517822, "learning_rate": 4.65399891677198e-05, "loss": 4.8603, "step": 7667 }, { "epoch": 0.06921827044592886, "grad_norm": 3.430913209915161, "learning_rate": 4.653953782271168e-05, "loss": 4.7963, "step": 7668 }, { "epoch": 0.06922729734609136, "grad_norm": 6.726027011871338, "learning_rate": 4.653908647770356e-05, "loss": 5.0997, "step": 7669 }, { "epoch": 0.06923632424625384, "grad_norm": 2.773326873779297, "learning_rate": 4.6538635132695434e-05, "loss": 4.4805, "step": 7670 }, { "epoch": 0.06924535114641632, "grad_norm": 3.188318967819214, "learning_rate": 4.653818378768731e-05, "loss": 5.1211, "step": 7671 }, { "epoch": 0.0692543780465788, "grad_norm": 3.8191065788269043, "learning_rate": 4.6537732442679185e-05, "loss": 4.304, "step": 7672 }, { "epoch": 0.06926340494674128, "grad_norm": 3.1749024391174316, "learning_rate": 4.653728109767106e-05, "loss": 4.3975, "step": 7673 }, { "epoch": 0.06927243184690378, "grad_norm": 4.037298202514648, "learning_rate": 4.653682975266294e-05, "loss": 4.9421, "step": 7674 }, { "epoch": 0.06928145874706626, "grad_norm": 4.589802265167236, "learning_rate": 4.653637840765481e-05, "loss": 4.9282, "step": 7675 }, { "epoch": 0.06929048564722874, "grad_norm": 3.3921868801116943, "learning_rate": 4.653592706264669e-05, "loss": 4.5911, "step": 7676 }, { "epoch": 0.06929951254739122, "grad_norm": 3.4203176498413086, "learning_rate": 4.653547571763857e-05, "loss": 4.6189, "step": 7677 }, { "epoch": 0.0693085394475537, "grad_norm": 3.1230406761169434, "learning_rate": 4.653502437263044e-05, "loss": 4.182, "step": 7678 }, { "epoch": 0.0693175663477162, "grad_norm": 2.152303457260132, "learning_rate": 4.653457302762232e-05, "loss": 4.6585, "step": 7679 }, { "epoch": 0.06932659324787868, "grad_norm": 3.846073627471924, "learning_rate": 4.653412168261419e-05, "loss": 4.7733, "step": 7680 }, { "epoch": 0.06933562014804116, "grad_norm": 4.218249797821045, "learning_rate": 4.6533670337606065e-05, "loss": 4.7442, "step": 7681 }, { "epoch": 0.06934464704820364, "grad_norm": 4.818662166595459, "learning_rate": 4.6533218992597944e-05, "loss": 4.2852, "step": 7682 }, { "epoch": 0.06935367394836613, "grad_norm": 3.3243610858917236, "learning_rate": 4.653276764758982e-05, "loss": 4.053, "step": 7683 }, { "epoch": 0.06936270084852862, "grad_norm": 5.343815803527832, "learning_rate": 4.6532316302581696e-05, "loss": 3.6139, "step": 7684 }, { "epoch": 0.0693717277486911, "grad_norm": 2.990851879119873, "learning_rate": 4.6531864957573575e-05, "loss": 4.4021, "step": 7685 }, { "epoch": 0.06938075464885358, "grad_norm": 3.1458263397216797, "learning_rate": 4.653141361256545e-05, "loss": 4.4261, "step": 7686 }, { "epoch": 0.06938978154901607, "grad_norm": 3.402510166168213, "learning_rate": 4.653096226755732e-05, "loss": 4.6661, "step": 7687 }, { "epoch": 0.06939880844917855, "grad_norm": 3.7655458450317383, "learning_rate": 4.65305109225492e-05, "loss": 5.2036, "step": 7688 }, { "epoch": 0.06940783534934103, "grad_norm": 3.046480655670166, "learning_rate": 4.653005957754107e-05, "loss": 4.8325, "step": 7689 }, { "epoch": 0.06941686224950352, "grad_norm": 3.5436501502990723, "learning_rate": 4.652960823253295e-05, "loss": 4.869, "step": 7690 }, { "epoch": 0.069425889149666, "grad_norm": 3.1373820304870605, "learning_rate": 4.652915688752483e-05, "loss": 4.4977, "step": 7691 }, { "epoch": 0.06943491604982849, "grad_norm": 2.841674327850342, "learning_rate": 4.65287055425167e-05, "loss": 4.8695, "step": 7692 }, { "epoch": 0.06944394294999097, "grad_norm": 3.0441367626190186, "learning_rate": 4.6528254197508575e-05, "loss": 4.72, "step": 7693 }, { "epoch": 0.06945296985015345, "grad_norm": 2.9621293544769287, "learning_rate": 4.6527802852500454e-05, "loss": 4.8819, "step": 7694 }, { "epoch": 0.06946199675031595, "grad_norm": 5.195305347442627, "learning_rate": 4.652735150749233e-05, "loss": 5.149, "step": 7695 }, { "epoch": 0.06947102365047843, "grad_norm": 3.7314891815185547, "learning_rate": 4.6526900162484206e-05, "loss": 4.2717, "step": 7696 }, { "epoch": 0.06948005055064091, "grad_norm": 3.4955883026123047, "learning_rate": 4.652644881747608e-05, "loss": 5.0433, "step": 7697 }, { "epoch": 0.06948907745080339, "grad_norm": 3.9124886989593506, "learning_rate": 4.652599747246796e-05, "loss": 4.6639, "step": 7698 }, { "epoch": 0.06949810435096587, "grad_norm": 4.152416229248047, "learning_rate": 4.652554612745984e-05, "loss": 5.1493, "step": 7699 }, { "epoch": 0.06950713125112837, "grad_norm": 3.283660888671875, "learning_rate": 4.65250947824517e-05, "loss": 4.5923, "step": 7700 }, { "epoch": 0.06951615815129085, "grad_norm": 2.9573264122009277, "learning_rate": 4.652464343744358e-05, "loss": 4.7348, "step": 7701 }, { "epoch": 0.06952518505145333, "grad_norm": 4.645166873931885, "learning_rate": 4.652419209243546e-05, "loss": 4.8641, "step": 7702 }, { "epoch": 0.06953421195161581, "grad_norm": 3.0488834381103516, "learning_rate": 4.6523740747427334e-05, "loss": 4.8521, "step": 7703 }, { "epoch": 0.0695432388517783, "grad_norm": 4.185957431793213, "learning_rate": 4.652328940241921e-05, "loss": 4.6355, "step": 7704 }, { "epoch": 0.06955226575194079, "grad_norm": 3.5556578636169434, "learning_rate": 4.652283805741109e-05, "loss": 4.4623, "step": 7705 }, { "epoch": 0.06956129265210327, "grad_norm": 3.218284845352173, "learning_rate": 4.6522386712402964e-05, "loss": 4.6536, "step": 7706 }, { "epoch": 0.06957031955226575, "grad_norm": 4.9566545486450195, "learning_rate": 4.652193536739484e-05, "loss": 3.2448, "step": 7707 }, { "epoch": 0.06957934645242823, "grad_norm": 2.968245267868042, "learning_rate": 4.6521484022386716e-05, "loss": 4.8292, "step": 7708 }, { "epoch": 0.06958837335259072, "grad_norm": 3.6220197677612305, "learning_rate": 4.652103267737859e-05, "loss": 4.2114, "step": 7709 }, { "epoch": 0.06959740025275321, "grad_norm": 3.9367034435272217, "learning_rate": 4.652058133237047e-05, "loss": 5.068, "step": 7710 }, { "epoch": 0.06960642715291569, "grad_norm": 3.130613088607788, "learning_rate": 4.652012998736234e-05, "loss": 4.4639, "step": 7711 }, { "epoch": 0.06961545405307817, "grad_norm": 5.414029598236084, "learning_rate": 4.651967864235422e-05, "loss": 5.3571, "step": 7712 }, { "epoch": 0.06962448095324066, "grad_norm": 3.385681390762329, "learning_rate": 4.65192272973461e-05, "loss": 4.1176, "step": 7713 }, { "epoch": 0.06963350785340314, "grad_norm": 3.7379353046417236, "learning_rate": 4.6518775952337965e-05, "loss": 5.2836, "step": 7714 }, { "epoch": 0.06964253475356563, "grad_norm": 3.295214891433716, "learning_rate": 4.6518324607329844e-05, "loss": 4.4622, "step": 7715 }, { "epoch": 0.06965156165372811, "grad_norm": 3.156055212020874, "learning_rate": 4.651787326232172e-05, "loss": 4.5229, "step": 7716 }, { "epoch": 0.0696605885538906, "grad_norm": 3.5972299575805664, "learning_rate": 4.6517421917313595e-05, "loss": 5.049, "step": 7717 }, { "epoch": 0.06966961545405308, "grad_norm": 3.232367753982544, "learning_rate": 4.6516970572305475e-05, "loss": 4.2752, "step": 7718 }, { "epoch": 0.06967864235421556, "grad_norm": 3.003488302230835, "learning_rate": 4.651651922729735e-05, "loss": 4.8859, "step": 7719 }, { "epoch": 0.06968766925437804, "grad_norm": 3.8710575103759766, "learning_rate": 4.651606788228922e-05, "loss": 4.6869, "step": 7720 }, { "epoch": 0.06969669615454054, "grad_norm": 3.360825300216675, "learning_rate": 4.65156165372811e-05, "loss": 4.7478, "step": 7721 }, { "epoch": 0.06970572305470302, "grad_norm": 3.7591049671173096, "learning_rate": 4.651516519227298e-05, "loss": 4.2155, "step": 7722 }, { "epoch": 0.0697147499548655, "grad_norm": 3.2412333488464355, "learning_rate": 4.651471384726485e-05, "loss": 4.8342, "step": 7723 }, { "epoch": 0.06972377685502798, "grad_norm": 3.216449737548828, "learning_rate": 4.651426250225673e-05, "loss": 3.9755, "step": 7724 }, { "epoch": 0.06973280375519046, "grad_norm": 4.555554389953613, "learning_rate": 4.65138111572486e-05, "loss": 5.1517, "step": 7725 }, { "epoch": 0.06974183065535296, "grad_norm": 3.6036224365234375, "learning_rate": 4.651335981224048e-05, "loss": 4.5789, "step": 7726 }, { "epoch": 0.06975085755551544, "grad_norm": 3.3130149841308594, "learning_rate": 4.6512908467232354e-05, "loss": 4.6339, "step": 7727 }, { "epoch": 0.06975988445567792, "grad_norm": 4.354923725128174, "learning_rate": 4.6512457122224226e-05, "loss": 4.1004, "step": 7728 }, { "epoch": 0.0697689113558404, "grad_norm": 3.3664467334747314, "learning_rate": 4.6512005777216106e-05, "loss": 4.1712, "step": 7729 }, { "epoch": 0.06977793825600288, "grad_norm": 3.0834314823150635, "learning_rate": 4.6511554432207985e-05, "loss": 4.9813, "step": 7730 }, { "epoch": 0.06978696515616538, "grad_norm": 3.0486972332000732, "learning_rate": 4.651110308719986e-05, "loss": 4.4853, "step": 7731 }, { "epoch": 0.06979599205632786, "grad_norm": 3.644610643386841, "learning_rate": 4.651065174219174e-05, "loss": 4.5401, "step": 7732 }, { "epoch": 0.06980501895649034, "grad_norm": 3.348792552947998, "learning_rate": 4.651020039718361e-05, "loss": 4.571, "step": 7733 }, { "epoch": 0.06981404585665282, "grad_norm": 5.260042667388916, "learning_rate": 4.650974905217548e-05, "loss": 3.2963, "step": 7734 }, { "epoch": 0.0698230727568153, "grad_norm": 2.2180263996124268, "learning_rate": 4.650929770716736e-05, "loss": 4.4757, "step": 7735 }, { "epoch": 0.0698320996569778, "grad_norm": 2.792567253112793, "learning_rate": 4.650884636215923e-05, "loss": 4.0886, "step": 7736 }, { "epoch": 0.06984112655714028, "grad_norm": 5.068848133087158, "learning_rate": 4.650839501715111e-05, "loss": 4.8561, "step": 7737 }, { "epoch": 0.06985015345730276, "grad_norm": 3.672086477279663, "learning_rate": 4.650794367214299e-05, "loss": 4.5948, "step": 7738 }, { "epoch": 0.06985918035746524, "grad_norm": 2.7411348819732666, "learning_rate": 4.6507492327134864e-05, "loss": 4.5159, "step": 7739 }, { "epoch": 0.06986820725762773, "grad_norm": 2.5988636016845703, "learning_rate": 4.650704098212674e-05, "loss": 4.264, "step": 7740 }, { "epoch": 0.06987723415779022, "grad_norm": 3.960176944732666, "learning_rate": 4.6506589637118616e-05, "loss": 3.8556, "step": 7741 }, { "epoch": 0.0698862610579527, "grad_norm": 3.53314471244812, "learning_rate": 4.650613829211049e-05, "loss": 4.5432, "step": 7742 }, { "epoch": 0.06989528795811518, "grad_norm": 3.8721351623535156, "learning_rate": 4.650568694710237e-05, "loss": 4.1173, "step": 7743 }, { "epoch": 0.06990431485827767, "grad_norm": 4.036937713623047, "learning_rate": 4.650523560209425e-05, "loss": 4.1092, "step": 7744 }, { "epoch": 0.06991334175844015, "grad_norm": 3.2153003215789795, "learning_rate": 4.650478425708612e-05, "loss": 4.5144, "step": 7745 }, { "epoch": 0.06992236865860263, "grad_norm": 3.3626959323883057, "learning_rate": 4.6504332912078e-05, "loss": 5.0301, "step": 7746 }, { "epoch": 0.06993139555876512, "grad_norm": 3.6943793296813965, "learning_rate": 4.650388156706987e-05, "loss": 5.1543, "step": 7747 }, { "epoch": 0.0699404224589276, "grad_norm": 2.9867353439331055, "learning_rate": 4.6503430222061744e-05, "loss": 4.6946, "step": 7748 }, { "epoch": 0.06994944935909009, "grad_norm": 3.107543706893921, "learning_rate": 4.650297887705362e-05, "loss": 4.7157, "step": 7749 }, { "epoch": 0.06995847625925257, "grad_norm": 4.054196834564209, "learning_rate": 4.6502527532045495e-05, "loss": 4.7549, "step": 7750 }, { "epoch": 0.06996750315941505, "grad_norm": 3.489647388458252, "learning_rate": 4.6502076187037374e-05, "loss": 4.5827, "step": 7751 }, { "epoch": 0.06997653005957755, "grad_norm": 3.9039878845214844, "learning_rate": 4.6501624842029254e-05, "loss": 4.6324, "step": 7752 }, { "epoch": 0.06998555695974003, "grad_norm": 3.360999345779419, "learning_rate": 4.6501173497021126e-05, "loss": 5.0021, "step": 7753 }, { "epoch": 0.06999458385990251, "grad_norm": 4.252448081970215, "learning_rate": 4.6500722152013e-05, "loss": 4.8531, "step": 7754 }, { "epoch": 0.07000361076006499, "grad_norm": 3.6680474281311035, "learning_rate": 4.650027080700488e-05, "loss": 4.5019, "step": 7755 }, { "epoch": 0.07001263766022747, "grad_norm": 2.720060110092163, "learning_rate": 4.649981946199675e-05, "loss": 4.9145, "step": 7756 }, { "epoch": 0.07002166456038997, "grad_norm": 4.014125347137451, "learning_rate": 4.649936811698863e-05, "loss": 4.849, "step": 7757 }, { "epoch": 0.07003069146055245, "grad_norm": 4.0762176513671875, "learning_rate": 4.64989167719805e-05, "loss": 4.4248, "step": 7758 }, { "epoch": 0.07003971836071493, "grad_norm": 2.9496002197265625, "learning_rate": 4.649846542697238e-05, "loss": 4.3477, "step": 7759 }, { "epoch": 0.07004874526087741, "grad_norm": 3.442981004714966, "learning_rate": 4.649801408196426e-05, "loss": 4.544, "step": 7760 }, { "epoch": 0.0700577721610399, "grad_norm": 3.877735137939453, "learning_rate": 4.6497562736956126e-05, "loss": 5.1192, "step": 7761 }, { "epoch": 0.07006679906120239, "grad_norm": 3.8222615718841553, "learning_rate": 4.6497111391948005e-05, "loss": 5.2434, "step": 7762 }, { "epoch": 0.07007582596136487, "grad_norm": 3.4687929153442383, "learning_rate": 4.6496660046939885e-05, "loss": 4.5252, "step": 7763 }, { "epoch": 0.07008485286152735, "grad_norm": 3.9365580081939697, "learning_rate": 4.649620870193176e-05, "loss": 4.2935, "step": 7764 }, { "epoch": 0.07009387976168983, "grad_norm": 5.365028381347656, "learning_rate": 4.6495757356923636e-05, "loss": 5.1288, "step": 7765 }, { "epoch": 0.07010290666185232, "grad_norm": 4.173686981201172, "learning_rate": 4.6495306011915516e-05, "loss": 4.7481, "step": 7766 }, { "epoch": 0.07011193356201481, "grad_norm": 3.0939435958862305, "learning_rate": 4.649485466690738e-05, "loss": 4.7062, "step": 7767 }, { "epoch": 0.07012096046217729, "grad_norm": 3.4838736057281494, "learning_rate": 4.649440332189926e-05, "loss": 5.1126, "step": 7768 }, { "epoch": 0.07012998736233977, "grad_norm": 3.0930354595184326, "learning_rate": 4.649395197689114e-05, "loss": 4.5906, "step": 7769 }, { "epoch": 0.07013901426250226, "grad_norm": 3.07559871673584, "learning_rate": 4.649350063188301e-05, "loss": 4.303, "step": 7770 }, { "epoch": 0.07014804116266474, "grad_norm": 3.2190027236938477, "learning_rate": 4.649304928687489e-05, "loss": 4.6374, "step": 7771 }, { "epoch": 0.07015706806282722, "grad_norm": 4.1995697021484375, "learning_rate": 4.6492597941866764e-05, "loss": 5.2236, "step": 7772 }, { "epoch": 0.07016609496298971, "grad_norm": 3.5986456871032715, "learning_rate": 4.649214659685864e-05, "loss": 5.2488, "step": 7773 }, { "epoch": 0.0701751218631522, "grad_norm": 4.808873653411865, "learning_rate": 4.6491695251850516e-05, "loss": 3.4677, "step": 7774 }, { "epoch": 0.07018414876331468, "grad_norm": 3.4856796264648438, "learning_rate": 4.649124390684239e-05, "loss": 4.4781, "step": 7775 }, { "epoch": 0.07019317566347716, "grad_norm": 4.022123336791992, "learning_rate": 4.649079256183427e-05, "loss": 4.5859, "step": 7776 }, { "epoch": 0.07020220256363964, "grad_norm": 3.425287961959839, "learning_rate": 4.649034121682615e-05, "loss": 5.087, "step": 7777 }, { "epoch": 0.07021122946380214, "grad_norm": 3.7662193775177, "learning_rate": 4.648988987181802e-05, "loss": 4.416, "step": 7778 }, { "epoch": 0.07022025636396462, "grad_norm": 3.0201776027679443, "learning_rate": 4.64894385268099e-05, "loss": 4.3041, "step": 7779 }, { "epoch": 0.0702292832641271, "grad_norm": 3.1100378036499023, "learning_rate": 4.648898718180177e-05, "loss": 4.169, "step": 7780 }, { "epoch": 0.07023831016428958, "grad_norm": 3.053410530090332, "learning_rate": 4.648853583679364e-05, "loss": 4.8195, "step": 7781 }, { "epoch": 0.07024733706445206, "grad_norm": 3.493191719055176, "learning_rate": 4.648808449178552e-05, "loss": 4.5793, "step": 7782 }, { "epoch": 0.07025636396461456, "grad_norm": 3.5612878799438477, "learning_rate": 4.64876331467774e-05, "loss": 5.0702, "step": 7783 }, { "epoch": 0.07026539086477704, "grad_norm": 2.869893789291382, "learning_rate": 4.6487181801769274e-05, "loss": 4.5704, "step": 7784 }, { "epoch": 0.07027441776493952, "grad_norm": 3.5324764251708984, "learning_rate": 4.6486730456761153e-05, "loss": 4.6171, "step": 7785 }, { "epoch": 0.070283444665102, "grad_norm": 3.2923898696899414, "learning_rate": 4.6486279111753026e-05, "loss": 4.3791, "step": 7786 }, { "epoch": 0.07029247156526448, "grad_norm": 3.422818660736084, "learning_rate": 4.64858277667449e-05, "loss": 4.5548, "step": 7787 }, { "epoch": 0.07030149846542698, "grad_norm": 3.957524061203003, "learning_rate": 4.648537642173678e-05, "loss": 4.8013, "step": 7788 }, { "epoch": 0.07031052536558946, "grad_norm": 5.968430995941162, "learning_rate": 4.648492507672865e-05, "loss": 4.5319, "step": 7789 }, { "epoch": 0.07031955226575194, "grad_norm": 3.644038677215576, "learning_rate": 4.648447373172053e-05, "loss": 5.1655, "step": 7790 }, { "epoch": 0.07032857916591442, "grad_norm": 3.758678674697876, "learning_rate": 4.648402238671241e-05, "loss": 4.6276, "step": 7791 }, { "epoch": 0.0703376060660769, "grad_norm": 3.859309673309326, "learning_rate": 4.648357104170428e-05, "loss": 4.8605, "step": 7792 }, { "epoch": 0.0703466329662394, "grad_norm": 3.3748581409454346, "learning_rate": 4.648311969669616e-05, "loss": 4.6955, "step": 7793 }, { "epoch": 0.07035565986640188, "grad_norm": 4.0228962898254395, "learning_rate": 4.648266835168803e-05, "loss": 4.9467, "step": 7794 }, { "epoch": 0.07036468676656436, "grad_norm": 3.9346978664398193, "learning_rate": 4.6482217006679905e-05, "loss": 4.7655, "step": 7795 }, { "epoch": 0.07037371366672684, "grad_norm": 3.5113558769226074, "learning_rate": 4.6481765661671784e-05, "loss": 4.7281, "step": 7796 }, { "epoch": 0.07038274056688933, "grad_norm": 3.252901315689087, "learning_rate": 4.648131431666366e-05, "loss": 3.5866, "step": 7797 }, { "epoch": 0.07039176746705181, "grad_norm": 3.4444403648376465, "learning_rate": 4.6480862971655536e-05, "loss": 5.1327, "step": 7798 }, { "epoch": 0.0704007943672143, "grad_norm": 3.1259751319885254, "learning_rate": 4.6480411626647415e-05, "loss": 4.4981, "step": 7799 }, { "epoch": 0.07040982126737678, "grad_norm": 2.70987868309021, "learning_rate": 4.647996028163929e-05, "loss": 5.296, "step": 7800 }, { "epoch": 0.07041884816753927, "grad_norm": 3.813991069793701, "learning_rate": 4.647950893663116e-05, "loss": 4.5717, "step": 7801 }, { "epoch": 0.07042787506770175, "grad_norm": 3.7582085132598877, "learning_rate": 4.647905759162304e-05, "loss": 4.0614, "step": 7802 }, { "epoch": 0.07043690196786423, "grad_norm": 3.3137598037719727, "learning_rate": 4.647860624661491e-05, "loss": 5.3714, "step": 7803 }, { "epoch": 0.07044592886802672, "grad_norm": 2.5699501037597656, "learning_rate": 4.647815490160679e-05, "loss": 4.7854, "step": 7804 }, { "epoch": 0.0704549557681892, "grad_norm": 3.1681113243103027, "learning_rate": 4.647770355659867e-05, "loss": 5.001, "step": 7805 }, { "epoch": 0.07046398266835169, "grad_norm": 3.311861515045166, "learning_rate": 4.647725221159054e-05, "loss": 5.0611, "step": 7806 }, { "epoch": 0.07047300956851417, "grad_norm": 3.9721884727478027, "learning_rate": 4.6476800866582415e-05, "loss": 4.6269, "step": 7807 }, { "epoch": 0.07048203646867665, "grad_norm": 3.8676397800445557, "learning_rate": 4.6476349521574295e-05, "loss": 4.579, "step": 7808 }, { "epoch": 0.07049106336883915, "grad_norm": 3.0808937549591064, "learning_rate": 4.647589817656617e-05, "loss": 4.4997, "step": 7809 }, { "epoch": 0.07050009026900163, "grad_norm": 3.6813955307006836, "learning_rate": 4.6475446831558046e-05, "loss": 3.9768, "step": 7810 }, { "epoch": 0.07050911716916411, "grad_norm": 4.681650638580322, "learning_rate": 4.647499548654992e-05, "loss": 4.4988, "step": 7811 }, { "epoch": 0.07051814406932659, "grad_norm": 3.2389042377471924, "learning_rate": 4.64745441415418e-05, "loss": 4.434, "step": 7812 }, { "epoch": 0.07052717096948907, "grad_norm": 3.388603448867798, "learning_rate": 4.647409279653368e-05, "loss": 4.6339, "step": 7813 }, { "epoch": 0.07053619786965157, "grad_norm": 2.7908570766448975, "learning_rate": 4.647364145152554e-05, "loss": 4.7813, "step": 7814 }, { "epoch": 0.07054522476981405, "grad_norm": 3.0295462608337402, "learning_rate": 4.647319010651742e-05, "loss": 4.6353, "step": 7815 }, { "epoch": 0.07055425166997653, "grad_norm": 3.0213403701782227, "learning_rate": 4.64727387615093e-05, "loss": 4.2686, "step": 7816 }, { "epoch": 0.07056327857013901, "grad_norm": 3.823981523513794, "learning_rate": 4.6472287416501174e-05, "loss": 4.0025, "step": 7817 }, { "epoch": 0.0705723054703015, "grad_norm": 5.955183982849121, "learning_rate": 4.647183607149305e-05, "loss": 4.7521, "step": 7818 }, { "epoch": 0.07058133237046399, "grad_norm": 3.5869812965393066, "learning_rate": 4.6471384726484926e-05, "loss": 4.9866, "step": 7819 }, { "epoch": 0.07059035927062647, "grad_norm": 3.4500105381011963, "learning_rate": 4.6470933381476805e-05, "loss": 5.2388, "step": 7820 }, { "epoch": 0.07059938617078895, "grad_norm": 3.013502597808838, "learning_rate": 4.647048203646868e-05, "loss": 4.0613, "step": 7821 }, { "epoch": 0.07060841307095143, "grad_norm": 4.281284332275391, "learning_rate": 4.647003069146055e-05, "loss": 4.4928, "step": 7822 }, { "epoch": 0.07061743997111392, "grad_norm": 3.3294906616210938, "learning_rate": 4.646957934645243e-05, "loss": 4.9687, "step": 7823 }, { "epoch": 0.0706264668712764, "grad_norm": 3.4905388355255127, "learning_rate": 4.646912800144431e-05, "loss": 4.2325, "step": 7824 }, { "epoch": 0.07063549377143889, "grad_norm": 2.9851715564727783, "learning_rate": 4.646867665643618e-05, "loss": 4.5228, "step": 7825 }, { "epoch": 0.07064452067160137, "grad_norm": 3.523472309112549, "learning_rate": 4.646822531142806e-05, "loss": 4.2625, "step": 7826 }, { "epoch": 0.07065354757176386, "grad_norm": 3.9339191913604736, "learning_rate": 4.646777396641994e-05, "loss": 4.6629, "step": 7827 }, { "epoch": 0.07066257447192634, "grad_norm": 3.523761510848999, "learning_rate": 4.6467322621411805e-05, "loss": 4.8651, "step": 7828 }, { "epoch": 0.07067160137208882, "grad_norm": 3.372018337249756, "learning_rate": 4.6466871276403684e-05, "loss": 4.4641, "step": 7829 }, { "epoch": 0.07068062827225131, "grad_norm": 2.87391996383667, "learning_rate": 4.6466419931395563e-05, "loss": 5.3808, "step": 7830 }, { "epoch": 0.0706896551724138, "grad_norm": 4.340638160705566, "learning_rate": 4.6465968586387436e-05, "loss": 3.6195, "step": 7831 }, { "epoch": 0.07069868207257628, "grad_norm": 4.072720527648926, "learning_rate": 4.6465517241379315e-05, "loss": 4.8935, "step": 7832 }, { "epoch": 0.07070770897273876, "grad_norm": 5.985753059387207, "learning_rate": 4.646506589637119e-05, "loss": 3.9878, "step": 7833 }, { "epoch": 0.07071673587290124, "grad_norm": 5.609668254852295, "learning_rate": 4.646461455136306e-05, "loss": 4.3869, "step": 7834 }, { "epoch": 0.07072576277306374, "grad_norm": 3.3093764781951904, "learning_rate": 4.646416320635494e-05, "loss": 4.6852, "step": 7835 }, { "epoch": 0.07073478967322622, "grad_norm": 3.34818696975708, "learning_rate": 4.646371186134681e-05, "loss": 4.3999, "step": 7836 }, { "epoch": 0.0707438165733887, "grad_norm": 4.372857570648193, "learning_rate": 4.646326051633869e-05, "loss": 4.5099, "step": 7837 }, { "epoch": 0.07075284347355118, "grad_norm": 3.192845344543457, "learning_rate": 4.646280917133057e-05, "loss": 4.4399, "step": 7838 }, { "epoch": 0.07076187037371366, "grad_norm": 3.076380968093872, "learning_rate": 4.646235782632244e-05, "loss": 4.6397, "step": 7839 }, { "epoch": 0.07077089727387616, "grad_norm": 2.762308359146118, "learning_rate": 4.646190648131432e-05, "loss": 4.9154, "step": 7840 }, { "epoch": 0.07077992417403864, "grad_norm": 2.7689054012298584, "learning_rate": 4.6461455136306194e-05, "loss": 4.94, "step": 7841 }, { "epoch": 0.07078895107420112, "grad_norm": 3.339214324951172, "learning_rate": 4.646100379129807e-05, "loss": 4.9148, "step": 7842 }, { "epoch": 0.0707979779743636, "grad_norm": 3.1347200870513916, "learning_rate": 4.6460552446289946e-05, "loss": 3.9341, "step": 7843 }, { "epoch": 0.07080700487452608, "grad_norm": 4.358320713043213, "learning_rate": 4.6460101101281825e-05, "loss": 5.03, "step": 7844 }, { "epoch": 0.07081603177468858, "grad_norm": 3.1043522357940674, "learning_rate": 4.64596497562737e-05, "loss": 4.4065, "step": 7845 }, { "epoch": 0.07082505867485106, "grad_norm": 3.2924625873565674, "learning_rate": 4.645919841126558e-05, "loss": 4.7055, "step": 7846 }, { "epoch": 0.07083408557501354, "grad_norm": 3.289524555206299, "learning_rate": 4.645874706625745e-05, "loss": 4.9967, "step": 7847 }, { "epoch": 0.07084311247517602, "grad_norm": 3.414774179458618, "learning_rate": 4.645829572124932e-05, "loss": 4.6518, "step": 7848 }, { "epoch": 0.0708521393753385, "grad_norm": 2.601155996322632, "learning_rate": 4.64578443762412e-05, "loss": 4.8551, "step": 7849 }, { "epoch": 0.07086116627550099, "grad_norm": 3.7721378803253174, "learning_rate": 4.6457393031233074e-05, "loss": 4.5839, "step": 7850 }, { "epoch": 0.07087019317566348, "grad_norm": 3.668950319290161, "learning_rate": 4.645694168622495e-05, "loss": 4.8428, "step": 7851 }, { "epoch": 0.07087922007582596, "grad_norm": 2.8577091693878174, "learning_rate": 4.645649034121683e-05, "loss": 4.4999, "step": 7852 }, { "epoch": 0.07088824697598844, "grad_norm": 3.3299455642700195, "learning_rate": 4.6456038996208705e-05, "loss": 5.0561, "step": 7853 }, { "epoch": 0.07089727387615093, "grad_norm": 3.848283529281616, "learning_rate": 4.645558765120058e-05, "loss": 4.1234, "step": 7854 }, { "epoch": 0.07090630077631341, "grad_norm": 3.01816725730896, "learning_rate": 4.6455136306192456e-05, "loss": 4.7182, "step": 7855 }, { "epoch": 0.0709153276764759, "grad_norm": 3.1882266998291016, "learning_rate": 4.645468496118433e-05, "loss": 4.4402, "step": 7856 }, { "epoch": 0.07092435457663838, "grad_norm": 4.117707252502441, "learning_rate": 4.645423361617621e-05, "loss": 4.9094, "step": 7857 }, { "epoch": 0.07093338147680087, "grad_norm": 4.17268180847168, "learning_rate": 4.645378227116808e-05, "loss": 4.8628, "step": 7858 }, { "epoch": 0.07094240837696335, "grad_norm": 5.851168632507324, "learning_rate": 4.645333092615996e-05, "loss": 4.919, "step": 7859 }, { "epoch": 0.07095143527712583, "grad_norm": 3.369016647338867, "learning_rate": 4.645287958115184e-05, "loss": 4.4351, "step": 7860 }, { "epoch": 0.07096046217728832, "grad_norm": 3.5559144020080566, "learning_rate": 4.6452428236143705e-05, "loss": 4.6135, "step": 7861 }, { "epoch": 0.0709694890774508, "grad_norm": 3.190335273742676, "learning_rate": 4.6451976891135584e-05, "loss": 4.8345, "step": 7862 }, { "epoch": 0.07097851597761329, "grad_norm": 4.143226623535156, "learning_rate": 4.645152554612746e-05, "loss": 5.0971, "step": 7863 }, { "epoch": 0.07098754287777577, "grad_norm": 4.669896602630615, "learning_rate": 4.6451074201119336e-05, "loss": 3.9156, "step": 7864 }, { "epoch": 0.07099656977793825, "grad_norm": 3.0413992404937744, "learning_rate": 4.6450622856111215e-05, "loss": 4.8262, "step": 7865 }, { "epoch": 0.07100559667810075, "grad_norm": 7.9502973556518555, "learning_rate": 4.6450171511103094e-05, "loss": 4.5117, "step": 7866 }, { "epoch": 0.07101462357826323, "grad_norm": 2.7501182556152344, "learning_rate": 4.644972016609497e-05, "loss": 4.4988, "step": 7867 }, { "epoch": 0.07102365047842571, "grad_norm": 7.101004123687744, "learning_rate": 4.644926882108684e-05, "loss": 4.1184, "step": 7868 }, { "epoch": 0.07103267737858819, "grad_norm": 4.67785120010376, "learning_rate": 4.644881747607872e-05, "loss": 3.5136, "step": 7869 }, { "epoch": 0.07104170427875067, "grad_norm": 3.2710344791412354, "learning_rate": 4.644836613107059e-05, "loss": 3.8381, "step": 7870 }, { "epoch": 0.07105073117891317, "grad_norm": 3.7741706371307373, "learning_rate": 4.644791478606247e-05, "loss": 4.7532, "step": 7871 }, { "epoch": 0.07105975807907565, "grad_norm": 3.9534757137298584, "learning_rate": 4.644746344105434e-05, "loss": 4.2564, "step": 7872 }, { "epoch": 0.07106878497923813, "grad_norm": 4.574255466461182, "learning_rate": 4.644701209604622e-05, "loss": 4.7146, "step": 7873 }, { "epoch": 0.07107781187940061, "grad_norm": 5.358782768249512, "learning_rate": 4.64465607510381e-05, "loss": 4.9749, "step": 7874 }, { "epoch": 0.0710868387795631, "grad_norm": 5.011303424835205, "learning_rate": 4.644610940602997e-05, "loss": 4.3384, "step": 7875 }, { "epoch": 0.07109586567972558, "grad_norm": 4.182582378387451, "learning_rate": 4.6445658061021846e-05, "loss": 4.3866, "step": 7876 }, { "epoch": 0.07110489257988807, "grad_norm": 4.74317741394043, "learning_rate": 4.6445206716013725e-05, "loss": 3.4567, "step": 7877 }, { "epoch": 0.07111391948005055, "grad_norm": 3.787306308746338, "learning_rate": 4.64447553710056e-05, "loss": 4.9688, "step": 7878 }, { "epoch": 0.07112294638021303, "grad_norm": 3.1085736751556396, "learning_rate": 4.644430402599748e-05, "loss": 4.6176, "step": 7879 }, { "epoch": 0.07113197328037552, "grad_norm": 4.093998432159424, "learning_rate": 4.644385268098935e-05, "loss": 5.191, "step": 7880 }, { "epoch": 0.071141000180538, "grad_norm": 3.3415045738220215, "learning_rate": 4.644340133598122e-05, "loss": 4.3708, "step": 7881 }, { "epoch": 0.07115002708070049, "grad_norm": 3.6447322368621826, "learning_rate": 4.64429499909731e-05, "loss": 4.5211, "step": 7882 }, { "epoch": 0.07115905398086297, "grad_norm": 2.971323251724243, "learning_rate": 4.644249864596498e-05, "loss": 4.3393, "step": 7883 }, { "epoch": 0.07116808088102546, "grad_norm": 4.470567226409912, "learning_rate": 4.644204730095685e-05, "loss": 4.9548, "step": 7884 }, { "epoch": 0.07117710778118794, "grad_norm": 4.502221584320068, "learning_rate": 4.644159595594873e-05, "loss": 5.0356, "step": 7885 }, { "epoch": 0.07118613468135042, "grad_norm": 4.973266124725342, "learning_rate": 4.6441144610940604e-05, "loss": 3.3501, "step": 7886 }, { "epoch": 0.07119516158151291, "grad_norm": 3.0899293422698975, "learning_rate": 4.6440693265932484e-05, "loss": 4.9534, "step": 7887 }, { "epoch": 0.0712041884816754, "grad_norm": 4.239581108093262, "learning_rate": 4.6440241920924356e-05, "loss": 4.7072, "step": 7888 }, { "epoch": 0.07121321538183788, "grad_norm": 3.3415606021881104, "learning_rate": 4.643979057591623e-05, "loss": 4.5842, "step": 7889 }, { "epoch": 0.07122224228200036, "grad_norm": 5.133006572723389, "learning_rate": 4.643933923090811e-05, "loss": 5.2707, "step": 7890 }, { "epoch": 0.07123126918216284, "grad_norm": 3.004148006439209, "learning_rate": 4.643888788589999e-05, "loss": 4.2501, "step": 7891 }, { "epoch": 0.07124029608232534, "grad_norm": 2.798809289932251, "learning_rate": 4.643843654089186e-05, "loss": 4.8065, "step": 7892 }, { "epoch": 0.07124932298248782, "grad_norm": 6.498408317565918, "learning_rate": 4.643798519588374e-05, "loss": 4.5504, "step": 7893 }, { "epoch": 0.0712583498826503, "grad_norm": 4.466810703277588, "learning_rate": 4.643753385087561e-05, "loss": 4.1099, "step": 7894 }, { "epoch": 0.07126737678281278, "grad_norm": 3.8058981895446777, "learning_rate": 4.6437082505867484e-05, "loss": 4.0503, "step": 7895 }, { "epoch": 0.07127640368297526, "grad_norm": 5.681200981140137, "learning_rate": 4.643663116085936e-05, "loss": 4.5667, "step": 7896 }, { "epoch": 0.07128543058313776, "grad_norm": 3.0493266582489014, "learning_rate": 4.6436179815851235e-05, "loss": 4.4308, "step": 7897 }, { "epoch": 0.07129445748330024, "grad_norm": 3.2443718910217285, "learning_rate": 4.6435728470843115e-05, "loss": 4.3738, "step": 7898 }, { "epoch": 0.07130348438346272, "grad_norm": 5.034900188446045, "learning_rate": 4.6435277125834994e-05, "loss": 4.1872, "step": 7899 }, { "epoch": 0.0713125112836252, "grad_norm": 4.767020225524902, "learning_rate": 4.6434825780826866e-05, "loss": 5.0916, "step": 7900 }, { "epoch": 0.07132153818378768, "grad_norm": 3.8902599811553955, "learning_rate": 4.643437443581874e-05, "loss": 4.6553, "step": 7901 }, { "epoch": 0.07133056508395016, "grad_norm": 6.66071891784668, "learning_rate": 4.643392309081062e-05, "loss": 4.9988, "step": 7902 }, { "epoch": 0.07133959198411266, "grad_norm": 4.559886932373047, "learning_rate": 4.643347174580249e-05, "loss": 4.8308, "step": 7903 }, { "epoch": 0.07134861888427514, "grad_norm": 3.5976457595825195, "learning_rate": 4.643302040079437e-05, "loss": 5.0188, "step": 7904 }, { "epoch": 0.07135764578443762, "grad_norm": 3.7478854656219482, "learning_rate": 4.643256905578625e-05, "loss": 4.6267, "step": 7905 }, { "epoch": 0.0713666726846001, "grad_norm": 2.8456499576568604, "learning_rate": 4.643211771077812e-05, "loss": 4.8848, "step": 7906 }, { "epoch": 0.07137569958476259, "grad_norm": 3.5690431594848633, "learning_rate": 4.643166636577e-05, "loss": 3.9019, "step": 7907 }, { "epoch": 0.07138472648492508, "grad_norm": 2.6730031967163086, "learning_rate": 4.643121502076187e-05, "loss": 4.2301, "step": 7908 }, { "epoch": 0.07139375338508756, "grad_norm": 3.5357418060302734, "learning_rate": 4.6430763675753746e-05, "loss": 4.9133, "step": 7909 }, { "epoch": 0.07140278028525004, "grad_norm": 3.453096628189087, "learning_rate": 4.6430312330745625e-05, "loss": 5.1232, "step": 7910 }, { "epoch": 0.07141180718541253, "grad_norm": 3.490478038787842, "learning_rate": 4.64298609857375e-05, "loss": 4.964, "step": 7911 }, { "epoch": 0.07142083408557501, "grad_norm": 5.949339389801025, "learning_rate": 4.6429409640729377e-05, "loss": 4.7769, "step": 7912 }, { "epoch": 0.0714298609857375, "grad_norm": 3.593154191970825, "learning_rate": 4.6428958295721256e-05, "loss": 4.4062, "step": 7913 }, { "epoch": 0.07143888788589998, "grad_norm": 3.3774707317352295, "learning_rate": 4.642850695071313e-05, "loss": 4.3769, "step": 7914 }, { "epoch": 0.07144791478606247, "grad_norm": 3.6473281383514404, "learning_rate": 4.6428055605705e-05, "loss": 4.6992, "step": 7915 }, { "epoch": 0.07145694168622495, "grad_norm": 5.073661804199219, "learning_rate": 4.642760426069688e-05, "loss": 4.7421, "step": 7916 }, { "epoch": 0.07146596858638743, "grad_norm": 3.179551601409912, "learning_rate": 4.642715291568875e-05, "loss": 4.9068, "step": 7917 }, { "epoch": 0.07147499548654992, "grad_norm": 3.30920147895813, "learning_rate": 4.642670157068063e-05, "loss": 4.9779, "step": 7918 }, { "epoch": 0.0714840223867124, "grad_norm": 2.8079915046691895, "learning_rate": 4.6426250225672504e-05, "loss": 4.8635, "step": 7919 }, { "epoch": 0.07149304928687489, "grad_norm": 3.496767520904541, "learning_rate": 4.6425798880664383e-05, "loss": 4.0692, "step": 7920 }, { "epoch": 0.07150207618703737, "grad_norm": 3.3662774562835693, "learning_rate": 4.642534753565626e-05, "loss": 4.2334, "step": 7921 }, { "epoch": 0.07151110308719985, "grad_norm": 7.381502151489258, "learning_rate": 4.642489619064813e-05, "loss": 3.8053, "step": 7922 }, { "epoch": 0.07152012998736235, "grad_norm": 3.13713002204895, "learning_rate": 4.642444484564001e-05, "loss": 4.9849, "step": 7923 }, { "epoch": 0.07152915688752483, "grad_norm": 3.7693166732788086, "learning_rate": 4.642399350063189e-05, "loss": 4.6154, "step": 7924 }, { "epoch": 0.07153818378768731, "grad_norm": 2.86843204498291, "learning_rate": 4.642354215562376e-05, "loss": 5.3534, "step": 7925 }, { "epoch": 0.07154721068784979, "grad_norm": 4.180068016052246, "learning_rate": 4.642309081061564e-05, "loss": 5.3394, "step": 7926 }, { "epoch": 0.07155623758801227, "grad_norm": 4.291769027709961, "learning_rate": 4.642263946560752e-05, "loss": 4.9169, "step": 7927 }, { "epoch": 0.07156526448817475, "grad_norm": 3.148301601409912, "learning_rate": 4.6422188120599383e-05, "loss": 4.7741, "step": 7928 }, { "epoch": 0.07157429138833725, "grad_norm": 2.3156230449676514, "learning_rate": 4.642173677559126e-05, "loss": 5.0076, "step": 7929 }, { "epoch": 0.07158331828849973, "grad_norm": 3.012939691543579, "learning_rate": 4.642128543058314e-05, "loss": 4.9578, "step": 7930 }, { "epoch": 0.07159234518866221, "grad_norm": 2.9334640502929688, "learning_rate": 4.6420834085575014e-05, "loss": 4.7118, "step": 7931 }, { "epoch": 0.0716013720888247, "grad_norm": 3.281142234802246, "learning_rate": 4.6420382740566894e-05, "loss": 4.5426, "step": 7932 }, { "epoch": 0.07161039898898718, "grad_norm": 3.9189722537994385, "learning_rate": 4.6419931395558766e-05, "loss": 4.6274, "step": 7933 }, { "epoch": 0.07161942588914967, "grad_norm": 2.9314239025115967, "learning_rate": 4.6419480050550645e-05, "loss": 4.3704, "step": 7934 }, { "epoch": 0.07162845278931215, "grad_norm": 3.7526092529296875, "learning_rate": 4.641902870554252e-05, "loss": 5.3137, "step": 7935 }, { "epoch": 0.07163747968947463, "grad_norm": 3.786235809326172, "learning_rate": 4.641857736053439e-05, "loss": 4.4736, "step": 7936 }, { "epoch": 0.07164650658963712, "grad_norm": 3.379024028778076, "learning_rate": 4.641812601552627e-05, "loss": 5.2332, "step": 7937 }, { "epoch": 0.0716555334897996, "grad_norm": 3.3696236610412598, "learning_rate": 4.641767467051815e-05, "loss": 4.5992, "step": 7938 }, { "epoch": 0.07166456038996209, "grad_norm": 4.044990062713623, "learning_rate": 4.641722332551002e-05, "loss": 3.8274, "step": 7939 }, { "epoch": 0.07167358729012457, "grad_norm": 4.063887596130371, "learning_rate": 4.64167719805019e-05, "loss": 5.0827, "step": 7940 }, { "epoch": 0.07168261419028706, "grad_norm": 3.5082945823669434, "learning_rate": 4.641632063549377e-05, "loss": 4.235, "step": 7941 }, { "epoch": 0.07169164109044954, "grad_norm": 3.461413621902466, "learning_rate": 4.6415869290485645e-05, "loss": 4.2492, "step": 7942 }, { "epoch": 0.07170066799061202, "grad_norm": 3.0172040462493896, "learning_rate": 4.6415417945477525e-05, "loss": 4.2211, "step": 7943 }, { "epoch": 0.07170969489077451, "grad_norm": 3.026756763458252, "learning_rate": 4.6414966600469404e-05, "loss": 4.5991, "step": 7944 }, { "epoch": 0.071718721790937, "grad_norm": 3.5589842796325684, "learning_rate": 4.6414515255461276e-05, "loss": 4.3087, "step": 7945 }, { "epoch": 0.07172774869109948, "grad_norm": 2.674478054046631, "learning_rate": 4.6414063910453156e-05, "loss": 4.8879, "step": 7946 }, { "epoch": 0.07173677559126196, "grad_norm": 3.421995162963867, "learning_rate": 4.641361256544503e-05, "loss": 4.9927, "step": 7947 }, { "epoch": 0.07174580249142444, "grad_norm": 4.033524990081787, "learning_rate": 4.64131612204369e-05, "loss": 4.9071, "step": 7948 }, { "epoch": 0.07175482939158694, "grad_norm": 2.9392457008361816, "learning_rate": 4.641270987542878e-05, "loss": 4.7144, "step": 7949 }, { "epoch": 0.07176385629174942, "grad_norm": 4.183493137359619, "learning_rate": 4.641225853042065e-05, "loss": 4.9852, "step": 7950 }, { "epoch": 0.0717728831919119, "grad_norm": 3.4477715492248535, "learning_rate": 4.641180718541253e-05, "loss": 5.3112, "step": 7951 }, { "epoch": 0.07178191009207438, "grad_norm": 4.514275074005127, "learning_rate": 4.641135584040441e-05, "loss": 4.6428, "step": 7952 }, { "epoch": 0.07179093699223686, "grad_norm": 3.33396315574646, "learning_rate": 4.641090449539628e-05, "loss": 4.9329, "step": 7953 }, { "epoch": 0.07179996389239934, "grad_norm": 2.985574245452881, "learning_rate": 4.641045315038816e-05, "loss": 5.0809, "step": 7954 }, { "epoch": 0.07180899079256184, "grad_norm": 3.693103313446045, "learning_rate": 4.6410001805380035e-05, "loss": 5.2511, "step": 7955 }, { "epoch": 0.07181801769272432, "grad_norm": 2.96694016456604, "learning_rate": 4.640955046037191e-05, "loss": 4.4233, "step": 7956 }, { "epoch": 0.0718270445928868, "grad_norm": 4.732241153717041, "learning_rate": 4.6409099115363787e-05, "loss": 4.9763, "step": 7957 }, { "epoch": 0.07183607149304928, "grad_norm": 3.3612003326416016, "learning_rate": 4.640864777035566e-05, "loss": 4.6681, "step": 7958 }, { "epoch": 0.07184509839321176, "grad_norm": 3.9829280376434326, "learning_rate": 4.640819642534754e-05, "loss": 4.4388, "step": 7959 }, { "epoch": 0.07185412529337426, "grad_norm": 4.627073287963867, "learning_rate": 4.640774508033942e-05, "loss": 4.7948, "step": 7960 }, { "epoch": 0.07186315219353674, "grad_norm": 3.173351764678955, "learning_rate": 4.640729373533129e-05, "loss": 4.2772, "step": 7961 }, { "epoch": 0.07187217909369922, "grad_norm": 4.2145490646362305, "learning_rate": 4.640684239032316e-05, "loss": 4.6079, "step": 7962 }, { "epoch": 0.0718812059938617, "grad_norm": 2.7221286296844482, "learning_rate": 4.640639104531504e-05, "loss": 4.9634, "step": 7963 }, { "epoch": 0.07189023289402419, "grad_norm": 4.165765285491943, "learning_rate": 4.6405939700306914e-05, "loss": 4.2254, "step": 7964 }, { "epoch": 0.07189925979418668, "grad_norm": 2.573673963546753, "learning_rate": 4.6405488355298793e-05, "loss": 5.0083, "step": 7965 }, { "epoch": 0.07190828669434916, "grad_norm": 3.0631542205810547, "learning_rate": 4.640503701029067e-05, "loss": 5.1084, "step": 7966 }, { "epoch": 0.07191731359451164, "grad_norm": 3.071662187576294, "learning_rate": 4.6404585665282545e-05, "loss": 4.7869, "step": 7967 }, { "epoch": 0.07192634049467413, "grad_norm": 3.6302542686462402, "learning_rate": 4.6404134320274424e-05, "loss": 5.0716, "step": 7968 }, { "epoch": 0.07193536739483661, "grad_norm": 3.9497599601745605, "learning_rate": 4.64036829752663e-05, "loss": 4.691, "step": 7969 }, { "epoch": 0.0719443942949991, "grad_norm": 3.469299793243408, "learning_rate": 4.640323163025817e-05, "loss": 4.8987, "step": 7970 }, { "epoch": 0.07195342119516158, "grad_norm": 2.807058334350586, "learning_rate": 4.640278028525005e-05, "loss": 4.704, "step": 7971 }, { "epoch": 0.07196244809532407, "grad_norm": 3.1679813861846924, "learning_rate": 4.640232894024192e-05, "loss": 4.7703, "step": 7972 }, { "epoch": 0.07197147499548655, "grad_norm": 3.9990270137786865, "learning_rate": 4.64018775952338e-05, "loss": 5.0638, "step": 7973 }, { "epoch": 0.07198050189564903, "grad_norm": 4.4442925453186035, "learning_rate": 4.640142625022568e-05, "loss": 4.8721, "step": 7974 }, { "epoch": 0.07198952879581152, "grad_norm": 3.968702554702759, "learning_rate": 4.6400974905217545e-05, "loss": 4.1152, "step": 7975 }, { "epoch": 0.071998555695974, "grad_norm": 3.457015037536621, "learning_rate": 4.6400523560209424e-05, "loss": 4.8176, "step": 7976 }, { "epoch": 0.07200758259613649, "grad_norm": 3.9703481197357178, "learning_rate": 4.6400072215201304e-05, "loss": 4.6472, "step": 7977 }, { "epoch": 0.07201660949629897, "grad_norm": 2.3613922595977783, "learning_rate": 4.6399620870193176e-05, "loss": 4.7643, "step": 7978 }, { "epoch": 0.07202563639646145, "grad_norm": 3.994720458984375, "learning_rate": 4.6399169525185055e-05, "loss": 5.078, "step": 7979 }, { "epoch": 0.07203466329662393, "grad_norm": 2.7087085247039795, "learning_rate": 4.639871818017693e-05, "loss": 4.2929, "step": 7980 }, { "epoch": 0.07204369019678643, "grad_norm": 3.0466012954711914, "learning_rate": 4.639826683516881e-05, "loss": 5.0306, "step": 7981 }, { "epoch": 0.07205271709694891, "grad_norm": 3.2604293823242188, "learning_rate": 4.639781549016068e-05, "loss": 4.598, "step": 7982 }, { "epoch": 0.07206174399711139, "grad_norm": 3.472783088684082, "learning_rate": 4.639736414515255e-05, "loss": 4.9933, "step": 7983 }, { "epoch": 0.07207077089727387, "grad_norm": 4.340880870819092, "learning_rate": 4.639691280014443e-05, "loss": 5.3168, "step": 7984 }, { "epoch": 0.07207979779743635, "grad_norm": 3.4798989295959473, "learning_rate": 4.639646145513631e-05, "loss": 5.0532, "step": 7985 }, { "epoch": 0.07208882469759885, "grad_norm": 3.361210346221924, "learning_rate": 4.639601011012818e-05, "loss": 4.7082, "step": 7986 }, { "epoch": 0.07209785159776133, "grad_norm": 4.386703968048096, "learning_rate": 4.639555876512006e-05, "loss": 5.2269, "step": 7987 }, { "epoch": 0.07210687849792381, "grad_norm": 3.161046028137207, "learning_rate": 4.639510742011194e-05, "loss": 4.4154, "step": 7988 }, { "epoch": 0.0721159053980863, "grad_norm": 4.060218334197998, "learning_rate": 4.639465607510381e-05, "loss": 4.3522, "step": 7989 }, { "epoch": 0.07212493229824878, "grad_norm": 3.457138776779175, "learning_rate": 4.6394204730095686e-05, "loss": 3.9369, "step": 7990 }, { "epoch": 0.07213395919841127, "grad_norm": 3.428636074066162, "learning_rate": 4.6393753385087566e-05, "loss": 4.2349, "step": 7991 }, { "epoch": 0.07214298609857375, "grad_norm": 3.7188363075256348, "learning_rate": 4.639330204007944e-05, "loss": 4.145, "step": 7992 }, { "epoch": 0.07215201299873623, "grad_norm": 3.182194709777832, "learning_rate": 4.639285069507132e-05, "loss": 4.2561, "step": 7993 }, { "epoch": 0.07216103989889872, "grad_norm": 3.1015632152557373, "learning_rate": 4.639239935006319e-05, "loss": 4.211, "step": 7994 }, { "epoch": 0.0721700667990612, "grad_norm": 3.5736377239227295, "learning_rate": 4.639194800505506e-05, "loss": 4.3561, "step": 7995 }, { "epoch": 0.07217909369922369, "grad_norm": 2.518465280532837, "learning_rate": 4.639149666004694e-05, "loss": 4.8976, "step": 7996 }, { "epoch": 0.07218812059938617, "grad_norm": 3.0190510749816895, "learning_rate": 4.6391045315038814e-05, "loss": 4.4891, "step": 7997 }, { "epoch": 0.07219714749954866, "grad_norm": 3.3173351287841797, "learning_rate": 4.639059397003069e-05, "loss": 4.4342, "step": 7998 }, { "epoch": 0.07220617439971114, "grad_norm": 3.7634615898132324, "learning_rate": 4.639014262502257e-05, "loss": 5.5679, "step": 7999 }, { "epoch": 0.07221520129987362, "grad_norm": 3.0336334705352783, "learning_rate": 4.6389691280014445e-05, "loss": 4.687, "step": 8000 }, { "epoch": 0.07222422820003611, "grad_norm": 3.314568281173706, "learning_rate": 4.6389239935006324e-05, "loss": 4.6032, "step": 8001 }, { "epoch": 0.0722332551001986, "grad_norm": 3.467961311340332, "learning_rate": 4.6388788589998197e-05, "loss": 4.8942, "step": 8002 }, { "epoch": 0.07224228200036108, "grad_norm": 3.4711081981658936, "learning_rate": 4.638833724499007e-05, "loss": 4.4551, "step": 8003 }, { "epoch": 0.07225130890052356, "grad_norm": 3.6310126781463623, "learning_rate": 4.638788589998195e-05, "loss": 5.1916, "step": 8004 }, { "epoch": 0.07226033580068604, "grad_norm": 4.913187503814697, "learning_rate": 4.638743455497383e-05, "loss": 4.8951, "step": 8005 }, { "epoch": 0.07226936270084852, "grad_norm": 5.051694869995117, "learning_rate": 4.63869832099657e-05, "loss": 5.2217, "step": 8006 }, { "epoch": 0.07227838960101102, "grad_norm": 2.6820991039276123, "learning_rate": 4.638653186495758e-05, "loss": 5.022, "step": 8007 }, { "epoch": 0.0722874165011735, "grad_norm": 3.8824150562286377, "learning_rate": 4.638608051994945e-05, "loss": 4.2936, "step": 8008 }, { "epoch": 0.07229644340133598, "grad_norm": 2.798643112182617, "learning_rate": 4.6385629174941324e-05, "loss": 5.2362, "step": 8009 }, { "epoch": 0.07230547030149846, "grad_norm": 2.4147303104400635, "learning_rate": 4.6385177829933203e-05, "loss": 4.964, "step": 8010 }, { "epoch": 0.07231449720166094, "grad_norm": 3.274144172668457, "learning_rate": 4.6384726484925076e-05, "loss": 4.7884, "step": 8011 }, { "epoch": 0.07232352410182344, "grad_norm": 3.1618497371673584, "learning_rate": 4.6384275139916955e-05, "loss": 4.1282, "step": 8012 }, { "epoch": 0.07233255100198592, "grad_norm": 3.5304431915283203, "learning_rate": 4.6383823794908834e-05, "loss": 4.7189, "step": 8013 }, { "epoch": 0.0723415779021484, "grad_norm": 4.095658302307129, "learning_rate": 4.638337244990071e-05, "loss": 4.5831, "step": 8014 }, { "epoch": 0.07235060480231088, "grad_norm": 3.6437342166900635, "learning_rate": 4.638292110489258e-05, "loss": 4.5417, "step": 8015 }, { "epoch": 0.07235963170247336, "grad_norm": 3.4019932746887207, "learning_rate": 4.638246975988446e-05, "loss": 4.9295, "step": 8016 }, { "epoch": 0.07236865860263586, "grad_norm": 3.2924883365631104, "learning_rate": 4.638201841487633e-05, "loss": 4.3493, "step": 8017 }, { "epoch": 0.07237768550279834, "grad_norm": 3.243999719619751, "learning_rate": 4.638156706986821e-05, "loss": 4.6329, "step": 8018 }, { "epoch": 0.07238671240296082, "grad_norm": 3.3836846351623535, "learning_rate": 4.638111572486008e-05, "loss": 5.0558, "step": 8019 }, { "epoch": 0.0723957393031233, "grad_norm": 3.23339581489563, "learning_rate": 4.638066437985196e-05, "loss": 4.8469, "step": 8020 }, { "epoch": 0.07240476620328579, "grad_norm": 5.289975643157959, "learning_rate": 4.638021303484384e-05, "loss": 3.6141, "step": 8021 }, { "epoch": 0.07241379310344828, "grad_norm": 2.6582541465759277, "learning_rate": 4.637976168983571e-05, "loss": 5.2175, "step": 8022 }, { "epoch": 0.07242282000361076, "grad_norm": 3.3521604537963867, "learning_rate": 4.6379310344827586e-05, "loss": 4.3995, "step": 8023 }, { "epoch": 0.07243184690377324, "grad_norm": 3.6278436183929443, "learning_rate": 4.6378858999819465e-05, "loss": 4.4655, "step": 8024 }, { "epoch": 0.07244087380393573, "grad_norm": 2.8708865642547607, "learning_rate": 4.637840765481134e-05, "loss": 4.3159, "step": 8025 }, { "epoch": 0.07244990070409821, "grad_norm": 3.54347562789917, "learning_rate": 4.637795630980322e-05, "loss": 4.6103, "step": 8026 }, { "epoch": 0.0724589276042607, "grad_norm": 3.3708837032318115, "learning_rate": 4.6377504964795096e-05, "loss": 4.6262, "step": 8027 }, { "epoch": 0.07246795450442318, "grad_norm": 4.492128372192383, "learning_rate": 4.637705361978697e-05, "loss": 4.9624, "step": 8028 }, { "epoch": 0.07247698140458567, "grad_norm": 3.6006789207458496, "learning_rate": 4.637660227477884e-05, "loss": 4.4768, "step": 8029 }, { "epoch": 0.07248600830474815, "grad_norm": 3.2628839015960693, "learning_rate": 4.637615092977072e-05, "loss": 4.7695, "step": 8030 }, { "epoch": 0.07249503520491063, "grad_norm": 3.6955983638763428, "learning_rate": 4.637569958476259e-05, "loss": 4.5892, "step": 8031 }, { "epoch": 0.07250406210507311, "grad_norm": 2.5752274990081787, "learning_rate": 4.637524823975447e-05, "loss": 4.728, "step": 8032 }, { "epoch": 0.0725130890052356, "grad_norm": 4.033056259155273, "learning_rate": 4.6374796894746345e-05, "loss": 4.594, "step": 8033 }, { "epoch": 0.07252211590539809, "grad_norm": 3.989469289779663, "learning_rate": 4.6374345549738224e-05, "loss": 4.7333, "step": 8034 }, { "epoch": 0.07253114280556057, "grad_norm": 2.63667368888855, "learning_rate": 4.63738942047301e-05, "loss": 4.5642, "step": 8035 }, { "epoch": 0.07254016970572305, "grad_norm": 2.895679473876953, "learning_rate": 4.637344285972197e-05, "loss": 4.8063, "step": 8036 }, { "epoch": 0.07254919660588553, "grad_norm": 6.982465744018555, "learning_rate": 4.637299151471385e-05, "loss": 4.9101, "step": 8037 }, { "epoch": 0.07255822350604803, "grad_norm": 3.4655911922454834, "learning_rate": 4.637254016970573e-05, "loss": 4.2279, "step": 8038 }, { "epoch": 0.07256725040621051, "grad_norm": 3.6506009101867676, "learning_rate": 4.63720888246976e-05, "loss": 5.0089, "step": 8039 }, { "epoch": 0.07257627730637299, "grad_norm": 4.479267120361328, "learning_rate": 4.637163747968948e-05, "loss": 4.8987, "step": 8040 }, { "epoch": 0.07258530420653547, "grad_norm": 3.686216354370117, "learning_rate": 4.637118613468135e-05, "loss": 3.8383, "step": 8041 }, { "epoch": 0.07259433110669795, "grad_norm": 3.4457194805145264, "learning_rate": 4.6370734789673224e-05, "loss": 5.1331, "step": 8042 }, { "epoch": 0.07260335800686045, "grad_norm": 2.933932065963745, "learning_rate": 4.63702834446651e-05, "loss": 5.3781, "step": 8043 }, { "epoch": 0.07261238490702293, "grad_norm": 3.963005781173706, "learning_rate": 4.636983209965698e-05, "loss": 4.9159, "step": 8044 }, { "epoch": 0.07262141180718541, "grad_norm": 2.7366855144500732, "learning_rate": 4.6369380754648855e-05, "loss": 4.8896, "step": 8045 }, { "epoch": 0.0726304387073479, "grad_norm": 4.8877997398376465, "learning_rate": 4.6368929409640734e-05, "loss": 4.937, "step": 8046 }, { "epoch": 0.07263946560751038, "grad_norm": 2.807555675506592, "learning_rate": 4.6368478064632607e-05, "loss": 4.6689, "step": 8047 }, { "epoch": 0.07264849250767287, "grad_norm": 3.2205231189727783, "learning_rate": 4.6368026719624486e-05, "loss": 5.6207, "step": 8048 }, { "epoch": 0.07265751940783535, "grad_norm": 3.4158260822296143, "learning_rate": 4.636757537461636e-05, "loss": 4.197, "step": 8049 }, { "epoch": 0.07266654630799783, "grad_norm": 3.599465847015381, "learning_rate": 4.636712402960823e-05, "loss": 4.64, "step": 8050 }, { "epoch": 0.07267557320816032, "grad_norm": 3.025569438934326, "learning_rate": 4.636667268460011e-05, "loss": 4.7087, "step": 8051 }, { "epoch": 0.0726846001083228, "grad_norm": 2.7665467262268066, "learning_rate": 4.636622133959199e-05, "loss": 4.524, "step": 8052 }, { "epoch": 0.07269362700848529, "grad_norm": 3.3786978721618652, "learning_rate": 4.636576999458386e-05, "loss": 4.426, "step": 8053 }, { "epoch": 0.07270265390864777, "grad_norm": 3.29411244392395, "learning_rate": 4.636531864957574e-05, "loss": 4.5917, "step": 8054 }, { "epoch": 0.07271168080881026, "grad_norm": 2.7950215339660645, "learning_rate": 4.6364867304567613e-05, "loss": 4.4628, "step": 8055 }, { "epoch": 0.07272070770897274, "grad_norm": 4.198369979858398, "learning_rate": 4.6364415959559486e-05, "loss": 4.8913, "step": 8056 }, { "epoch": 0.07272973460913522, "grad_norm": 3.3225347995758057, "learning_rate": 4.6363964614551365e-05, "loss": 4.291, "step": 8057 }, { "epoch": 0.0727387615092977, "grad_norm": 3.1638755798339844, "learning_rate": 4.636351326954324e-05, "loss": 4.3394, "step": 8058 }, { "epoch": 0.0727477884094602, "grad_norm": 3.4006693363189697, "learning_rate": 4.636306192453512e-05, "loss": 5.037, "step": 8059 }, { "epoch": 0.07275681530962268, "grad_norm": 2.5799896717071533, "learning_rate": 4.6362610579526996e-05, "loss": 5.0889, "step": 8060 }, { "epoch": 0.07276584220978516, "grad_norm": 4.220076084136963, "learning_rate": 4.636215923451887e-05, "loss": 4.9228, "step": 8061 }, { "epoch": 0.07277486910994764, "grad_norm": 3.371936559677124, "learning_rate": 4.636170788951074e-05, "loss": 4.4491, "step": 8062 }, { "epoch": 0.07278389601011012, "grad_norm": 3.634941577911377, "learning_rate": 4.636125654450262e-05, "loss": 3.7447, "step": 8063 }, { "epoch": 0.07279292291027262, "grad_norm": 2.392643451690674, "learning_rate": 4.636080519949449e-05, "loss": 4.664, "step": 8064 }, { "epoch": 0.0728019498104351, "grad_norm": 3.96110463142395, "learning_rate": 4.636035385448637e-05, "loss": 4.4926, "step": 8065 }, { "epoch": 0.07281097671059758, "grad_norm": 2.5451571941375732, "learning_rate": 4.635990250947825e-05, "loss": 5.4993, "step": 8066 }, { "epoch": 0.07282000361076006, "grad_norm": 3.387220859527588, "learning_rate": 4.6359451164470124e-05, "loss": 4.76, "step": 8067 }, { "epoch": 0.07282903051092254, "grad_norm": 6.5696516036987305, "learning_rate": 4.6358999819462e-05, "loss": 4.5359, "step": 8068 }, { "epoch": 0.07283805741108504, "grad_norm": 3.6440985202789307, "learning_rate": 4.6358548474453875e-05, "loss": 4.3957, "step": 8069 }, { "epoch": 0.07284708431124752, "grad_norm": 2.356387138366699, "learning_rate": 4.635809712944575e-05, "loss": 4.3627, "step": 8070 }, { "epoch": 0.07285611121141, "grad_norm": 3.4449801445007324, "learning_rate": 4.635764578443763e-05, "loss": 4.9722, "step": 8071 }, { "epoch": 0.07286513811157248, "grad_norm": 4.0324578285217285, "learning_rate": 4.63571944394295e-05, "loss": 4.2446, "step": 8072 }, { "epoch": 0.07287416501173496, "grad_norm": 3.2162177562713623, "learning_rate": 4.635674309442138e-05, "loss": 4.2806, "step": 8073 }, { "epoch": 0.07288319191189746, "grad_norm": 3.6575324535369873, "learning_rate": 4.635629174941326e-05, "loss": 4.5216, "step": 8074 }, { "epoch": 0.07289221881205994, "grad_norm": 2.8716373443603516, "learning_rate": 4.635584040440513e-05, "loss": 4.9373, "step": 8075 }, { "epoch": 0.07290124571222242, "grad_norm": 3.896867275238037, "learning_rate": 4.6355389059397e-05, "loss": 4.5102, "step": 8076 }, { "epoch": 0.0729102726123849, "grad_norm": 3.0315871238708496, "learning_rate": 4.635493771438888e-05, "loss": 4.2596, "step": 8077 }, { "epoch": 0.07291929951254739, "grad_norm": 2.982604503631592, "learning_rate": 4.6354486369380755e-05, "loss": 5.0734, "step": 8078 }, { "epoch": 0.07292832641270988, "grad_norm": 3.555320978164673, "learning_rate": 4.6354035024372634e-05, "loss": 4.6362, "step": 8079 }, { "epoch": 0.07293735331287236, "grad_norm": 3.177945613861084, "learning_rate": 4.6353583679364506e-05, "loss": 5.0094, "step": 8080 }, { "epoch": 0.07294638021303484, "grad_norm": 3.011488914489746, "learning_rate": 4.6353132334356386e-05, "loss": 4.6683, "step": 8081 }, { "epoch": 0.07295540711319733, "grad_norm": 3.5136141777038574, "learning_rate": 4.6352680989348265e-05, "loss": 4.2689, "step": 8082 }, { "epoch": 0.07296443401335981, "grad_norm": 3.6220505237579346, "learning_rate": 4.635222964434013e-05, "loss": 4.4362, "step": 8083 }, { "epoch": 0.0729734609135223, "grad_norm": 3.7605135440826416, "learning_rate": 4.635177829933201e-05, "loss": 4.2651, "step": 8084 }, { "epoch": 0.07298248781368478, "grad_norm": 3.5053625106811523, "learning_rate": 4.635132695432389e-05, "loss": 5.3561, "step": 8085 }, { "epoch": 0.07299151471384727, "grad_norm": 4.07689094543457, "learning_rate": 4.635087560931576e-05, "loss": 4.5539, "step": 8086 }, { "epoch": 0.07300054161400975, "grad_norm": 4.325682163238525, "learning_rate": 4.635042426430764e-05, "loss": 4.4484, "step": 8087 }, { "epoch": 0.07300956851417223, "grad_norm": 4.393901824951172, "learning_rate": 4.634997291929952e-05, "loss": 3.2611, "step": 8088 }, { "epoch": 0.07301859541433471, "grad_norm": 3.3431296348571777, "learning_rate": 4.6349521574291386e-05, "loss": 3.9856, "step": 8089 }, { "epoch": 0.0730276223144972, "grad_norm": 4.064981937408447, "learning_rate": 4.6349070229283265e-05, "loss": 5.5505, "step": 8090 }, { "epoch": 0.07303664921465969, "grad_norm": 3.3154447078704834, "learning_rate": 4.6348618884275144e-05, "loss": 4.4544, "step": 8091 }, { "epoch": 0.07304567611482217, "grad_norm": 3.0488202571868896, "learning_rate": 4.6348167539267017e-05, "loss": 5.0484, "step": 8092 }, { "epoch": 0.07305470301498465, "grad_norm": 3.8816442489624023, "learning_rate": 4.6347716194258896e-05, "loss": 4.9736, "step": 8093 }, { "epoch": 0.07306372991514713, "grad_norm": 3.2053017616271973, "learning_rate": 4.634726484925077e-05, "loss": 5.1553, "step": 8094 }, { "epoch": 0.07307275681530963, "grad_norm": 3.5565080642700195, "learning_rate": 4.634681350424265e-05, "loss": 4.4312, "step": 8095 }, { "epoch": 0.07308178371547211, "grad_norm": 3.5768113136291504, "learning_rate": 4.634636215923452e-05, "loss": 4.9716, "step": 8096 }, { "epoch": 0.07309081061563459, "grad_norm": 2.86781644821167, "learning_rate": 4.634591081422639e-05, "loss": 4.452, "step": 8097 }, { "epoch": 0.07309983751579707, "grad_norm": 3.0259199142456055, "learning_rate": 4.634545946921827e-05, "loss": 4.9547, "step": 8098 }, { "epoch": 0.07310886441595955, "grad_norm": 4.311614036560059, "learning_rate": 4.634500812421015e-05, "loss": 3.8419, "step": 8099 }, { "epoch": 0.07311789131612205, "grad_norm": 3.972733974456787, "learning_rate": 4.6344556779202023e-05, "loss": 4.8222, "step": 8100 }, { "epoch": 0.07312691821628453, "grad_norm": 3.4041683673858643, "learning_rate": 4.63441054341939e-05, "loss": 4.7407, "step": 8101 }, { "epoch": 0.07313594511644701, "grad_norm": 3.402061700820923, "learning_rate": 4.6343654089185775e-05, "loss": 4.6416, "step": 8102 }, { "epoch": 0.0731449720166095, "grad_norm": 2.805098295211792, "learning_rate": 4.634320274417765e-05, "loss": 4.9495, "step": 8103 }, { "epoch": 0.07315399891677198, "grad_norm": 3.594187021255493, "learning_rate": 4.634275139916953e-05, "loss": 4.707, "step": 8104 }, { "epoch": 0.07316302581693447, "grad_norm": 4.774258136749268, "learning_rate": 4.6342300054161406e-05, "loss": 4.6532, "step": 8105 }, { "epoch": 0.07317205271709695, "grad_norm": 2.5134994983673096, "learning_rate": 4.634184870915328e-05, "loss": 4.331, "step": 8106 }, { "epoch": 0.07318107961725943, "grad_norm": 3.390061378479004, "learning_rate": 4.634139736414516e-05, "loss": 5.0448, "step": 8107 }, { "epoch": 0.07319010651742192, "grad_norm": 3.0464818477630615, "learning_rate": 4.634094601913703e-05, "loss": 5.009, "step": 8108 }, { "epoch": 0.0731991334175844, "grad_norm": 2.8134100437164307, "learning_rate": 4.63404946741289e-05, "loss": 4.8603, "step": 8109 }, { "epoch": 0.07320816031774689, "grad_norm": 3.7000274658203125, "learning_rate": 4.634004332912078e-05, "loss": 5.0466, "step": 8110 }, { "epoch": 0.07321718721790937, "grad_norm": 3.6589536666870117, "learning_rate": 4.6339591984112654e-05, "loss": 4.3938, "step": 8111 }, { "epoch": 0.07322621411807186, "grad_norm": 3.0328242778778076, "learning_rate": 4.6339140639104534e-05, "loss": 4.6802, "step": 8112 }, { "epoch": 0.07323524101823434, "grad_norm": 3.423401117324829, "learning_rate": 4.633868929409641e-05, "loss": 4.805, "step": 8113 }, { "epoch": 0.07324426791839682, "grad_norm": 5.156978130340576, "learning_rate": 4.6338237949088285e-05, "loss": 4.8826, "step": 8114 }, { "epoch": 0.0732532948185593, "grad_norm": 3.813478469848633, "learning_rate": 4.6337786604080165e-05, "loss": 4.8026, "step": 8115 }, { "epoch": 0.0732623217187218, "grad_norm": 3.3973073959350586, "learning_rate": 4.633733525907204e-05, "loss": 4.9428, "step": 8116 }, { "epoch": 0.07327134861888428, "grad_norm": 3.0642192363739014, "learning_rate": 4.633688391406391e-05, "loss": 4.7357, "step": 8117 }, { "epoch": 0.07328037551904676, "grad_norm": 3.155580997467041, "learning_rate": 4.633643256905579e-05, "loss": 4.5471, "step": 8118 }, { "epoch": 0.07328940241920924, "grad_norm": 3.4643688201904297, "learning_rate": 4.633598122404766e-05, "loss": 4.7162, "step": 8119 }, { "epoch": 0.07329842931937172, "grad_norm": 4.197239875793457, "learning_rate": 4.633552987903954e-05, "loss": 4.2321, "step": 8120 }, { "epoch": 0.07330745621953422, "grad_norm": 5.288405895233154, "learning_rate": 4.633507853403142e-05, "loss": 4.7458, "step": 8121 }, { "epoch": 0.0733164831196967, "grad_norm": 4.3480377197265625, "learning_rate": 4.633462718902329e-05, "loss": 4.5724, "step": 8122 }, { "epoch": 0.07332551001985918, "grad_norm": 3.4617667198181152, "learning_rate": 4.6334175844015165e-05, "loss": 4.8429, "step": 8123 }, { "epoch": 0.07333453692002166, "grad_norm": 4.585937023162842, "learning_rate": 4.6333724499007044e-05, "loss": 5.1084, "step": 8124 }, { "epoch": 0.07334356382018414, "grad_norm": 2.7609806060791016, "learning_rate": 4.6333273153998916e-05, "loss": 4.5677, "step": 8125 }, { "epoch": 0.07335259072034664, "grad_norm": 3.9488027095794678, "learning_rate": 4.6332821808990796e-05, "loss": 4.5026, "step": 8126 }, { "epoch": 0.07336161762050912, "grad_norm": 3.449235200881958, "learning_rate": 4.6332370463982675e-05, "loss": 5.2583, "step": 8127 }, { "epoch": 0.0733706445206716, "grad_norm": 4.504028797149658, "learning_rate": 4.633191911897455e-05, "loss": 4.4609, "step": 8128 }, { "epoch": 0.07337967142083408, "grad_norm": 3.6375808715820312, "learning_rate": 4.6331467773966427e-05, "loss": 4.7639, "step": 8129 }, { "epoch": 0.07338869832099656, "grad_norm": 3.0569939613342285, "learning_rate": 4.63310164289583e-05, "loss": 5.0423, "step": 8130 }, { "epoch": 0.07339772522115906, "grad_norm": 3.1740217208862305, "learning_rate": 4.633056508395017e-05, "loss": 5.2484, "step": 8131 }, { "epoch": 0.07340675212132154, "grad_norm": 2.8526406288146973, "learning_rate": 4.633011373894205e-05, "loss": 4.6558, "step": 8132 }, { "epoch": 0.07341577902148402, "grad_norm": 4.395828723907471, "learning_rate": 4.632966239393392e-05, "loss": 4.5608, "step": 8133 }, { "epoch": 0.0734248059216465, "grad_norm": 6.485781669616699, "learning_rate": 4.63292110489258e-05, "loss": 3.8315, "step": 8134 }, { "epoch": 0.07343383282180899, "grad_norm": 2.997360944747925, "learning_rate": 4.632875970391768e-05, "loss": 4.0508, "step": 8135 }, { "epoch": 0.07344285972197148, "grad_norm": 4.44881010055542, "learning_rate": 4.632830835890955e-05, "loss": 4.5633, "step": 8136 }, { "epoch": 0.07345188662213396, "grad_norm": 3.942418098449707, "learning_rate": 4.6327857013901427e-05, "loss": 4.7985, "step": 8137 }, { "epoch": 0.07346091352229644, "grad_norm": 3.0486888885498047, "learning_rate": 4.6327405668893306e-05, "loss": 4.9669, "step": 8138 }, { "epoch": 0.07346994042245893, "grad_norm": 3.7698116302490234, "learning_rate": 4.632695432388518e-05, "loss": 4.4507, "step": 8139 }, { "epoch": 0.07347896732262141, "grad_norm": 3.573911666870117, "learning_rate": 4.632650297887706e-05, "loss": 3.5006, "step": 8140 }, { "epoch": 0.07348799422278389, "grad_norm": 3.364243745803833, "learning_rate": 4.632605163386893e-05, "loss": 4.8419, "step": 8141 }, { "epoch": 0.07349702112294638, "grad_norm": 3.1601951122283936, "learning_rate": 4.632560028886081e-05, "loss": 4.6899, "step": 8142 }, { "epoch": 0.07350604802310887, "grad_norm": 2.809058427810669, "learning_rate": 4.632514894385268e-05, "loss": 4.913, "step": 8143 }, { "epoch": 0.07351507492327135, "grad_norm": 3.5274579524993896, "learning_rate": 4.6324697598844554e-05, "loss": 4.3216, "step": 8144 }, { "epoch": 0.07352410182343383, "grad_norm": 3.260862350463867, "learning_rate": 4.6324246253836433e-05, "loss": 4.9002, "step": 8145 }, { "epoch": 0.07353312872359631, "grad_norm": 3.1534578800201416, "learning_rate": 4.632379490882831e-05, "loss": 5.289, "step": 8146 }, { "epoch": 0.0735421556237588, "grad_norm": 3.9727895259857178, "learning_rate": 4.6323343563820185e-05, "loss": 4.7494, "step": 8147 }, { "epoch": 0.07355118252392129, "grad_norm": 2.7701354026794434, "learning_rate": 4.6322892218812064e-05, "loss": 4.4204, "step": 8148 }, { "epoch": 0.07356020942408377, "grad_norm": 4.062902450561523, "learning_rate": 4.6322440873803944e-05, "loss": 4.7294, "step": 8149 }, { "epoch": 0.07356923632424625, "grad_norm": 3.8418707847595215, "learning_rate": 4.632198952879581e-05, "loss": 4.3514, "step": 8150 }, { "epoch": 0.07357826322440873, "grad_norm": 5.098934173583984, "learning_rate": 4.632153818378769e-05, "loss": 4.3971, "step": 8151 }, { "epoch": 0.07358729012457123, "grad_norm": 3.432979106903076, "learning_rate": 4.632108683877957e-05, "loss": 3.904, "step": 8152 }, { "epoch": 0.07359631702473371, "grad_norm": 4.939630031585693, "learning_rate": 4.632063549377144e-05, "loss": 4.5842, "step": 8153 }, { "epoch": 0.07360534392489619, "grad_norm": 4.242260932922363, "learning_rate": 4.632018414876332e-05, "loss": 4.7749, "step": 8154 }, { "epoch": 0.07361437082505867, "grad_norm": 3.430743932723999, "learning_rate": 4.631973280375519e-05, "loss": 4.553, "step": 8155 }, { "epoch": 0.07362339772522115, "grad_norm": 2.943819522857666, "learning_rate": 4.6319281458747064e-05, "loss": 4.2397, "step": 8156 }, { "epoch": 0.07363242462538365, "grad_norm": 4.298511981964111, "learning_rate": 4.6318830113738944e-05, "loss": 5.215, "step": 8157 }, { "epoch": 0.07364145152554613, "grad_norm": 4.538706302642822, "learning_rate": 4.6318378768730816e-05, "loss": 3.3704, "step": 8158 }, { "epoch": 0.07365047842570861, "grad_norm": 3.540351390838623, "learning_rate": 4.6317927423722695e-05, "loss": 4.7961, "step": 8159 }, { "epoch": 0.0736595053258711, "grad_norm": 3.6722347736358643, "learning_rate": 4.6317476078714575e-05, "loss": 4.3352, "step": 8160 }, { "epoch": 0.07366853222603358, "grad_norm": 3.1141397953033447, "learning_rate": 4.631702473370645e-05, "loss": 4.1617, "step": 8161 }, { "epoch": 0.07367755912619607, "grad_norm": 3.691164016723633, "learning_rate": 4.6316573388698326e-05, "loss": 4.4031, "step": 8162 }, { "epoch": 0.07368658602635855, "grad_norm": 3.7285866737365723, "learning_rate": 4.63161220436902e-05, "loss": 5.1228, "step": 8163 }, { "epoch": 0.07369561292652103, "grad_norm": 4.366866111755371, "learning_rate": 4.631567069868207e-05, "loss": 5.6102, "step": 8164 }, { "epoch": 0.07370463982668352, "grad_norm": 4.593288898468018, "learning_rate": 4.631521935367395e-05, "loss": 4.1922, "step": 8165 }, { "epoch": 0.073713666726846, "grad_norm": 3.493905544281006, "learning_rate": 4.631476800866583e-05, "loss": 4.4229, "step": 8166 }, { "epoch": 0.07372269362700848, "grad_norm": 3.666672706604004, "learning_rate": 4.63143166636577e-05, "loss": 3.0374, "step": 8167 }, { "epoch": 0.07373172052717097, "grad_norm": 3.626862049102783, "learning_rate": 4.631386531864958e-05, "loss": 4.3417, "step": 8168 }, { "epoch": 0.07374074742733346, "grad_norm": 2.9640092849731445, "learning_rate": 4.6313413973641454e-05, "loss": 4.965, "step": 8169 }, { "epoch": 0.07374977432749594, "grad_norm": 8.888297080993652, "learning_rate": 4.6312962628633326e-05, "loss": 3.6236, "step": 8170 }, { "epoch": 0.07375880122765842, "grad_norm": 3.0870087146759033, "learning_rate": 4.6312511283625206e-05, "loss": 4.4943, "step": 8171 }, { "epoch": 0.0737678281278209, "grad_norm": 3.1512889862060547, "learning_rate": 4.631205993861708e-05, "loss": 4.7966, "step": 8172 }, { "epoch": 0.0737768550279834, "grad_norm": 3.1637516021728516, "learning_rate": 4.631160859360896e-05, "loss": 4.3917, "step": 8173 }, { "epoch": 0.07378588192814588, "grad_norm": 3.9898324012756348, "learning_rate": 4.6311157248600837e-05, "loss": 5.1679, "step": 8174 }, { "epoch": 0.07379490882830836, "grad_norm": 3.7464499473571777, "learning_rate": 4.631070590359271e-05, "loss": 5.3138, "step": 8175 }, { "epoch": 0.07380393572847084, "grad_norm": 2.191575527191162, "learning_rate": 4.631025455858458e-05, "loss": 4.3814, "step": 8176 }, { "epoch": 0.07381296262863332, "grad_norm": 3.6892502307891846, "learning_rate": 4.630980321357646e-05, "loss": 5.2014, "step": 8177 }, { "epoch": 0.07382198952879582, "grad_norm": 3.146557331085205, "learning_rate": 4.630935186856833e-05, "loss": 4.323, "step": 8178 }, { "epoch": 0.0738310164289583, "grad_norm": 3.6409478187561035, "learning_rate": 4.630890052356021e-05, "loss": 4.5692, "step": 8179 }, { "epoch": 0.07384004332912078, "grad_norm": 4.453668117523193, "learning_rate": 4.6308449178552085e-05, "loss": 4.8749, "step": 8180 }, { "epoch": 0.07384907022928326, "grad_norm": 3.4981558322906494, "learning_rate": 4.6307997833543964e-05, "loss": 4.9487, "step": 8181 }, { "epoch": 0.07385809712944574, "grad_norm": 3.7348668575286865, "learning_rate": 4.630754648853584e-05, "loss": 4.0571, "step": 8182 }, { "epoch": 0.07386712402960824, "grad_norm": 4.67234468460083, "learning_rate": 4.630709514352771e-05, "loss": 4.3116, "step": 8183 }, { "epoch": 0.07387615092977072, "grad_norm": 3.4724748134613037, "learning_rate": 4.630664379851959e-05, "loss": 4.8646, "step": 8184 }, { "epoch": 0.0738851778299332, "grad_norm": 3.306396722793579, "learning_rate": 4.630619245351147e-05, "loss": 4.0668, "step": 8185 }, { "epoch": 0.07389420473009568, "grad_norm": 3.9905240535736084, "learning_rate": 4.630574110850334e-05, "loss": 4.5123, "step": 8186 }, { "epoch": 0.07390323163025816, "grad_norm": 3.3825628757476807, "learning_rate": 4.630528976349522e-05, "loss": 4.7183, "step": 8187 }, { "epoch": 0.07391225853042066, "grad_norm": 3.779982328414917, "learning_rate": 4.63048384184871e-05, "loss": 4.9079, "step": 8188 }, { "epoch": 0.07392128543058314, "grad_norm": 3.6501455307006836, "learning_rate": 4.630438707347897e-05, "loss": 4.5684, "step": 8189 }, { "epoch": 0.07393031233074562, "grad_norm": 3.401156187057495, "learning_rate": 4.630393572847084e-05, "loss": 4.6199, "step": 8190 }, { "epoch": 0.0739393392309081, "grad_norm": 4.473331451416016, "learning_rate": 4.630348438346272e-05, "loss": 4.2858, "step": 8191 }, { "epoch": 0.07394836613107059, "grad_norm": 3.509124755859375, "learning_rate": 4.6303033038454595e-05, "loss": 4.4871, "step": 8192 }, { "epoch": 0.07395739303123307, "grad_norm": 3.9048871994018555, "learning_rate": 4.6302581693446474e-05, "loss": 4.3161, "step": 8193 }, { "epoch": 0.07396641993139556, "grad_norm": 3.4025282859802246, "learning_rate": 4.630213034843835e-05, "loss": 4.5782, "step": 8194 }, { "epoch": 0.07397544683155804, "grad_norm": 2.9957962036132812, "learning_rate": 4.6301679003430226e-05, "loss": 3.9186, "step": 8195 }, { "epoch": 0.07398447373172053, "grad_norm": 3.898449182510376, "learning_rate": 4.6301227658422105e-05, "loss": 4.5952, "step": 8196 }, { "epoch": 0.07399350063188301, "grad_norm": 3.101191759109497, "learning_rate": 4.630077631341397e-05, "loss": 4.7061, "step": 8197 }, { "epoch": 0.07400252753204549, "grad_norm": 2.8399059772491455, "learning_rate": 4.630032496840585e-05, "loss": 4.8312, "step": 8198 }, { "epoch": 0.07401155443220798, "grad_norm": 3.1424577236175537, "learning_rate": 4.629987362339773e-05, "loss": 4.0015, "step": 8199 }, { "epoch": 0.07402058133237047, "grad_norm": 3.6248831748962402, "learning_rate": 4.62994222783896e-05, "loss": 4.5601, "step": 8200 }, { "epoch": 0.07402960823253295, "grad_norm": 3.142610549926758, "learning_rate": 4.629897093338148e-05, "loss": 4.604, "step": 8201 }, { "epoch": 0.07403863513269543, "grad_norm": 3.3641581535339355, "learning_rate": 4.6298519588373354e-05, "loss": 4.3592, "step": 8202 }, { "epoch": 0.07404766203285791, "grad_norm": 4.546741962432861, "learning_rate": 4.6298068243365226e-05, "loss": 3.3672, "step": 8203 }, { "epoch": 0.0740566889330204, "grad_norm": 3.5222384929656982, "learning_rate": 4.6297616898357105e-05, "loss": 4.2776, "step": 8204 }, { "epoch": 0.07406571583318289, "grad_norm": 5.178517818450928, "learning_rate": 4.6297165553348985e-05, "loss": 5.071, "step": 8205 }, { "epoch": 0.07407474273334537, "grad_norm": 3.1274654865264893, "learning_rate": 4.629671420834086e-05, "loss": 4.3028, "step": 8206 }, { "epoch": 0.07408376963350785, "grad_norm": 3.8485734462738037, "learning_rate": 4.6296262863332736e-05, "loss": 4.8144, "step": 8207 }, { "epoch": 0.07409279653367033, "grad_norm": 4.436586380004883, "learning_rate": 4.629581151832461e-05, "loss": 4.918, "step": 8208 }, { "epoch": 0.07410182343383283, "grad_norm": 4.233644962310791, "learning_rate": 4.629536017331649e-05, "loss": 3.54, "step": 8209 }, { "epoch": 0.07411085033399531, "grad_norm": 3.7141213417053223, "learning_rate": 4.629490882830836e-05, "loss": 4.8143, "step": 8210 }, { "epoch": 0.07411987723415779, "grad_norm": 3.623608112335205, "learning_rate": 4.629445748330023e-05, "loss": 4.6818, "step": 8211 }, { "epoch": 0.07412890413432027, "grad_norm": 3.5425126552581787, "learning_rate": 4.629400613829211e-05, "loss": 4.6601, "step": 8212 }, { "epoch": 0.07413793103448275, "grad_norm": 4.037015914916992, "learning_rate": 4.629355479328399e-05, "loss": 4.5675, "step": 8213 }, { "epoch": 0.07414695793464525, "grad_norm": 3.3139748573303223, "learning_rate": 4.6293103448275864e-05, "loss": 4.2162, "step": 8214 }, { "epoch": 0.07415598483480773, "grad_norm": 3.185302972793579, "learning_rate": 4.629265210326774e-05, "loss": 4.6319, "step": 8215 }, { "epoch": 0.07416501173497021, "grad_norm": 2.9921276569366455, "learning_rate": 4.6292200758259616e-05, "loss": 4.1339, "step": 8216 }, { "epoch": 0.0741740386351327, "grad_norm": 3.6392276287078857, "learning_rate": 4.629174941325149e-05, "loss": 4.7173, "step": 8217 }, { "epoch": 0.07418306553529518, "grad_norm": 3.4988577365875244, "learning_rate": 4.629129806824337e-05, "loss": 4.709, "step": 8218 }, { "epoch": 0.07419209243545766, "grad_norm": 2.788655996322632, "learning_rate": 4.629084672323524e-05, "loss": 4.9014, "step": 8219 }, { "epoch": 0.07420111933562015, "grad_norm": 3.6460630893707275, "learning_rate": 4.629039537822712e-05, "loss": 4.3907, "step": 8220 }, { "epoch": 0.07421014623578263, "grad_norm": 3.6016812324523926, "learning_rate": 4.6289944033219e-05, "loss": 4.3433, "step": 8221 }, { "epoch": 0.07421917313594512, "grad_norm": 4.421736717224121, "learning_rate": 4.628949268821087e-05, "loss": 4.55, "step": 8222 }, { "epoch": 0.0742282000361076, "grad_norm": 4.28110408782959, "learning_rate": 4.628904134320274e-05, "loss": 4.453, "step": 8223 }, { "epoch": 0.07423722693627008, "grad_norm": 2.892091751098633, "learning_rate": 4.628858999819462e-05, "loss": 4.7672, "step": 8224 }, { "epoch": 0.07424625383643257, "grad_norm": 3.995738983154297, "learning_rate": 4.6288138653186495e-05, "loss": 4.881, "step": 8225 }, { "epoch": 0.07425528073659506, "grad_norm": 2.6401467323303223, "learning_rate": 4.6287687308178374e-05, "loss": 4.1492, "step": 8226 }, { "epoch": 0.07426430763675754, "grad_norm": 2.6437220573425293, "learning_rate": 4.628723596317025e-05, "loss": 3.9224, "step": 8227 }, { "epoch": 0.07427333453692002, "grad_norm": 3.5523383617401123, "learning_rate": 4.6286784618162126e-05, "loss": 5.0959, "step": 8228 }, { "epoch": 0.0742823614370825, "grad_norm": 4.227432727813721, "learning_rate": 4.6286333273154005e-05, "loss": 5.0741, "step": 8229 }, { "epoch": 0.074291388337245, "grad_norm": 3.338059902191162, "learning_rate": 4.628588192814588e-05, "loss": 5.2009, "step": 8230 }, { "epoch": 0.07430041523740748, "grad_norm": 5.396544933319092, "learning_rate": 4.628543058313775e-05, "loss": 4.5348, "step": 8231 }, { "epoch": 0.07430944213756996, "grad_norm": 3.8097550868988037, "learning_rate": 4.628497923812963e-05, "loss": 5.1711, "step": 8232 }, { "epoch": 0.07431846903773244, "grad_norm": 3.822953939437866, "learning_rate": 4.62845278931215e-05, "loss": 4.4272, "step": 8233 }, { "epoch": 0.07432749593789492, "grad_norm": 3.194514274597168, "learning_rate": 4.628407654811338e-05, "loss": 4.1468, "step": 8234 }, { "epoch": 0.07433652283805742, "grad_norm": 3.8343281745910645, "learning_rate": 4.628362520310526e-05, "loss": 3.1528, "step": 8235 }, { "epoch": 0.0743455497382199, "grad_norm": 4.327957630157471, "learning_rate": 4.628317385809713e-05, "loss": 4.9143, "step": 8236 }, { "epoch": 0.07435457663838238, "grad_norm": 3.6133463382720947, "learning_rate": 4.6282722513089005e-05, "loss": 4.3262, "step": 8237 }, { "epoch": 0.07436360353854486, "grad_norm": 3.3281140327453613, "learning_rate": 4.6282271168080884e-05, "loss": 4.2276, "step": 8238 }, { "epoch": 0.07437263043870734, "grad_norm": 3.473405361175537, "learning_rate": 4.628181982307276e-05, "loss": 4.4407, "step": 8239 }, { "epoch": 0.07438165733886984, "grad_norm": 5.437019348144531, "learning_rate": 4.6281368478064636e-05, "loss": 4.1919, "step": 8240 }, { "epoch": 0.07439068423903232, "grad_norm": 2.9513936042785645, "learning_rate": 4.628091713305651e-05, "loss": 4.0566, "step": 8241 }, { "epoch": 0.0743997111391948, "grad_norm": 4.603085041046143, "learning_rate": 4.628046578804839e-05, "loss": 4.7885, "step": 8242 }, { "epoch": 0.07440873803935728, "grad_norm": 3.8465495109558105, "learning_rate": 4.628001444304027e-05, "loss": 4.3567, "step": 8243 }, { "epoch": 0.07441776493951976, "grad_norm": 3.690699577331543, "learning_rate": 4.627956309803213e-05, "loss": 4.0849, "step": 8244 }, { "epoch": 0.07442679183968225, "grad_norm": 4.268357753753662, "learning_rate": 4.627911175302401e-05, "loss": 5.6924, "step": 8245 }, { "epoch": 0.07443581873984474, "grad_norm": 3.125121593475342, "learning_rate": 4.627866040801589e-05, "loss": 5.0163, "step": 8246 }, { "epoch": 0.07444484564000722, "grad_norm": 5.079281330108643, "learning_rate": 4.6278209063007764e-05, "loss": 3.7234, "step": 8247 }, { "epoch": 0.0744538725401697, "grad_norm": 3.232224464416504, "learning_rate": 4.627775771799964e-05, "loss": 4.7401, "step": 8248 }, { "epoch": 0.07446289944033219, "grad_norm": 2.881441831588745, "learning_rate": 4.627730637299152e-05, "loss": 5.5071, "step": 8249 }, { "epoch": 0.07447192634049467, "grad_norm": 3.132047653198242, "learning_rate": 4.627685502798339e-05, "loss": 4.6103, "step": 8250 }, { "epoch": 0.07448095324065716, "grad_norm": 3.250298500061035, "learning_rate": 4.627640368297527e-05, "loss": 4.8534, "step": 8251 }, { "epoch": 0.07448998014081964, "grad_norm": 4.177177906036377, "learning_rate": 4.6275952337967146e-05, "loss": 4.7032, "step": 8252 }, { "epoch": 0.07449900704098213, "grad_norm": 3.3854246139526367, "learning_rate": 4.627550099295902e-05, "loss": 3.8964, "step": 8253 }, { "epoch": 0.07450803394114461, "grad_norm": 4.631128311157227, "learning_rate": 4.62750496479509e-05, "loss": 3.9259, "step": 8254 }, { "epoch": 0.07451706084130709, "grad_norm": 4.43096399307251, "learning_rate": 4.627459830294277e-05, "loss": 4.8931, "step": 8255 }, { "epoch": 0.07452608774146958, "grad_norm": 3.3077309131622314, "learning_rate": 4.627414695793465e-05, "loss": 4.5933, "step": 8256 }, { "epoch": 0.07453511464163207, "grad_norm": 3.250535011291504, "learning_rate": 4.627369561292652e-05, "loss": 4.7585, "step": 8257 }, { "epoch": 0.07454414154179455, "grad_norm": 4.072688102722168, "learning_rate": 4.6273244267918395e-05, "loss": 4.1363, "step": 8258 }, { "epoch": 0.07455316844195703, "grad_norm": 4.02022647857666, "learning_rate": 4.6272792922910274e-05, "loss": 4.9912, "step": 8259 }, { "epoch": 0.07456219534211951, "grad_norm": 3.3189284801483154, "learning_rate": 4.627234157790215e-05, "loss": 4.8115, "step": 8260 }, { "epoch": 0.074571222242282, "grad_norm": 4.2395243644714355, "learning_rate": 4.6271890232894026e-05, "loss": 4.7574, "step": 8261 }, { "epoch": 0.07458024914244449, "grad_norm": 3.736349105834961, "learning_rate": 4.6271438887885905e-05, "loss": 5.4032, "step": 8262 }, { "epoch": 0.07458927604260697, "grad_norm": 3.7683770656585693, "learning_rate": 4.627098754287778e-05, "loss": 4.8946, "step": 8263 }, { "epoch": 0.07459830294276945, "grad_norm": 2.4529178142547607, "learning_rate": 4.627053619786965e-05, "loss": 5.1844, "step": 8264 }, { "epoch": 0.07460732984293193, "grad_norm": 3.530352830886841, "learning_rate": 4.627008485286153e-05, "loss": 4.6201, "step": 8265 }, { "epoch": 0.07461635674309443, "grad_norm": 3.427509307861328, "learning_rate": 4.626963350785341e-05, "loss": 4.6912, "step": 8266 }, { "epoch": 0.07462538364325691, "grad_norm": 3.0262646675109863, "learning_rate": 4.626918216284528e-05, "loss": 4.8946, "step": 8267 }, { "epoch": 0.07463441054341939, "grad_norm": 2.6047251224517822, "learning_rate": 4.626873081783716e-05, "loss": 4.3009, "step": 8268 }, { "epoch": 0.07464343744358187, "grad_norm": 3.637054920196533, "learning_rate": 4.626827947282903e-05, "loss": 4.5286, "step": 8269 }, { "epoch": 0.07465246434374435, "grad_norm": 4.3090643882751465, "learning_rate": 4.6267828127820905e-05, "loss": 4.8731, "step": 8270 }, { "epoch": 0.07466149124390684, "grad_norm": 2.868421792984009, "learning_rate": 4.6267376782812784e-05, "loss": 5.0997, "step": 8271 }, { "epoch": 0.07467051814406933, "grad_norm": 4.208822250366211, "learning_rate": 4.6266925437804657e-05, "loss": 4.2284, "step": 8272 }, { "epoch": 0.07467954504423181, "grad_norm": 4.731010437011719, "learning_rate": 4.6266474092796536e-05, "loss": 5.3491, "step": 8273 }, { "epoch": 0.0746885719443943, "grad_norm": 3.0424301624298096, "learning_rate": 4.6266022747788415e-05, "loss": 4.9526, "step": 8274 }, { "epoch": 0.07469759884455678, "grad_norm": 3.070258140563965, "learning_rate": 4.626557140278029e-05, "loss": 4.4685, "step": 8275 }, { "epoch": 0.07470662574471926, "grad_norm": 3.307788133621216, "learning_rate": 4.626512005777217e-05, "loss": 4.8947, "step": 8276 }, { "epoch": 0.07471565264488175, "grad_norm": 3.4385745525360107, "learning_rate": 4.626466871276404e-05, "loss": 3.6787, "step": 8277 }, { "epoch": 0.07472467954504423, "grad_norm": 3.1957688331604004, "learning_rate": 4.626421736775591e-05, "loss": 4.899, "step": 8278 }, { "epoch": 0.07473370644520672, "grad_norm": 2.9348533153533936, "learning_rate": 4.626376602274779e-05, "loss": 4.7257, "step": 8279 }, { "epoch": 0.0747427333453692, "grad_norm": 3.3077070713043213, "learning_rate": 4.626331467773966e-05, "loss": 4.9935, "step": 8280 }, { "epoch": 0.07475176024553168, "grad_norm": 3.6753435134887695, "learning_rate": 4.626286333273154e-05, "loss": 4.7846, "step": 8281 }, { "epoch": 0.07476078714569417, "grad_norm": 7.716330051422119, "learning_rate": 4.626241198772342e-05, "loss": 4.5391, "step": 8282 }, { "epoch": 0.07476981404585666, "grad_norm": 2.7808353900909424, "learning_rate": 4.6261960642715294e-05, "loss": 4.4368, "step": 8283 }, { "epoch": 0.07477884094601914, "grad_norm": 3.0107948780059814, "learning_rate": 4.626150929770717e-05, "loss": 4.7158, "step": 8284 }, { "epoch": 0.07478786784618162, "grad_norm": 3.320638656616211, "learning_rate": 4.6261057952699046e-05, "loss": 5.199, "step": 8285 }, { "epoch": 0.0747968947463441, "grad_norm": 3.3139426708221436, "learning_rate": 4.626060660769092e-05, "loss": 4.6547, "step": 8286 }, { "epoch": 0.0748059216465066, "grad_norm": 4.265475749969482, "learning_rate": 4.62601552626828e-05, "loss": 4.3664, "step": 8287 }, { "epoch": 0.07481494854666908, "grad_norm": 2.970686674118042, "learning_rate": 4.625970391767468e-05, "loss": 4.7907, "step": 8288 }, { "epoch": 0.07482397544683156, "grad_norm": 2.953031301498413, "learning_rate": 4.625925257266655e-05, "loss": 4.4362, "step": 8289 }, { "epoch": 0.07483300234699404, "grad_norm": 2.7146763801574707, "learning_rate": 4.625880122765843e-05, "loss": 5.1697, "step": 8290 }, { "epoch": 0.07484202924715652, "grad_norm": 3.443817615509033, "learning_rate": 4.62583498826503e-05, "loss": 5.1616, "step": 8291 }, { "epoch": 0.07485105614731902, "grad_norm": 3.2681000232696533, "learning_rate": 4.6257898537642174e-05, "loss": 4.7723, "step": 8292 }, { "epoch": 0.0748600830474815, "grad_norm": 3.0536558628082275, "learning_rate": 4.625744719263405e-05, "loss": 4.9338, "step": 8293 }, { "epoch": 0.07486910994764398, "grad_norm": 3.0676820278167725, "learning_rate": 4.6256995847625925e-05, "loss": 4.532, "step": 8294 }, { "epoch": 0.07487813684780646, "grad_norm": 3.2688183784484863, "learning_rate": 4.6256544502617805e-05, "loss": 4.5149, "step": 8295 }, { "epoch": 0.07488716374796894, "grad_norm": 3.822629690170288, "learning_rate": 4.6256093157609684e-05, "loss": 4.5577, "step": 8296 }, { "epoch": 0.07489619064813142, "grad_norm": 4.366918087005615, "learning_rate": 4.625564181260155e-05, "loss": 4.474, "step": 8297 }, { "epoch": 0.07490521754829392, "grad_norm": 3.5134646892547607, "learning_rate": 4.625519046759343e-05, "loss": 4.6366, "step": 8298 }, { "epoch": 0.0749142444484564, "grad_norm": 3.069972276687622, "learning_rate": 4.625473912258531e-05, "loss": 4.9579, "step": 8299 }, { "epoch": 0.07492327134861888, "grad_norm": 3.8860368728637695, "learning_rate": 4.625428777757718e-05, "loss": 4.3303, "step": 8300 }, { "epoch": 0.07493229824878136, "grad_norm": 3.678077220916748, "learning_rate": 4.625383643256906e-05, "loss": 4.9194, "step": 8301 }, { "epoch": 0.07494132514894385, "grad_norm": 3.1692018508911133, "learning_rate": 4.625338508756093e-05, "loss": 4.7183, "step": 8302 }, { "epoch": 0.07495035204910634, "grad_norm": 3.138141393661499, "learning_rate": 4.625293374255281e-05, "loss": 4.9593, "step": 8303 }, { "epoch": 0.07495937894926882, "grad_norm": 3.568398952484131, "learning_rate": 4.6252482397544684e-05, "loss": 4.4034, "step": 8304 }, { "epoch": 0.0749684058494313, "grad_norm": 5.8114094734191895, "learning_rate": 4.6252031052536556e-05, "loss": 4.3295, "step": 8305 }, { "epoch": 0.07497743274959379, "grad_norm": 3.287593364715576, "learning_rate": 4.6251579707528436e-05, "loss": 4.3225, "step": 8306 }, { "epoch": 0.07498645964975627, "grad_norm": 4.324321746826172, "learning_rate": 4.6251128362520315e-05, "loss": 4.5576, "step": 8307 }, { "epoch": 0.07499548654991876, "grad_norm": 3.934300661087036, "learning_rate": 4.625067701751219e-05, "loss": 4.1168, "step": 8308 }, { "epoch": 0.07500451345008124, "grad_norm": 3.209214210510254, "learning_rate": 4.6250225672504066e-05, "loss": 4.4249, "step": 8309 }, { "epoch": 0.07501354035024373, "grad_norm": 3.0061285495758057, "learning_rate": 4.6249774327495946e-05, "loss": 5.292, "step": 8310 }, { "epoch": 0.07502256725040621, "grad_norm": 3.001983880996704, "learning_rate": 4.624932298248781e-05, "loss": 4.408, "step": 8311 }, { "epoch": 0.07503159415056869, "grad_norm": 3.268056631088257, "learning_rate": 4.624887163747969e-05, "loss": 4.2227, "step": 8312 }, { "epoch": 0.07504062105073118, "grad_norm": 3.185771942138672, "learning_rate": 4.624842029247157e-05, "loss": 5.1112, "step": 8313 }, { "epoch": 0.07504964795089367, "grad_norm": 4.012211799621582, "learning_rate": 4.624796894746344e-05, "loss": 4.642, "step": 8314 }, { "epoch": 0.07505867485105615, "grad_norm": 2.907374620437622, "learning_rate": 4.624751760245532e-05, "loss": 4.264, "step": 8315 }, { "epoch": 0.07506770175121863, "grad_norm": 3.0152900218963623, "learning_rate": 4.6247066257447194e-05, "loss": 4.9856, "step": 8316 }, { "epoch": 0.07507672865138111, "grad_norm": 3.747774839401245, "learning_rate": 4.6246614912439067e-05, "loss": 4.4949, "step": 8317 }, { "epoch": 0.0750857555515436, "grad_norm": 3.3906097412109375, "learning_rate": 4.6246163567430946e-05, "loss": 5.0709, "step": 8318 }, { "epoch": 0.07509478245170609, "grad_norm": 2.8061935901641846, "learning_rate": 4.624571222242282e-05, "loss": 4.4533, "step": 8319 }, { "epoch": 0.07510380935186857, "grad_norm": 3.1000783443450928, "learning_rate": 4.62452608774147e-05, "loss": 4.3568, "step": 8320 }, { "epoch": 0.07511283625203105, "grad_norm": 3.2145209312438965, "learning_rate": 4.624480953240658e-05, "loss": 5.1926, "step": 8321 }, { "epoch": 0.07512186315219353, "grad_norm": 2.7893259525299072, "learning_rate": 4.624435818739845e-05, "loss": 4.8572, "step": 8322 }, { "epoch": 0.07513089005235601, "grad_norm": 7.8501081466674805, "learning_rate": 4.624390684239033e-05, "loss": 4.9888, "step": 8323 }, { "epoch": 0.07513991695251851, "grad_norm": 2.8814103603363037, "learning_rate": 4.62434554973822e-05, "loss": 4.5992, "step": 8324 }, { "epoch": 0.07514894385268099, "grad_norm": 4.418187141418457, "learning_rate": 4.624300415237407e-05, "loss": 4.2986, "step": 8325 }, { "epoch": 0.07515797075284347, "grad_norm": 3.787379026412964, "learning_rate": 4.624255280736595e-05, "loss": 4.9107, "step": 8326 }, { "epoch": 0.07516699765300595, "grad_norm": 4.033364772796631, "learning_rate": 4.624210146235783e-05, "loss": 4.9688, "step": 8327 }, { "epoch": 0.07517602455316844, "grad_norm": 2.597264528274536, "learning_rate": 4.6241650117349704e-05, "loss": 4.9922, "step": 8328 }, { "epoch": 0.07518505145333093, "grad_norm": 3.220013380050659, "learning_rate": 4.6241198772341584e-05, "loss": 4.1538, "step": 8329 }, { "epoch": 0.07519407835349341, "grad_norm": 3.754060745239258, "learning_rate": 4.6240747427333456e-05, "loss": 5.0007, "step": 8330 }, { "epoch": 0.0752031052536559, "grad_norm": 4.521826267242432, "learning_rate": 4.624029608232533e-05, "loss": 4.0331, "step": 8331 }, { "epoch": 0.07521213215381838, "grad_norm": 3.520587205886841, "learning_rate": 4.623984473731721e-05, "loss": 4.9557, "step": 8332 }, { "epoch": 0.07522115905398086, "grad_norm": 4.542074203491211, "learning_rate": 4.623939339230908e-05, "loss": 5.0354, "step": 8333 }, { "epoch": 0.07523018595414335, "grad_norm": 3.080671548843384, "learning_rate": 4.623894204730096e-05, "loss": 5.0533, "step": 8334 }, { "epoch": 0.07523921285430583, "grad_norm": 3.456998348236084, "learning_rate": 4.623849070229284e-05, "loss": 3.504, "step": 8335 }, { "epoch": 0.07524823975446832, "grad_norm": 3.819103956222534, "learning_rate": 4.623803935728471e-05, "loss": 4.6386, "step": 8336 }, { "epoch": 0.0752572666546308, "grad_norm": 3.0601558685302734, "learning_rate": 4.623758801227659e-05, "loss": 4.8378, "step": 8337 }, { "epoch": 0.07526629355479328, "grad_norm": 3.1887400150299072, "learning_rate": 4.623713666726846e-05, "loss": 4.8732, "step": 8338 }, { "epoch": 0.07527532045495577, "grad_norm": 4.215680122375488, "learning_rate": 4.6236685322260335e-05, "loss": 4.6115, "step": 8339 }, { "epoch": 0.07528434735511826, "grad_norm": 3.625426769256592, "learning_rate": 4.6236233977252215e-05, "loss": 4.9391, "step": 8340 }, { "epoch": 0.07529337425528074, "grad_norm": 2.905801773071289, "learning_rate": 4.623578263224409e-05, "loss": 4.9155, "step": 8341 }, { "epoch": 0.07530240115544322, "grad_norm": 3.8327951431274414, "learning_rate": 4.6235331287235966e-05, "loss": 4.1158, "step": 8342 }, { "epoch": 0.0753114280556057, "grad_norm": 3.5469188690185547, "learning_rate": 4.6234879942227845e-05, "loss": 5.2512, "step": 8343 }, { "epoch": 0.0753204549557682, "grad_norm": 3.3356857299804688, "learning_rate": 4.623442859721971e-05, "loss": 4.57, "step": 8344 }, { "epoch": 0.07532948185593068, "grad_norm": 3.7506518363952637, "learning_rate": 4.623397725221159e-05, "loss": 4.9895, "step": 8345 }, { "epoch": 0.07533850875609316, "grad_norm": 2.4001357555389404, "learning_rate": 4.623352590720347e-05, "loss": 4.5597, "step": 8346 }, { "epoch": 0.07534753565625564, "grad_norm": 4.3353495597839355, "learning_rate": 4.623307456219534e-05, "loss": 4.38, "step": 8347 }, { "epoch": 0.07535656255641812, "grad_norm": 2.8255865573883057, "learning_rate": 4.623262321718722e-05, "loss": 4.0378, "step": 8348 }, { "epoch": 0.0753655894565806, "grad_norm": 2.889606475830078, "learning_rate": 4.62321718721791e-05, "loss": 4.5719, "step": 8349 }, { "epoch": 0.0753746163567431, "grad_norm": 5.290503978729248, "learning_rate": 4.623172052717097e-05, "loss": 4.6801, "step": 8350 }, { "epoch": 0.07538364325690558, "grad_norm": 3.61333966255188, "learning_rate": 4.6231269182162846e-05, "loss": 4.8165, "step": 8351 }, { "epoch": 0.07539267015706806, "grad_norm": 3.220548629760742, "learning_rate": 4.6230817837154725e-05, "loss": 4.692, "step": 8352 }, { "epoch": 0.07540169705723054, "grad_norm": 3.1883704662323, "learning_rate": 4.62303664921466e-05, "loss": 4.2675, "step": 8353 }, { "epoch": 0.07541072395739302, "grad_norm": 3.1827950477600098, "learning_rate": 4.6229915147138476e-05, "loss": 4.1714, "step": 8354 }, { "epoch": 0.07541975085755552, "grad_norm": 3.766854763031006, "learning_rate": 4.622946380213035e-05, "loss": 4.9266, "step": 8355 }, { "epoch": 0.075428777757718, "grad_norm": 3.568243980407715, "learning_rate": 4.622901245712223e-05, "loss": 4.487, "step": 8356 }, { "epoch": 0.07543780465788048, "grad_norm": 3.8779799938201904, "learning_rate": 4.622856111211411e-05, "loss": 5.0971, "step": 8357 }, { "epoch": 0.07544683155804296, "grad_norm": 5.31936502456665, "learning_rate": 4.622810976710597e-05, "loss": 4.6648, "step": 8358 }, { "epoch": 0.07545585845820545, "grad_norm": 4.63920259475708, "learning_rate": 4.622765842209785e-05, "loss": 3.6673, "step": 8359 }, { "epoch": 0.07546488535836794, "grad_norm": 3.614971876144409, "learning_rate": 4.622720707708973e-05, "loss": 4.553, "step": 8360 }, { "epoch": 0.07547391225853042, "grad_norm": 4.487197399139404, "learning_rate": 4.6226755732081604e-05, "loss": 5.0781, "step": 8361 }, { "epoch": 0.0754829391586929, "grad_norm": 4.009776592254639, "learning_rate": 4.622630438707348e-05, "loss": 4.1233, "step": 8362 }, { "epoch": 0.07549196605885539, "grad_norm": 3.2990915775299072, "learning_rate": 4.6225853042065356e-05, "loss": 4.4977, "step": 8363 }, { "epoch": 0.07550099295901787, "grad_norm": 3.3267359733581543, "learning_rate": 4.622540169705723e-05, "loss": 4.1426, "step": 8364 }, { "epoch": 0.07551001985918036, "grad_norm": 3.506291389465332, "learning_rate": 4.622495035204911e-05, "loss": 5.0251, "step": 8365 }, { "epoch": 0.07551904675934284, "grad_norm": 4.283559799194336, "learning_rate": 4.622449900704099e-05, "loss": 4.0402, "step": 8366 }, { "epoch": 0.07552807365950533, "grad_norm": 3.9356725215911865, "learning_rate": 4.622404766203286e-05, "loss": 3.8772, "step": 8367 }, { "epoch": 0.07553710055966781, "grad_norm": 2.969181776046753, "learning_rate": 4.622359631702474e-05, "loss": 4.8329, "step": 8368 }, { "epoch": 0.07554612745983029, "grad_norm": 3.7814974784851074, "learning_rate": 4.622314497201661e-05, "loss": 4.6402, "step": 8369 }, { "epoch": 0.07555515435999278, "grad_norm": 3.522308826446533, "learning_rate": 4.622269362700849e-05, "loss": 4.3041, "step": 8370 }, { "epoch": 0.07556418126015527, "grad_norm": 2.7445292472839355, "learning_rate": 4.622224228200036e-05, "loss": 4.3405, "step": 8371 }, { "epoch": 0.07557320816031775, "grad_norm": 3.000324010848999, "learning_rate": 4.6221790936992235e-05, "loss": 3.9752, "step": 8372 }, { "epoch": 0.07558223506048023, "grad_norm": 3.396246910095215, "learning_rate": 4.6221339591984114e-05, "loss": 4.2775, "step": 8373 }, { "epoch": 0.07559126196064271, "grad_norm": 3.4063453674316406, "learning_rate": 4.6220888246975994e-05, "loss": 4.8208, "step": 8374 }, { "epoch": 0.07560028886080519, "grad_norm": 4.295266151428223, "learning_rate": 4.6220436901967866e-05, "loss": 4.7497, "step": 8375 }, { "epoch": 0.07560931576096769, "grad_norm": 2.700685501098633, "learning_rate": 4.6219985556959745e-05, "loss": 5.0369, "step": 8376 }, { "epoch": 0.07561834266113017, "grad_norm": 3.670332670211792, "learning_rate": 4.621953421195162e-05, "loss": 4.5282, "step": 8377 }, { "epoch": 0.07562736956129265, "grad_norm": 3.8997485637664795, "learning_rate": 4.621908286694349e-05, "loss": 4.2464, "step": 8378 }, { "epoch": 0.07563639646145513, "grad_norm": 4.1525044441223145, "learning_rate": 4.621863152193537e-05, "loss": 4.8325, "step": 8379 }, { "epoch": 0.07564542336161761, "grad_norm": 3.7143547534942627, "learning_rate": 4.621818017692724e-05, "loss": 4.4223, "step": 8380 }, { "epoch": 0.07565445026178011, "grad_norm": 4.235045433044434, "learning_rate": 4.621772883191912e-05, "loss": 4.0664, "step": 8381 }, { "epoch": 0.07566347716194259, "grad_norm": 3.0414047241210938, "learning_rate": 4.6217277486911e-05, "loss": 4.2153, "step": 8382 }, { "epoch": 0.07567250406210507, "grad_norm": 2.633939266204834, "learning_rate": 4.621682614190287e-05, "loss": 4.7586, "step": 8383 }, { "epoch": 0.07568153096226755, "grad_norm": 3.4342501163482666, "learning_rate": 4.6216374796894745e-05, "loss": 4.3927, "step": 8384 }, { "epoch": 0.07569055786243004, "grad_norm": 2.8498146533966064, "learning_rate": 4.6215923451886625e-05, "loss": 4.5006, "step": 8385 }, { "epoch": 0.07569958476259253, "grad_norm": 3.3327744007110596, "learning_rate": 4.62154721068785e-05, "loss": 5.5567, "step": 8386 }, { "epoch": 0.07570861166275501, "grad_norm": 5.4035325050354, "learning_rate": 4.6215020761870376e-05, "loss": 5.1748, "step": 8387 }, { "epoch": 0.0757176385629175, "grad_norm": 3.2753522396087646, "learning_rate": 4.6214569416862255e-05, "loss": 4.8664, "step": 8388 }, { "epoch": 0.07572666546307998, "grad_norm": 2.699531316757202, "learning_rate": 4.621411807185413e-05, "loss": 4.8229, "step": 8389 }, { "epoch": 0.07573569236324246, "grad_norm": 3.172381639480591, "learning_rate": 4.621366672684601e-05, "loss": 5.2257, "step": 8390 }, { "epoch": 0.07574471926340495, "grad_norm": 3.543714761734009, "learning_rate": 4.621321538183788e-05, "loss": 4.0124, "step": 8391 }, { "epoch": 0.07575374616356743, "grad_norm": 5.525576591491699, "learning_rate": 4.621276403682975e-05, "loss": 4.4773, "step": 8392 }, { "epoch": 0.07576277306372992, "grad_norm": 3.054703950881958, "learning_rate": 4.621231269182163e-05, "loss": 4.3288, "step": 8393 }, { "epoch": 0.0757717999638924, "grad_norm": 3.7674074172973633, "learning_rate": 4.6211861346813504e-05, "loss": 4.9115, "step": 8394 }, { "epoch": 0.07578082686405488, "grad_norm": 4.038536548614502, "learning_rate": 4.621141000180538e-05, "loss": 5.0166, "step": 8395 }, { "epoch": 0.07578985376421737, "grad_norm": 3.3729934692382812, "learning_rate": 4.621095865679726e-05, "loss": 4.7288, "step": 8396 }, { "epoch": 0.07579888066437986, "grad_norm": 4.940866470336914, "learning_rate": 4.6210507311789135e-05, "loss": 4.5914, "step": 8397 }, { "epoch": 0.07580790756454234, "grad_norm": 3.7481939792633057, "learning_rate": 4.621005596678101e-05, "loss": 4.5212, "step": 8398 }, { "epoch": 0.07581693446470482, "grad_norm": 3.3823323249816895, "learning_rate": 4.6209604621772886e-05, "loss": 4.7257, "step": 8399 }, { "epoch": 0.0758259613648673, "grad_norm": 3.4137122631073, "learning_rate": 4.620915327676476e-05, "loss": 4.4884, "step": 8400 }, { "epoch": 0.07583498826502978, "grad_norm": 3.4978396892547607, "learning_rate": 4.620870193175664e-05, "loss": 4.1509, "step": 8401 }, { "epoch": 0.07584401516519228, "grad_norm": 5.147857189178467, "learning_rate": 4.620825058674851e-05, "loss": 3.9881, "step": 8402 }, { "epoch": 0.07585304206535476, "grad_norm": 4.295525074005127, "learning_rate": 4.620779924174039e-05, "loss": 4.1648, "step": 8403 }, { "epoch": 0.07586206896551724, "grad_norm": 3.2672576904296875, "learning_rate": 4.620734789673227e-05, "loss": 4.1412, "step": 8404 }, { "epoch": 0.07587109586567972, "grad_norm": 3.336008310317993, "learning_rate": 4.6206896551724135e-05, "loss": 4.5287, "step": 8405 }, { "epoch": 0.0758801227658422, "grad_norm": 4.296131134033203, "learning_rate": 4.6206445206716014e-05, "loss": 3.5798, "step": 8406 }, { "epoch": 0.0758891496660047, "grad_norm": 3.2978765964508057, "learning_rate": 4.620599386170789e-05, "loss": 4.5475, "step": 8407 }, { "epoch": 0.07589817656616718, "grad_norm": 3.3472790718078613, "learning_rate": 4.6205542516699766e-05, "loss": 5.0649, "step": 8408 }, { "epoch": 0.07590720346632966, "grad_norm": 2.9044687747955322, "learning_rate": 4.6205091171691645e-05, "loss": 4.5337, "step": 8409 }, { "epoch": 0.07591623036649214, "grad_norm": 4.0697174072265625, "learning_rate": 4.6204639826683524e-05, "loss": 4.3803, "step": 8410 }, { "epoch": 0.07592525726665462, "grad_norm": 4.362511157989502, "learning_rate": 4.620418848167539e-05, "loss": 5.328, "step": 8411 }, { "epoch": 0.07593428416681712, "grad_norm": 3.278099298477173, "learning_rate": 4.620373713666727e-05, "loss": 4.8501, "step": 8412 }, { "epoch": 0.0759433110669796, "grad_norm": 2.903261184692383, "learning_rate": 4.620328579165915e-05, "loss": 5.3479, "step": 8413 }, { "epoch": 0.07595233796714208, "grad_norm": 3.7327935695648193, "learning_rate": 4.620283444665102e-05, "loss": 5.0818, "step": 8414 }, { "epoch": 0.07596136486730456, "grad_norm": 3.390939950942993, "learning_rate": 4.62023831016429e-05, "loss": 5.4779, "step": 8415 }, { "epoch": 0.07597039176746705, "grad_norm": 3.007843494415283, "learning_rate": 4.620193175663477e-05, "loss": 4.7185, "step": 8416 }, { "epoch": 0.07597941866762954, "grad_norm": 4.246171951293945, "learning_rate": 4.620148041162665e-05, "loss": 4.5624, "step": 8417 }, { "epoch": 0.07598844556779202, "grad_norm": 4.14610481262207, "learning_rate": 4.6201029066618524e-05, "loss": 4.8128, "step": 8418 }, { "epoch": 0.0759974724679545, "grad_norm": 3.5928056240081787, "learning_rate": 4.62005777216104e-05, "loss": 4.5707, "step": 8419 }, { "epoch": 0.07600649936811699, "grad_norm": 3.726180076599121, "learning_rate": 4.6200126376602276e-05, "loss": 4.7793, "step": 8420 }, { "epoch": 0.07601552626827947, "grad_norm": 3.429870367050171, "learning_rate": 4.6199675031594155e-05, "loss": 4.8977, "step": 8421 }, { "epoch": 0.07602455316844196, "grad_norm": 3.4060544967651367, "learning_rate": 4.619922368658603e-05, "loss": 4.9494, "step": 8422 }, { "epoch": 0.07603358006860444, "grad_norm": 4.383519172668457, "learning_rate": 4.619877234157791e-05, "loss": 3.4463, "step": 8423 }, { "epoch": 0.07604260696876693, "grad_norm": 3.378530263900757, "learning_rate": 4.619832099656978e-05, "loss": 4.5128, "step": 8424 }, { "epoch": 0.07605163386892941, "grad_norm": 3.7017431259155273, "learning_rate": 4.619786965156165e-05, "loss": 4.1263, "step": 8425 }, { "epoch": 0.07606066076909189, "grad_norm": 3.654564380645752, "learning_rate": 4.619741830655353e-05, "loss": 4.7464, "step": 8426 }, { "epoch": 0.07606968766925438, "grad_norm": 3.5689942836761475, "learning_rate": 4.619696696154541e-05, "loss": 4.7119, "step": 8427 }, { "epoch": 0.07607871456941687, "grad_norm": 4.044758319854736, "learning_rate": 4.619651561653728e-05, "loss": 4.5379, "step": 8428 }, { "epoch": 0.07608774146957935, "grad_norm": 3.9858925342559814, "learning_rate": 4.619606427152916e-05, "loss": 4.3183, "step": 8429 }, { "epoch": 0.07609676836974183, "grad_norm": 4.994046211242676, "learning_rate": 4.6195612926521035e-05, "loss": 4.4805, "step": 8430 }, { "epoch": 0.07610579526990431, "grad_norm": 3.938060760498047, "learning_rate": 4.619516158151291e-05, "loss": 4.3866, "step": 8431 }, { "epoch": 0.07611482217006679, "grad_norm": 3.621530055999756, "learning_rate": 4.6194710236504786e-05, "loss": 3.9523, "step": 8432 }, { "epoch": 0.07612384907022929, "grad_norm": 4.180856227874756, "learning_rate": 4.619425889149666e-05, "loss": 4.613, "step": 8433 }, { "epoch": 0.07613287597039177, "grad_norm": 3.8683393001556396, "learning_rate": 4.619380754648854e-05, "loss": 4.8553, "step": 8434 }, { "epoch": 0.07614190287055425, "grad_norm": 3.3970348834991455, "learning_rate": 4.619335620148042e-05, "loss": 4.0327, "step": 8435 }, { "epoch": 0.07615092977071673, "grad_norm": 3.294503688812256, "learning_rate": 4.619290485647229e-05, "loss": 4.4093, "step": 8436 }, { "epoch": 0.07615995667087921, "grad_norm": 3.073155403137207, "learning_rate": 4.619245351146417e-05, "loss": 4.7445, "step": 8437 }, { "epoch": 0.07616898357104171, "grad_norm": 2.788653612136841, "learning_rate": 4.619200216645604e-05, "loss": 4.6349, "step": 8438 }, { "epoch": 0.07617801047120419, "grad_norm": 4.6065521240234375, "learning_rate": 4.6191550821447914e-05, "loss": 4.3284, "step": 8439 }, { "epoch": 0.07618703737136667, "grad_norm": 4.49607515335083, "learning_rate": 4.619109947643979e-05, "loss": 4.967, "step": 8440 }, { "epoch": 0.07619606427152915, "grad_norm": 3.390793800354004, "learning_rate": 4.6190648131431666e-05, "loss": 4.4679, "step": 8441 }, { "epoch": 0.07620509117169164, "grad_norm": 2.8728785514831543, "learning_rate": 4.6190196786423545e-05, "loss": 5.1916, "step": 8442 }, { "epoch": 0.07621411807185413, "grad_norm": 3.2649362087249756, "learning_rate": 4.6189745441415424e-05, "loss": 4.914, "step": 8443 }, { "epoch": 0.07622314497201661, "grad_norm": 3.237096071243286, "learning_rate": 4.6189294096407296e-05, "loss": 4.261, "step": 8444 }, { "epoch": 0.0762321718721791, "grad_norm": 3.4507579803466797, "learning_rate": 4.618884275139917e-05, "loss": 4.5637, "step": 8445 }, { "epoch": 0.07624119877234158, "grad_norm": 3.217759847640991, "learning_rate": 4.618839140639105e-05, "loss": 5.6224, "step": 8446 }, { "epoch": 0.07625022567250406, "grad_norm": 3.249061107635498, "learning_rate": 4.618794006138292e-05, "loss": 4.846, "step": 8447 }, { "epoch": 0.07625925257266655, "grad_norm": 3.1574792861938477, "learning_rate": 4.61874887163748e-05, "loss": 4.5894, "step": 8448 }, { "epoch": 0.07626827947282903, "grad_norm": 3.8209588527679443, "learning_rate": 4.618703737136668e-05, "loss": 5.3567, "step": 8449 }, { "epoch": 0.07627730637299152, "grad_norm": 3.02091121673584, "learning_rate": 4.618658602635855e-05, "loss": 4.9679, "step": 8450 }, { "epoch": 0.076286333273154, "grad_norm": 4.738175868988037, "learning_rate": 4.618613468135043e-05, "loss": 4.8362, "step": 8451 }, { "epoch": 0.07629536017331648, "grad_norm": 4.950796127319336, "learning_rate": 4.61856833363423e-05, "loss": 5.0875, "step": 8452 }, { "epoch": 0.07630438707347897, "grad_norm": 3.2962398529052734, "learning_rate": 4.6185231991334176e-05, "loss": 4.5672, "step": 8453 }, { "epoch": 0.07631341397364146, "grad_norm": 3.649951696395874, "learning_rate": 4.6184780646326055e-05, "loss": 5.2174, "step": 8454 }, { "epoch": 0.07632244087380394, "grad_norm": 3.5408453941345215, "learning_rate": 4.618432930131793e-05, "loss": 4.9967, "step": 8455 }, { "epoch": 0.07633146777396642, "grad_norm": 3.7936506271362305, "learning_rate": 4.618387795630981e-05, "loss": 4.9596, "step": 8456 }, { "epoch": 0.0763404946741289, "grad_norm": 2.833495855331421, "learning_rate": 4.6183426611301686e-05, "loss": 4.6214, "step": 8457 }, { "epoch": 0.07634952157429138, "grad_norm": 3.778940439224243, "learning_rate": 4.618297526629355e-05, "loss": 4.0421, "step": 8458 }, { "epoch": 0.07635854847445388, "grad_norm": 2.8003432750701904, "learning_rate": 4.618252392128543e-05, "loss": 4.9908, "step": 8459 }, { "epoch": 0.07636757537461636, "grad_norm": 3.8358724117279053, "learning_rate": 4.618207257627731e-05, "loss": 4.3924, "step": 8460 }, { "epoch": 0.07637660227477884, "grad_norm": 2.39180588722229, "learning_rate": 4.618162123126918e-05, "loss": 4.7074, "step": 8461 }, { "epoch": 0.07638562917494132, "grad_norm": 5.4019012451171875, "learning_rate": 4.618116988626106e-05, "loss": 3.2154, "step": 8462 }, { "epoch": 0.0763946560751038, "grad_norm": 2.830706834793091, "learning_rate": 4.6180718541252934e-05, "loss": 4.8465, "step": 8463 }, { "epoch": 0.0764036829752663, "grad_norm": 4.538445472717285, "learning_rate": 4.6180267196244814e-05, "loss": 4.8574, "step": 8464 }, { "epoch": 0.07641270987542878, "grad_norm": 2.587660551071167, "learning_rate": 4.6179815851236686e-05, "loss": 4.4949, "step": 8465 }, { "epoch": 0.07642173677559126, "grad_norm": 3.5978198051452637, "learning_rate": 4.617936450622856e-05, "loss": 4.0838, "step": 8466 }, { "epoch": 0.07643076367575374, "grad_norm": 3.9003570079803467, "learning_rate": 4.617891316122044e-05, "loss": 4.4803, "step": 8467 }, { "epoch": 0.07643979057591622, "grad_norm": 3.5868430137634277, "learning_rate": 4.617846181621232e-05, "loss": 4.5277, "step": 8468 }, { "epoch": 0.07644881747607872, "grad_norm": 2.6293721199035645, "learning_rate": 4.617801047120419e-05, "loss": 4.8296, "step": 8469 }, { "epoch": 0.0764578443762412, "grad_norm": 3.7425990104675293, "learning_rate": 4.617755912619607e-05, "loss": 4.9077, "step": 8470 }, { "epoch": 0.07646687127640368, "grad_norm": 2.8764400482177734, "learning_rate": 4.617710778118795e-05, "loss": 5.1234, "step": 8471 }, { "epoch": 0.07647589817656616, "grad_norm": 2.664971351623535, "learning_rate": 4.6176656436179814e-05, "loss": 4.2671, "step": 8472 }, { "epoch": 0.07648492507672865, "grad_norm": 3.933666706085205, "learning_rate": 4.617620509117169e-05, "loss": 4.4605, "step": 8473 }, { "epoch": 0.07649395197689114, "grad_norm": 2.439633846282959, "learning_rate": 4.617575374616357e-05, "loss": 4.7135, "step": 8474 }, { "epoch": 0.07650297887705362, "grad_norm": 2.9199607372283936, "learning_rate": 4.6175302401155445e-05, "loss": 4.9554, "step": 8475 }, { "epoch": 0.0765120057772161, "grad_norm": 3.7692596912384033, "learning_rate": 4.6174851056147324e-05, "loss": 4.4356, "step": 8476 }, { "epoch": 0.07652103267737859, "grad_norm": 3.0104787349700928, "learning_rate": 4.6174399711139196e-05, "loss": 4.4807, "step": 8477 }, { "epoch": 0.07653005957754107, "grad_norm": 3.4412612915039062, "learning_rate": 4.617394836613107e-05, "loss": 4.5318, "step": 8478 }, { "epoch": 0.07653908647770356, "grad_norm": 4.101443767547607, "learning_rate": 4.617349702112295e-05, "loss": 4.9159, "step": 8479 }, { "epoch": 0.07654811337786604, "grad_norm": 2.6140425205230713, "learning_rate": 4.617304567611482e-05, "loss": 4.8263, "step": 8480 }, { "epoch": 0.07655714027802853, "grad_norm": 3.0410144329071045, "learning_rate": 4.61725943311067e-05, "loss": 4.819, "step": 8481 }, { "epoch": 0.07656616717819101, "grad_norm": 3.6752281188964844, "learning_rate": 4.617214298609858e-05, "loss": 4.9131, "step": 8482 }, { "epoch": 0.07657519407835349, "grad_norm": 4.19439172744751, "learning_rate": 4.617169164109045e-05, "loss": 4.4901, "step": 8483 }, { "epoch": 0.07658422097851597, "grad_norm": 3.455462694168091, "learning_rate": 4.617124029608233e-05, "loss": 3.9677, "step": 8484 }, { "epoch": 0.07659324787867847, "grad_norm": 4.632823944091797, "learning_rate": 4.61707889510742e-05, "loss": 4.6183, "step": 8485 }, { "epoch": 0.07660227477884095, "grad_norm": 4.91249942779541, "learning_rate": 4.6170337606066076e-05, "loss": 4.469, "step": 8486 }, { "epoch": 0.07661130167900343, "grad_norm": 3.8393914699554443, "learning_rate": 4.6169886261057955e-05, "loss": 4.5489, "step": 8487 }, { "epoch": 0.07662032857916591, "grad_norm": 2.999743938446045, "learning_rate": 4.6169434916049834e-05, "loss": 4.6001, "step": 8488 }, { "epoch": 0.07662935547932839, "grad_norm": 3.80873441696167, "learning_rate": 4.6168983571041706e-05, "loss": 4.8339, "step": 8489 }, { "epoch": 0.07663838237949089, "grad_norm": 4.7557291984558105, "learning_rate": 4.6168532226033586e-05, "loss": 3.8267, "step": 8490 }, { "epoch": 0.07664740927965337, "grad_norm": 2.550855875015259, "learning_rate": 4.616808088102546e-05, "loss": 4.813, "step": 8491 }, { "epoch": 0.07665643617981585, "grad_norm": 4.30673360824585, "learning_rate": 4.616762953601733e-05, "loss": 4.7972, "step": 8492 }, { "epoch": 0.07666546307997833, "grad_norm": 3.484344005584717, "learning_rate": 4.616717819100921e-05, "loss": 4.7981, "step": 8493 }, { "epoch": 0.07667448998014081, "grad_norm": 5.153425693511963, "learning_rate": 4.616672684600108e-05, "loss": 4.8734, "step": 8494 }, { "epoch": 0.07668351688030331, "grad_norm": 3.0697543621063232, "learning_rate": 4.616627550099296e-05, "loss": 5.1172, "step": 8495 }, { "epoch": 0.07669254378046579, "grad_norm": 4.067912578582764, "learning_rate": 4.616582415598484e-05, "loss": 4.4974, "step": 8496 }, { "epoch": 0.07670157068062827, "grad_norm": 4.170863628387451, "learning_rate": 4.616537281097671e-05, "loss": 4.3942, "step": 8497 }, { "epoch": 0.07671059758079075, "grad_norm": 4.3896636962890625, "learning_rate": 4.616492146596859e-05, "loss": 5.1191, "step": 8498 }, { "epoch": 0.07671962448095324, "grad_norm": 3.2240374088287354, "learning_rate": 4.6164470120960465e-05, "loss": 4.938, "step": 8499 }, { "epoch": 0.07672865138111573, "grad_norm": 3.6712646484375, "learning_rate": 4.616401877595234e-05, "loss": 5.1801, "step": 8500 }, { "epoch": 0.07673767828127821, "grad_norm": 3.6283607482910156, "learning_rate": 4.616356743094422e-05, "loss": 4.7339, "step": 8501 }, { "epoch": 0.0767467051814407, "grad_norm": 3.052584409713745, "learning_rate": 4.616311608593609e-05, "loss": 4.5894, "step": 8502 }, { "epoch": 0.07675573208160318, "grad_norm": 2.6217148303985596, "learning_rate": 4.616266474092797e-05, "loss": 4.7561, "step": 8503 }, { "epoch": 0.07676475898176566, "grad_norm": 3.492318868637085, "learning_rate": 4.616221339591985e-05, "loss": 4.4031, "step": 8504 }, { "epoch": 0.07677378588192815, "grad_norm": 3.684016466140747, "learning_rate": 4.616176205091171e-05, "loss": 4.5118, "step": 8505 }, { "epoch": 0.07678281278209063, "grad_norm": 3.3913278579711914, "learning_rate": 4.616131070590359e-05, "loss": 4.5077, "step": 8506 }, { "epoch": 0.07679183968225312, "grad_norm": 4.118086814880371, "learning_rate": 4.616085936089547e-05, "loss": 3.3852, "step": 8507 }, { "epoch": 0.0768008665824156, "grad_norm": 3.919027090072632, "learning_rate": 4.6160408015887344e-05, "loss": 3.9697, "step": 8508 }, { "epoch": 0.07680989348257808, "grad_norm": 3.5855860710144043, "learning_rate": 4.6159956670879224e-05, "loss": 4.9808, "step": 8509 }, { "epoch": 0.07681892038274056, "grad_norm": 2.4290590286254883, "learning_rate": 4.61595053258711e-05, "loss": 4.7197, "step": 8510 }, { "epoch": 0.07682794728290306, "grad_norm": 4.034630298614502, "learning_rate": 4.6159053980862975e-05, "loss": 4.3336, "step": 8511 }, { "epoch": 0.07683697418306554, "grad_norm": 3.3946807384490967, "learning_rate": 4.615860263585485e-05, "loss": 4.4522, "step": 8512 }, { "epoch": 0.07684600108322802, "grad_norm": 2.661675214767456, "learning_rate": 4.615815129084673e-05, "loss": 4.4264, "step": 8513 }, { "epoch": 0.0768550279833905, "grad_norm": 4.041332721710205, "learning_rate": 4.61576999458386e-05, "loss": 4.5845, "step": 8514 }, { "epoch": 0.07686405488355298, "grad_norm": 3.4140965938568115, "learning_rate": 4.615724860083048e-05, "loss": 4.7998, "step": 8515 }, { "epoch": 0.07687308178371548, "grad_norm": 3.552932024002075, "learning_rate": 4.615679725582235e-05, "loss": 4.1796, "step": 8516 }, { "epoch": 0.07688210868387796, "grad_norm": 4.4227705001831055, "learning_rate": 4.615634591081423e-05, "loss": 4.6721, "step": 8517 }, { "epoch": 0.07689113558404044, "grad_norm": 3.610650062561035, "learning_rate": 4.615589456580611e-05, "loss": 4.1613, "step": 8518 }, { "epoch": 0.07690016248420292, "grad_norm": 3.5953361988067627, "learning_rate": 4.6155443220797975e-05, "loss": 5.391, "step": 8519 }, { "epoch": 0.0769091893843654, "grad_norm": 3.286310911178589, "learning_rate": 4.6154991875789855e-05, "loss": 5.2292, "step": 8520 }, { "epoch": 0.0769182162845279, "grad_norm": 2.9360485076904297, "learning_rate": 4.6154540530781734e-05, "loss": 4.8755, "step": 8521 }, { "epoch": 0.07692724318469038, "grad_norm": 3.1168649196624756, "learning_rate": 4.6154089185773606e-05, "loss": 4.5046, "step": 8522 }, { "epoch": 0.07693627008485286, "grad_norm": 3.1514713764190674, "learning_rate": 4.6153637840765485e-05, "loss": 4.6153, "step": 8523 }, { "epoch": 0.07694529698501534, "grad_norm": 4.423250198364258, "learning_rate": 4.615318649575736e-05, "loss": 3.7939, "step": 8524 }, { "epoch": 0.07695432388517782, "grad_norm": 3.372762441635132, "learning_rate": 4.615273515074923e-05, "loss": 4.8096, "step": 8525 }, { "epoch": 0.07696335078534032, "grad_norm": 3.7619454860687256, "learning_rate": 4.615228380574111e-05, "loss": 4.7333, "step": 8526 }, { "epoch": 0.0769723776855028, "grad_norm": 3.5184900760650635, "learning_rate": 4.615183246073299e-05, "loss": 4.5741, "step": 8527 }, { "epoch": 0.07698140458566528, "grad_norm": 3.7536962032318115, "learning_rate": 4.615138111572486e-05, "loss": 4.4134, "step": 8528 }, { "epoch": 0.07699043148582776, "grad_norm": 3.648390531539917, "learning_rate": 4.615092977071674e-05, "loss": 4.2987, "step": 8529 }, { "epoch": 0.07699945838599025, "grad_norm": 3.0644595623016357, "learning_rate": 4.615047842570861e-05, "loss": 4.6221, "step": 8530 }, { "epoch": 0.07700848528615274, "grad_norm": 3.984295606613159, "learning_rate": 4.615002708070049e-05, "loss": 4.499, "step": 8531 }, { "epoch": 0.07701751218631522, "grad_norm": 4.109230041503906, "learning_rate": 4.6149575735692365e-05, "loss": 4.6036, "step": 8532 }, { "epoch": 0.0770265390864777, "grad_norm": 4.1538405418396, "learning_rate": 4.614912439068424e-05, "loss": 4.1277, "step": 8533 }, { "epoch": 0.07703556598664019, "grad_norm": 3.518209934234619, "learning_rate": 4.6148673045676116e-05, "loss": 4.9067, "step": 8534 }, { "epoch": 0.07704459288680267, "grad_norm": 3.23502516746521, "learning_rate": 4.6148221700667996e-05, "loss": 5.0154, "step": 8535 }, { "epoch": 0.07705361978696515, "grad_norm": 3.1085357666015625, "learning_rate": 4.614777035565987e-05, "loss": 3.8027, "step": 8536 }, { "epoch": 0.07706264668712764, "grad_norm": 2.8137621879577637, "learning_rate": 4.614731901065175e-05, "loss": 4.2804, "step": 8537 }, { "epoch": 0.07707167358729013, "grad_norm": 3.021599292755127, "learning_rate": 4.614686766564362e-05, "loss": 5.0093, "step": 8538 }, { "epoch": 0.07708070048745261, "grad_norm": 3.7684273719787598, "learning_rate": 4.614641632063549e-05, "loss": 4.7534, "step": 8539 }, { "epoch": 0.07708972738761509, "grad_norm": 7.830081939697266, "learning_rate": 4.614596497562737e-05, "loss": 4.6553, "step": 8540 }, { "epoch": 0.07709875428777757, "grad_norm": 2.751997947692871, "learning_rate": 4.6145513630619244e-05, "loss": 4.7509, "step": 8541 }, { "epoch": 0.07710778118794007, "grad_norm": 3.8753316402435303, "learning_rate": 4.614506228561112e-05, "loss": 4.9241, "step": 8542 }, { "epoch": 0.07711680808810255, "grad_norm": 3.431084394454956, "learning_rate": 4.6144610940603e-05, "loss": 4.469, "step": 8543 }, { "epoch": 0.07712583498826503, "grad_norm": 7.926413536071777, "learning_rate": 4.6144159595594875e-05, "loss": 5.014, "step": 8544 }, { "epoch": 0.07713486188842751, "grad_norm": 3.0451598167419434, "learning_rate": 4.614370825058675e-05, "loss": 5.0097, "step": 8545 }, { "epoch": 0.07714388878858999, "grad_norm": 2.8743069171905518, "learning_rate": 4.614325690557863e-05, "loss": 4.6237, "step": 8546 }, { "epoch": 0.07715291568875249, "grad_norm": 3.272475004196167, "learning_rate": 4.61428055605705e-05, "loss": 4.9938, "step": 8547 }, { "epoch": 0.07716194258891497, "grad_norm": 3.4785525798797607, "learning_rate": 4.614235421556238e-05, "loss": 4.7969, "step": 8548 }, { "epoch": 0.07717096948907745, "grad_norm": 3.181323289871216, "learning_rate": 4.614190287055426e-05, "loss": 4.5455, "step": 8549 }, { "epoch": 0.07717999638923993, "grad_norm": 3.145831346511841, "learning_rate": 4.614145152554613e-05, "loss": 4.6294, "step": 8550 }, { "epoch": 0.07718902328940241, "grad_norm": 4.141261100769043, "learning_rate": 4.614100018053801e-05, "loss": 4.6689, "step": 8551 }, { "epoch": 0.07719805018956491, "grad_norm": 6.043136119842529, "learning_rate": 4.614054883552988e-05, "loss": 4.9544, "step": 8552 }, { "epoch": 0.07720707708972739, "grad_norm": 2.803286552429199, "learning_rate": 4.6140097490521754e-05, "loss": 4.4952, "step": 8553 }, { "epoch": 0.07721610398988987, "grad_norm": 3.401031017303467, "learning_rate": 4.6139646145513634e-05, "loss": 4.769, "step": 8554 }, { "epoch": 0.07722513089005235, "grad_norm": 3.785757064819336, "learning_rate": 4.6139194800505506e-05, "loss": 4.803, "step": 8555 }, { "epoch": 0.07723415779021484, "grad_norm": 2.358795642852783, "learning_rate": 4.6138743455497385e-05, "loss": 5.2614, "step": 8556 }, { "epoch": 0.07724318469037733, "grad_norm": 3.897059202194214, "learning_rate": 4.6138292110489264e-05, "loss": 4.3061, "step": 8557 }, { "epoch": 0.07725221159053981, "grad_norm": 2.7452192306518555, "learning_rate": 4.613784076548114e-05, "loss": 5.4344, "step": 8558 }, { "epoch": 0.0772612384907023, "grad_norm": 5.119626045227051, "learning_rate": 4.613738942047301e-05, "loss": 3.9385, "step": 8559 }, { "epoch": 0.07727026539086478, "grad_norm": 3.761082887649536, "learning_rate": 4.613693807546489e-05, "loss": 4.5353, "step": 8560 }, { "epoch": 0.07727929229102726, "grad_norm": 4.040842056274414, "learning_rate": 4.613648673045676e-05, "loss": 5.0227, "step": 8561 }, { "epoch": 0.07728831919118974, "grad_norm": 2.7409274578094482, "learning_rate": 4.613603538544864e-05, "loss": 4.7501, "step": 8562 }, { "epoch": 0.07729734609135223, "grad_norm": 3.627225637435913, "learning_rate": 4.613558404044051e-05, "loss": 4.7701, "step": 8563 }, { "epoch": 0.07730637299151472, "grad_norm": 4.91810941696167, "learning_rate": 4.613513269543239e-05, "loss": 4.3601, "step": 8564 }, { "epoch": 0.0773153998916772, "grad_norm": 3.581272602081299, "learning_rate": 4.613468135042427e-05, "loss": 4.4923, "step": 8565 }, { "epoch": 0.07732442679183968, "grad_norm": 3.3212904930114746, "learning_rate": 4.613423000541614e-05, "loss": 4.6184, "step": 8566 }, { "epoch": 0.07733345369200216, "grad_norm": 3.6901371479034424, "learning_rate": 4.6133778660408016e-05, "loss": 4.5549, "step": 8567 }, { "epoch": 0.07734248059216466, "grad_norm": 3.966675281524658, "learning_rate": 4.6133327315399895e-05, "loss": 4.7629, "step": 8568 }, { "epoch": 0.07735150749232714, "grad_norm": 3.4227917194366455, "learning_rate": 4.613287597039177e-05, "loss": 4.6058, "step": 8569 }, { "epoch": 0.07736053439248962, "grad_norm": 2.7664384841918945, "learning_rate": 4.613242462538365e-05, "loss": 5.07, "step": 8570 }, { "epoch": 0.0773695612926521, "grad_norm": 4.022496700286865, "learning_rate": 4.6131973280375526e-05, "loss": 4.7963, "step": 8571 }, { "epoch": 0.07737858819281458, "grad_norm": 3.4660933017730713, "learning_rate": 4.613152193536739e-05, "loss": 4.5026, "step": 8572 }, { "epoch": 0.07738761509297708, "grad_norm": 3.7396411895751953, "learning_rate": 4.613107059035927e-05, "loss": 3.9605, "step": 8573 }, { "epoch": 0.07739664199313956, "grad_norm": 4.191160678863525, "learning_rate": 4.613061924535115e-05, "loss": 4.8558, "step": 8574 }, { "epoch": 0.07740566889330204, "grad_norm": 3.672625780105591, "learning_rate": 4.613016790034302e-05, "loss": 5.0818, "step": 8575 }, { "epoch": 0.07741469579346452, "grad_norm": 2.609966278076172, "learning_rate": 4.61297165553349e-05, "loss": 4.9968, "step": 8576 }, { "epoch": 0.077423722693627, "grad_norm": 4.221729278564453, "learning_rate": 4.6129265210326775e-05, "loss": 4.6604, "step": 8577 }, { "epoch": 0.0774327495937895, "grad_norm": 3.343106746673584, "learning_rate": 4.6128813865318654e-05, "loss": 4.7136, "step": 8578 }, { "epoch": 0.07744177649395198, "grad_norm": 3.097637414932251, "learning_rate": 4.6128362520310526e-05, "loss": 3.969, "step": 8579 }, { "epoch": 0.07745080339411446, "grad_norm": 4.084310531616211, "learning_rate": 4.61279111753024e-05, "loss": 4.3194, "step": 8580 }, { "epoch": 0.07745983029427694, "grad_norm": 2.920397996902466, "learning_rate": 4.612745983029428e-05, "loss": 4.7779, "step": 8581 }, { "epoch": 0.07746885719443942, "grad_norm": 3.467294454574585, "learning_rate": 4.612700848528616e-05, "loss": 4.9272, "step": 8582 }, { "epoch": 0.07747788409460192, "grad_norm": 3.554462194442749, "learning_rate": 4.612655714027803e-05, "loss": 4.6721, "step": 8583 }, { "epoch": 0.0774869109947644, "grad_norm": 3.777876615524292, "learning_rate": 4.612610579526991e-05, "loss": 3.4592, "step": 8584 }, { "epoch": 0.07749593789492688, "grad_norm": 4.082188606262207, "learning_rate": 4.612565445026178e-05, "loss": 4.2418, "step": 8585 }, { "epoch": 0.07750496479508936, "grad_norm": 6.171962261199951, "learning_rate": 4.6125203105253654e-05, "loss": 4.4762, "step": 8586 }, { "epoch": 0.07751399169525185, "grad_norm": 2.761932611465454, "learning_rate": 4.612475176024553e-05, "loss": 4.354, "step": 8587 }, { "epoch": 0.07752301859541433, "grad_norm": 2.9031825065612793, "learning_rate": 4.612430041523741e-05, "loss": 4.8432, "step": 8588 }, { "epoch": 0.07753204549557682, "grad_norm": 3.157318592071533, "learning_rate": 4.6123849070229285e-05, "loss": 3.8366, "step": 8589 }, { "epoch": 0.0775410723957393, "grad_norm": 3.038203716278076, "learning_rate": 4.6123397725221164e-05, "loss": 4.8752, "step": 8590 }, { "epoch": 0.07755009929590179, "grad_norm": 2.799213409423828, "learning_rate": 4.612294638021304e-05, "loss": 4.2987, "step": 8591 }, { "epoch": 0.07755912619606427, "grad_norm": 2.616433620452881, "learning_rate": 4.612249503520491e-05, "loss": 4.451, "step": 8592 }, { "epoch": 0.07756815309622675, "grad_norm": 2.57057523727417, "learning_rate": 4.612204369019679e-05, "loss": 4.3564, "step": 8593 }, { "epoch": 0.07757717999638924, "grad_norm": 3.4412412643432617, "learning_rate": 4.612159234518866e-05, "loss": 4.1409, "step": 8594 }, { "epoch": 0.07758620689655173, "grad_norm": 4.332107067108154, "learning_rate": 4.612114100018054e-05, "loss": 3.4849, "step": 8595 }, { "epoch": 0.07759523379671421, "grad_norm": 3.4660661220550537, "learning_rate": 4.612068965517242e-05, "loss": 4.2667, "step": 8596 }, { "epoch": 0.07760426069687669, "grad_norm": 3.152056932449341, "learning_rate": 4.612023831016429e-05, "loss": 4.2591, "step": 8597 }, { "epoch": 0.07761328759703917, "grad_norm": 3.6925415992736816, "learning_rate": 4.611978696515617e-05, "loss": 4.5159, "step": 8598 }, { "epoch": 0.07762231449720167, "grad_norm": 2.6547467708587646, "learning_rate": 4.6119335620148044e-05, "loss": 4.8302, "step": 8599 }, { "epoch": 0.07763134139736415, "grad_norm": 4.162605285644531, "learning_rate": 4.6118884275139916e-05, "loss": 4.41, "step": 8600 }, { "epoch": 0.07764036829752663, "grad_norm": 3.5452065467834473, "learning_rate": 4.6118432930131795e-05, "loss": 4.2979, "step": 8601 }, { "epoch": 0.07764939519768911, "grad_norm": 3.236290454864502, "learning_rate": 4.611798158512367e-05, "loss": 4.9939, "step": 8602 }, { "epoch": 0.07765842209785159, "grad_norm": 4.525350093841553, "learning_rate": 4.611753024011555e-05, "loss": 5.0986, "step": 8603 }, { "epoch": 0.07766744899801409, "grad_norm": 3.7653989791870117, "learning_rate": 4.6117078895107426e-05, "loss": 4.268, "step": 8604 }, { "epoch": 0.07767647589817657, "grad_norm": 3.988084554672241, "learning_rate": 4.61166275500993e-05, "loss": 4.892, "step": 8605 }, { "epoch": 0.07768550279833905, "grad_norm": 3.9726171493530273, "learning_rate": 4.611617620509117e-05, "loss": 4.762, "step": 8606 }, { "epoch": 0.07769452969850153, "grad_norm": 2.8339695930480957, "learning_rate": 4.611572486008305e-05, "loss": 4.9915, "step": 8607 }, { "epoch": 0.07770355659866401, "grad_norm": 3.6418652534484863, "learning_rate": 4.611527351507492e-05, "loss": 4.5359, "step": 8608 }, { "epoch": 0.07771258349882651, "grad_norm": 4.206101894378662, "learning_rate": 4.61148221700668e-05, "loss": 4.5614, "step": 8609 }, { "epoch": 0.07772161039898899, "grad_norm": 2.8680477142333984, "learning_rate": 4.611437082505868e-05, "loss": 4.7793, "step": 8610 }, { "epoch": 0.07773063729915147, "grad_norm": 5.405294895172119, "learning_rate": 4.6113919480050554e-05, "loss": 4.9296, "step": 8611 }, { "epoch": 0.07773966419931395, "grad_norm": 3.0570082664489746, "learning_rate": 4.611346813504243e-05, "loss": 4.2953, "step": 8612 }, { "epoch": 0.07774869109947644, "grad_norm": 3.7236382961273193, "learning_rate": 4.6113016790034305e-05, "loss": 4.1556, "step": 8613 }, { "epoch": 0.07775771799963892, "grad_norm": 3.7238121032714844, "learning_rate": 4.611256544502618e-05, "loss": 4.4928, "step": 8614 }, { "epoch": 0.07776674489980141, "grad_norm": 3.889888286590576, "learning_rate": 4.611211410001806e-05, "loss": 4.3477, "step": 8615 }, { "epoch": 0.0777757717999639, "grad_norm": 3.0137577056884766, "learning_rate": 4.611166275500993e-05, "loss": 4.991, "step": 8616 }, { "epoch": 0.07778479870012638, "grad_norm": 3.451049327850342, "learning_rate": 4.611121141000181e-05, "loss": 5.1164, "step": 8617 }, { "epoch": 0.07779382560028886, "grad_norm": 4.1976823806762695, "learning_rate": 4.611076006499369e-05, "loss": 4.5283, "step": 8618 }, { "epoch": 0.07780285250045134, "grad_norm": 3.6637959480285645, "learning_rate": 4.6110308719985554e-05, "loss": 4.7159, "step": 8619 }, { "epoch": 0.07781187940061383, "grad_norm": 5.433656692504883, "learning_rate": 4.610985737497743e-05, "loss": 3.633, "step": 8620 }, { "epoch": 0.07782090630077632, "grad_norm": 3.4736688137054443, "learning_rate": 4.610940602996931e-05, "loss": 4.9593, "step": 8621 }, { "epoch": 0.0778299332009388, "grad_norm": 4.309720039367676, "learning_rate": 4.6108954684961185e-05, "loss": 3.9077, "step": 8622 }, { "epoch": 0.07783896010110128, "grad_norm": 3.7339775562286377, "learning_rate": 4.6108503339953064e-05, "loss": 3.9025, "step": 8623 }, { "epoch": 0.07784798700126376, "grad_norm": 2.965514659881592, "learning_rate": 4.6108051994944936e-05, "loss": 5.0151, "step": 8624 }, { "epoch": 0.07785701390142626, "grad_norm": 2.6203267574310303, "learning_rate": 4.6107600649936816e-05, "loss": 4.9119, "step": 8625 }, { "epoch": 0.07786604080158874, "grad_norm": 4.188138008117676, "learning_rate": 4.610714930492869e-05, "loss": 4.8031, "step": 8626 }, { "epoch": 0.07787506770175122, "grad_norm": 3.3217074871063232, "learning_rate": 4.610669795992056e-05, "loss": 4.2379, "step": 8627 }, { "epoch": 0.0778840946019137, "grad_norm": 5.013567924499512, "learning_rate": 4.610624661491244e-05, "loss": 4.2147, "step": 8628 }, { "epoch": 0.07789312150207618, "grad_norm": 3.497678279876709, "learning_rate": 4.610579526990432e-05, "loss": 4.5887, "step": 8629 }, { "epoch": 0.07790214840223868, "grad_norm": 3.388526201248169, "learning_rate": 4.610534392489619e-05, "loss": 4.8514, "step": 8630 }, { "epoch": 0.07791117530240116, "grad_norm": 3.872326612472534, "learning_rate": 4.610489257988807e-05, "loss": 3.999, "step": 8631 }, { "epoch": 0.07792020220256364, "grad_norm": 3.2704391479492188, "learning_rate": 4.610444123487995e-05, "loss": 4.8061, "step": 8632 }, { "epoch": 0.07792922910272612, "grad_norm": 3.9285881519317627, "learning_rate": 4.6103989889871816e-05, "loss": 3.1616, "step": 8633 }, { "epoch": 0.0779382560028886, "grad_norm": 3.23795223236084, "learning_rate": 4.6103538544863695e-05, "loss": 4.3387, "step": 8634 }, { "epoch": 0.0779472829030511, "grad_norm": 3.43821120262146, "learning_rate": 4.6103087199855574e-05, "loss": 4.5155, "step": 8635 }, { "epoch": 0.07795630980321358, "grad_norm": 3.4058644771575928, "learning_rate": 4.610263585484745e-05, "loss": 4.9268, "step": 8636 }, { "epoch": 0.07796533670337606, "grad_norm": 4.172461032867432, "learning_rate": 4.6102184509839326e-05, "loss": 4.3522, "step": 8637 }, { "epoch": 0.07797436360353854, "grad_norm": 2.7356581687927246, "learning_rate": 4.61017331648312e-05, "loss": 4.455, "step": 8638 }, { "epoch": 0.07798339050370102, "grad_norm": 3.450639247894287, "learning_rate": 4.610128181982307e-05, "loss": 4.3701, "step": 8639 }, { "epoch": 0.0779924174038635, "grad_norm": 3.0760586261749268, "learning_rate": 4.610083047481495e-05, "loss": 5.3394, "step": 8640 }, { "epoch": 0.078001444304026, "grad_norm": 4.024689197540283, "learning_rate": 4.610037912980682e-05, "loss": 5.3378, "step": 8641 }, { "epoch": 0.07801047120418848, "grad_norm": 3.186922788619995, "learning_rate": 4.60999277847987e-05, "loss": 4.2579, "step": 8642 }, { "epoch": 0.07801949810435096, "grad_norm": 3.2815325260162354, "learning_rate": 4.609947643979058e-05, "loss": 4.8215, "step": 8643 }, { "epoch": 0.07802852500451345, "grad_norm": 3.384228467941284, "learning_rate": 4.6099025094782453e-05, "loss": 4.7241, "step": 8644 }, { "epoch": 0.07803755190467593, "grad_norm": 3.3389084339141846, "learning_rate": 4.609857374977433e-05, "loss": 4.8215, "step": 8645 }, { "epoch": 0.07804657880483842, "grad_norm": 4.278362274169922, "learning_rate": 4.6098122404766205e-05, "loss": 4.5428, "step": 8646 }, { "epoch": 0.0780556057050009, "grad_norm": 4.508191108703613, "learning_rate": 4.609767105975808e-05, "loss": 4.9512, "step": 8647 }, { "epoch": 0.07806463260516339, "grad_norm": 3.6000540256500244, "learning_rate": 4.609721971474996e-05, "loss": 4.8162, "step": 8648 }, { "epoch": 0.07807365950532587, "grad_norm": 3.8129186630249023, "learning_rate": 4.6096768369741836e-05, "loss": 3.9158, "step": 8649 }, { "epoch": 0.07808268640548835, "grad_norm": 3.692081928253174, "learning_rate": 4.609631702473371e-05, "loss": 4.3083, "step": 8650 }, { "epoch": 0.07809171330565085, "grad_norm": 3.13468337059021, "learning_rate": 4.609586567972559e-05, "loss": 4.6758, "step": 8651 }, { "epoch": 0.07810074020581333, "grad_norm": 2.654052257537842, "learning_rate": 4.609541433471746e-05, "loss": 4.6854, "step": 8652 }, { "epoch": 0.07810976710597581, "grad_norm": 3.6153242588043213, "learning_rate": 4.609496298970933e-05, "loss": 5.3015, "step": 8653 }, { "epoch": 0.07811879400613829, "grad_norm": 3.2269794940948486, "learning_rate": 4.609451164470121e-05, "loss": 5.3844, "step": 8654 }, { "epoch": 0.07812782090630077, "grad_norm": 3.809952735900879, "learning_rate": 4.6094060299693084e-05, "loss": 4.6742, "step": 8655 }, { "epoch": 0.07813684780646327, "grad_norm": 4.507537364959717, "learning_rate": 4.6093608954684964e-05, "loss": 4.5517, "step": 8656 }, { "epoch": 0.07814587470662575, "grad_norm": 3.3092424869537354, "learning_rate": 4.609315760967684e-05, "loss": 4.313, "step": 8657 }, { "epoch": 0.07815490160678823, "grad_norm": 3.5143513679504395, "learning_rate": 4.6092706264668715e-05, "loss": 4.9272, "step": 8658 }, { "epoch": 0.07816392850695071, "grad_norm": 2.8515615463256836, "learning_rate": 4.6092254919660595e-05, "loss": 4.4966, "step": 8659 }, { "epoch": 0.07817295540711319, "grad_norm": 3.274623394012451, "learning_rate": 4.609180357465247e-05, "loss": 4.6775, "step": 8660 }, { "epoch": 0.07818198230727569, "grad_norm": 3.5853781700134277, "learning_rate": 4.609135222964434e-05, "loss": 4.1546, "step": 8661 }, { "epoch": 0.07819100920743817, "grad_norm": 4.09815788269043, "learning_rate": 4.609090088463622e-05, "loss": 4.3487, "step": 8662 }, { "epoch": 0.07820003610760065, "grad_norm": 3.029200553894043, "learning_rate": 4.609044953962809e-05, "loss": 4.7785, "step": 8663 }, { "epoch": 0.07820906300776313, "grad_norm": 3.4498355388641357, "learning_rate": 4.608999819461997e-05, "loss": 4.2543, "step": 8664 }, { "epoch": 0.07821808990792561, "grad_norm": 3.848436117172241, "learning_rate": 4.608954684961185e-05, "loss": 4.6859, "step": 8665 }, { "epoch": 0.0782271168080881, "grad_norm": 4.00855016708374, "learning_rate": 4.6089095504603715e-05, "loss": 4.5271, "step": 8666 }, { "epoch": 0.07823614370825059, "grad_norm": 4.4756669998168945, "learning_rate": 4.6088644159595595e-05, "loss": 4.1461, "step": 8667 }, { "epoch": 0.07824517060841307, "grad_norm": 2.9095282554626465, "learning_rate": 4.6088192814587474e-05, "loss": 4.9563, "step": 8668 }, { "epoch": 0.07825419750857555, "grad_norm": 3.557291269302368, "learning_rate": 4.6087741469579346e-05, "loss": 4.0963, "step": 8669 }, { "epoch": 0.07826322440873804, "grad_norm": 4.03797721862793, "learning_rate": 4.6087290124571226e-05, "loss": 4.0574, "step": 8670 }, { "epoch": 0.07827225130890052, "grad_norm": 3.9535348415374756, "learning_rate": 4.6086838779563105e-05, "loss": 5.5146, "step": 8671 }, { "epoch": 0.07828127820906301, "grad_norm": 4.9861884117126465, "learning_rate": 4.608638743455498e-05, "loss": 5.2873, "step": 8672 }, { "epoch": 0.0782903051092255, "grad_norm": 3.15504789352417, "learning_rate": 4.608593608954685e-05, "loss": 4.6841, "step": 8673 }, { "epoch": 0.07829933200938798, "grad_norm": 5.091951847076416, "learning_rate": 4.608548474453873e-05, "loss": 4.5776, "step": 8674 }, { "epoch": 0.07830835890955046, "grad_norm": 2.972525119781494, "learning_rate": 4.60850333995306e-05, "loss": 4.9963, "step": 8675 }, { "epoch": 0.07831738580971294, "grad_norm": 3.3088202476501465, "learning_rate": 4.608458205452248e-05, "loss": 4.3825, "step": 8676 }, { "epoch": 0.07832641270987543, "grad_norm": 3.770674467086792, "learning_rate": 4.608413070951435e-05, "loss": 5.032, "step": 8677 }, { "epoch": 0.07833543961003792, "grad_norm": 3.454314947128296, "learning_rate": 4.608367936450623e-05, "loss": 4.6061, "step": 8678 }, { "epoch": 0.0783444665102004, "grad_norm": 2.8984122276306152, "learning_rate": 4.608322801949811e-05, "loss": 4.0161, "step": 8679 }, { "epoch": 0.07835349341036288, "grad_norm": 3.0921833515167236, "learning_rate": 4.608277667448998e-05, "loss": 4.3391, "step": 8680 }, { "epoch": 0.07836252031052536, "grad_norm": 3.8674917221069336, "learning_rate": 4.608232532948186e-05, "loss": 4.6137, "step": 8681 }, { "epoch": 0.07837154721068786, "grad_norm": 2.8022241592407227, "learning_rate": 4.6081873984473736e-05, "loss": 4.1856, "step": 8682 }, { "epoch": 0.07838057411085034, "grad_norm": 5.231094837188721, "learning_rate": 4.608142263946561e-05, "loss": 4.1577, "step": 8683 }, { "epoch": 0.07838960101101282, "grad_norm": 3.743816614151001, "learning_rate": 4.608097129445749e-05, "loss": 4.1853, "step": 8684 }, { "epoch": 0.0783986279111753, "grad_norm": 2.7544405460357666, "learning_rate": 4.608051994944936e-05, "loss": 4.5991, "step": 8685 }, { "epoch": 0.07840765481133778, "grad_norm": 4.855722427368164, "learning_rate": 4.608006860444123e-05, "loss": 4.3044, "step": 8686 }, { "epoch": 0.07841668171150028, "grad_norm": 2.887547731399536, "learning_rate": 4.607961725943311e-05, "loss": 5.0989, "step": 8687 }, { "epoch": 0.07842570861166276, "grad_norm": 3.071864604949951, "learning_rate": 4.607916591442499e-05, "loss": 4.7259, "step": 8688 }, { "epoch": 0.07843473551182524, "grad_norm": 4.118009567260742, "learning_rate": 4.6078714569416863e-05, "loss": 4.0554, "step": 8689 }, { "epoch": 0.07844376241198772, "grad_norm": 2.9968528747558594, "learning_rate": 4.607826322440874e-05, "loss": 4.5073, "step": 8690 }, { "epoch": 0.0784527893121502, "grad_norm": 3.4952635765075684, "learning_rate": 4.6077811879400615e-05, "loss": 4.6799, "step": 8691 }, { "epoch": 0.07846181621231268, "grad_norm": 3.2937135696411133, "learning_rate": 4.6077360534392494e-05, "loss": 4.243, "step": 8692 }, { "epoch": 0.07847084311247518, "grad_norm": 4.143538475036621, "learning_rate": 4.607690918938437e-05, "loss": 4.4371, "step": 8693 }, { "epoch": 0.07847987001263766, "grad_norm": 4.157354831695557, "learning_rate": 4.607645784437624e-05, "loss": 3.474, "step": 8694 }, { "epoch": 0.07848889691280014, "grad_norm": 3.599123001098633, "learning_rate": 4.607600649936812e-05, "loss": 4.6366, "step": 8695 }, { "epoch": 0.07849792381296263, "grad_norm": 3.0662901401519775, "learning_rate": 4.607555515436e-05, "loss": 4.3197, "step": 8696 }, { "epoch": 0.0785069507131251, "grad_norm": 3.085695743560791, "learning_rate": 4.607510380935187e-05, "loss": 4.3403, "step": 8697 }, { "epoch": 0.0785159776132876, "grad_norm": 3.415076971054077, "learning_rate": 4.607465246434375e-05, "loss": 5.4759, "step": 8698 }, { "epoch": 0.07852500451345008, "grad_norm": 2.7352614402770996, "learning_rate": 4.607420111933562e-05, "loss": 4.4949, "step": 8699 }, { "epoch": 0.07853403141361257, "grad_norm": 2.6228673458099365, "learning_rate": 4.6073749774327494e-05, "loss": 5.0996, "step": 8700 }, { "epoch": 0.07854305831377505, "grad_norm": 4.8819050788879395, "learning_rate": 4.6073298429319374e-05, "loss": 4.5164, "step": 8701 }, { "epoch": 0.07855208521393753, "grad_norm": 3.631704807281494, "learning_rate": 4.6072847084311246e-05, "loss": 5.0654, "step": 8702 }, { "epoch": 0.07856111211410002, "grad_norm": 3.3740406036376953, "learning_rate": 4.6072395739303125e-05, "loss": 4.3857, "step": 8703 }, { "epoch": 0.0785701390142625, "grad_norm": 3.9110260009765625, "learning_rate": 4.6071944394295005e-05, "loss": 5.1042, "step": 8704 }, { "epoch": 0.07857916591442499, "grad_norm": 4.710090637207031, "learning_rate": 4.607149304928688e-05, "loss": 3.4627, "step": 8705 }, { "epoch": 0.07858819281458747, "grad_norm": 4.120253562927246, "learning_rate": 4.6071041704278756e-05, "loss": 4.5419, "step": 8706 }, { "epoch": 0.07859721971474995, "grad_norm": 4.345451354980469, "learning_rate": 4.607059035927063e-05, "loss": 3.9226, "step": 8707 }, { "epoch": 0.07860624661491245, "grad_norm": 4.217996597290039, "learning_rate": 4.60701390142625e-05, "loss": 4.1198, "step": 8708 }, { "epoch": 0.07861527351507493, "grad_norm": 5.233577251434326, "learning_rate": 4.606968766925438e-05, "loss": 5.1298, "step": 8709 }, { "epoch": 0.07862430041523741, "grad_norm": 5.415334701538086, "learning_rate": 4.606923632424626e-05, "loss": 4.5366, "step": 8710 }, { "epoch": 0.07863332731539989, "grad_norm": 3.1076626777648926, "learning_rate": 4.606878497923813e-05, "loss": 4.372, "step": 8711 }, { "epoch": 0.07864235421556237, "grad_norm": 3.4769842624664307, "learning_rate": 4.606833363423001e-05, "loss": 4.9917, "step": 8712 }, { "epoch": 0.07865138111572487, "grad_norm": 3.063324213027954, "learning_rate": 4.6067882289221884e-05, "loss": 4.7683, "step": 8713 }, { "epoch": 0.07866040801588735, "grad_norm": 3.9866154193878174, "learning_rate": 4.6067430944213756e-05, "loss": 3.7858, "step": 8714 }, { "epoch": 0.07866943491604983, "grad_norm": 3.0731287002563477, "learning_rate": 4.6066979599205636e-05, "loss": 5.0705, "step": 8715 }, { "epoch": 0.07867846181621231, "grad_norm": 3.7845358848571777, "learning_rate": 4.606652825419751e-05, "loss": 4.5155, "step": 8716 }, { "epoch": 0.07868748871637479, "grad_norm": 2.4036672115325928, "learning_rate": 4.606607690918939e-05, "loss": 4.6557, "step": 8717 }, { "epoch": 0.07869651561653727, "grad_norm": 2.53293514251709, "learning_rate": 4.6065625564181267e-05, "loss": 5.076, "step": 8718 }, { "epoch": 0.07870554251669977, "grad_norm": 3.486884117126465, "learning_rate": 4.606517421917314e-05, "loss": 4.7874, "step": 8719 }, { "epoch": 0.07871456941686225, "grad_norm": 3.352027177810669, "learning_rate": 4.606472287416501e-05, "loss": 4.9717, "step": 8720 }, { "epoch": 0.07872359631702473, "grad_norm": 3.1472854614257812, "learning_rate": 4.606427152915689e-05, "loss": 4.2251, "step": 8721 }, { "epoch": 0.07873262321718721, "grad_norm": 4.606432914733887, "learning_rate": 4.606382018414876e-05, "loss": 3.1579, "step": 8722 }, { "epoch": 0.0787416501173497, "grad_norm": 3.3294076919555664, "learning_rate": 4.606336883914064e-05, "loss": 4.6159, "step": 8723 }, { "epoch": 0.07875067701751219, "grad_norm": 3.9286491870880127, "learning_rate": 4.6062917494132515e-05, "loss": 4.2727, "step": 8724 }, { "epoch": 0.07875970391767467, "grad_norm": 3.3461294174194336, "learning_rate": 4.6062466149124394e-05, "loss": 4.3444, "step": 8725 }, { "epoch": 0.07876873081783715, "grad_norm": 3.88246488571167, "learning_rate": 4.6062014804116273e-05, "loss": 4.4937, "step": 8726 }, { "epoch": 0.07877775771799964, "grad_norm": 5.575805187225342, "learning_rate": 4.606156345910814e-05, "loss": 3.9723, "step": 8727 }, { "epoch": 0.07878678461816212, "grad_norm": 3.1257243156433105, "learning_rate": 4.606111211410002e-05, "loss": 4.5342, "step": 8728 }, { "epoch": 0.07879581151832461, "grad_norm": 3.984527826309204, "learning_rate": 4.60606607690919e-05, "loss": 4.8723, "step": 8729 }, { "epoch": 0.0788048384184871, "grad_norm": 5.265547275543213, "learning_rate": 4.606020942408377e-05, "loss": 4.6649, "step": 8730 }, { "epoch": 0.07881386531864958, "grad_norm": 4.3784260749816895, "learning_rate": 4.605975807907565e-05, "loss": 4.5424, "step": 8731 }, { "epoch": 0.07882289221881206, "grad_norm": 3.485675811767578, "learning_rate": 4.605930673406753e-05, "loss": 4.2511, "step": 8732 }, { "epoch": 0.07883191911897454, "grad_norm": 2.950127363204956, "learning_rate": 4.6058855389059394e-05, "loss": 4.5062, "step": 8733 }, { "epoch": 0.07884094601913703, "grad_norm": 3.3927526473999023, "learning_rate": 4.6058404044051273e-05, "loss": 4.8171, "step": 8734 }, { "epoch": 0.07884997291929952, "grad_norm": 3.026465892791748, "learning_rate": 4.605795269904315e-05, "loss": 4.466, "step": 8735 }, { "epoch": 0.078858999819462, "grad_norm": 2.773268699645996, "learning_rate": 4.6057501354035025e-05, "loss": 4.2956, "step": 8736 }, { "epoch": 0.07886802671962448, "grad_norm": 3.8084158897399902, "learning_rate": 4.6057050009026904e-05, "loss": 4.4238, "step": 8737 }, { "epoch": 0.07887705361978696, "grad_norm": 3.7521798610687256, "learning_rate": 4.605659866401878e-05, "loss": 4.401, "step": 8738 }, { "epoch": 0.07888608051994946, "grad_norm": 4.5450921058654785, "learning_rate": 4.6056147319010656e-05, "loss": 4.7618, "step": 8739 }, { "epoch": 0.07889510742011194, "grad_norm": 2.717172861099243, "learning_rate": 4.605569597400253e-05, "loss": 4.7454, "step": 8740 }, { "epoch": 0.07890413432027442, "grad_norm": 3.301138401031494, "learning_rate": 4.60552446289944e-05, "loss": 4.9886, "step": 8741 }, { "epoch": 0.0789131612204369, "grad_norm": 3.006185531616211, "learning_rate": 4.605479328398628e-05, "loss": 4.8235, "step": 8742 }, { "epoch": 0.07892218812059938, "grad_norm": 2.9167964458465576, "learning_rate": 4.605434193897816e-05, "loss": 4.7036, "step": 8743 }, { "epoch": 0.07893121502076186, "grad_norm": 2.8924970626831055, "learning_rate": 4.605389059397003e-05, "loss": 4.8884, "step": 8744 }, { "epoch": 0.07894024192092436, "grad_norm": 3.8036086559295654, "learning_rate": 4.605343924896191e-05, "loss": 4.8555, "step": 8745 }, { "epoch": 0.07894926882108684, "grad_norm": 4.289439678192139, "learning_rate": 4.6052987903953784e-05, "loss": 4.8603, "step": 8746 }, { "epoch": 0.07895829572124932, "grad_norm": 3.995892286300659, "learning_rate": 4.6052536558945656e-05, "loss": 4.5071, "step": 8747 }, { "epoch": 0.0789673226214118, "grad_norm": 3.5516974925994873, "learning_rate": 4.6052085213937535e-05, "loss": 4.5248, "step": 8748 }, { "epoch": 0.07897634952157429, "grad_norm": 2.6799254417419434, "learning_rate": 4.6051633868929415e-05, "loss": 5.0501, "step": 8749 }, { "epoch": 0.07898537642173678, "grad_norm": 2.3323936462402344, "learning_rate": 4.605118252392129e-05, "loss": 4.9839, "step": 8750 }, { "epoch": 0.07899440332189926, "grad_norm": 2.9286861419677734, "learning_rate": 4.6050731178913166e-05, "loss": 3.8062, "step": 8751 }, { "epoch": 0.07900343022206174, "grad_norm": 2.523449182510376, "learning_rate": 4.605027983390504e-05, "loss": 4.9338, "step": 8752 }, { "epoch": 0.07901245712222423, "grad_norm": 3.628303289413452, "learning_rate": 4.604982848889691e-05, "loss": 4.4413, "step": 8753 }, { "epoch": 0.0790214840223867, "grad_norm": 3.5011117458343506, "learning_rate": 4.604937714388879e-05, "loss": 4.8903, "step": 8754 }, { "epoch": 0.0790305109225492, "grad_norm": 3.120032548904419, "learning_rate": 4.604892579888066e-05, "loss": 4.5197, "step": 8755 }, { "epoch": 0.07903953782271168, "grad_norm": 4.453729629516602, "learning_rate": 4.604847445387254e-05, "loss": 4.7701, "step": 8756 }, { "epoch": 0.07904856472287417, "grad_norm": 2.4073216915130615, "learning_rate": 4.604802310886442e-05, "loss": 4.8859, "step": 8757 }, { "epoch": 0.07905759162303665, "grad_norm": 3.094651699066162, "learning_rate": 4.6047571763856294e-05, "loss": 4.2964, "step": 8758 }, { "epoch": 0.07906661852319913, "grad_norm": 3.791736602783203, "learning_rate": 4.604712041884817e-05, "loss": 4.8981, "step": 8759 }, { "epoch": 0.07907564542336162, "grad_norm": 3.5625855922698975, "learning_rate": 4.6046669073840046e-05, "loss": 4.3905, "step": 8760 }, { "epoch": 0.0790846723235241, "grad_norm": 3.6423842906951904, "learning_rate": 4.604621772883192e-05, "loss": 4.5619, "step": 8761 }, { "epoch": 0.07909369922368659, "grad_norm": 2.697244644165039, "learning_rate": 4.60457663838238e-05, "loss": 5.2013, "step": 8762 }, { "epoch": 0.07910272612384907, "grad_norm": 3.8087852001190186, "learning_rate": 4.604531503881567e-05, "loss": 4.862, "step": 8763 }, { "epoch": 0.07911175302401155, "grad_norm": 2.957843780517578, "learning_rate": 4.604486369380755e-05, "loss": 4.7245, "step": 8764 }, { "epoch": 0.07912077992417405, "grad_norm": 3.214876890182495, "learning_rate": 4.604441234879943e-05, "loss": 4.3604, "step": 8765 }, { "epoch": 0.07912980682433653, "grad_norm": 3.2059848308563232, "learning_rate": 4.60439610037913e-05, "loss": 4.7166, "step": 8766 }, { "epoch": 0.07913883372449901, "grad_norm": 3.6371243000030518, "learning_rate": 4.604350965878317e-05, "loss": 4.7035, "step": 8767 }, { "epoch": 0.07914786062466149, "grad_norm": 3.1805880069732666, "learning_rate": 4.604305831377505e-05, "loss": 4.8734, "step": 8768 }, { "epoch": 0.07915688752482397, "grad_norm": 3.319985866546631, "learning_rate": 4.6042606968766925e-05, "loss": 4.3496, "step": 8769 }, { "epoch": 0.07916591442498647, "grad_norm": 3.326615571975708, "learning_rate": 4.6042155623758804e-05, "loss": 4.6824, "step": 8770 }, { "epoch": 0.07917494132514895, "grad_norm": 3.4696857929229736, "learning_rate": 4.6041704278750683e-05, "loss": 4.8723, "step": 8771 }, { "epoch": 0.07918396822531143, "grad_norm": 3.094740629196167, "learning_rate": 4.6041252933742556e-05, "loss": 4.4558, "step": 8772 }, { "epoch": 0.07919299512547391, "grad_norm": 3.4840567111968994, "learning_rate": 4.6040801588734435e-05, "loss": 4.9232, "step": 8773 }, { "epoch": 0.07920202202563639, "grad_norm": 3.7591564655303955, "learning_rate": 4.604035024372631e-05, "loss": 4.4923, "step": 8774 }, { "epoch": 0.07921104892579887, "grad_norm": 3.5690064430236816, "learning_rate": 4.603989889871818e-05, "loss": 4.2705, "step": 8775 }, { "epoch": 0.07922007582596137, "grad_norm": 2.116692066192627, "learning_rate": 4.603944755371006e-05, "loss": 4.8937, "step": 8776 }, { "epoch": 0.07922910272612385, "grad_norm": 3.88688588142395, "learning_rate": 4.603899620870193e-05, "loss": 4.4184, "step": 8777 }, { "epoch": 0.07923812962628633, "grad_norm": 2.9072964191436768, "learning_rate": 4.603854486369381e-05, "loss": 5.0507, "step": 8778 }, { "epoch": 0.07924715652644881, "grad_norm": 4.722231388092041, "learning_rate": 4.603809351868569e-05, "loss": 3.9566, "step": 8779 }, { "epoch": 0.0792561834266113, "grad_norm": 2.7136621475219727, "learning_rate": 4.6037642173677556e-05, "loss": 5.1806, "step": 8780 }, { "epoch": 0.07926521032677379, "grad_norm": 3.6339380741119385, "learning_rate": 4.6037190828669435e-05, "loss": 4.314, "step": 8781 }, { "epoch": 0.07927423722693627, "grad_norm": 3.3370015621185303, "learning_rate": 4.6036739483661314e-05, "loss": 4.2464, "step": 8782 }, { "epoch": 0.07928326412709875, "grad_norm": 2.3493618965148926, "learning_rate": 4.603628813865319e-05, "loss": 4.91, "step": 8783 }, { "epoch": 0.07929229102726124, "grad_norm": 3.2883570194244385, "learning_rate": 4.6035836793645066e-05, "loss": 4.823, "step": 8784 }, { "epoch": 0.07930131792742372, "grad_norm": 7.144737243652344, "learning_rate": 4.603538544863694e-05, "loss": 4.6544, "step": 8785 }, { "epoch": 0.07931034482758621, "grad_norm": 3.7793989181518555, "learning_rate": 4.603493410362882e-05, "loss": 5.1031, "step": 8786 }, { "epoch": 0.0793193717277487, "grad_norm": 2.914496660232544, "learning_rate": 4.603448275862069e-05, "loss": 4.6126, "step": 8787 }, { "epoch": 0.07932839862791118, "grad_norm": 4.709386348724365, "learning_rate": 4.603403141361256e-05, "loss": 4.7164, "step": 8788 }, { "epoch": 0.07933742552807366, "grad_norm": 2.8674280643463135, "learning_rate": 4.603358006860444e-05, "loss": 5.1038, "step": 8789 }, { "epoch": 0.07934645242823614, "grad_norm": 3.590038299560547, "learning_rate": 4.603312872359632e-05, "loss": 4.9213, "step": 8790 }, { "epoch": 0.07935547932839863, "grad_norm": 3.2860329151153564, "learning_rate": 4.6032677378588194e-05, "loss": 4.8258, "step": 8791 }, { "epoch": 0.07936450622856112, "grad_norm": 3.554851531982422, "learning_rate": 4.603222603358007e-05, "loss": 5.0748, "step": 8792 }, { "epoch": 0.0793735331287236, "grad_norm": 3.272711753845215, "learning_rate": 4.603177468857195e-05, "loss": 4.5919, "step": 8793 }, { "epoch": 0.07938256002888608, "grad_norm": 3.4914610385894775, "learning_rate": 4.603132334356382e-05, "loss": 4.6269, "step": 8794 }, { "epoch": 0.07939158692904856, "grad_norm": 3.098301410675049, "learning_rate": 4.60308719985557e-05, "loss": 4.3142, "step": 8795 }, { "epoch": 0.07940061382921106, "grad_norm": 3.8022379875183105, "learning_rate": 4.6030420653547576e-05, "loss": 4.965, "step": 8796 }, { "epoch": 0.07940964072937354, "grad_norm": 3.4610707759857178, "learning_rate": 4.602996930853945e-05, "loss": 3.9212, "step": 8797 }, { "epoch": 0.07941866762953602, "grad_norm": 3.1334540843963623, "learning_rate": 4.602951796353133e-05, "loss": 4.4324, "step": 8798 }, { "epoch": 0.0794276945296985, "grad_norm": 4.117389678955078, "learning_rate": 4.60290666185232e-05, "loss": 5.0795, "step": 8799 }, { "epoch": 0.07943672142986098, "grad_norm": 3.372175455093384, "learning_rate": 4.602861527351507e-05, "loss": 5.0132, "step": 8800 }, { "epoch": 0.07944574833002346, "grad_norm": 3.0320746898651123, "learning_rate": 4.602816392850695e-05, "loss": 4.5385, "step": 8801 }, { "epoch": 0.07945477523018596, "grad_norm": 4.773359298706055, "learning_rate": 4.6027712583498825e-05, "loss": 4.8038, "step": 8802 }, { "epoch": 0.07946380213034844, "grad_norm": 2.9823832511901855, "learning_rate": 4.6027261238490704e-05, "loss": 4.1984, "step": 8803 }, { "epoch": 0.07947282903051092, "grad_norm": 2.9532763957977295, "learning_rate": 4.602680989348258e-05, "loss": 5.0767, "step": 8804 }, { "epoch": 0.0794818559306734, "grad_norm": 3.808899164199829, "learning_rate": 4.6026358548474456e-05, "loss": 4.1952, "step": 8805 }, { "epoch": 0.07949088283083589, "grad_norm": 3.1438281536102295, "learning_rate": 4.6025907203466335e-05, "loss": 4.3147, "step": 8806 }, { "epoch": 0.07949990973099838, "grad_norm": 4.571322917938232, "learning_rate": 4.602545585845821e-05, "loss": 4.5573, "step": 8807 }, { "epoch": 0.07950893663116086, "grad_norm": 4.688242435455322, "learning_rate": 4.602500451345008e-05, "loss": 3.3976, "step": 8808 }, { "epoch": 0.07951796353132334, "grad_norm": 2.6962263584136963, "learning_rate": 4.602455316844196e-05, "loss": 4.0017, "step": 8809 }, { "epoch": 0.07952699043148583, "grad_norm": 4.771925449371338, "learning_rate": 4.602410182343384e-05, "loss": 4.8252, "step": 8810 }, { "epoch": 0.0795360173316483, "grad_norm": 2.8264987468719482, "learning_rate": 4.602365047842571e-05, "loss": 4.6139, "step": 8811 }, { "epoch": 0.0795450442318108, "grad_norm": 3.4081172943115234, "learning_rate": 4.602319913341759e-05, "loss": 4.4877, "step": 8812 }, { "epoch": 0.07955407113197328, "grad_norm": 3.770811080932617, "learning_rate": 4.602274778840946e-05, "loss": 4.6541, "step": 8813 }, { "epoch": 0.07956309803213577, "grad_norm": 3.5710222721099854, "learning_rate": 4.6022296443401335e-05, "loss": 4.5065, "step": 8814 }, { "epoch": 0.07957212493229825, "grad_norm": 3.5332515239715576, "learning_rate": 4.6021845098393214e-05, "loss": 4.5538, "step": 8815 }, { "epoch": 0.07958115183246073, "grad_norm": 3.961684465408325, "learning_rate": 4.602139375338509e-05, "loss": 4.6229, "step": 8816 }, { "epoch": 0.07959017873262322, "grad_norm": 2.7464473247528076, "learning_rate": 4.6020942408376966e-05, "loss": 4.8986, "step": 8817 }, { "epoch": 0.0795992056327857, "grad_norm": 2.3875198364257812, "learning_rate": 4.6020491063368845e-05, "loss": 4.6091, "step": 8818 }, { "epoch": 0.07960823253294819, "grad_norm": 3.16776967048645, "learning_rate": 4.602003971836072e-05, "loss": 5.0067, "step": 8819 }, { "epoch": 0.07961725943311067, "grad_norm": 3.5814411640167236, "learning_rate": 4.60195883733526e-05, "loss": 4.8775, "step": 8820 }, { "epoch": 0.07962628633327315, "grad_norm": 3.535783052444458, "learning_rate": 4.601913702834447e-05, "loss": 3.3479, "step": 8821 }, { "epoch": 0.07963531323343565, "grad_norm": 3.8913533687591553, "learning_rate": 4.601868568333634e-05, "loss": 3.7201, "step": 8822 }, { "epoch": 0.07964434013359813, "grad_norm": 4.579864501953125, "learning_rate": 4.601823433832822e-05, "loss": 3.7971, "step": 8823 }, { "epoch": 0.07965336703376061, "grad_norm": 3.148185968399048, "learning_rate": 4.6017782993320093e-05, "loss": 4.6667, "step": 8824 }, { "epoch": 0.07966239393392309, "grad_norm": 3.832174301147461, "learning_rate": 4.601733164831197e-05, "loss": 4.9025, "step": 8825 }, { "epoch": 0.07967142083408557, "grad_norm": 3.4533751010894775, "learning_rate": 4.601688030330385e-05, "loss": 4.4481, "step": 8826 }, { "epoch": 0.07968044773424805, "grad_norm": 3.3186047077178955, "learning_rate": 4.601642895829572e-05, "loss": 4.2525, "step": 8827 }, { "epoch": 0.07968947463441055, "grad_norm": 2.647104024887085, "learning_rate": 4.60159776132876e-05, "loss": 4.4628, "step": 8828 }, { "epoch": 0.07969850153457303, "grad_norm": 2.968662738800049, "learning_rate": 4.6015526268279476e-05, "loss": 4.5339, "step": 8829 }, { "epoch": 0.07970752843473551, "grad_norm": 4.700820446014404, "learning_rate": 4.601507492327135e-05, "loss": 4.8721, "step": 8830 }, { "epoch": 0.07971655533489799, "grad_norm": 3.7163219451904297, "learning_rate": 4.601462357826323e-05, "loss": 4.2945, "step": 8831 }, { "epoch": 0.07972558223506047, "grad_norm": 3.180464267730713, "learning_rate": 4.601417223325511e-05, "loss": 4.148, "step": 8832 }, { "epoch": 0.07973460913522297, "grad_norm": 3.484344720840454, "learning_rate": 4.601372088824698e-05, "loss": 4.6447, "step": 8833 }, { "epoch": 0.07974363603538545, "grad_norm": 3.0980398654937744, "learning_rate": 4.601326954323885e-05, "loss": 4.8789, "step": 8834 }, { "epoch": 0.07975266293554793, "grad_norm": 3.0669608116149902, "learning_rate": 4.601281819823073e-05, "loss": 4.1742, "step": 8835 }, { "epoch": 0.07976168983571041, "grad_norm": 4.791337490081787, "learning_rate": 4.6012366853222604e-05, "loss": 3.4329, "step": 8836 }, { "epoch": 0.0797707167358729, "grad_norm": 3.808992624282837, "learning_rate": 4.601191550821448e-05, "loss": 5.1825, "step": 8837 }, { "epoch": 0.07977974363603539, "grad_norm": 3.026953935623169, "learning_rate": 4.6011464163206355e-05, "loss": 4.0667, "step": 8838 }, { "epoch": 0.07978877053619787, "grad_norm": 3.482034683227539, "learning_rate": 4.6011012818198235e-05, "loss": 4.4901, "step": 8839 }, { "epoch": 0.07979779743636035, "grad_norm": 3.8284502029418945, "learning_rate": 4.6010561473190114e-05, "loss": 4.9219, "step": 8840 }, { "epoch": 0.07980682433652284, "grad_norm": 3.4650604724884033, "learning_rate": 4.601011012818198e-05, "loss": 4.3856, "step": 8841 }, { "epoch": 0.07981585123668532, "grad_norm": 5.4365763664245605, "learning_rate": 4.600965878317386e-05, "loss": 3.1692, "step": 8842 }, { "epoch": 0.07982487813684781, "grad_norm": 3.1371798515319824, "learning_rate": 4.600920743816574e-05, "loss": 4.6219, "step": 8843 }, { "epoch": 0.0798339050370103, "grad_norm": 3.0529141426086426, "learning_rate": 4.600875609315761e-05, "loss": 4.6982, "step": 8844 }, { "epoch": 0.07984293193717278, "grad_norm": 2.849915027618408, "learning_rate": 4.600830474814949e-05, "loss": 4.5279, "step": 8845 }, { "epoch": 0.07985195883733526, "grad_norm": 4.4575276374816895, "learning_rate": 4.600785340314136e-05, "loss": 4.4935, "step": 8846 }, { "epoch": 0.07986098573749774, "grad_norm": 3.4994466304779053, "learning_rate": 4.6007402058133235e-05, "loss": 4.6687, "step": 8847 }, { "epoch": 0.07987001263766023, "grad_norm": 4.2870330810546875, "learning_rate": 4.6006950713125114e-05, "loss": 4.5269, "step": 8848 }, { "epoch": 0.07987903953782272, "grad_norm": 3.1536359786987305, "learning_rate": 4.600649936811699e-05, "loss": 4.5061, "step": 8849 }, { "epoch": 0.0798880664379852, "grad_norm": 3.0695903301239014, "learning_rate": 4.6006048023108866e-05, "loss": 4.434, "step": 8850 }, { "epoch": 0.07989709333814768, "grad_norm": 2.8379688262939453, "learning_rate": 4.6005596678100745e-05, "loss": 3.9854, "step": 8851 }, { "epoch": 0.07990612023831016, "grad_norm": 3.2616260051727295, "learning_rate": 4.600514533309262e-05, "loss": 4.4349, "step": 8852 }, { "epoch": 0.07991514713847264, "grad_norm": 3.8207192420959473, "learning_rate": 4.6004693988084497e-05, "loss": 4.3675, "step": 8853 }, { "epoch": 0.07992417403863514, "grad_norm": 5.695288181304932, "learning_rate": 4.600424264307637e-05, "loss": 4.9621, "step": 8854 }, { "epoch": 0.07993320093879762, "grad_norm": 4.080031394958496, "learning_rate": 4.600379129806824e-05, "loss": 4.5462, "step": 8855 }, { "epoch": 0.0799422278389601, "grad_norm": 3.204091787338257, "learning_rate": 4.600333995306012e-05, "loss": 5.5997, "step": 8856 }, { "epoch": 0.07995125473912258, "grad_norm": 3.9235126972198486, "learning_rate": 4.6002888608052e-05, "loss": 4.8991, "step": 8857 }, { "epoch": 0.07996028163928506, "grad_norm": 3.521153211593628, "learning_rate": 4.600243726304387e-05, "loss": 4.8005, "step": 8858 }, { "epoch": 0.07996930853944756, "grad_norm": 2.8282876014709473, "learning_rate": 4.600198591803575e-05, "loss": 5.3805, "step": 8859 }, { "epoch": 0.07997833543961004, "grad_norm": 4.868592262268066, "learning_rate": 4.6001534573027624e-05, "loss": 4.4963, "step": 8860 }, { "epoch": 0.07998736233977252, "grad_norm": 3.6492393016815186, "learning_rate": 4.60010832280195e-05, "loss": 4.1139, "step": 8861 }, { "epoch": 0.079996389239935, "grad_norm": 2.9570276737213135, "learning_rate": 4.6000631883011376e-05, "loss": 4.965, "step": 8862 }, { "epoch": 0.08000541614009749, "grad_norm": 3.043013572692871, "learning_rate": 4.600018053800325e-05, "loss": 4.5314, "step": 8863 }, { "epoch": 0.08001444304025998, "grad_norm": 3.1937310695648193, "learning_rate": 4.599972919299513e-05, "loss": 4.8691, "step": 8864 }, { "epoch": 0.08002346994042246, "grad_norm": 3.4558634757995605, "learning_rate": 4.599927784798701e-05, "loss": 3.237, "step": 8865 }, { "epoch": 0.08003249684058494, "grad_norm": 3.675328016281128, "learning_rate": 4.599882650297888e-05, "loss": 4.4577, "step": 8866 }, { "epoch": 0.08004152374074743, "grad_norm": 2.5836873054504395, "learning_rate": 4.599837515797076e-05, "loss": 4.5309, "step": 8867 }, { "epoch": 0.0800505506409099, "grad_norm": 3.9224514961242676, "learning_rate": 4.599792381296263e-05, "loss": 4.3598, "step": 8868 }, { "epoch": 0.0800595775410724, "grad_norm": 3.3198955059051514, "learning_rate": 4.5997472467954503e-05, "loss": 4.2627, "step": 8869 }, { "epoch": 0.08006860444123488, "grad_norm": 3.3490114212036133, "learning_rate": 4.599702112294638e-05, "loss": 4.4962, "step": 8870 }, { "epoch": 0.08007763134139737, "grad_norm": 3.3802664279937744, "learning_rate": 4.599656977793826e-05, "loss": 4.3101, "step": 8871 }, { "epoch": 0.08008665824155985, "grad_norm": 3.1004977226257324, "learning_rate": 4.5996118432930134e-05, "loss": 4.46, "step": 8872 }, { "epoch": 0.08009568514172233, "grad_norm": 3.4895386695861816, "learning_rate": 4.5995667087922014e-05, "loss": 5.0124, "step": 8873 }, { "epoch": 0.08010471204188482, "grad_norm": 3.5723459720611572, "learning_rate": 4.5995215742913886e-05, "loss": 4.7816, "step": 8874 }, { "epoch": 0.0801137389420473, "grad_norm": 5.398276329040527, "learning_rate": 4.599476439790576e-05, "loss": 4.3389, "step": 8875 }, { "epoch": 0.08012276584220979, "grad_norm": 3.0953328609466553, "learning_rate": 4.599431305289764e-05, "loss": 4.7905, "step": 8876 }, { "epoch": 0.08013179274237227, "grad_norm": 3.848341941833496, "learning_rate": 4.599386170788951e-05, "loss": 4.5766, "step": 8877 }, { "epoch": 0.08014081964253475, "grad_norm": 3.7534778118133545, "learning_rate": 4.599341036288139e-05, "loss": 4.2878, "step": 8878 }, { "epoch": 0.08014984654269723, "grad_norm": 3.389444351196289, "learning_rate": 4.599295901787327e-05, "loss": 4.8254, "step": 8879 }, { "epoch": 0.08015887344285973, "grad_norm": 2.8429489135742188, "learning_rate": 4.599250767286514e-05, "loss": 4.8699, "step": 8880 }, { "epoch": 0.08016790034302221, "grad_norm": 3.5343666076660156, "learning_rate": 4.5992056327857014e-05, "loss": 5.0262, "step": 8881 }, { "epoch": 0.08017692724318469, "grad_norm": 3.600339889526367, "learning_rate": 4.599160498284889e-05, "loss": 4.1105, "step": 8882 }, { "epoch": 0.08018595414334717, "grad_norm": 5.088688850402832, "learning_rate": 4.5991153637840765e-05, "loss": 3.6157, "step": 8883 }, { "epoch": 0.08019498104350965, "grad_norm": 2.9458866119384766, "learning_rate": 4.5990702292832645e-05, "loss": 4.6702, "step": 8884 }, { "epoch": 0.08020400794367215, "grad_norm": 3.2039544582366943, "learning_rate": 4.599025094782452e-05, "loss": 4.3339, "step": 8885 }, { "epoch": 0.08021303484383463, "grad_norm": 4.92914342880249, "learning_rate": 4.5989799602816396e-05, "loss": 4.2439, "step": 8886 }, { "epoch": 0.08022206174399711, "grad_norm": 6.861179828643799, "learning_rate": 4.5989348257808276e-05, "loss": 2.7617, "step": 8887 }, { "epoch": 0.08023108864415959, "grad_norm": 3.269822597503662, "learning_rate": 4.598889691280014e-05, "loss": 4.1463, "step": 8888 }, { "epoch": 0.08024011554432207, "grad_norm": 3.281557559967041, "learning_rate": 4.598844556779202e-05, "loss": 4.9534, "step": 8889 }, { "epoch": 0.08024914244448457, "grad_norm": 2.6720829010009766, "learning_rate": 4.59879942227839e-05, "loss": 4.5204, "step": 8890 }, { "epoch": 0.08025816934464705, "grad_norm": 4.1138715744018555, "learning_rate": 4.598754287777577e-05, "loss": 4.4961, "step": 8891 }, { "epoch": 0.08026719624480953, "grad_norm": 3.623304843902588, "learning_rate": 4.598709153276765e-05, "loss": 4.9003, "step": 8892 }, { "epoch": 0.08027622314497201, "grad_norm": 3.3064560890197754, "learning_rate": 4.598664018775953e-05, "loss": 5.0793, "step": 8893 }, { "epoch": 0.0802852500451345, "grad_norm": 2.7756855487823486, "learning_rate": 4.5986188842751396e-05, "loss": 4.7829, "step": 8894 }, { "epoch": 0.08029427694529699, "grad_norm": 4.049150466918945, "learning_rate": 4.5985737497743276e-05, "loss": 5.4493, "step": 8895 }, { "epoch": 0.08030330384545947, "grad_norm": 3.357661008834839, "learning_rate": 4.5985286152735155e-05, "loss": 4.7578, "step": 8896 }, { "epoch": 0.08031233074562195, "grad_norm": 2.8440916538238525, "learning_rate": 4.598483480772703e-05, "loss": 4.8238, "step": 8897 }, { "epoch": 0.08032135764578444, "grad_norm": 3.5202291011810303, "learning_rate": 4.5984383462718907e-05, "loss": 5.3008, "step": 8898 }, { "epoch": 0.08033038454594692, "grad_norm": 3.394838333129883, "learning_rate": 4.598393211771078e-05, "loss": 4.8662, "step": 8899 }, { "epoch": 0.08033941144610941, "grad_norm": 3.3171706199645996, "learning_rate": 4.598348077270266e-05, "loss": 4.5905, "step": 8900 }, { "epoch": 0.0803484383462719, "grad_norm": 3.5908472537994385, "learning_rate": 4.598302942769453e-05, "loss": 4.1422, "step": 8901 }, { "epoch": 0.08035746524643438, "grad_norm": 5.353067398071289, "learning_rate": 4.59825780826864e-05, "loss": 4.2694, "step": 8902 }, { "epoch": 0.08036649214659686, "grad_norm": 4.32619571685791, "learning_rate": 4.598212673767828e-05, "loss": 3.9326, "step": 8903 }, { "epoch": 0.08037551904675934, "grad_norm": 3.4337339401245117, "learning_rate": 4.598167539267016e-05, "loss": 4.8428, "step": 8904 }, { "epoch": 0.08038454594692182, "grad_norm": 2.705092430114746, "learning_rate": 4.5981224047662034e-05, "loss": 4.9447, "step": 8905 }, { "epoch": 0.08039357284708432, "grad_norm": 2.6295392513275146, "learning_rate": 4.5980772702653913e-05, "loss": 4.4022, "step": 8906 }, { "epoch": 0.0804025997472468, "grad_norm": 3.0905497074127197, "learning_rate": 4.5980321357645786e-05, "loss": 4.6132, "step": 8907 }, { "epoch": 0.08041162664740928, "grad_norm": 3.7067923545837402, "learning_rate": 4.597987001263766e-05, "loss": 4.8517, "step": 8908 }, { "epoch": 0.08042065354757176, "grad_norm": 4.013866901397705, "learning_rate": 4.597941866762954e-05, "loss": 4.9825, "step": 8909 }, { "epoch": 0.08042968044773424, "grad_norm": 3.2210755348205566, "learning_rate": 4.597896732262142e-05, "loss": 4.4786, "step": 8910 }, { "epoch": 0.08043870734789674, "grad_norm": 3.229100227355957, "learning_rate": 4.597851597761329e-05, "loss": 4.081, "step": 8911 }, { "epoch": 0.08044773424805922, "grad_norm": 2.774752378463745, "learning_rate": 4.597806463260517e-05, "loss": 4.887, "step": 8912 }, { "epoch": 0.0804567611482217, "grad_norm": 3.632943868637085, "learning_rate": 4.597761328759704e-05, "loss": 4.6698, "step": 8913 }, { "epoch": 0.08046578804838418, "grad_norm": 3.9207711219787598, "learning_rate": 4.5977161942588913e-05, "loss": 4.7726, "step": 8914 }, { "epoch": 0.08047481494854666, "grad_norm": 4.1603522300720215, "learning_rate": 4.597671059758079e-05, "loss": 4.6171, "step": 8915 }, { "epoch": 0.08048384184870916, "grad_norm": 5.112011432647705, "learning_rate": 4.5976259252572665e-05, "loss": 4.8713, "step": 8916 }, { "epoch": 0.08049286874887164, "grad_norm": 3.163679838180542, "learning_rate": 4.5975807907564544e-05, "loss": 4.5698, "step": 8917 }, { "epoch": 0.08050189564903412, "grad_norm": 3.248197555541992, "learning_rate": 4.5975356562556424e-05, "loss": 4.5138, "step": 8918 }, { "epoch": 0.0805109225491966, "grad_norm": 4.151552677154541, "learning_rate": 4.5974905217548296e-05, "loss": 4.9183, "step": 8919 }, { "epoch": 0.08051994944935909, "grad_norm": 3.1965861320495605, "learning_rate": 4.5974453872540175e-05, "loss": 4.7787, "step": 8920 }, { "epoch": 0.08052897634952158, "grad_norm": 3.3045542240142822, "learning_rate": 4.597400252753205e-05, "loss": 4.7763, "step": 8921 }, { "epoch": 0.08053800324968406, "grad_norm": 3.6504201889038086, "learning_rate": 4.597355118252392e-05, "loss": 4.0944, "step": 8922 }, { "epoch": 0.08054703014984654, "grad_norm": 3.6447863578796387, "learning_rate": 4.59730998375158e-05, "loss": 5.3296, "step": 8923 }, { "epoch": 0.08055605705000903, "grad_norm": 3.1607022285461426, "learning_rate": 4.597264849250767e-05, "loss": 4.8861, "step": 8924 }, { "epoch": 0.0805650839501715, "grad_norm": 4.510249614715576, "learning_rate": 4.597219714749955e-05, "loss": 4.9999, "step": 8925 }, { "epoch": 0.080574110850334, "grad_norm": 2.545060873031616, "learning_rate": 4.597174580249143e-05, "loss": 4.4504, "step": 8926 }, { "epoch": 0.08058313775049648, "grad_norm": 2.966315984725952, "learning_rate": 4.59712944574833e-05, "loss": 4.5535, "step": 8927 }, { "epoch": 0.08059216465065897, "grad_norm": 3.327500820159912, "learning_rate": 4.5970843112475175e-05, "loss": 4.7079, "step": 8928 }, { "epoch": 0.08060119155082145, "grad_norm": 3.149686574935913, "learning_rate": 4.5970391767467055e-05, "loss": 4.6209, "step": 8929 }, { "epoch": 0.08061021845098393, "grad_norm": 3.794217109680176, "learning_rate": 4.596994042245893e-05, "loss": 4.662, "step": 8930 }, { "epoch": 0.08061924535114641, "grad_norm": 3.4173405170440674, "learning_rate": 4.5969489077450806e-05, "loss": 4.8695, "step": 8931 }, { "epoch": 0.0806282722513089, "grad_norm": 4.515817642211914, "learning_rate": 4.5969037732442686e-05, "loss": 5.4456, "step": 8932 }, { "epoch": 0.08063729915147139, "grad_norm": 2.9874320030212402, "learning_rate": 4.596858638743456e-05, "loss": 4.3351, "step": 8933 }, { "epoch": 0.08064632605163387, "grad_norm": 2.770454168319702, "learning_rate": 4.596813504242644e-05, "loss": 4.8785, "step": 8934 }, { "epoch": 0.08065535295179635, "grad_norm": 3.737736463546753, "learning_rate": 4.596768369741831e-05, "loss": 4.7485, "step": 8935 }, { "epoch": 0.08066437985195883, "grad_norm": 3.399905204772949, "learning_rate": 4.596723235241018e-05, "loss": 4.4837, "step": 8936 }, { "epoch": 0.08067340675212133, "grad_norm": 5.320535182952881, "learning_rate": 4.596678100740206e-05, "loss": 4.8937, "step": 8937 }, { "epoch": 0.08068243365228381, "grad_norm": 4.0995306968688965, "learning_rate": 4.5966329662393934e-05, "loss": 5.3374, "step": 8938 }, { "epoch": 0.08069146055244629, "grad_norm": 4.496901035308838, "learning_rate": 4.596587831738581e-05, "loss": 5.2496, "step": 8939 }, { "epoch": 0.08070048745260877, "grad_norm": 4.872959613800049, "learning_rate": 4.596542697237769e-05, "loss": 3.5768, "step": 8940 }, { "epoch": 0.08070951435277125, "grad_norm": 3.147564649581909, "learning_rate": 4.596497562736956e-05, "loss": 4.1795, "step": 8941 }, { "epoch": 0.08071854125293375, "grad_norm": 6.3127760887146, "learning_rate": 4.596452428236144e-05, "loss": 4.0106, "step": 8942 }, { "epoch": 0.08072756815309623, "grad_norm": 3.425642728805542, "learning_rate": 4.5964072937353317e-05, "loss": 4.3575, "step": 8943 }, { "epoch": 0.08073659505325871, "grad_norm": 3.480285406112671, "learning_rate": 4.596362159234519e-05, "loss": 4.8769, "step": 8944 }, { "epoch": 0.08074562195342119, "grad_norm": 3.000239610671997, "learning_rate": 4.596317024733707e-05, "loss": 4.5884, "step": 8945 }, { "epoch": 0.08075464885358367, "grad_norm": 4.532136917114258, "learning_rate": 4.596271890232894e-05, "loss": 4.8958, "step": 8946 }, { "epoch": 0.08076367575374617, "grad_norm": 3.5816328525543213, "learning_rate": 4.596226755732082e-05, "loss": 4.7501, "step": 8947 }, { "epoch": 0.08077270265390865, "grad_norm": 4.476059913635254, "learning_rate": 4.596181621231269e-05, "loss": 4.1778, "step": 8948 }, { "epoch": 0.08078172955407113, "grad_norm": 3.0371241569519043, "learning_rate": 4.5961364867304565e-05, "loss": 4.5627, "step": 8949 }, { "epoch": 0.08079075645423361, "grad_norm": 3.404287576675415, "learning_rate": 4.5960913522296444e-05, "loss": 5.2879, "step": 8950 }, { "epoch": 0.0807997833543961, "grad_norm": 3.819366693496704, "learning_rate": 4.5960462177288323e-05, "loss": 4.6674, "step": 8951 }, { "epoch": 0.08080881025455859, "grad_norm": 4.070602893829346, "learning_rate": 4.5960010832280196e-05, "loss": 3.7312, "step": 8952 }, { "epoch": 0.08081783715472107, "grad_norm": 3.3986289501190186, "learning_rate": 4.5959559487272075e-05, "loss": 4.2423, "step": 8953 }, { "epoch": 0.08082686405488355, "grad_norm": 3.4971015453338623, "learning_rate": 4.5959108142263954e-05, "loss": 4.8086, "step": 8954 }, { "epoch": 0.08083589095504604, "grad_norm": 3.71022629737854, "learning_rate": 4.595865679725582e-05, "loss": 4.0718, "step": 8955 }, { "epoch": 0.08084491785520852, "grad_norm": 2.5290284156799316, "learning_rate": 4.59582054522477e-05, "loss": 5.1214, "step": 8956 }, { "epoch": 0.080853944755371, "grad_norm": 3.314131498336792, "learning_rate": 4.595775410723958e-05, "loss": 4.637, "step": 8957 }, { "epoch": 0.0808629716555335, "grad_norm": 3.261054277420044, "learning_rate": 4.595730276223145e-05, "loss": 4.9016, "step": 8958 }, { "epoch": 0.08087199855569598, "grad_norm": 3.9218435287475586, "learning_rate": 4.595685141722333e-05, "loss": 4.6223, "step": 8959 }, { "epoch": 0.08088102545585846, "grad_norm": 3.5524842739105225, "learning_rate": 4.59564000722152e-05, "loss": 4.6342, "step": 8960 }, { "epoch": 0.08089005235602094, "grad_norm": 4.06544303894043, "learning_rate": 4.5955948727207075e-05, "loss": 5.2012, "step": 8961 }, { "epoch": 0.08089907925618342, "grad_norm": 4.87388277053833, "learning_rate": 4.5955497382198954e-05, "loss": 4.8073, "step": 8962 }, { "epoch": 0.08090810615634592, "grad_norm": 3.590055465698242, "learning_rate": 4.595504603719083e-05, "loss": 5.2653, "step": 8963 }, { "epoch": 0.0809171330565084, "grad_norm": 3.2006924152374268, "learning_rate": 4.5954594692182706e-05, "loss": 4.3449, "step": 8964 }, { "epoch": 0.08092615995667088, "grad_norm": 4.774459362030029, "learning_rate": 4.5954143347174585e-05, "loss": 3.7381, "step": 8965 }, { "epoch": 0.08093518685683336, "grad_norm": 4.079883575439453, "learning_rate": 4.595369200216646e-05, "loss": 4.4898, "step": 8966 }, { "epoch": 0.08094421375699584, "grad_norm": 3.5936167240142822, "learning_rate": 4.595324065715834e-05, "loss": 4.7216, "step": 8967 }, { "epoch": 0.08095324065715834, "grad_norm": 2.9697911739349365, "learning_rate": 4.595278931215021e-05, "loss": 4.1512, "step": 8968 }, { "epoch": 0.08096226755732082, "grad_norm": 2.987821578979492, "learning_rate": 4.595233796714208e-05, "loss": 5.0517, "step": 8969 }, { "epoch": 0.0809712944574833, "grad_norm": 3.886005163192749, "learning_rate": 4.595188662213396e-05, "loss": 3.6533, "step": 8970 }, { "epoch": 0.08098032135764578, "grad_norm": 3.8334789276123047, "learning_rate": 4.595143527712584e-05, "loss": 4.5284, "step": 8971 }, { "epoch": 0.08098934825780826, "grad_norm": 3.2138900756835938, "learning_rate": 4.595098393211771e-05, "loss": 4.485, "step": 8972 }, { "epoch": 0.08099837515797076, "grad_norm": 3.1844167709350586, "learning_rate": 4.595053258710959e-05, "loss": 4.2202, "step": 8973 }, { "epoch": 0.08100740205813324, "grad_norm": 4.168237209320068, "learning_rate": 4.5950081242101465e-05, "loss": 4.076, "step": 8974 }, { "epoch": 0.08101642895829572, "grad_norm": 3.674834966659546, "learning_rate": 4.594962989709334e-05, "loss": 4.5109, "step": 8975 }, { "epoch": 0.0810254558584582, "grad_norm": 3.273747682571411, "learning_rate": 4.5949178552085216e-05, "loss": 4.27, "step": 8976 }, { "epoch": 0.08103448275862069, "grad_norm": 2.932218551635742, "learning_rate": 4.594872720707709e-05, "loss": 4.7734, "step": 8977 }, { "epoch": 0.08104350965878318, "grad_norm": 3.523253917694092, "learning_rate": 4.594827586206897e-05, "loss": 5.1068, "step": 8978 }, { "epoch": 0.08105253655894566, "grad_norm": 3.7041871547698975, "learning_rate": 4.594782451706085e-05, "loss": 4.8587, "step": 8979 }, { "epoch": 0.08106156345910814, "grad_norm": 3.8705437183380127, "learning_rate": 4.594737317205272e-05, "loss": 4.9123, "step": 8980 }, { "epoch": 0.08107059035927063, "grad_norm": 3.3052077293395996, "learning_rate": 4.59469218270446e-05, "loss": 4.6632, "step": 8981 }, { "epoch": 0.0810796172594331, "grad_norm": 2.943406820297241, "learning_rate": 4.594647048203647e-05, "loss": 3.944, "step": 8982 }, { "epoch": 0.08108864415959559, "grad_norm": 3.712432622909546, "learning_rate": 4.5946019137028344e-05, "loss": 4.3898, "step": 8983 }, { "epoch": 0.08109767105975808, "grad_norm": 3.712965965270996, "learning_rate": 4.594556779202022e-05, "loss": 4.4707, "step": 8984 }, { "epoch": 0.08110669795992057, "grad_norm": 2.656414270401001, "learning_rate": 4.5945116447012096e-05, "loss": 4.968, "step": 8985 }, { "epoch": 0.08111572486008305, "grad_norm": 5.478793621063232, "learning_rate": 4.5944665102003975e-05, "loss": 4.3835, "step": 8986 }, { "epoch": 0.08112475176024553, "grad_norm": 3.846215009689331, "learning_rate": 4.5944213756995854e-05, "loss": 4.3529, "step": 8987 }, { "epoch": 0.08113377866040801, "grad_norm": 3.6359879970550537, "learning_rate": 4.594376241198772e-05, "loss": 4.8144, "step": 8988 }, { "epoch": 0.0811428055605705, "grad_norm": 4.261408805847168, "learning_rate": 4.59433110669796e-05, "loss": 4.2245, "step": 8989 }, { "epoch": 0.08115183246073299, "grad_norm": 4.07118034362793, "learning_rate": 4.594285972197148e-05, "loss": 4.6261, "step": 8990 }, { "epoch": 0.08116085936089547, "grad_norm": 3.0082437992095947, "learning_rate": 4.594240837696335e-05, "loss": 4.3543, "step": 8991 }, { "epoch": 0.08116988626105795, "grad_norm": 3.530611515045166, "learning_rate": 4.594195703195523e-05, "loss": 4.8979, "step": 8992 }, { "epoch": 0.08117891316122043, "grad_norm": 4.094677448272705, "learning_rate": 4.594150568694711e-05, "loss": 5.0172, "step": 8993 }, { "epoch": 0.08118794006138293, "grad_norm": 3.8357667922973633, "learning_rate": 4.594105434193898e-05, "loss": 5.0485, "step": 8994 }, { "epoch": 0.08119696696154541, "grad_norm": 4.137411594390869, "learning_rate": 4.5940602996930854e-05, "loss": 4.7544, "step": 8995 }, { "epoch": 0.08120599386170789, "grad_norm": 3.5549886226654053, "learning_rate": 4.594015165192273e-05, "loss": 4.4114, "step": 8996 }, { "epoch": 0.08121502076187037, "grad_norm": 4.038547515869141, "learning_rate": 4.5939700306914606e-05, "loss": 5.3494, "step": 8997 }, { "epoch": 0.08122404766203285, "grad_norm": 3.2011947631835938, "learning_rate": 4.5939248961906485e-05, "loss": 4.6563, "step": 8998 }, { "epoch": 0.08123307456219535, "grad_norm": 2.659808397293091, "learning_rate": 4.593879761689836e-05, "loss": 4.9316, "step": 8999 }, { "epoch": 0.08124210146235783, "grad_norm": 5.442005157470703, "learning_rate": 4.593834627189024e-05, "loss": 5.3053, "step": 9000 }, { "epoch": 0.08125112836252031, "grad_norm": 3.7910995483398438, "learning_rate": 4.5937894926882116e-05, "loss": 4.6618, "step": 9001 }, { "epoch": 0.08126015526268279, "grad_norm": 3.386833429336548, "learning_rate": 4.593744358187398e-05, "loss": 4.7652, "step": 9002 }, { "epoch": 0.08126918216284527, "grad_norm": 3.389788866043091, "learning_rate": 4.593699223686586e-05, "loss": 4.8174, "step": 9003 }, { "epoch": 0.08127820906300777, "grad_norm": 2.408843994140625, "learning_rate": 4.593654089185774e-05, "loss": 5.206, "step": 9004 }, { "epoch": 0.08128723596317025, "grad_norm": 6.396061420440674, "learning_rate": 4.593608954684961e-05, "loss": 4.3573, "step": 9005 }, { "epoch": 0.08129626286333273, "grad_norm": 2.949424982070923, "learning_rate": 4.593563820184149e-05, "loss": 5.0764, "step": 9006 }, { "epoch": 0.08130528976349521, "grad_norm": 3.42582368850708, "learning_rate": 4.5935186856833364e-05, "loss": 5.226, "step": 9007 }, { "epoch": 0.0813143166636577, "grad_norm": 3.3942887783050537, "learning_rate": 4.593473551182524e-05, "loss": 4.7221, "step": 9008 }, { "epoch": 0.08132334356382018, "grad_norm": 2.45991587638855, "learning_rate": 4.5934284166817116e-05, "loss": 4.3082, "step": 9009 }, { "epoch": 0.08133237046398267, "grad_norm": 3.445539951324463, "learning_rate": 4.5933832821808995e-05, "loss": 4.2617, "step": 9010 }, { "epoch": 0.08134139736414515, "grad_norm": 4.28886079788208, "learning_rate": 4.593338147680087e-05, "loss": 4.4877, "step": 9011 }, { "epoch": 0.08135042426430764, "grad_norm": 3.2105655670166016, "learning_rate": 4.593293013179275e-05, "loss": 4.5556, "step": 9012 }, { "epoch": 0.08135945116447012, "grad_norm": 3.688676357269287, "learning_rate": 4.593247878678462e-05, "loss": 5.2927, "step": 9013 }, { "epoch": 0.0813684780646326, "grad_norm": 3.3554317951202393, "learning_rate": 4.59320274417765e-05, "loss": 4.5622, "step": 9014 }, { "epoch": 0.0813775049647951, "grad_norm": 3.456866502761841, "learning_rate": 4.593157609676837e-05, "loss": 4.7841, "step": 9015 }, { "epoch": 0.08138653186495758, "grad_norm": 3.9532008171081543, "learning_rate": 4.5931124751760244e-05, "loss": 4.4062, "step": 9016 }, { "epoch": 0.08139555876512006, "grad_norm": 3.250485897064209, "learning_rate": 4.593067340675212e-05, "loss": 4.8616, "step": 9017 }, { "epoch": 0.08140458566528254, "grad_norm": 6.2174553871154785, "learning_rate": 4.5930222061744e-05, "loss": 4.6495, "step": 9018 }, { "epoch": 0.08141361256544502, "grad_norm": 2.8337807655334473, "learning_rate": 4.5929770716735875e-05, "loss": 4.8538, "step": 9019 }, { "epoch": 0.08142263946560752, "grad_norm": 2.903256893157959, "learning_rate": 4.5929319371727754e-05, "loss": 4.3435, "step": 9020 }, { "epoch": 0.08143166636577, "grad_norm": 3.7113656997680664, "learning_rate": 4.5928868026719626e-05, "loss": 4.7641, "step": 9021 }, { "epoch": 0.08144069326593248, "grad_norm": 3.118246078491211, "learning_rate": 4.59284166817115e-05, "loss": 4.1248, "step": 9022 }, { "epoch": 0.08144972016609496, "grad_norm": 3.8013646602630615, "learning_rate": 4.592796533670338e-05, "loss": 4.8257, "step": 9023 }, { "epoch": 0.08145874706625744, "grad_norm": 5.5789713859558105, "learning_rate": 4.592751399169525e-05, "loss": 4.5861, "step": 9024 }, { "epoch": 0.08146777396641994, "grad_norm": 3.7142751216888428, "learning_rate": 4.592706264668713e-05, "loss": 4.8173, "step": 9025 }, { "epoch": 0.08147680086658242, "grad_norm": 4.451744556427002, "learning_rate": 4.592661130167901e-05, "loss": 4.8528, "step": 9026 }, { "epoch": 0.0814858277667449, "grad_norm": 3.2182741165161133, "learning_rate": 4.592615995667088e-05, "loss": 4.5082, "step": 9027 }, { "epoch": 0.08149485466690738, "grad_norm": 4.5883636474609375, "learning_rate": 4.592570861166276e-05, "loss": 5.276, "step": 9028 }, { "epoch": 0.08150388156706986, "grad_norm": 2.910778284072876, "learning_rate": 4.592525726665463e-05, "loss": 4.4401, "step": 9029 }, { "epoch": 0.08151290846723236, "grad_norm": 2.80416202545166, "learning_rate": 4.5924805921646506e-05, "loss": 4.773, "step": 9030 }, { "epoch": 0.08152193536739484, "grad_norm": 3.60610032081604, "learning_rate": 4.5924354576638385e-05, "loss": 4.3184, "step": 9031 }, { "epoch": 0.08153096226755732, "grad_norm": 4.678083419799805, "learning_rate": 4.5923903231630264e-05, "loss": 4.6863, "step": 9032 }, { "epoch": 0.0815399891677198, "grad_norm": 4.180092811584473, "learning_rate": 4.5923451886622137e-05, "loss": 4.8922, "step": 9033 }, { "epoch": 0.08154901606788229, "grad_norm": 3.740867853164673, "learning_rate": 4.5923000541614016e-05, "loss": 4.5272, "step": 9034 }, { "epoch": 0.08155804296804477, "grad_norm": 3.5612244606018066, "learning_rate": 4.592254919660589e-05, "loss": 4.6449, "step": 9035 }, { "epoch": 0.08156706986820726, "grad_norm": 2.736252784729004, "learning_rate": 4.592209785159776e-05, "loss": 5.4304, "step": 9036 }, { "epoch": 0.08157609676836974, "grad_norm": 3.1842339038848877, "learning_rate": 4.592164650658964e-05, "loss": 4.6372, "step": 9037 }, { "epoch": 0.08158512366853223, "grad_norm": 3.990708112716675, "learning_rate": 4.592119516158151e-05, "loss": 3.6194, "step": 9038 }, { "epoch": 0.0815941505686947, "grad_norm": 2.810920238494873, "learning_rate": 4.592074381657339e-05, "loss": 4.6229, "step": 9039 }, { "epoch": 0.08160317746885719, "grad_norm": 4.53670597076416, "learning_rate": 4.592029247156527e-05, "loss": 4.9159, "step": 9040 }, { "epoch": 0.08161220436901968, "grad_norm": 3.5782883167266846, "learning_rate": 4.591984112655714e-05, "loss": 4.6342, "step": 9041 }, { "epoch": 0.08162123126918217, "grad_norm": 3.080836057662964, "learning_rate": 4.5919389781549016e-05, "loss": 5.2755, "step": 9042 }, { "epoch": 0.08163025816934465, "grad_norm": 4.298844337463379, "learning_rate": 4.5918938436540895e-05, "loss": 5.1187, "step": 9043 }, { "epoch": 0.08163928506950713, "grad_norm": 3.8643782138824463, "learning_rate": 4.591848709153277e-05, "loss": 4.396, "step": 9044 }, { "epoch": 0.08164831196966961, "grad_norm": 4.208352565765381, "learning_rate": 4.591803574652465e-05, "loss": 4.6876, "step": 9045 }, { "epoch": 0.0816573388698321, "grad_norm": 4.249100208282471, "learning_rate": 4.591758440151652e-05, "loss": 4.5178, "step": 9046 }, { "epoch": 0.08166636576999459, "grad_norm": 3.4843032360076904, "learning_rate": 4.59171330565084e-05, "loss": 5.3743, "step": 9047 }, { "epoch": 0.08167539267015707, "grad_norm": 4.030287742614746, "learning_rate": 4.591668171150028e-05, "loss": 4.4989, "step": 9048 }, { "epoch": 0.08168441957031955, "grad_norm": 3.3176698684692383, "learning_rate": 4.5916230366492143e-05, "loss": 4.5786, "step": 9049 }, { "epoch": 0.08169344647048203, "grad_norm": 3.7316884994506836, "learning_rate": 4.591577902148402e-05, "loss": 4.0855, "step": 9050 }, { "epoch": 0.08170247337064453, "grad_norm": 4.514349460601807, "learning_rate": 4.59153276764759e-05, "loss": 4.7527, "step": 9051 }, { "epoch": 0.08171150027080701, "grad_norm": 3.041505813598633, "learning_rate": 4.5914876331467774e-05, "loss": 4.3277, "step": 9052 }, { "epoch": 0.08172052717096949, "grad_norm": 3.1072282791137695, "learning_rate": 4.5914424986459654e-05, "loss": 4.0535, "step": 9053 }, { "epoch": 0.08172955407113197, "grad_norm": 3.53206467628479, "learning_rate": 4.591397364145153e-05, "loss": 4.0514, "step": 9054 }, { "epoch": 0.08173858097129445, "grad_norm": 4.624290466308594, "learning_rate": 4.59135222964434e-05, "loss": 3.7342, "step": 9055 }, { "epoch": 0.08174760787145695, "grad_norm": 3.475264549255371, "learning_rate": 4.591307095143528e-05, "loss": 4.7602, "step": 9056 }, { "epoch": 0.08175663477161943, "grad_norm": 3.159724473953247, "learning_rate": 4.591261960642716e-05, "loss": 4.9725, "step": 9057 }, { "epoch": 0.08176566167178191, "grad_norm": 4.208782196044922, "learning_rate": 4.591216826141903e-05, "loss": 5.3136, "step": 9058 }, { "epoch": 0.08177468857194439, "grad_norm": 4.365444660186768, "learning_rate": 4.591171691641091e-05, "loss": 4.2156, "step": 9059 }, { "epoch": 0.08178371547210687, "grad_norm": 3.6092405319213867, "learning_rate": 4.591126557140278e-05, "loss": 4.4273, "step": 9060 }, { "epoch": 0.08179274237226936, "grad_norm": 3.556124448776245, "learning_rate": 4.591081422639466e-05, "loss": 4.4513, "step": 9061 }, { "epoch": 0.08180176927243185, "grad_norm": 3.3608384132385254, "learning_rate": 4.591036288138653e-05, "loss": 4.805, "step": 9062 }, { "epoch": 0.08181079617259433, "grad_norm": 3.2672104835510254, "learning_rate": 4.5909911536378405e-05, "loss": 4.6754, "step": 9063 }, { "epoch": 0.08181982307275681, "grad_norm": 3.3534271717071533, "learning_rate": 4.5909460191370285e-05, "loss": 4.1131, "step": 9064 }, { "epoch": 0.0818288499729193, "grad_norm": 3.3287484645843506, "learning_rate": 4.5909008846362164e-05, "loss": 4.8362, "step": 9065 }, { "epoch": 0.08183787687308178, "grad_norm": 3.6722464561462402, "learning_rate": 4.5908557501354036e-05, "loss": 4.5346, "step": 9066 }, { "epoch": 0.08184690377324427, "grad_norm": 4.009917736053467, "learning_rate": 4.5908106156345916e-05, "loss": 4.4664, "step": 9067 }, { "epoch": 0.08185593067340675, "grad_norm": 3.81235408782959, "learning_rate": 4.590765481133779e-05, "loss": 3.7271, "step": 9068 }, { "epoch": 0.08186495757356924, "grad_norm": 3.76406192779541, "learning_rate": 4.590720346632966e-05, "loss": 4.9852, "step": 9069 }, { "epoch": 0.08187398447373172, "grad_norm": 2.8247005939483643, "learning_rate": 4.590675212132154e-05, "loss": 4.7474, "step": 9070 }, { "epoch": 0.0818830113738942, "grad_norm": 7.133955478668213, "learning_rate": 4.590630077631342e-05, "loss": 3.8405, "step": 9071 }, { "epoch": 0.0818920382740567, "grad_norm": 6.401009559631348, "learning_rate": 4.590584943130529e-05, "loss": 4.9758, "step": 9072 }, { "epoch": 0.08190106517421918, "grad_norm": 3.7394022941589355, "learning_rate": 4.590539808629717e-05, "loss": 4.4465, "step": 9073 }, { "epoch": 0.08191009207438166, "grad_norm": 3.0534632205963135, "learning_rate": 4.590494674128904e-05, "loss": 4.3481, "step": 9074 }, { "epoch": 0.08191911897454414, "grad_norm": 4.288756370544434, "learning_rate": 4.590449539628092e-05, "loss": 4.8278, "step": 9075 }, { "epoch": 0.08192814587470662, "grad_norm": 3.7225804328918457, "learning_rate": 4.5904044051272795e-05, "loss": 4.8116, "step": 9076 }, { "epoch": 0.08193717277486912, "grad_norm": 4.4937615394592285, "learning_rate": 4.590359270626467e-05, "loss": 4.8716, "step": 9077 }, { "epoch": 0.0819461996750316, "grad_norm": 2.9356436729431152, "learning_rate": 4.5903141361256547e-05, "loss": 4.5881, "step": 9078 }, { "epoch": 0.08195522657519408, "grad_norm": 4.310064792633057, "learning_rate": 4.5902690016248426e-05, "loss": 4.5522, "step": 9079 }, { "epoch": 0.08196425347535656, "grad_norm": 3.4004299640655518, "learning_rate": 4.59022386712403e-05, "loss": 4.663, "step": 9080 }, { "epoch": 0.08197328037551904, "grad_norm": 3.1075351238250732, "learning_rate": 4.590178732623218e-05, "loss": 4.5991, "step": 9081 }, { "epoch": 0.08198230727568154, "grad_norm": 2.848651885986328, "learning_rate": 4.590133598122405e-05, "loss": 5.2922, "step": 9082 }, { "epoch": 0.08199133417584402, "grad_norm": 4.18644380569458, "learning_rate": 4.590088463621592e-05, "loss": 4.8841, "step": 9083 }, { "epoch": 0.0820003610760065, "grad_norm": 3.4772064685821533, "learning_rate": 4.59004332912078e-05, "loss": 4.1754, "step": 9084 }, { "epoch": 0.08200938797616898, "grad_norm": 3.1366710662841797, "learning_rate": 4.5899981946199674e-05, "loss": 4.4236, "step": 9085 }, { "epoch": 0.08201841487633146, "grad_norm": 4.55519962310791, "learning_rate": 4.589953060119155e-05, "loss": 5.2975, "step": 9086 }, { "epoch": 0.08202744177649395, "grad_norm": 3.773822546005249, "learning_rate": 4.589907925618343e-05, "loss": 4.2702, "step": 9087 }, { "epoch": 0.08203646867665644, "grad_norm": 4.592790603637695, "learning_rate": 4.5898627911175305e-05, "loss": 3.6198, "step": 9088 }, { "epoch": 0.08204549557681892, "grad_norm": 3.900613784790039, "learning_rate": 4.589817656616718e-05, "loss": 4.1249, "step": 9089 }, { "epoch": 0.0820545224769814, "grad_norm": 3.6931474208831787, "learning_rate": 4.589772522115906e-05, "loss": 4.5083, "step": 9090 }, { "epoch": 0.08206354937714389, "grad_norm": 3.3227179050445557, "learning_rate": 4.589727387615093e-05, "loss": 4.7003, "step": 9091 }, { "epoch": 0.08207257627730637, "grad_norm": 2.9920051097869873, "learning_rate": 4.589682253114281e-05, "loss": 3.859, "step": 9092 }, { "epoch": 0.08208160317746886, "grad_norm": 5.420119762420654, "learning_rate": 4.589637118613469e-05, "loss": 4.0098, "step": 9093 }, { "epoch": 0.08209063007763134, "grad_norm": 3.222092866897583, "learning_rate": 4.589591984112656e-05, "loss": 4.3871, "step": 9094 }, { "epoch": 0.08209965697779383, "grad_norm": 3.281264066696167, "learning_rate": 4.589546849611844e-05, "loss": 4.9561, "step": 9095 }, { "epoch": 0.0821086838779563, "grad_norm": 3.234044313430786, "learning_rate": 4.589501715111031e-05, "loss": 4.78, "step": 9096 }, { "epoch": 0.08211771077811879, "grad_norm": 4.679818630218506, "learning_rate": 4.5894565806102184e-05, "loss": 4.7887, "step": 9097 }, { "epoch": 0.08212673767828128, "grad_norm": 2.3843400478363037, "learning_rate": 4.5894114461094064e-05, "loss": 5.1426, "step": 9098 }, { "epoch": 0.08213576457844377, "grad_norm": 3.211642265319824, "learning_rate": 4.5893663116085936e-05, "loss": 4.2533, "step": 9099 }, { "epoch": 0.08214479147860625, "grad_norm": 2.862980842590332, "learning_rate": 4.5893211771077815e-05, "loss": 5.2215, "step": 9100 }, { "epoch": 0.08215381837876873, "grad_norm": 3.440366268157959, "learning_rate": 4.5892760426069695e-05, "loss": 4.704, "step": 9101 }, { "epoch": 0.08216284527893121, "grad_norm": 4.113619327545166, "learning_rate": 4.589230908106156e-05, "loss": 4.267, "step": 9102 }, { "epoch": 0.0821718721790937, "grad_norm": 4.365836143493652, "learning_rate": 4.589185773605344e-05, "loss": 4.7426, "step": 9103 }, { "epoch": 0.08218089907925619, "grad_norm": 5.225337982177734, "learning_rate": 4.589140639104532e-05, "loss": 5.0187, "step": 9104 }, { "epoch": 0.08218992597941867, "grad_norm": 3.060375452041626, "learning_rate": 4.589095504603719e-05, "loss": 4.5567, "step": 9105 }, { "epoch": 0.08219895287958115, "grad_norm": 3.9150500297546387, "learning_rate": 4.589050370102907e-05, "loss": 4.3328, "step": 9106 }, { "epoch": 0.08220797977974363, "grad_norm": 3.019855499267578, "learning_rate": 4.589005235602094e-05, "loss": 4.6548, "step": 9107 }, { "epoch": 0.08221700667990613, "grad_norm": 3.9298064708709717, "learning_rate": 4.588960101101282e-05, "loss": 4.7599, "step": 9108 }, { "epoch": 0.08222603358006861, "grad_norm": 3.3620245456695557, "learning_rate": 4.5889149666004695e-05, "loss": 5.4262, "step": 9109 }, { "epoch": 0.08223506048023109, "grad_norm": 3.701493501663208, "learning_rate": 4.588869832099657e-05, "loss": 5.1601, "step": 9110 }, { "epoch": 0.08224408738039357, "grad_norm": 3.7048628330230713, "learning_rate": 4.5888246975988446e-05, "loss": 4.6797, "step": 9111 }, { "epoch": 0.08225311428055605, "grad_norm": 3.4543936252593994, "learning_rate": 4.5887795630980326e-05, "loss": 4.4585, "step": 9112 }, { "epoch": 0.08226214118071853, "grad_norm": 3.388927698135376, "learning_rate": 4.58873442859722e-05, "loss": 4.5952, "step": 9113 }, { "epoch": 0.08227116808088103, "grad_norm": 3.898043394088745, "learning_rate": 4.588689294096408e-05, "loss": 4.5499, "step": 9114 }, { "epoch": 0.08228019498104351, "grad_norm": 4.001118183135986, "learning_rate": 4.5886441595955956e-05, "loss": 4.5194, "step": 9115 }, { "epoch": 0.08228922188120599, "grad_norm": 3.1468288898468018, "learning_rate": 4.588599025094782e-05, "loss": 5.0548, "step": 9116 }, { "epoch": 0.08229824878136847, "grad_norm": 3.528324842453003, "learning_rate": 4.58855389059397e-05, "loss": 4.161, "step": 9117 }, { "epoch": 0.08230727568153096, "grad_norm": 4.28719425201416, "learning_rate": 4.588508756093158e-05, "loss": 2.9935, "step": 9118 }, { "epoch": 0.08231630258169345, "grad_norm": 3.6971778869628906, "learning_rate": 4.588463621592345e-05, "loss": 4.5627, "step": 9119 }, { "epoch": 0.08232532948185593, "grad_norm": 3.2715206146240234, "learning_rate": 4.588418487091533e-05, "loss": 3.5205, "step": 9120 }, { "epoch": 0.08233435638201841, "grad_norm": 4.973825454711914, "learning_rate": 4.5883733525907205e-05, "loss": 4.0251, "step": 9121 }, { "epoch": 0.0823433832821809, "grad_norm": 4.975940227508545, "learning_rate": 4.588328218089908e-05, "loss": 5.5527, "step": 9122 }, { "epoch": 0.08235241018234338, "grad_norm": 3.3695740699768066, "learning_rate": 4.5882830835890957e-05, "loss": 4.5445, "step": 9123 }, { "epoch": 0.08236143708250587, "grad_norm": 3.2724742889404297, "learning_rate": 4.588237949088283e-05, "loss": 4.8753, "step": 9124 }, { "epoch": 0.08237046398266835, "grad_norm": 3.380176067352295, "learning_rate": 4.588192814587471e-05, "loss": 4.3431, "step": 9125 }, { "epoch": 0.08237949088283084, "grad_norm": 4.539141654968262, "learning_rate": 4.588147680086659e-05, "loss": 4.8759, "step": 9126 }, { "epoch": 0.08238851778299332, "grad_norm": 3.5280628204345703, "learning_rate": 4.588102545585846e-05, "loss": 5.0368, "step": 9127 }, { "epoch": 0.0823975446831558, "grad_norm": 3.8470420837402344, "learning_rate": 4.588057411085034e-05, "loss": 4.1321, "step": 9128 }, { "epoch": 0.0824065715833183, "grad_norm": 3.6146485805511475, "learning_rate": 4.588012276584221e-05, "loss": 4.3192, "step": 9129 }, { "epoch": 0.08241559848348078, "grad_norm": 3.6237173080444336, "learning_rate": 4.5879671420834084e-05, "loss": 5.2966, "step": 9130 }, { "epoch": 0.08242462538364326, "grad_norm": 3.4636471271514893, "learning_rate": 4.587922007582596e-05, "loss": 5.0816, "step": 9131 }, { "epoch": 0.08243365228380574, "grad_norm": 3.957134962081909, "learning_rate": 4.587876873081784e-05, "loss": 4.4544, "step": 9132 }, { "epoch": 0.08244267918396822, "grad_norm": 3.7375845909118652, "learning_rate": 4.5878317385809715e-05, "loss": 4.3935, "step": 9133 }, { "epoch": 0.08245170608413072, "grad_norm": 3.5260119438171387, "learning_rate": 4.5877866040801594e-05, "loss": 4.3817, "step": 9134 }, { "epoch": 0.0824607329842932, "grad_norm": 3.899672031402588, "learning_rate": 4.587741469579347e-05, "loss": 4.8283, "step": 9135 }, { "epoch": 0.08246975988445568, "grad_norm": 2.9408280849456787, "learning_rate": 4.587696335078534e-05, "loss": 3.9036, "step": 9136 }, { "epoch": 0.08247878678461816, "grad_norm": 4.2749810218811035, "learning_rate": 4.587651200577722e-05, "loss": 4.2715, "step": 9137 }, { "epoch": 0.08248781368478064, "grad_norm": 3.6952006816864014, "learning_rate": 4.587606066076909e-05, "loss": 5.0623, "step": 9138 }, { "epoch": 0.08249684058494314, "grad_norm": 4.292487621307373, "learning_rate": 4.587560931576097e-05, "loss": 4.1163, "step": 9139 }, { "epoch": 0.08250586748510562, "grad_norm": 3.710653066635132, "learning_rate": 4.587515797075285e-05, "loss": 4.5677, "step": 9140 }, { "epoch": 0.0825148943852681, "grad_norm": 3.1696438789367676, "learning_rate": 4.587470662574472e-05, "loss": 4.6371, "step": 9141 }, { "epoch": 0.08252392128543058, "grad_norm": 3.7949607372283936, "learning_rate": 4.58742552807366e-05, "loss": 5.0511, "step": 9142 }, { "epoch": 0.08253294818559306, "grad_norm": 5.541371822357178, "learning_rate": 4.5873803935728474e-05, "loss": 4.7021, "step": 9143 }, { "epoch": 0.08254197508575555, "grad_norm": 4.366859436035156, "learning_rate": 4.5873352590720346e-05, "loss": 4.9412, "step": 9144 }, { "epoch": 0.08255100198591804, "grad_norm": 4.016740322113037, "learning_rate": 4.5872901245712225e-05, "loss": 4.6487, "step": 9145 }, { "epoch": 0.08256002888608052, "grad_norm": 2.7716469764709473, "learning_rate": 4.58724499007041e-05, "loss": 4.6155, "step": 9146 }, { "epoch": 0.082569055786243, "grad_norm": 3.3309645652770996, "learning_rate": 4.587199855569598e-05, "loss": 4.9199, "step": 9147 }, { "epoch": 0.08257808268640549, "grad_norm": 3.495168447494507, "learning_rate": 4.5871547210687856e-05, "loss": 4.611, "step": 9148 }, { "epoch": 0.08258710958656797, "grad_norm": 3.269786834716797, "learning_rate": 4.587109586567972e-05, "loss": 4.5989, "step": 9149 }, { "epoch": 0.08259613648673046, "grad_norm": 4.088781356811523, "learning_rate": 4.58706445206716e-05, "loss": 3.9366, "step": 9150 }, { "epoch": 0.08260516338689294, "grad_norm": 3.9129374027252197, "learning_rate": 4.587019317566348e-05, "loss": 4.6984, "step": 9151 }, { "epoch": 0.08261419028705543, "grad_norm": 3.112846851348877, "learning_rate": 4.586974183065535e-05, "loss": 4.7247, "step": 9152 }, { "epoch": 0.0826232171872179, "grad_norm": 4.134344100952148, "learning_rate": 4.586929048564723e-05, "loss": 4.3677, "step": 9153 }, { "epoch": 0.08263224408738039, "grad_norm": 4.417530536651611, "learning_rate": 4.586883914063911e-05, "loss": 4.3139, "step": 9154 }, { "epoch": 0.08264127098754288, "grad_norm": 3.6657023429870605, "learning_rate": 4.5868387795630984e-05, "loss": 4.598, "step": 9155 }, { "epoch": 0.08265029788770537, "grad_norm": 3.4072582721710205, "learning_rate": 4.5867936450622856e-05, "loss": 4.4386, "step": 9156 }, { "epoch": 0.08265932478786785, "grad_norm": 4.228630542755127, "learning_rate": 4.5867485105614736e-05, "loss": 4.8929, "step": 9157 }, { "epoch": 0.08266835168803033, "grad_norm": 2.981283187866211, "learning_rate": 4.586703376060661e-05, "loss": 4.3322, "step": 9158 }, { "epoch": 0.08267737858819281, "grad_norm": 3.2891883850097656, "learning_rate": 4.586658241559849e-05, "loss": 4.4634, "step": 9159 }, { "epoch": 0.0826864054883553, "grad_norm": 2.788315773010254, "learning_rate": 4.586613107059036e-05, "loss": 4.4898, "step": 9160 }, { "epoch": 0.08269543238851779, "grad_norm": 4.464210510253906, "learning_rate": 4.586567972558224e-05, "loss": 3.5721, "step": 9161 }, { "epoch": 0.08270445928868027, "grad_norm": 3.8098201751708984, "learning_rate": 4.586522838057412e-05, "loss": 4.8827, "step": 9162 }, { "epoch": 0.08271348618884275, "grad_norm": 3.1489336490631104, "learning_rate": 4.5864777035565984e-05, "loss": 4.8026, "step": 9163 }, { "epoch": 0.08272251308900523, "grad_norm": 3.531602144241333, "learning_rate": 4.586432569055786e-05, "loss": 4.3018, "step": 9164 }, { "epoch": 0.08273153998916773, "grad_norm": 3.6807098388671875, "learning_rate": 4.586387434554974e-05, "loss": 4.2529, "step": 9165 }, { "epoch": 0.08274056688933021, "grad_norm": 2.937702178955078, "learning_rate": 4.5863423000541615e-05, "loss": 5.4927, "step": 9166 }, { "epoch": 0.08274959378949269, "grad_norm": 4.3624186515808105, "learning_rate": 4.5862971655533494e-05, "loss": 4.2606, "step": 9167 }, { "epoch": 0.08275862068965517, "grad_norm": 3.817913770675659, "learning_rate": 4.5862520310525367e-05, "loss": 4.3527, "step": 9168 }, { "epoch": 0.08276764758981765, "grad_norm": 2.777540683746338, "learning_rate": 4.586206896551724e-05, "loss": 5.1893, "step": 9169 }, { "epoch": 0.08277667448998013, "grad_norm": 4.561122417449951, "learning_rate": 4.586161762050912e-05, "loss": 4.3224, "step": 9170 }, { "epoch": 0.08278570139014263, "grad_norm": 2.778815984725952, "learning_rate": 4.5861166275501e-05, "loss": 5.0769, "step": 9171 }, { "epoch": 0.08279472829030511, "grad_norm": 3.3249149322509766, "learning_rate": 4.586071493049287e-05, "loss": 4.3039, "step": 9172 }, { "epoch": 0.08280375519046759, "grad_norm": 3.6469497680664062, "learning_rate": 4.586026358548475e-05, "loss": 4.8433, "step": 9173 }, { "epoch": 0.08281278209063007, "grad_norm": 3.1099278926849365, "learning_rate": 4.585981224047662e-05, "loss": 5.2727, "step": 9174 }, { "epoch": 0.08282180899079256, "grad_norm": 2.876260280609131, "learning_rate": 4.58593608954685e-05, "loss": 4.6532, "step": 9175 }, { "epoch": 0.08283083589095505, "grad_norm": 3.186142683029175, "learning_rate": 4.585890955046037e-05, "loss": 5.1464, "step": 9176 }, { "epoch": 0.08283986279111753, "grad_norm": 4.91285514831543, "learning_rate": 4.5858458205452246e-05, "loss": 5.1949, "step": 9177 }, { "epoch": 0.08284888969128001, "grad_norm": 3.545297384262085, "learning_rate": 4.5858006860444125e-05, "loss": 4.3494, "step": 9178 }, { "epoch": 0.0828579165914425, "grad_norm": 3.774939775466919, "learning_rate": 4.5857555515436004e-05, "loss": 4.487, "step": 9179 }, { "epoch": 0.08286694349160498, "grad_norm": 2.8199501037597656, "learning_rate": 4.585710417042788e-05, "loss": 4.9534, "step": 9180 }, { "epoch": 0.08287597039176747, "grad_norm": 3.8198721408843994, "learning_rate": 4.5856652825419756e-05, "loss": 4.9551, "step": 9181 }, { "epoch": 0.08288499729192995, "grad_norm": 3.649829626083374, "learning_rate": 4.585620148041163e-05, "loss": 4.9392, "step": 9182 }, { "epoch": 0.08289402419209244, "grad_norm": 4.267587661743164, "learning_rate": 4.58557501354035e-05, "loss": 4.9417, "step": 9183 }, { "epoch": 0.08290305109225492, "grad_norm": 3.365527391433716, "learning_rate": 4.585529879039538e-05, "loss": 4.3562, "step": 9184 }, { "epoch": 0.0829120779924174, "grad_norm": 3.3607444763183594, "learning_rate": 4.585484744538725e-05, "loss": 4.3413, "step": 9185 }, { "epoch": 0.0829211048925799, "grad_norm": 4.121659278869629, "learning_rate": 4.585439610037913e-05, "loss": 5.298, "step": 9186 }, { "epoch": 0.08293013179274238, "grad_norm": 2.526752471923828, "learning_rate": 4.585394475537101e-05, "loss": 4.4681, "step": 9187 }, { "epoch": 0.08293915869290486, "grad_norm": 2.8688204288482666, "learning_rate": 4.5853493410362884e-05, "loss": 4.9043, "step": 9188 }, { "epoch": 0.08294818559306734, "grad_norm": 2.831266403198242, "learning_rate": 4.585304206535476e-05, "loss": 4.8147, "step": 9189 }, { "epoch": 0.08295721249322982, "grad_norm": 3.1273140907287598, "learning_rate": 4.5852590720346635e-05, "loss": 4.6421, "step": 9190 }, { "epoch": 0.08296623939339232, "grad_norm": 3.509871482849121, "learning_rate": 4.585213937533851e-05, "loss": 4.3723, "step": 9191 }, { "epoch": 0.0829752662935548, "grad_norm": 2.63781476020813, "learning_rate": 4.585168803033039e-05, "loss": 4.4655, "step": 9192 }, { "epoch": 0.08298429319371728, "grad_norm": 3.415276050567627, "learning_rate": 4.5851236685322266e-05, "loss": 4.3774, "step": 9193 }, { "epoch": 0.08299332009387976, "grad_norm": 3.5542562007904053, "learning_rate": 4.585078534031414e-05, "loss": 4.4355, "step": 9194 }, { "epoch": 0.08300234699404224, "grad_norm": 5.231680870056152, "learning_rate": 4.585033399530602e-05, "loss": 5.2266, "step": 9195 }, { "epoch": 0.08301137389420472, "grad_norm": 3.085850238800049, "learning_rate": 4.584988265029789e-05, "loss": 4.3758, "step": 9196 }, { "epoch": 0.08302040079436722, "grad_norm": 3.3776965141296387, "learning_rate": 4.584943130528976e-05, "loss": 4.7464, "step": 9197 }, { "epoch": 0.0830294276945297, "grad_norm": 4.337138652801514, "learning_rate": 4.584897996028164e-05, "loss": 3.2711, "step": 9198 }, { "epoch": 0.08303845459469218, "grad_norm": 3.8106675148010254, "learning_rate": 4.5848528615273515e-05, "loss": 3.7259, "step": 9199 }, { "epoch": 0.08304748149485466, "grad_norm": 2.952002763748169, "learning_rate": 4.5848077270265394e-05, "loss": 4.7374, "step": 9200 }, { "epoch": 0.08305650839501715, "grad_norm": 3.3380355834960938, "learning_rate": 4.584762592525727e-05, "loss": 4.3426, "step": 9201 }, { "epoch": 0.08306553529517964, "grad_norm": 4.3744659423828125, "learning_rate": 4.5847174580249146e-05, "loss": 3.8181, "step": 9202 }, { "epoch": 0.08307456219534212, "grad_norm": 3.956031560897827, "learning_rate": 4.584672323524102e-05, "loss": 5.2044, "step": 9203 }, { "epoch": 0.0830835890955046, "grad_norm": 3.221597194671631, "learning_rate": 4.58462718902329e-05, "loss": 5.1099, "step": 9204 }, { "epoch": 0.08309261599566709, "grad_norm": 3.8081769943237305, "learning_rate": 4.584582054522477e-05, "loss": 5.2618, "step": 9205 }, { "epoch": 0.08310164289582957, "grad_norm": 3.3810312747955322, "learning_rate": 4.584536920021665e-05, "loss": 4.5799, "step": 9206 }, { "epoch": 0.08311066979599206, "grad_norm": 3.092808723449707, "learning_rate": 4.584491785520852e-05, "loss": 5.0594, "step": 9207 }, { "epoch": 0.08311969669615454, "grad_norm": 3.0942940711975098, "learning_rate": 4.58444665102004e-05, "loss": 4.8023, "step": 9208 }, { "epoch": 0.08312872359631703, "grad_norm": 3.16839337348938, "learning_rate": 4.584401516519228e-05, "loss": 4.8863, "step": 9209 }, { "epoch": 0.0831377504964795, "grad_norm": 3.614607572555542, "learning_rate": 4.5843563820184146e-05, "loss": 4.3624, "step": 9210 }, { "epoch": 0.08314677739664199, "grad_norm": 3.6638643741607666, "learning_rate": 4.5843112475176025e-05, "loss": 4.0845, "step": 9211 }, { "epoch": 0.08315580429680448, "grad_norm": 3.49680757522583, "learning_rate": 4.5842661130167904e-05, "loss": 4.7487, "step": 9212 }, { "epoch": 0.08316483119696697, "grad_norm": 3.6114609241485596, "learning_rate": 4.5842209785159777e-05, "loss": 4.525, "step": 9213 }, { "epoch": 0.08317385809712945, "grad_norm": 3.261068105697632, "learning_rate": 4.5841758440151656e-05, "loss": 4.7145, "step": 9214 }, { "epoch": 0.08318288499729193, "grad_norm": 3.4896349906921387, "learning_rate": 4.5841307095143535e-05, "loss": 4.6966, "step": 9215 }, { "epoch": 0.08319191189745441, "grad_norm": 3.69079327583313, "learning_rate": 4.58408557501354e-05, "loss": 4.6316, "step": 9216 }, { "epoch": 0.0832009387976169, "grad_norm": 3.1513054370880127, "learning_rate": 4.584040440512728e-05, "loss": 5.0252, "step": 9217 }, { "epoch": 0.08320996569777939, "grad_norm": 3.8536674976348877, "learning_rate": 4.583995306011916e-05, "loss": 4.9, "step": 9218 }, { "epoch": 0.08321899259794187, "grad_norm": 4.1235198974609375, "learning_rate": 4.583950171511103e-05, "loss": 5.0122, "step": 9219 }, { "epoch": 0.08322801949810435, "grad_norm": 3.2659459114074707, "learning_rate": 4.583905037010291e-05, "loss": 4.304, "step": 9220 }, { "epoch": 0.08323704639826683, "grad_norm": 4.62279748916626, "learning_rate": 4.583859902509478e-05, "loss": 3.4967, "step": 9221 }, { "epoch": 0.08324607329842931, "grad_norm": 3.4735209941864014, "learning_rate": 4.583814768008666e-05, "loss": 4.5007, "step": 9222 }, { "epoch": 0.08325510019859181, "grad_norm": 5.1178789138793945, "learning_rate": 4.5837696335078535e-05, "loss": 4.1805, "step": 9223 }, { "epoch": 0.08326412709875429, "grad_norm": 3.6665842533111572, "learning_rate": 4.583724499007041e-05, "loss": 4.0416, "step": 9224 }, { "epoch": 0.08327315399891677, "grad_norm": 3.8543026447296143, "learning_rate": 4.583679364506229e-05, "loss": 5.0466, "step": 9225 }, { "epoch": 0.08328218089907925, "grad_norm": 2.8175556659698486, "learning_rate": 4.5836342300054166e-05, "loss": 4.9273, "step": 9226 }, { "epoch": 0.08329120779924173, "grad_norm": 3.566596269607544, "learning_rate": 4.583589095504604e-05, "loss": 4.4608, "step": 9227 }, { "epoch": 0.08330023469940423, "grad_norm": 3.561802864074707, "learning_rate": 4.583543961003792e-05, "loss": 4.3107, "step": 9228 }, { "epoch": 0.08330926159956671, "grad_norm": 2.8850622177124023, "learning_rate": 4.583498826502979e-05, "loss": 5.1674, "step": 9229 }, { "epoch": 0.08331828849972919, "grad_norm": 4.294450759887695, "learning_rate": 4.583453692002166e-05, "loss": 5.0876, "step": 9230 }, { "epoch": 0.08332731539989167, "grad_norm": 3.874929666519165, "learning_rate": 4.583408557501354e-05, "loss": 4.551, "step": 9231 }, { "epoch": 0.08333634230005416, "grad_norm": 3.818455934524536, "learning_rate": 4.583363423000542e-05, "loss": 4.196, "step": 9232 }, { "epoch": 0.08334536920021665, "grad_norm": 2.825103282928467, "learning_rate": 4.5833182884997294e-05, "loss": 5.1079, "step": 9233 }, { "epoch": 0.08335439610037913, "grad_norm": 3.985304594039917, "learning_rate": 4.583273153998917e-05, "loss": 4.0276, "step": 9234 }, { "epoch": 0.08336342300054161, "grad_norm": 5.070952892303467, "learning_rate": 4.5832280194981045e-05, "loss": 3.6895, "step": 9235 }, { "epoch": 0.0833724499007041, "grad_norm": 4.4231977462768555, "learning_rate": 4.5831828849972925e-05, "loss": 3.6475, "step": 9236 }, { "epoch": 0.08338147680086658, "grad_norm": 2.6529879570007324, "learning_rate": 4.58313775049648e-05, "loss": 5.2474, "step": 9237 }, { "epoch": 0.08339050370102907, "grad_norm": 3.4409842491149902, "learning_rate": 4.583092615995667e-05, "loss": 5.0523, "step": 9238 }, { "epoch": 0.08339953060119155, "grad_norm": 3.6092844009399414, "learning_rate": 4.583047481494855e-05, "loss": 4.9268, "step": 9239 }, { "epoch": 0.08340855750135404, "grad_norm": 3.7952775955200195, "learning_rate": 4.583002346994043e-05, "loss": 4.9726, "step": 9240 }, { "epoch": 0.08341758440151652, "grad_norm": 2.930284023284912, "learning_rate": 4.58295721249323e-05, "loss": 4.4107, "step": 9241 }, { "epoch": 0.083426611301679, "grad_norm": 3.818483591079712, "learning_rate": 4.582912077992418e-05, "loss": 4.6652, "step": 9242 }, { "epoch": 0.0834356382018415, "grad_norm": 4.073602199554443, "learning_rate": 4.582866943491605e-05, "loss": 4.8732, "step": 9243 }, { "epoch": 0.08344466510200398, "grad_norm": 4.057222366333008, "learning_rate": 4.5828218089907925e-05, "loss": 4.2854, "step": 9244 }, { "epoch": 0.08345369200216646, "grad_norm": 3.288313865661621, "learning_rate": 4.5827766744899804e-05, "loss": 4.8048, "step": 9245 }, { "epoch": 0.08346271890232894, "grad_norm": 3.0310041904449463, "learning_rate": 4.5827315399891676e-05, "loss": 5.0675, "step": 9246 }, { "epoch": 0.08347174580249142, "grad_norm": 3.7593579292297363, "learning_rate": 4.5826864054883556e-05, "loss": 5.0984, "step": 9247 }, { "epoch": 0.0834807727026539, "grad_norm": 2.889160394668579, "learning_rate": 4.5826412709875435e-05, "loss": 5.1144, "step": 9248 }, { "epoch": 0.0834897996028164, "grad_norm": 3.056196689605713, "learning_rate": 4.582596136486731e-05, "loss": 4.2483, "step": 9249 }, { "epoch": 0.08349882650297888, "grad_norm": 3.6963369846343994, "learning_rate": 4.582551001985918e-05, "loss": 4.7241, "step": 9250 }, { "epoch": 0.08350785340314136, "grad_norm": 3.168457269668579, "learning_rate": 4.582505867485106e-05, "loss": 4.6153, "step": 9251 }, { "epoch": 0.08351688030330384, "grad_norm": 3.850355386734009, "learning_rate": 4.582460732984293e-05, "loss": 4.6392, "step": 9252 }, { "epoch": 0.08352590720346632, "grad_norm": 3.9715397357940674, "learning_rate": 4.582415598483481e-05, "loss": 3.8873, "step": 9253 }, { "epoch": 0.08353493410362882, "grad_norm": 4.035181999206543, "learning_rate": 4.582370463982669e-05, "loss": 4.5996, "step": 9254 }, { "epoch": 0.0835439610037913, "grad_norm": 2.968900680541992, "learning_rate": 4.582325329481856e-05, "loss": 5.0958, "step": 9255 }, { "epoch": 0.08355298790395378, "grad_norm": 4.1490559577941895, "learning_rate": 4.582280194981044e-05, "loss": 5.0351, "step": 9256 }, { "epoch": 0.08356201480411626, "grad_norm": 3.3893051147460938, "learning_rate": 4.5822350604802314e-05, "loss": 4.267, "step": 9257 }, { "epoch": 0.08357104170427875, "grad_norm": 3.6895058155059814, "learning_rate": 4.5821899259794187e-05, "loss": 3.7588, "step": 9258 }, { "epoch": 0.08358006860444124, "grad_norm": 3.3085083961486816, "learning_rate": 4.5821447914786066e-05, "loss": 4.2681, "step": 9259 }, { "epoch": 0.08358909550460372, "grad_norm": 3.720427989959717, "learning_rate": 4.582099656977794e-05, "loss": 4.6392, "step": 9260 }, { "epoch": 0.0835981224047662, "grad_norm": 3.5219502449035645, "learning_rate": 4.582054522476982e-05, "loss": 4.8766, "step": 9261 }, { "epoch": 0.08360714930492869, "grad_norm": 3.032241106033325, "learning_rate": 4.58200938797617e-05, "loss": 5.2215, "step": 9262 }, { "epoch": 0.08361617620509117, "grad_norm": 3.1896989345550537, "learning_rate": 4.581964253475356e-05, "loss": 4.7698, "step": 9263 }, { "epoch": 0.08362520310525366, "grad_norm": 4.417269706726074, "learning_rate": 4.581919118974544e-05, "loss": 5.0078, "step": 9264 }, { "epoch": 0.08363423000541614, "grad_norm": 3.056361436843872, "learning_rate": 4.581873984473732e-05, "loss": 5.0141, "step": 9265 }, { "epoch": 0.08364325690557863, "grad_norm": 2.880917549133301, "learning_rate": 4.581828849972919e-05, "loss": 5.0656, "step": 9266 }, { "epoch": 0.0836522838057411, "grad_norm": 3.1562530994415283, "learning_rate": 4.581783715472107e-05, "loss": 4.792, "step": 9267 }, { "epoch": 0.08366131070590359, "grad_norm": 3.627586841583252, "learning_rate": 4.5817385809712945e-05, "loss": 4.144, "step": 9268 }, { "epoch": 0.08367033760606608, "grad_norm": 3.6146934032440186, "learning_rate": 4.5816934464704824e-05, "loss": 5.3355, "step": 9269 }, { "epoch": 0.08367936450622857, "grad_norm": 9.520204544067383, "learning_rate": 4.58164831196967e-05, "loss": 5.1876, "step": 9270 }, { "epoch": 0.08368839140639105, "grad_norm": 3.3866219520568848, "learning_rate": 4.581603177468857e-05, "loss": 4.9335, "step": 9271 }, { "epoch": 0.08369741830655353, "grad_norm": 3.414428949356079, "learning_rate": 4.581558042968045e-05, "loss": 5.0321, "step": 9272 }, { "epoch": 0.08370644520671601, "grad_norm": 2.3487801551818848, "learning_rate": 4.581512908467233e-05, "loss": 4.4599, "step": 9273 }, { "epoch": 0.08371547210687849, "grad_norm": 3.2987940311431885, "learning_rate": 4.58146777396642e-05, "loss": 4.2962, "step": 9274 }, { "epoch": 0.08372449900704099, "grad_norm": 3.6204020977020264, "learning_rate": 4.581422639465608e-05, "loss": 4.8015, "step": 9275 }, { "epoch": 0.08373352590720347, "grad_norm": 2.971592664718628, "learning_rate": 4.581377504964796e-05, "loss": 4.683, "step": 9276 }, { "epoch": 0.08374255280736595, "grad_norm": 2.243483066558838, "learning_rate": 4.5813323704639824e-05, "loss": 4.7615, "step": 9277 }, { "epoch": 0.08375157970752843, "grad_norm": 3.800424814224243, "learning_rate": 4.5812872359631704e-05, "loss": 4.5753, "step": 9278 }, { "epoch": 0.08376060660769091, "grad_norm": 3.0930464267730713, "learning_rate": 4.581242101462358e-05, "loss": 5.0428, "step": 9279 }, { "epoch": 0.08376963350785341, "grad_norm": 5.8234968185424805, "learning_rate": 4.5811969669615455e-05, "loss": 5.3578, "step": 9280 }, { "epoch": 0.08377866040801589, "grad_norm": 4.167551040649414, "learning_rate": 4.5811518324607335e-05, "loss": 4.4868, "step": 9281 }, { "epoch": 0.08378768730817837, "grad_norm": 2.6515393257141113, "learning_rate": 4.581106697959921e-05, "loss": 5.2262, "step": 9282 }, { "epoch": 0.08379671420834085, "grad_norm": 3.935392379760742, "learning_rate": 4.581061563459108e-05, "loss": 5.2186, "step": 9283 }, { "epoch": 0.08380574110850333, "grad_norm": 3.456479549407959, "learning_rate": 4.581016428958296e-05, "loss": 4.7131, "step": 9284 }, { "epoch": 0.08381476800866583, "grad_norm": 3.938880681991577, "learning_rate": 4.580971294457483e-05, "loss": 4.9642, "step": 9285 }, { "epoch": 0.08382379490882831, "grad_norm": 2.5692906379699707, "learning_rate": 4.580926159956671e-05, "loss": 4.9476, "step": 9286 }, { "epoch": 0.08383282180899079, "grad_norm": 3.5588550567626953, "learning_rate": 4.580881025455859e-05, "loss": 4.6678, "step": 9287 }, { "epoch": 0.08384184870915327, "grad_norm": 3.6393818855285645, "learning_rate": 4.580835890955046e-05, "loss": 4.357, "step": 9288 }, { "epoch": 0.08385087560931576, "grad_norm": 2.609683036804199, "learning_rate": 4.580790756454234e-05, "loss": 5.5098, "step": 9289 }, { "epoch": 0.08385990250947825, "grad_norm": 2.999518394470215, "learning_rate": 4.5807456219534214e-05, "loss": 4.5615, "step": 9290 }, { "epoch": 0.08386892940964073, "grad_norm": 3.2937448024749756, "learning_rate": 4.5807004874526086e-05, "loss": 4.0238, "step": 9291 }, { "epoch": 0.08387795630980321, "grad_norm": 3.3266239166259766, "learning_rate": 4.5806553529517966e-05, "loss": 4.4211, "step": 9292 }, { "epoch": 0.0838869832099657, "grad_norm": 3.3369321823120117, "learning_rate": 4.5806102184509845e-05, "loss": 4.49, "step": 9293 }, { "epoch": 0.08389601011012818, "grad_norm": 4.874824523925781, "learning_rate": 4.580565083950172e-05, "loss": 2.8252, "step": 9294 }, { "epoch": 0.08390503701029067, "grad_norm": 3.898911237716675, "learning_rate": 4.5805199494493596e-05, "loss": 4.412, "step": 9295 }, { "epoch": 0.08391406391045315, "grad_norm": 3.288479804992676, "learning_rate": 4.580474814948547e-05, "loss": 4.4684, "step": 9296 }, { "epoch": 0.08392309081061564, "grad_norm": 3.017930030822754, "learning_rate": 4.580429680447734e-05, "loss": 4.5417, "step": 9297 }, { "epoch": 0.08393211771077812, "grad_norm": 4.118541717529297, "learning_rate": 4.580384545946922e-05, "loss": 3.9746, "step": 9298 }, { "epoch": 0.0839411446109406, "grad_norm": 4.7967143058776855, "learning_rate": 4.580339411446109e-05, "loss": 4.8599, "step": 9299 }, { "epoch": 0.08395017151110308, "grad_norm": 3.603492021560669, "learning_rate": 4.580294276945297e-05, "loss": 4.3989, "step": 9300 }, { "epoch": 0.08395919841126558, "grad_norm": 3.5866858959198, "learning_rate": 4.580249142444485e-05, "loss": 5.5336, "step": 9301 }, { "epoch": 0.08396822531142806, "grad_norm": 3.1190030574798584, "learning_rate": 4.5802040079436724e-05, "loss": 4.5063, "step": 9302 }, { "epoch": 0.08397725221159054, "grad_norm": 7.638747215270996, "learning_rate": 4.58015887344286e-05, "loss": 5.1614, "step": 9303 }, { "epoch": 0.08398627911175302, "grad_norm": 4.032633304595947, "learning_rate": 4.5801137389420476e-05, "loss": 4.9821, "step": 9304 }, { "epoch": 0.0839953060119155, "grad_norm": 4.1487932205200195, "learning_rate": 4.580068604441235e-05, "loss": 4.1232, "step": 9305 }, { "epoch": 0.084004332912078, "grad_norm": 2.3674535751342773, "learning_rate": 4.580023469940423e-05, "loss": 4.7934, "step": 9306 }, { "epoch": 0.08401335981224048, "grad_norm": 3.5683562755584717, "learning_rate": 4.57997833543961e-05, "loss": 5.1435, "step": 9307 }, { "epoch": 0.08402238671240296, "grad_norm": 2.9905543327331543, "learning_rate": 4.579933200938798e-05, "loss": 4.9901, "step": 9308 }, { "epoch": 0.08403141361256544, "grad_norm": 3.3186779022216797, "learning_rate": 4.579888066437986e-05, "loss": 4.4939, "step": 9309 }, { "epoch": 0.08404044051272792, "grad_norm": 3.328392744064331, "learning_rate": 4.5798429319371724e-05, "loss": 4.9246, "step": 9310 }, { "epoch": 0.08404946741289042, "grad_norm": 3.49273419380188, "learning_rate": 4.57979779743636e-05, "loss": 4.9437, "step": 9311 }, { "epoch": 0.0840584943130529, "grad_norm": 3.7313997745513916, "learning_rate": 4.579752662935548e-05, "loss": 5.2117, "step": 9312 }, { "epoch": 0.08406752121321538, "grad_norm": 3.92545747756958, "learning_rate": 4.5797075284347355e-05, "loss": 4.8135, "step": 9313 }, { "epoch": 0.08407654811337786, "grad_norm": 3.406682014465332, "learning_rate": 4.5796623939339234e-05, "loss": 4.6963, "step": 9314 }, { "epoch": 0.08408557501354035, "grad_norm": 3.760483980178833, "learning_rate": 4.5796172594331114e-05, "loss": 4.4923, "step": 9315 }, { "epoch": 0.08409460191370284, "grad_norm": 2.7872912883758545, "learning_rate": 4.5795721249322986e-05, "loss": 5.1688, "step": 9316 }, { "epoch": 0.08410362881386532, "grad_norm": 3.101973533630371, "learning_rate": 4.579526990431486e-05, "loss": 4.8197, "step": 9317 }, { "epoch": 0.0841126557140278, "grad_norm": 3.434544563293457, "learning_rate": 4.579481855930674e-05, "loss": 3.7883, "step": 9318 }, { "epoch": 0.08412168261419029, "grad_norm": 2.971223831176758, "learning_rate": 4.579436721429861e-05, "loss": 5.4021, "step": 9319 }, { "epoch": 0.08413070951435277, "grad_norm": 4.341183662414551, "learning_rate": 4.579391586929049e-05, "loss": 4.5474, "step": 9320 }, { "epoch": 0.08413973641451526, "grad_norm": 3.7410504817962646, "learning_rate": 4.579346452428236e-05, "loss": 4.1801, "step": 9321 }, { "epoch": 0.08414876331467774, "grad_norm": 4.128664016723633, "learning_rate": 4.579301317927424e-05, "loss": 4.2429, "step": 9322 }, { "epoch": 0.08415779021484023, "grad_norm": 2.990220308303833, "learning_rate": 4.579256183426612e-05, "loss": 4.9654, "step": 9323 }, { "epoch": 0.0841668171150027, "grad_norm": 4.9899420738220215, "learning_rate": 4.5792110489257986e-05, "loss": 5.3446, "step": 9324 }, { "epoch": 0.08417584401516519, "grad_norm": 3.7718565464019775, "learning_rate": 4.5791659144249865e-05, "loss": 5.0153, "step": 9325 }, { "epoch": 0.08418487091532767, "grad_norm": 3.9486215114593506, "learning_rate": 4.5791207799241745e-05, "loss": 4.0317, "step": 9326 }, { "epoch": 0.08419389781549017, "grad_norm": 2.6281685829162598, "learning_rate": 4.579075645423362e-05, "loss": 4.8522, "step": 9327 }, { "epoch": 0.08420292471565265, "grad_norm": 2.9933018684387207, "learning_rate": 4.5790305109225496e-05, "loss": 5.1436, "step": 9328 }, { "epoch": 0.08421195161581513, "grad_norm": 5.859213352203369, "learning_rate": 4.578985376421737e-05, "loss": 4.9417, "step": 9329 }, { "epoch": 0.08422097851597761, "grad_norm": 3.212935447692871, "learning_rate": 4.578940241920924e-05, "loss": 4.65, "step": 9330 }, { "epoch": 0.08423000541614009, "grad_norm": 3.238377809524536, "learning_rate": 4.578895107420112e-05, "loss": 4.6934, "step": 9331 }, { "epoch": 0.08423903231630259, "grad_norm": 3.6328353881835938, "learning_rate": 4.578849972919299e-05, "loss": 3.3061, "step": 9332 }, { "epoch": 0.08424805921646507, "grad_norm": 2.4422767162323, "learning_rate": 4.578804838418487e-05, "loss": 4.5702, "step": 9333 }, { "epoch": 0.08425708611662755, "grad_norm": 3.0815646648406982, "learning_rate": 4.578759703917675e-05, "loss": 4.8888, "step": 9334 }, { "epoch": 0.08426611301679003, "grad_norm": 3.9165806770324707, "learning_rate": 4.5787145694168624e-05, "loss": 4.2519, "step": 9335 }, { "epoch": 0.08427513991695251, "grad_norm": 3.288194417953491, "learning_rate": 4.57866943491605e-05, "loss": 4.6327, "step": 9336 }, { "epoch": 0.08428416681711501, "grad_norm": 3.902405023574829, "learning_rate": 4.5786243004152376e-05, "loss": 4.8192, "step": 9337 }, { "epoch": 0.08429319371727749, "grad_norm": 3.517892837524414, "learning_rate": 4.578579165914425e-05, "loss": 4.4414, "step": 9338 }, { "epoch": 0.08430222061743997, "grad_norm": 3.5480763912200928, "learning_rate": 4.578534031413613e-05, "loss": 4.7908, "step": 9339 }, { "epoch": 0.08431124751760245, "grad_norm": 3.832246780395508, "learning_rate": 4.5784888969128006e-05, "loss": 4.4615, "step": 9340 }, { "epoch": 0.08432027441776493, "grad_norm": 3.4326746463775635, "learning_rate": 4.578443762411988e-05, "loss": 4.321, "step": 9341 }, { "epoch": 0.08432930131792743, "grad_norm": 2.8122060298919678, "learning_rate": 4.578398627911176e-05, "loss": 4.3335, "step": 9342 }, { "epoch": 0.08433832821808991, "grad_norm": 3.3867499828338623, "learning_rate": 4.578353493410363e-05, "loss": 4.6966, "step": 9343 }, { "epoch": 0.08434735511825239, "grad_norm": 5.802635669708252, "learning_rate": 4.57830835890955e-05, "loss": 4.802, "step": 9344 }, { "epoch": 0.08435638201841487, "grad_norm": 3.229015827178955, "learning_rate": 4.578263224408738e-05, "loss": 5.4944, "step": 9345 }, { "epoch": 0.08436540891857736, "grad_norm": 3.4060518741607666, "learning_rate": 4.5782180899079255e-05, "loss": 5.2397, "step": 9346 }, { "epoch": 0.08437443581873985, "grad_norm": 4.750657081604004, "learning_rate": 4.5781729554071134e-05, "loss": 3.955, "step": 9347 }, { "epoch": 0.08438346271890233, "grad_norm": 2.9514384269714355, "learning_rate": 4.578127820906301e-05, "loss": 4.4768, "step": 9348 }, { "epoch": 0.08439248961906481, "grad_norm": 3.236419677734375, "learning_rate": 4.5780826864054886e-05, "loss": 4.5157, "step": 9349 }, { "epoch": 0.0844015165192273, "grad_norm": 3.687864303588867, "learning_rate": 4.5780375519046765e-05, "loss": 4.4089, "step": 9350 }, { "epoch": 0.08441054341938978, "grad_norm": 2.3502564430236816, "learning_rate": 4.577992417403864e-05, "loss": 4.5783, "step": 9351 }, { "epoch": 0.08441957031955226, "grad_norm": 3.609393835067749, "learning_rate": 4.577947282903051e-05, "loss": 4.7023, "step": 9352 }, { "epoch": 0.08442859721971475, "grad_norm": 3.678269863128662, "learning_rate": 4.577902148402239e-05, "loss": 4.3369, "step": 9353 }, { "epoch": 0.08443762411987724, "grad_norm": 2.8481738567352295, "learning_rate": 4.577857013901427e-05, "loss": 4.7594, "step": 9354 }, { "epoch": 0.08444665102003972, "grad_norm": 3.454249858856201, "learning_rate": 4.577811879400614e-05, "loss": 4.9424, "step": 9355 }, { "epoch": 0.0844556779202022, "grad_norm": 3.502769947052002, "learning_rate": 4.577766744899802e-05, "loss": 4.3731, "step": 9356 }, { "epoch": 0.08446470482036468, "grad_norm": 3.7692995071411133, "learning_rate": 4.577721610398989e-05, "loss": 4.7965, "step": 9357 }, { "epoch": 0.08447373172052718, "grad_norm": 3.3615782260894775, "learning_rate": 4.5776764758981765e-05, "loss": 4.8917, "step": 9358 }, { "epoch": 0.08448275862068966, "grad_norm": 3.591984987258911, "learning_rate": 4.5776313413973644e-05, "loss": 4.2388, "step": 9359 }, { "epoch": 0.08449178552085214, "grad_norm": 3.272751569747925, "learning_rate": 4.577586206896552e-05, "loss": 4.8351, "step": 9360 }, { "epoch": 0.08450081242101462, "grad_norm": 2.5503461360931396, "learning_rate": 4.5775410723957396e-05, "loss": 5.0153, "step": 9361 }, { "epoch": 0.0845098393211771, "grad_norm": 3.2788374423980713, "learning_rate": 4.5774959378949275e-05, "loss": 5.1125, "step": 9362 }, { "epoch": 0.0845188662213396, "grad_norm": 2.785367965698242, "learning_rate": 4.577450803394115e-05, "loss": 4.6898, "step": 9363 }, { "epoch": 0.08452789312150208, "grad_norm": 5.3138556480407715, "learning_rate": 4.577405668893302e-05, "loss": 4.9479, "step": 9364 }, { "epoch": 0.08453692002166456, "grad_norm": 3.734105110168457, "learning_rate": 4.57736053439249e-05, "loss": 4.976, "step": 9365 }, { "epoch": 0.08454594692182704, "grad_norm": 3.4471662044525146, "learning_rate": 4.577315399891677e-05, "loss": 4.5652, "step": 9366 }, { "epoch": 0.08455497382198952, "grad_norm": 3.3419301509857178, "learning_rate": 4.577270265390865e-05, "loss": 4.7914, "step": 9367 }, { "epoch": 0.08456400072215202, "grad_norm": 2.605656623840332, "learning_rate": 4.5772251308900524e-05, "loss": 5.1709, "step": 9368 }, { "epoch": 0.0845730276223145, "grad_norm": 5.124144554138184, "learning_rate": 4.57717999638924e-05, "loss": 4.6967, "step": 9369 }, { "epoch": 0.08458205452247698, "grad_norm": 2.9280669689178467, "learning_rate": 4.577134861888428e-05, "loss": 4.6877, "step": 9370 }, { "epoch": 0.08459108142263946, "grad_norm": 3.1069324016571045, "learning_rate": 4.577089727387615e-05, "loss": 4.3451, "step": 9371 }, { "epoch": 0.08460010832280195, "grad_norm": 3.8564696311950684, "learning_rate": 4.577044592886803e-05, "loss": 3.7958, "step": 9372 }, { "epoch": 0.08460913522296444, "grad_norm": 3.2139337062835693, "learning_rate": 4.5769994583859906e-05, "loss": 4.2547, "step": 9373 }, { "epoch": 0.08461816212312692, "grad_norm": 5.273467540740967, "learning_rate": 4.576954323885178e-05, "loss": 4.4264, "step": 9374 }, { "epoch": 0.0846271890232894, "grad_norm": 3.6240696907043457, "learning_rate": 4.576909189384366e-05, "loss": 4.4798, "step": 9375 }, { "epoch": 0.08463621592345189, "grad_norm": 3.572307825088501, "learning_rate": 4.576864054883554e-05, "loss": 4.8483, "step": 9376 }, { "epoch": 0.08464524282361437, "grad_norm": 3.0886106491088867, "learning_rate": 4.57681892038274e-05, "loss": 4.5928, "step": 9377 }, { "epoch": 0.08465426972377685, "grad_norm": 4.619015693664551, "learning_rate": 4.576773785881928e-05, "loss": 3.5717, "step": 9378 }, { "epoch": 0.08466329662393934, "grad_norm": 3.7960870265960693, "learning_rate": 4.576728651381116e-05, "loss": 4.6015, "step": 9379 }, { "epoch": 0.08467232352410183, "grad_norm": 2.478015661239624, "learning_rate": 4.5766835168803034e-05, "loss": 4.6583, "step": 9380 }, { "epoch": 0.0846813504242643, "grad_norm": 4.938941955566406, "learning_rate": 4.576638382379491e-05, "loss": 3.9173, "step": 9381 }, { "epoch": 0.08469037732442679, "grad_norm": 2.9952588081359863, "learning_rate": 4.5765932478786785e-05, "loss": 5.2302, "step": 9382 }, { "epoch": 0.08469940422458927, "grad_norm": 3.3946611881256104, "learning_rate": 4.5765481133778665e-05, "loss": 4.4577, "step": 9383 }, { "epoch": 0.08470843112475177, "grad_norm": 4.862112998962402, "learning_rate": 4.576502978877054e-05, "loss": 4.2316, "step": 9384 }, { "epoch": 0.08471745802491425, "grad_norm": 3.4901325702667236, "learning_rate": 4.576457844376241e-05, "loss": 4.6182, "step": 9385 }, { "epoch": 0.08472648492507673, "grad_norm": 3.375711441040039, "learning_rate": 4.576412709875429e-05, "loss": 4.6518, "step": 9386 }, { "epoch": 0.08473551182523921, "grad_norm": 3.8566598892211914, "learning_rate": 4.576367575374617e-05, "loss": 4.4905, "step": 9387 }, { "epoch": 0.08474453872540169, "grad_norm": 4.889719486236572, "learning_rate": 4.576322440873804e-05, "loss": 4.7493, "step": 9388 }, { "epoch": 0.08475356562556419, "grad_norm": 3.4338347911834717, "learning_rate": 4.576277306372992e-05, "loss": 4.37, "step": 9389 }, { "epoch": 0.08476259252572667, "grad_norm": 2.4910852909088135, "learning_rate": 4.576232171872179e-05, "loss": 4.7283, "step": 9390 }, { "epoch": 0.08477161942588915, "grad_norm": 3.73921799659729, "learning_rate": 4.5761870373713665e-05, "loss": 4.421, "step": 9391 }, { "epoch": 0.08478064632605163, "grad_norm": 3.120450258255005, "learning_rate": 4.5761419028705544e-05, "loss": 4.4583, "step": 9392 }, { "epoch": 0.08478967322621411, "grad_norm": 3.9524528980255127, "learning_rate": 4.576096768369742e-05, "loss": 4.4781, "step": 9393 }, { "epoch": 0.08479870012637661, "grad_norm": 2.7729122638702393, "learning_rate": 4.5760516338689296e-05, "loss": 5.0961, "step": 9394 }, { "epoch": 0.08480772702653909, "grad_norm": 3.206784963607788, "learning_rate": 4.5760064993681175e-05, "loss": 4.8912, "step": 9395 }, { "epoch": 0.08481675392670157, "grad_norm": 3.144868850708008, "learning_rate": 4.575961364867305e-05, "loss": 5.1633, "step": 9396 }, { "epoch": 0.08482578082686405, "grad_norm": 4.259343147277832, "learning_rate": 4.575916230366493e-05, "loss": 4.2777, "step": 9397 }, { "epoch": 0.08483480772702653, "grad_norm": 4.074955940246582, "learning_rate": 4.57587109586568e-05, "loss": 4.9528, "step": 9398 }, { "epoch": 0.08484383462718903, "grad_norm": 3.409667730331421, "learning_rate": 4.575825961364867e-05, "loss": 5.1334, "step": 9399 }, { "epoch": 0.08485286152735151, "grad_norm": 3.8863043785095215, "learning_rate": 4.575780826864055e-05, "loss": 4.9834, "step": 9400 }, { "epoch": 0.08486188842751399, "grad_norm": 3.6028060913085938, "learning_rate": 4.575735692363243e-05, "loss": 4.9388, "step": 9401 }, { "epoch": 0.08487091532767647, "grad_norm": 4.400666236877441, "learning_rate": 4.57569055786243e-05, "loss": 3.4075, "step": 9402 }, { "epoch": 0.08487994222783896, "grad_norm": 3.585475206375122, "learning_rate": 4.575645423361618e-05, "loss": 4.5735, "step": 9403 }, { "epoch": 0.08488896912800144, "grad_norm": 3.8412272930145264, "learning_rate": 4.5756002888608054e-05, "loss": 4.3666, "step": 9404 }, { "epoch": 0.08489799602816393, "grad_norm": 3.5350897312164307, "learning_rate": 4.575555154359993e-05, "loss": 4.9813, "step": 9405 }, { "epoch": 0.08490702292832641, "grad_norm": 5.968494892120361, "learning_rate": 4.5755100198591806e-05, "loss": 3.9118, "step": 9406 }, { "epoch": 0.0849160498284889, "grad_norm": 3.2768290042877197, "learning_rate": 4.575464885358368e-05, "loss": 4.5804, "step": 9407 }, { "epoch": 0.08492507672865138, "grad_norm": 3.1065967082977295, "learning_rate": 4.575419750857556e-05, "loss": 4.6107, "step": 9408 }, { "epoch": 0.08493410362881386, "grad_norm": 4.3112030029296875, "learning_rate": 4.575374616356744e-05, "loss": 4.513, "step": 9409 }, { "epoch": 0.08494313052897635, "grad_norm": 4.02972936630249, "learning_rate": 4.575329481855931e-05, "loss": 4.3378, "step": 9410 }, { "epoch": 0.08495215742913884, "grad_norm": 3.0315091609954834, "learning_rate": 4.575284347355118e-05, "loss": 5.1022, "step": 9411 }, { "epoch": 0.08496118432930132, "grad_norm": 2.8324198722839355, "learning_rate": 4.575239212854306e-05, "loss": 4.7589, "step": 9412 }, { "epoch": 0.0849702112294638, "grad_norm": 3.0263938903808594, "learning_rate": 4.5751940783534934e-05, "loss": 4.3114, "step": 9413 }, { "epoch": 0.08497923812962628, "grad_norm": 3.07946515083313, "learning_rate": 4.575148943852681e-05, "loss": 4.5494, "step": 9414 }, { "epoch": 0.08498826502978878, "grad_norm": 3.6562037467956543, "learning_rate": 4.575103809351869e-05, "loss": 4.5992, "step": 9415 }, { "epoch": 0.08499729192995126, "grad_norm": 3.5779263973236084, "learning_rate": 4.5750586748510564e-05, "loss": 4.8002, "step": 9416 }, { "epoch": 0.08500631883011374, "grad_norm": 4.950777053833008, "learning_rate": 4.5750135403502444e-05, "loss": 4.6312, "step": 9417 }, { "epoch": 0.08501534573027622, "grad_norm": 3.8890280723571777, "learning_rate": 4.5749684058494316e-05, "loss": 5.0692, "step": 9418 }, { "epoch": 0.0850243726304387, "grad_norm": 3.147627353668213, "learning_rate": 4.574923271348619e-05, "loss": 4.1417, "step": 9419 }, { "epoch": 0.0850333995306012, "grad_norm": 3.9826035499572754, "learning_rate": 4.574878136847807e-05, "loss": 4.5814, "step": 9420 }, { "epoch": 0.08504242643076368, "grad_norm": 4.25468111038208, "learning_rate": 4.574833002346994e-05, "loss": 4.4072, "step": 9421 }, { "epoch": 0.08505145333092616, "grad_norm": 2.9247825145721436, "learning_rate": 4.574787867846182e-05, "loss": 4.2918, "step": 9422 }, { "epoch": 0.08506048023108864, "grad_norm": 4.001223087310791, "learning_rate": 4.57474273334537e-05, "loss": 4.4639, "step": 9423 }, { "epoch": 0.08506950713125112, "grad_norm": 3.207496404647827, "learning_rate": 4.5746975988445565e-05, "loss": 4.5262, "step": 9424 }, { "epoch": 0.08507853403141362, "grad_norm": 3.9299166202545166, "learning_rate": 4.5746524643437444e-05, "loss": 4.3818, "step": 9425 }, { "epoch": 0.0850875609315761, "grad_norm": 3.9430630207061768, "learning_rate": 4.574607329842932e-05, "loss": 5.7604, "step": 9426 }, { "epoch": 0.08509658783173858, "grad_norm": 3.0778119564056396, "learning_rate": 4.5745621953421195e-05, "loss": 5.0103, "step": 9427 }, { "epoch": 0.08510561473190106, "grad_norm": 3.2158384323120117, "learning_rate": 4.5745170608413075e-05, "loss": 4.592, "step": 9428 }, { "epoch": 0.08511464163206355, "grad_norm": 3.5082170963287354, "learning_rate": 4.574471926340495e-05, "loss": 5.145, "step": 9429 }, { "epoch": 0.08512366853222603, "grad_norm": 3.8127176761627197, "learning_rate": 4.5744267918396826e-05, "loss": 4.7677, "step": 9430 }, { "epoch": 0.08513269543238852, "grad_norm": 3.646320343017578, "learning_rate": 4.57438165733887e-05, "loss": 4.2844, "step": 9431 }, { "epoch": 0.085141722332551, "grad_norm": 3.911464214324951, "learning_rate": 4.574336522838057e-05, "loss": 3.1268, "step": 9432 }, { "epoch": 0.08515074923271349, "grad_norm": 3.567678451538086, "learning_rate": 4.574291388337245e-05, "loss": 4.403, "step": 9433 }, { "epoch": 0.08515977613287597, "grad_norm": 3.6886329650878906, "learning_rate": 4.574246253836433e-05, "loss": 4.6309, "step": 9434 }, { "epoch": 0.08516880303303845, "grad_norm": 4.181800842285156, "learning_rate": 4.57420111933562e-05, "loss": 3.5249, "step": 9435 }, { "epoch": 0.08517782993320094, "grad_norm": 3.040599822998047, "learning_rate": 4.574155984834808e-05, "loss": 4.6986, "step": 9436 }, { "epoch": 0.08518685683336343, "grad_norm": 6.904181003570557, "learning_rate": 4.574110850333996e-05, "loss": 3.6751, "step": 9437 }, { "epoch": 0.0851958837335259, "grad_norm": 3.2409396171569824, "learning_rate": 4.5740657158331826e-05, "loss": 4.7598, "step": 9438 }, { "epoch": 0.08520491063368839, "grad_norm": 3.5376687049865723, "learning_rate": 4.5740205813323706e-05, "loss": 4.5158, "step": 9439 }, { "epoch": 0.08521393753385087, "grad_norm": 3.8879785537719727, "learning_rate": 4.5739754468315585e-05, "loss": 4.4639, "step": 9440 }, { "epoch": 0.08522296443401337, "grad_norm": 3.718017578125, "learning_rate": 4.573930312330746e-05, "loss": 4.4703, "step": 9441 }, { "epoch": 0.08523199133417585, "grad_norm": 3.163128137588501, "learning_rate": 4.573885177829934e-05, "loss": 4.0583, "step": 9442 }, { "epoch": 0.08524101823433833, "grad_norm": 3.745180606842041, "learning_rate": 4.573840043329121e-05, "loss": 4.8338, "step": 9443 }, { "epoch": 0.08525004513450081, "grad_norm": 2.8410580158233643, "learning_rate": 4.573794908828309e-05, "loss": 5.0119, "step": 9444 }, { "epoch": 0.08525907203466329, "grad_norm": 2.8190324306488037, "learning_rate": 4.573749774327496e-05, "loss": 5.021, "step": 9445 }, { "epoch": 0.08526809893482579, "grad_norm": 2.522160768508911, "learning_rate": 4.573704639826683e-05, "loss": 4.7512, "step": 9446 }, { "epoch": 0.08527712583498827, "grad_norm": 2.7061870098114014, "learning_rate": 4.573659505325871e-05, "loss": 5.0937, "step": 9447 }, { "epoch": 0.08528615273515075, "grad_norm": 2.8435816764831543, "learning_rate": 4.573614370825059e-05, "loss": 4.7423, "step": 9448 }, { "epoch": 0.08529517963531323, "grad_norm": 2.6878297328948975, "learning_rate": 4.5735692363242464e-05, "loss": 4.713, "step": 9449 }, { "epoch": 0.08530420653547571, "grad_norm": 3.409170150756836, "learning_rate": 4.5735241018234343e-05, "loss": 4.3614, "step": 9450 }, { "epoch": 0.08531323343563821, "grad_norm": 2.994452714920044, "learning_rate": 4.5734789673226216e-05, "loss": 4.3569, "step": 9451 }, { "epoch": 0.08532226033580069, "grad_norm": 4.070236682891846, "learning_rate": 4.573433832821809e-05, "loss": 4.9008, "step": 9452 }, { "epoch": 0.08533128723596317, "grad_norm": 4.242562294006348, "learning_rate": 4.573388698320997e-05, "loss": 4.91, "step": 9453 }, { "epoch": 0.08534031413612565, "grad_norm": 3.680236339569092, "learning_rate": 4.573343563820185e-05, "loss": 4.804, "step": 9454 }, { "epoch": 0.08534934103628813, "grad_norm": 3.884479284286499, "learning_rate": 4.573298429319372e-05, "loss": 5.1362, "step": 9455 }, { "epoch": 0.08535836793645062, "grad_norm": 3.6631455421447754, "learning_rate": 4.57325329481856e-05, "loss": 4.3289, "step": 9456 }, { "epoch": 0.08536739483661311, "grad_norm": 3.150360107421875, "learning_rate": 4.573208160317747e-05, "loss": 4.4944, "step": 9457 }, { "epoch": 0.08537642173677559, "grad_norm": 3.720628023147583, "learning_rate": 4.5731630258169344e-05, "loss": 5.2614, "step": 9458 }, { "epoch": 0.08538544863693807, "grad_norm": 3.1202855110168457, "learning_rate": 4.573117891316122e-05, "loss": 4.6038, "step": 9459 }, { "epoch": 0.08539447553710056, "grad_norm": 3.119624376296997, "learning_rate": 4.5730727568153095e-05, "loss": 5.0838, "step": 9460 }, { "epoch": 0.08540350243726304, "grad_norm": 3.5221989154815674, "learning_rate": 4.5730276223144974e-05, "loss": 5.1631, "step": 9461 }, { "epoch": 0.08541252933742553, "grad_norm": 3.9406888484954834, "learning_rate": 4.5729824878136854e-05, "loss": 5.3038, "step": 9462 }, { "epoch": 0.08542155623758801, "grad_norm": 2.5981688499450684, "learning_rate": 4.5729373533128726e-05, "loss": 4.8798, "step": 9463 }, { "epoch": 0.0854305831377505, "grad_norm": 3.152198553085327, "learning_rate": 4.5728922188120605e-05, "loss": 5.1304, "step": 9464 }, { "epoch": 0.08543961003791298, "grad_norm": 3.387530565261841, "learning_rate": 4.572847084311248e-05, "loss": 4.9458, "step": 9465 }, { "epoch": 0.08544863693807546, "grad_norm": 3.555844783782959, "learning_rate": 4.572801949810435e-05, "loss": 4.1248, "step": 9466 }, { "epoch": 0.08545766383823795, "grad_norm": 3.3634841442108154, "learning_rate": 4.572756815309623e-05, "loss": 4.2492, "step": 9467 }, { "epoch": 0.08546669073840044, "grad_norm": 4.239567279815674, "learning_rate": 4.57271168080881e-05, "loss": 3.7136, "step": 9468 }, { "epoch": 0.08547571763856292, "grad_norm": 2.844858169555664, "learning_rate": 4.572666546307998e-05, "loss": 4.4338, "step": 9469 }, { "epoch": 0.0854847445387254, "grad_norm": 3.3200581073760986, "learning_rate": 4.572621411807186e-05, "loss": 4.6061, "step": 9470 }, { "epoch": 0.08549377143888788, "grad_norm": 3.6618239879608154, "learning_rate": 4.5725762773063726e-05, "loss": 5.0789, "step": 9471 }, { "epoch": 0.08550279833905038, "grad_norm": 5.865142345428467, "learning_rate": 4.5725311428055605e-05, "loss": 4.6587, "step": 9472 }, { "epoch": 0.08551182523921286, "grad_norm": 3.1795785427093506, "learning_rate": 4.5724860083047485e-05, "loss": 5.0032, "step": 9473 }, { "epoch": 0.08552085213937534, "grad_norm": 2.981419801712036, "learning_rate": 4.572440873803936e-05, "loss": 4.3577, "step": 9474 }, { "epoch": 0.08552987903953782, "grad_norm": 4.083171367645264, "learning_rate": 4.5723957393031236e-05, "loss": 4.1594, "step": 9475 }, { "epoch": 0.0855389059397003, "grad_norm": 3.6961021423339844, "learning_rate": 4.5723506048023116e-05, "loss": 4.5515, "step": 9476 }, { "epoch": 0.0855479328398628, "grad_norm": 4.335766315460205, "learning_rate": 4.572305470301499e-05, "loss": 3.5786, "step": 9477 }, { "epoch": 0.08555695974002528, "grad_norm": 2.7359158992767334, "learning_rate": 4.572260335800686e-05, "loss": 4.8854, "step": 9478 }, { "epoch": 0.08556598664018776, "grad_norm": 3.400550365447998, "learning_rate": 4.572215201299874e-05, "loss": 4.5307, "step": 9479 }, { "epoch": 0.08557501354035024, "grad_norm": 4.4149065017700195, "learning_rate": 4.572170066799061e-05, "loss": 4.0351, "step": 9480 }, { "epoch": 0.08558404044051272, "grad_norm": 3.0172808170318604, "learning_rate": 4.572124932298249e-05, "loss": 5.2849, "step": 9481 }, { "epoch": 0.08559306734067522, "grad_norm": 4.396260738372803, "learning_rate": 4.5720797977974364e-05, "loss": 4.6386, "step": 9482 }, { "epoch": 0.0856020942408377, "grad_norm": 3.7592804431915283, "learning_rate": 4.572034663296624e-05, "loss": 4.3886, "step": 9483 }, { "epoch": 0.08561112114100018, "grad_norm": 3.81111478805542, "learning_rate": 4.571989528795812e-05, "loss": 4.2357, "step": 9484 }, { "epoch": 0.08562014804116266, "grad_norm": 3.8399760723114014, "learning_rate": 4.571944394294999e-05, "loss": 4.8748, "step": 9485 }, { "epoch": 0.08562917494132515, "grad_norm": 4.219907760620117, "learning_rate": 4.571899259794187e-05, "loss": 4.9459, "step": 9486 }, { "epoch": 0.08563820184148763, "grad_norm": 3.181124210357666, "learning_rate": 4.571854125293375e-05, "loss": 3.9745, "step": 9487 }, { "epoch": 0.08564722874165012, "grad_norm": 4.3018388748168945, "learning_rate": 4.571808990792562e-05, "loss": 4.5466, "step": 9488 }, { "epoch": 0.0856562556418126, "grad_norm": 3.3556582927703857, "learning_rate": 4.57176385629175e-05, "loss": 4.8681, "step": 9489 }, { "epoch": 0.08566528254197509, "grad_norm": 2.9718377590179443, "learning_rate": 4.571718721790937e-05, "loss": 5.082, "step": 9490 }, { "epoch": 0.08567430944213757, "grad_norm": 3.804682970046997, "learning_rate": 4.571673587290124e-05, "loss": 4.7299, "step": 9491 }, { "epoch": 0.08568333634230005, "grad_norm": 4.192939758300781, "learning_rate": 4.571628452789312e-05, "loss": 3.6736, "step": 9492 }, { "epoch": 0.08569236324246254, "grad_norm": 3.0716426372528076, "learning_rate": 4.5715833182884995e-05, "loss": 4.2316, "step": 9493 }, { "epoch": 0.08570139014262503, "grad_norm": 4.212525367736816, "learning_rate": 4.5715381837876874e-05, "loss": 4.8856, "step": 9494 }, { "epoch": 0.0857104170427875, "grad_norm": 3.392481565475464, "learning_rate": 4.5714930492868753e-05, "loss": 4.8846, "step": 9495 }, { "epoch": 0.08571944394294999, "grad_norm": 3.4898128509521484, "learning_rate": 4.5714479147860626e-05, "loss": 4.5328, "step": 9496 }, { "epoch": 0.08572847084311247, "grad_norm": 3.372821092605591, "learning_rate": 4.5714027802852505e-05, "loss": 4.7228, "step": 9497 }, { "epoch": 0.08573749774327497, "grad_norm": 4.027344226837158, "learning_rate": 4.571357645784438e-05, "loss": 4.4312, "step": 9498 }, { "epoch": 0.08574652464343745, "grad_norm": 3.4873769283294678, "learning_rate": 4.571312511283625e-05, "loss": 4.5602, "step": 9499 }, { "epoch": 0.08575555154359993, "grad_norm": 3.9206860065460205, "learning_rate": 4.571267376782813e-05, "loss": 5.3282, "step": 9500 }, { "epoch": 0.08576457844376241, "grad_norm": 3.1304054260253906, "learning_rate": 4.571222242282001e-05, "loss": 4.8275, "step": 9501 }, { "epoch": 0.08577360534392489, "grad_norm": 3.5888872146606445, "learning_rate": 4.571177107781188e-05, "loss": 4.7417, "step": 9502 }, { "epoch": 0.08578263224408739, "grad_norm": 2.7459588050842285, "learning_rate": 4.571131973280376e-05, "loss": 4.8579, "step": 9503 }, { "epoch": 0.08579165914424987, "grad_norm": 3.2341809272766113, "learning_rate": 4.571086838779563e-05, "loss": 4.6113, "step": 9504 }, { "epoch": 0.08580068604441235, "grad_norm": 3.4679925441741943, "learning_rate": 4.5710417042787505e-05, "loss": 5.407, "step": 9505 }, { "epoch": 0.08580971294457483, "grad_norm": 4.461577415466309, "learning_rate": 4.5709965697779384e-05, "loss": 3.7644, "step": 9506 }, { "epoch": 0.08581873984473731, "grad_norm": 2.824235200881958, "learning_rate": 4.570951435277126e-05, "loss": 4.7841, "step": 9507 }, { "epoch": 0.08582776674489981, "grad_norm": 3.0342721939086914, "learning_rate": 4.5709063007763136e-05, "loss": 4.9813, "step": 9508 }, { "epoch": 0.08583679364506229, "grad_norm": 2.8392207622528076, "learning_rate": 4.5708611662755015e-05, "loss": 4.4777, "step": 9509 }, { "epoch": 0.08584582054522477, "grad_norm": 3.3627192974090576, "learning_rate": 4.570816031774689e-05, "loss": 4.3351, "step": 9510 }, { "epoch": 0.08585484744538725, "grad_norm": 3.984687566757202, "learning_rate": 4.570770897273877e-05, "loss": 3.9536, "step": 9511 }, { "epoch": 0.08586387434554973, "grad_norm": 3.290271759033203, "learning_rate": 4.570725762773064e-05, "loss": 4.9554, "step": 9512 }, { "epoch": 0.08587290124571222, "grad_norm": 3.6987788677215576, "learning_rate": 4.570680628272251e-05, "loss": 4.2582, "step": 9513 }, { "epoch": 0.08588192814587471, "grad_norm": 4.673306465148926, "learning_rate": 4.570635493771439e-05, "loss": 5.0403, "step": 9514 }, { "epoch": 0.08589095504603719, "grad_norm": 4.750516891479492, "learning_rate": 4.570590359270627e-05, "loss": 3.4023, "step": 9515 }, { "epoch": 0.08589998194619967, "grad_norm": 3.342235803604126, "learning_rate": 4.570545224769814e-05, "loss": 4.2297, "step": 9516 }, { "epoch": 0.08590900884636216, "grad_norm": 3.5672366619110107, "learning_rate": 4.570500090269002e-05, "loss": 3.7268, "step": 9517 }, { "epoch": 0.08591803574652464, "grad_norm": 4.9233880043029785, "learning_rate": 4.5704549557681895e-05, "loss": 4.362, "step": 9518 }, { "epoch": 0.08592706264668713, "grad_norm": 4.133143901824951, "learning_rate": 4.570409821267377e-05, "loss": 4.6841, "step": 9519 }, { "epoch": 0.08593608954684961, "grad_norm": 4.7006330490112305, "learning_rate": 4.5703646867665646e-05, "loss": 4.3397, "step": 9520 }, { "epoch": 0.0859451164470121, "grad_norm": 3.264369010925293, "learning_rate": 4.570319552265752e-05, "loss": 4.2183, "step": 9521 }, { "epoch": 0.08595414334717458, "grad_norm": 3.873702049255371, "learning_rate": 4.57027441776494e-05, "loss": 4.9285, "step": 9522 }, { "epoch": 0.08596317024733706, "grad_norm": 3.5122323036193848, "learning_rate": 4.570229283264128e-05, "loss": 4.9077, "step": 9523 }, { "epoch": 0.08597219714749955, "grad_norm": 3.3227698802948, "learning_rate": 4.570184148763315e-05, "loss": 3.9466, "step": 9524 }, { "epoch": 0.08598122404766204, "grad_norm": 4.227372646331787, "learning_rate": 4.570139014262502e-05, "loss": 5.0478, "step": 9525 }, { "epoch": 0.08599025094782452, "grad_norm": 3.586838722229004, "learning_rate": 4.57009387976169e-05, "loss": 3.9381, "step": 9526 }, { "epoch": 0.085999277847987, "grad_norm": 3.3876430988311768, "learning_rate": 4.5700487452608774e-05, "loss": 5.0629, "step": 9527 }, { "epoch": 0.08600830474814948, "grad_norm": 4.514501094818115, "learning_rate": 4.570003610760065e-05, "loss": 4.964, "step": 9528 }, { "epoch": 0.08601733164831198, "grad_norm": 2.9152486324310303, "learning_rate": 4.5699584762592526e-05, "loss": 4.7029, "step": 9529 }, { "epoch": 0.08602635854847446, "grad_norm": 2.8658196926116943, "learning_rate": 4.5699133417584405e-05, "loss": 5.2612, "step": 9530 }, { "epoch": 0.08603538544863694, "grad_norm": 3.265010356903076, "learning_rate": 4.5698682072576284e-05, "loss": 4.3151, "step": 9531 }, { "epoch": 0.08604441234879942, "grad_norm": 3.7046236991882324, "learning_rate": 4.569823072756815e-05, "loss": 3.9525, "step": 9532 }, { "epoch": 0.0860534392489619, "grad_norm": 3.7860987186431885, "learning_rate": 4.569777938256003e-05, "loss": 4.7028, "step": 9533 }, { "epoch": 0.0860624661491244, "grad_norm": 3.9119441509246826, "learning_rate": 4.569732803755191e-05, "loss": 4.3931, "step": 9534 }, { "epoch": 0.08607149304928688, "grad_norm": 3.448882579803467, "learning_rate": 4.569687669254378e-05, "loss": 3.6621, "step": 9535 }, { "epoch": 0.08608051994944936, "grad_norm": 3.596201181411743, "learning_rate": 4.569642534753566e-05, "loss": 4.2545, "step": 9536 }, { "epoch": 0.08608954684961184, "grad_norm": 3.1549994945526123, "learning_rate": 4.569597400252754e-05, "loss": 4.4309, "step": 9537 }, { "epoch": 0.08609857374977432, "grad_norm": 3.8490686416625977, "learning_rate": 4.5695522657519405e-05, "loss": 4.441, "step": 9538 }, { "epoch": 0.0861076006499368, "grad_norm": 3.523972749710083, "learning_rate": 4.5695071312511284e-05, "loss": 4.2404, "step": 9539 }, { "epoch": 0.0861166275500993, "grad_norm": 3.032294750213623, "learning_rate": 4.5694619967503163e-05, "loss": 4.4759, "step": 9540 }, { "epoch": 0.08612565445026178, "grad_norm": 3.299574613571167, "learning_rate": 4.5694168622495036e-05, "loss": 4.6234, "step": 9541 }, { "epoch": 0.08613468135042426, "grad_norm": 4.360349655151367, "learning_rate": 4.5693717277486915e-05, "loss": 3.8184, "step": 9542 }, { "epoch": 0.08614370825058675, "grad_norm": 3.2636661529541016, "learning_rate": 4.569326593247879e-05, "loss": 4.3832, "step": 9543 }, { "epoch": 0.08615273515074923, "grad_norm": 3.108332872390747, "learning_rate": 4.569281458747067e-05, "loss": 4.9107, "step": 9544 }, { "epoch": 0.08616176205091172, "grad_norm": 4.000767230987549, "learning_rate": 4.569236324246254e-05, "loss": 4.2149, "step": 9545 }, { "epoch": 0.0861707889510742, "grad_norm": 3.327544689178467, "learning_rate": 4.569191189745441e-05, "loss": 4.7534, "step": 9546 }, { "epoch": 0.08617981585123669, "grad_norm": 3.7327215671539307, "learning_rate": 4.569146055244629e-05, "loss": 3.7344, "step": 9547 }, { "epoch": 0.08618884275139917, "grad_norm": 3.5178303718566895, "learning_rate": 4.569100920743817e-05, "loss": 4.5735, "step": 9548 }, { "epoch": 0.08619786965156165, "grad_norm": 6.272181987762451, "learning_rate": 4.569055786243004e-05, "loss": 4.4011, "step": 9549 }, { "epoch": 0.08620689655172414, "grad_norm": 2.5658509731292725, "learning_rate": 4.569010651742192e-05, "loss": 5.3789, "step": 9550 }, { "epoch": 0.08621592345188663, "grad_norm": 3.7304434776306152, "learning_rate": 4.5689655172413794e-05, "loss": 4.5233, "step": 9551 }, { "epoch": 0.0862249503520491, "grad_norm": 4.167622089385986, "learning_rate": 4.568920382740567e-05, "loss": 4.8373, "step": 9552 }, { "epoch": 0.08623397725221159, "grad_norm": 3.9880762100219727, "learning_rate": 4.5688752482397546e-05, "loss": 5.0802, "step": 9553 }, { "epoch": 0.08624300415237407, "grad_norm": 4.659292697906494, "learning_rate": 4.5688301137389425e-05, "loss": 4.4345, "step": 9554 }, { "epoch": 0.08625203105253657, "grad_norm": 2.9697091579437256, "learning_rate": 4.56878497923813e-05, "loss": 4.5949, "step": 9555 }, { "epoch": 0.08626105795269905, "grad_norm": 3.161565065383911, "learning_rate": 4.568739844737318e-05, "loss": 4.7888, "step": 9556 }, { "epoch": 0.08627008485286153, "grad_norm": 3.334900140762329, "learning_rate": 4.568694710236505e-05, "loss": 5.1182, "step": 9557 }, { "epoch": 0.08627911175302401, "grad_norm": 3.2705304622650146, "learning_rate": 4.568649575735693e-05, "loss": 4.8783, "step": 9558 }, { "epoch": 0.08628813865318649, "grad_norm": 4.320333957672119, "learning_rate": 4.56860444123488e-05, "loss": 4.1099, "step": 9559 }, { "epoch": 0.08629716555334899, "grad_norm": 2.8887996673583984, "learning_rate": 4.5685593067340674e-05, "loss": 5.1242, "step": 9560 }, { "epoch": 0.08630619245351147, "grad_norm": 3.43121600151062, "learning_rate": 4.568514172233255e-05, "loss": 3.5681, "step": 9561 }, { "epoch": 0.08631521935367395, "grad_norm": 3.3345015048980713, "learning_rate": 4.568469037732443e-05, "loss": 4.3167, "step": 9562 }, { "epoch": 0.08632424625383643, "grad_norm": 3.1731321811676025, "learning_rate": 4.5684239032316305e-05, "loss": 4.1176, "step": 9563 }, { "epoch": 0.08633327315399891, "grad_norm": 3.5829720497131348, "learning_rate": 4.5683787687308184e-05, "loss": 4.8346, "step": 9564 }, { "epoch": 0.0863423000541614, "grad_norm": 3.278916120529175, "learning_rate": 4.5683336342300056e-05, "loss": 4.8183, "step": 9565 }, { "epoch": 0.08635132695432389, "grad_norm": 3.3102786540985107, "learning_rate": 4.568288499729193e-05, "loss": 4.3879, "step": 9566 }, { "epoch": 0.08636035385448637, "grad_norm": 2.301360607147217, "learning_rate": 4.568243365228381e-05, "loss": 4.7468, "step": 9567 }, { "epoch": 0.08636938075464885, "grad_norm": 3.20991587638855, "learning_rate": 4.568198230727568e-05, "loss": 4.4602, "step": 9568 }, { "epoch": 0.08637840765481133, "grad_norm": 3.0540504455566406, "learning_rate": 4.568153096226756e-05, "loss": 5.5484, "step": 9569 }, { "epoch": 0.08638743455497382, "grad_norm": 3.652050256729126, "learning_rate": 4.568107961725944e-05, "loss": 4.8165, "step": 9570 }, { "epoch": 0.08639646145513631, "grad_norm": 2.6452064514160156, "learning_rate": 4.568062827225131e-05, "loss": 5.1907, "step": 9571 }, { "epoch": 0.08640548835529879, "grad_norm": 3.3838891983032227, "learning_rate": 4.5680176927243184e-05, "loss": 4.5035, "step": 9572 }, { "epoch": 0.08641451525546127, "grad_norm": 3.8304383754730225, "learning_rate": 4.567972558223506e-05, "loss": 4.5563, "step": 9573 }, { "epoch": 0.08642354215562376, "grad_norm": 3.0124545097351074, "learning_rate": 4.5679274237226936e-05, "loss": 4.4558, "step": 9574 }, { "epoch": 0.08643256905578624, "grad_norm": 3.6746292114257812, "learning_rate": 4.5678822892218815e-05, "loss": 4.3588, "step": 9575 }, { "epoch": 0.08644159595594873, "grad_norm": 3.667759656906128, "learning_rate": 4.5678371547210694e-05, "loss": 3.9114, "step": 9576 }, { "epoch": 0.08645062285611121, "grad_norm": 3.348238229751587, "learning_rate": 4.567792020220257e-05, "loss": 4.2701, "step": 9577 }, { "epoch": 0.0864596497562737, "grad_norm": 4.104750156402588, "learning_rate": 4.5677468857194446e-05, "loss": 4.0527, "step": 9578 }, { "epoch": 0.08646867665643618, "grad_norm": 2.775864839553833, "learning_rate": 4.567701751218632e-05, "loss": 4.8333, "step": 9579 }, { "epoch": 0.08647770355659866, "grad_norm": 3.6703035831451416, "learning_rate": 4.567656616717819e-05, "loss": 4.9506, "step": 9580 }, { "epoch": 0.08648673045676115, "grad_norm": 3.9847970008850098, "learning_rate": 4.567611482217007e-05, "loss": 4.7142, "step": 9581 }, { "epoch": 0.08649575735692364, "grad_norm": 3.819103240966797, "learning_rate": 4.567566347716194e-05, "loss": 4.9176, "step": 9582 }, { "epoch": 0.08650478425708612, "grad_norm": 3.1480252742767334, "learning_rate": 4.567521213215382e-05, "loss": 4.4514, "step": 9583 }, { "epoch": 0.0865138111572486, "grad_norm": 3.5221517086029053, "learning_rate": 4.56747607871457e-05, "loss": 5.0118, "step": 9584 }, { "epoch": 0.08652283805741108, "grad_norm": 3.2927026748657227, "learning_rate": 4.567430944213757e-05, "loss": 4.2504, "step": 9585 }, { "epoch": 0.08653186495757358, "grad_norm": 4.629014015197754, "learning_rate": 4.5673858097129446e-05, "loss": 4.0927, "step": 9586 }, { "epoch": 0.08654089185773606, "grad_norm": 3.054732322692871, "learning_rate": 4.5673406752121325e-05, "loss": 5.0861, "step": 9587 }, { "epoch": 0.08654991875789854, "grad_norm": 5.107945442199707, "learning_rate": 4.56729554071132e-05, "loss": 4.7054, "step": 9588 }, { "epoch": 0.08655894565806102, "grad_norm": 3.676177740097046, "learning_rate": 4.567250406210508e-05, "loss": 4.6667, "step": 9589 }, { "epoch": 0.0865679725582235, "grad_norm": 3.843916654586792, "learning_rate": 4.567205271709695e-05, "loss": 4.2705, "step": 9590 }, { "epoch": 0.08657699945838598, "grad_norm": 3.2603061199188232, "learning_rate": 4.567160137208883e-05, "loss": 5.1572, "step": 9591 }, { "epoch": 0.08658602635854848, "grad_norm": 3.5095221996307373, "learning_rate": 4.56711500270807e-05, "loss": 5.137, "step": 9592 }, { "epoch": 0.08659505325871096, "grad_norm": 3.600372791290283, "learning_rate": 4.5670698682072574e-05, "loss": 5.2374, "step": 9593 }, { "epoch": 0.08660408015887344, "grad_norm": 3.2127928733825684, "learning_rate": 4.567024733706445e-05, "loss": 4.5266, "step": 9594 }, { "epoch": 0.08661310705903592, "grad_norm": 3.03096866607666, "learning_rate": 4.566979599205633e-05, "loss": 4.3307, "step": 9595 }, { "epoch": 0.0866221339591984, "grad_norm": 3.032707929611206, "learning_rate": 4.5669344647048204e-05, "loss": 4.7785, "step": 9596 }, { "epoch": 0.0866311608593609, "grad_norm": 3.3781471252441406, "learning_rate": 4.5668893302040084e-05, "loss": 4.5737, "step": 9597 }, { "epoch": 0.08664018775952338, "grad_norm": 2.9625139236450195, "learning_rate": 4.566844195703196e-05, "loss": 4.1314, "step": 9598 }, { "epoch": 0.08664921465968586, "grad_norm": 6.097740650177002, "learning_rate": 4.566799061202383e-05, "loss": 3.6316, "step": 9599 }, { "epoch": 0.08665824155984835, "grad_norm": 3.5136210918426514, "learning_rate": 4.566753926701571e-05, "loss": 4.5203, "step": 9600 }, { "epoch": 0.08666726846001083, "grad_norm": 3.2083187103271484, "learning_rate": 4.566708792200759e-05, "loss": 4.6646, "step": 9601 }, { "epoch": 0.08667629536017332, "grad_norm": 2.555069923400879, "learning_rate": 4.566663657699946e-05, "loss": 4.7679, "step": 9602 }, { "epoch": 0.0866853222603358, "grad_norm": 3.894113063812256, "learning_rate": 4.566618523199134e-05, "loss": 4.1164, "step": 9603 }, { "epoch": 0.08669434916049829, "grad_norm": 5.339234352111816, "learning_rate": 4.566573388698321e-05, "loss": 4.4381, "step": 9604 }, { "epoch": 0.08670337606066077, "grad_norm": 3.852566957473755, "learning_rate": 4.566528254197509e-05, "loss": 4.7041, "step": 9605 }, { "epoch": 0.08671240296082325, "grad_norm": 3.4306821823120117, "learning_rate": 4.566483119696696e-05, "loss": 4.889, "step": 9606 }, { "epoch": 0.08672142986098574, "grad_norm": 3.216682195663452, "learning_rate": 4.5664379851958835e-05, "loss": 4.4229, "step": 9607 }, { "epoch": 0.08673045676114823, "grad_norm": 3.7186005115509033, "learning_rate": 4.5663928506950715e-05, "loss": 4.775, "step": 9608 }, { "epoch": 0.0867394836613107, "grad_norm": 3.6449949741363525, "learning_rate": 4.5663477161942594e-05, "loss": 4.4673, "step": 9609 }, { "epoch": 0.08674851056147319, "grad_norm": 3.8491904735565186, "learning_rate": 4.5663025816934466e-05, "loss": 4.8566, "step": 9610 }, { "epoch": 0.08675753746163567, "grad_norm": 4.2638115882873535, "learning_rate": 4.5662574471926346e-05, "loss": 4.1598, "step": 9611 }, { "epoch": 0.08676656436179817, "grad_norm": 3.562653064727783, "learning_rate": 4.566212312691822e-05, "loss": 4.701, "step": 9612 }, { "epoch": 0.08677559126196065, "grad_norm": 3.111647605895996, "learning_rate": 4.566167178191009e-05, "loss": 4.9263, "step": 9613 }, { "epoch": 0.08678461816212313, "grad_norm": 2.9133636951446533, "learning_rate": 4.566122043690197e-05, "loss": 5.1988, "step": 9614 }, { "epoch": 0.08679364506228561, "grad_norm": 3.73117733001709, "learning_rate": 4.566076909189385e-05, "loss": 4.7286, "step": 9615 }, { "epoch": 0.08680267196244809, "grad_norm": 3.5345206260681152, "learning_rate": 4.566031774688572e-05, "loss": 4.7219, "step": 9616 }, { "epoch": 0.08681169886261057, "grad_norm": 6.024276256561279, "learning_rate": 4.56598664018776e-05, "loss": 4.5822, "step": 9617 }, { "epoch": 0.08682072576277307, "grad_norm": 4.6278557777404785, "learning_rate": 4.565941505686947e-05, "loss": 4.1613, "step": 9618 }, { "epoch": 0.08682975266293555, "grad_norm": 4.034207820892334, "learning_rate": 4.5658963711861346e-05, "loss": 3.5608, "step": 9619 }, { "epoch": 0.08683877956309803, "grad_norm": 3.9672017097473145, "learning_rate": 4.5658512366853225e-05, "loss": 4.4145, "step": 9620 }, { "epoch": 0.08684780646326051, "grad_norm": 3.211639404296875, "learning_rate": 4.56580610218451e-05, "loss": 4.611, "step": 9621 }, { "epoch": 0.086856833363423, "grad_norm": 2.701982021331787, "learning_rate": 4.565760967683698e-05, "loss": 4.9754, "step": 9622 }, { "epoch": 0.08686586026358549, "grad_norm": 2.6291301250457764, "learning_rate": 4.5657158331828856e-05, "loss": 5.0978, "step": 9623 }, { "epoch": 0.08687488716374797, "grad_norm": 3.9564173221588135, "learning_rate": 4.565670698682073e-05, "loss": 4.6576, "step": 9624 }, { "epoch": 0.08688391406391045, "grad_norm": 3.770070791244507, "learning_rate": 4.565625564181261e-05, "loss": 4.8237, "step": 9625 }, { "epoch": 0.08689294096407293, "grad_norm": 3.3609800338745117, "learning_rate": 4.565580429680448e-05, "loss": 5.0595, "step": 9626 }, { "epoch": 0.08690196786423542, "grad_norm": 3.1661503314971924, "learning_rate": 4.565535295179635e-05, "loss": 4.8725, "step": 9627 }, { "epoch": 0.08691099476439791, "grad_norm": 2.9673712253570557, "learning_rate": 4.565490160678823e-05, "loss": 4.5785, "step": 9628 }, { "epoch": 0.08692002166456039, "grad_norm": 3.2414846420288086, "learning_rate": 4.5654450261780104e-05, "loss": 5.0073, "step": 9629 }, { "epoch": 0.08692904856472287, "grad_norm": 2.859745740890503, "learning_rate": 4.5653998916771983e-05, "loss": 4.6664, "step": 9630 }, { "epoch": 0.08693807546488536, "grad_norm": 3.416537284851074, "learning_rate": 4.565354757176386e-05, "loss": 4.9555, "step": 9631 }, { "epoch": 0.08694710236504784, "grad_norm": 3.335181951522827, "learning_rate": 4.565309622675573e-05, "loss": 4.7665, "step": 9632 }, { "epoch": 0.08695612926521033, "grad_norm": 4.989749431610107, "learning_rate": 4.565264488174761e-05, "loss": 4.4071, "step": 9633 }, { "epoch": 0.08696515616537281, "grad_norm": 2.977201223373413, "learning_rate": 4.565219353673949e-05, "loss": 4.5974, "step": 9634 }, { "epoch": 0.0869741830655353, "grad_norm": 4.016733169555664, "learning_rate": 4.565174219173136e-05, "loss": 4.1757, "step": 9635 }, { "epoch": 0.08698320996569778, "grad_norm": 3.90441632270813, "learning_rate": 4.565129084672324e-05, "loss": 5.1542, "step": 9636 }, { "epoch": 0.08699223686586026, "grad_norm": 3.4605817794799805, "learning_rate": 4.565083950171512e-05, "loss": 5.0291, "step": 9637 }, { "epoch": 0.08700126376602275, "grad_norm": 4.17263650894165, "learning_rate": 4.565038815670699e-05, "loss": 4.7298, "step": 9638 }, { "epoch": 0.08701029066618524, "grad_norm": 4.076101779937744, "learning_rate": 4.564993681169886e-05, "loss": 4.7333, "step": 9639 }, { "epoch": 0.08701931756634772, "grad_norm": 3.253690242767334, "learning_rate": 4.564948546669074e-05, "loss": 4.2828, "step": 9640 }, { "epoch": 0.0870283444665102, "grad_norm": 4.578309535980225, "learning_rate": 4.5649034121682614e-05, "loss": 5.0728, "step": 9641 }, { "epoch": 0.08703737136667268, "grad_norm": 5.096127033233643, "learning_rate": 4.5648582776674494e-05, "loss": 4.0282, "step": 9642 }, { "epoch": 0.08704639826683516, "grad_norm": 3.7384681701660156, "learning_rate": 4.5648131431666366e-05, "loss": 4.3615, "step": 9643 }, { "epoch": 0.08705542516699766, "grad_norm": 2.6861846446990967, "learning_rate": 4.5647680086658245e-05, "loss": 5.1365, "step": 9644 }, { "epoch": 0.08706445206716014, "grad_norm": 2.7169137001037598, "learning_rate": 4.5647228741650125e-05, "loss": 5.0291, "step": 9645 }, { "epoch": 0.08707347896732262, "grad_norm": 3.6770520210266113, "learning_rate": 4.564677739664199e-05, "loss": 4.7138, "step": 9646 }, { "epoch": 0.0870825058674851, "grad_norm": 3.426588773727417, "learning_rate": 4.564632605163387e-05, "loss": 4.0583, "step": 9647 }, { "epoch": 0.08709153276764758, "grad_norm": 3.3476288318634033, "learning_rate": 4.564587470662575e-05, "loss": 5.1007, "step": 9648 }, { "epoch": 0.08710055966781008, "grad_norm": 3.3457846641540527, "learning_rate": 4.564542336161762e-05, "loss": 4.6676, "step": 9649 }, { "epoch": 0.08710958656797256, "grad_norm": 3.857569694519043, "learning_rate": 4.56449720166095e-05, "loss": 5.1212, "step": 9650 }, { "epoch": 0.08711861346813504, "grad_norm": 4.251889705657959, "learning_rate": 4.564452067160137e-05, "loss": 4.9291, "step": 9651 }, { "epoch": 0.08712764036829752, "grad_norm": 3.8579366207122803, "learning_rate": 4.5644069326593245e-05, "loss": 4.8377, "step": 9652 }, { "epoch": 0.08713666726846, "grad_norm": 3.601771831512451, "learning_rate": 4.5643617981585125e-05, "loss": 4.6837, "step": 9653 }, { "epoch": 0.0871456941686225, "grad_norm": 2.938420295715332, "learning_rate": 4.5643166636577e-05, "loss": 4.3528, "step": 9654 }, { "epoch": 0.08715472106878498, "grad_norm": 2.7659428119659424, "learning_rate": 4.5642715291568876e-05, "loss": 4.8852, "step": 9655 }, { "epoch": 0.08716374796894746, "grad_norm": 3.6808018684387207, "learning_rate": 4.5642263946560756e-05, "loss": 5.0098, "step": 9656 }, { "epoch": 0.08717277486910995, "grad_norm": 2.9315719604492188, "learning_rate": 4.564181260155263e-05, "loss": 4.7102, "step": 9657 }, { "epoch": 0.08718180176927243, "grad_norm": 3.473557710647583, "learning_rate": 4.564136125654451e-05, "loss": 4.2911, "step": 9658 }, { "epoch": 0.08719082866943492, "grad_norm": 2.7110490798950195, "learning_rate": 4.564090991153638e-05, "loss": 4.301, "step": 9659 }, { "epoch": 0.0871998555695974, "grad_norm": 4.894271373748779, "learning_rate": 4.564045856652825e-05, "loss": 3.0679, "step": 9660 }, { "epoch": 0.08720888246975989, "grad_norm": 3.1735613346099854, "learning_rate": 4.564000722152013e-05, "loss": 5.0948, "step": 9661 }, { "epoch": 0.08721790936992237, "grad_norm": 3.961186408996582, "learning_rate": 4.563955587651201e-05, "loss": 4.8905, "step": 9662 }, { "epoch": 0.08722693627008485, "grad_norm": 3.420206069946289, "learning_rate": 4.563910453150388e-05, "loss": 4.1614, "step": 9663 }, { "epoch": 0.08723596317024734, "grad_norm": 5.004724025726318, "learning_rate": 4.563865318649576e-05, "loss": 5.4833, "step": 9664 }, { "epoch": 0.08724499007040983, "grad_norm": 3.103459596633911, "learning_rate": 4.5638201841487635e-05, "loss": 4.3238, "step": 9665 }, { "epoch": 0.0872540169705723, "grad_norm": 3.7591731548309326, "learning_rate": 4.563775049647951e-05, "loss": 4.2745, "step": 9666 }, { "epoch": 0.08726304387073479, "grad_norm": 3.8188860416412354, "learning_rate": 4.563729915147139e-05, "loss": 4.6466, "step": 9667 }, { "epoch": 0.08727207077089727, "grad_norm": 3.133376359939575, "learning_rate": 4.563684780646326e-05, "loss": 4.9597, "step": 9668 }, { "epoch": 0.08728109767105975, "grad_norm": 3.3907711505889893, "learning_rate": 4.563639646145514e-05, "loss": 5.2858, "step": 9669 }, { "epoch": 0.08729012457122225, "grad_norm": 4.173959255218506, "learning_rate": 4.563594511644702e-05, "loss": 4.4227, "step": 9670 }, { "epoch": 0.08729915147138473, "grad_norm": 3.7304840087890625, "learning_rate": 4.563549377143889e-05, "loss": 4.7184, "step": 9671 }, { "epoch": 0.08730817837154721, "grad_norm": 2.995201826095581, "learning_rate": 4.563504242643077e-05, "loss": 5.0069, "step": 9672 }, { "epoch": 0.08731720527170969, "grad_norm": 2.8698794841766357, "learning_rate": 4.563459108142264e-05, "loss": 4.3732, "step": 9673 }, { "epoch": 0.08732623217187217, "grad_norm": 3.0718393325805664, "learning_rate": 4.5634139736414514e-05, "loss": 5.0386, "step": 9674 }, { "epoch": 0.08733525907203467, "grad_norm": 3.2419400215148926, "learning_rate": 4.5633688391406393e-05, "loss": 5.1596, "step": 9675 }, { "epoch": 0.08734428597219715, "grad_norm": 3.490792751312256, "learning_rate": 4.563323704639827e-05, "loss": 4.0548, "step": 9676 }, { "epoch": 0.08735331287235963, "grad_norm": 3.7912793159484863, "learning_rate": 4.5632785701390145e-05, "loss": 5.019, "step": 9677 }, { "epoch": 0.08736233977252211, "grad_norm": 3.9007997512817383, "learning_rate": 4.5632334356382024e-05, "loss": 4.3066, "step": 9678 }, { "epoch": 0.0873713666726846, "grad_norm": 3.68764328956604, "learning_rate": 4.56318830113739e-05, "loss": 4.9848, "step": 9679 }, { "epoch": 0.08738039357284709, "grad_norm": 3.771367073059082, "learning_rate": 4.563143166636577e-05, "loss": 3.8458, "step": 9680 }, { "epoch": 0.08738942047300957, "grad_norm": 2.8358590602874756, "learning_rate": 4.563098032135765e-05, "loss": 5.5308, "step": 9681 }, { "epoch": 0.08739844737317205, "grad_norm": 4.220605850219727, "learning_rate": 4.563052897634952e-05, "loss": 4.5417, "step": 9682 }, { "epoch": 0.08740747427333453, "grad_norm": 3.1058566570281982, "learning_rate": 4.56300776313414e-05, "loss": 4.4976, "step": 9683 }, { "epoch": 0.08741650117349702, "grad_norm": 3.8499183654785156, "learning_rate": 4.562962628633328e-05, "loss": 5.0014, "step": 9684 }, { "epoch": 0.08742552807365951, "grad_norm": 3.592519998550415, "learning_rate": 4.562917494132515e-05, "loss": 4.7382, "step": 9685 }, { "epoch": 0.08743455497382199, "grad_norm": 3.7698299884796143, "learning_rate": 4.5628723596317024e-05, "loss": 4.8305, "step": 9686 }, { "epoch": 0.08744358187398447, "grad_norm": 3.7306196689605713, "learning_rate": 4.5628272251308904e-05, "loss": 5.5546, "step": 9687 }, { "epoch": 0.08745260877414696, "grad_norm": 3.307543992996216, "learning_rate": 4.5627820906300776e-05, "loss": 4.7831, "step": 9688 }, { "epoch": 0.08746163567430944, "grad_norm": 2.930325984954834, "learning_rate": 4.5627369561292655e-05, "loss": 5.0164, "step": 9689 }, { "epoch": 0.08747066257447193, "grad_norm": 3.5391170978546143, "learning_rate": 4.562691821628453e-05, "loss": 4.8656, "step": 9690 }, { "epoch": 0.08747968947463441, "grad_norm": 2.7677197456359863, "learning_rate": 4.562646687127641e-05, "loss": 5.4516, "step": 9691 }, { "epoch": 0.0874887163747969, "grad_norm": 3.6987414360046387, "learning_rate": 4.5626015526268286e-05, "loss": 4.8971, "step": 9692 }, { "epoch": 0.08749774327495938, "grad_norm": 3.721468210220337, "learning_rate": 4.562556418126015e-05, "loss": 4.4351, "step": 9693 }, { "epoch": 0.08750677017512186, "grad_norm": 3.702550172805786, "learning_rate": 4.562511283625203e-05, "loss": 5.0487, "step": 9694 }, { "epoch": 0.08751579707528434, "grad_norm": 3.9820024967193604, "learning_rate": 4.562466149124391e-05, "loss": 4.9164, "step": 9695 }, { "epoch": 0.08752482397544684, "grad_norm": 3.2977914810180664, "learning_rate": 4.562421014623578e-05, "loss": 4.9966, "step": 9696 }, { "epoch": 0.08753385087560932, "grad_norm": 3.5843160152435303, "learning_rate": 4.562375880122766e-05, "loss": 4.895, "step": 9697 }, { "epoch": 0.0875428777757718, "grad_norm": 3.9880151748657227, "learning_rate": 4.562330745621954e-05, "loss": 4.5034, "step": 9698 }, { "epoch": 0.08755190467593428, "grad_norm": 3.9501290321350098, "learning_rate": 4.562285611121141e-05, "loss": 4.3637, "step": 9699 }, { "epoch": 0.08756093157609676, "grad_norm": 2.9212186336517334, "learning_rate": 4.5622404766203286e-05, "loss": 4.567, "step": 9700 }, { "epoch": 0.08756995847625926, "grad_norm": 3.1107823848724365, "learning_rate": 4.5621953421195166e-05, "loss": 4.2689, "step": 9701 }, { "epoch": 0.08757898537642174, "grad_norm": 2.5731394290924072, "learning_rate": 4.562150207618704e-05, "loss": 4.9105, "step": 9702 }, { "epoch": 0.08758801227658422, "grad_norm": 3.0184388160705566, "learning_rate": 4.562105073117892e-05, "loss": 4.8147, "step": 9703 }, { "epoch": 0.0875970391767467, "grad_norm": 8.377710342407227, "learning_rate": 4.562059938617079e-05, "loss": 3.8682, "step": 9704 }, { "epoch": 0.08760606607690918, "grad_norm": 2.8914084434509277, "learning_rate": 4.562014804116267e-05, "loss": 4.4509, "step": 9705 }, { "epoch": 0.08761509297707168, "grad_norm": 3.383092164993286, "learning_rate": 4.561969669615454e-05, "loss": 4.5373, "step": 9706 }, { "epoch": 0.08762411987723416, "grad_norm": 3.476111888885498, "learning_rate": 4.5619245351146414e-05, "loss": 4.3564, "step": 9707 }, { "epoch": 0.08763314677739664, "grad_norm": 3.6877076625823975, "learning_rate": 4.561879400613829e-05, "loss": 4.4336, "step": 9708 }, { "epoch": 0.08764217367755912, "grad_norm": 3.099109172821045, "learning_rate": 4.561834266113017e-05, "loss": 4.8692, "step": 9709 }, { "epoch": 0.0876512005777216, "grad_norm": 3.2232675552368164, "learning_rate": 4.5617891316122045e-05, "loss": 4.672, "step": 9710 }, { "epoch": 0.0876602274778841, "grad_norm": 3.8505258560180664, "learning_rate": 4.5617439971113924e-05, "loss": 4.8492, "step": 9711 }, { "epoch": 0.08766925437804658, "grad_norm": 3.9837517738342285, "learning_rate": 4.56169886261058e-05, "loss": 4.5929, "step": 9712 }, { "epoch": 0.08767828127820906, "grad_norm": 3.3650598526000977, "learning_rate": 4.561653728109767e-05, "loss": 4.2015, "step": 9713 }, { "epoch": 0.08768730817837155, "grad_norm": 4.571897506713867, "learning_rate": 4.561608593608955e-05, "loss": 4.9581, "step": 9714 }, { "epoch": 0.08769633507853403, "grad_norm": 4.518139839172363, "learning_rate": 4.561563459108143e-05, "loss": 4.6805, "step": 9715 }, { "epoch": 0.08770536197869652, "grad_norm": 2.999354839324951, "learning_rate": 4.56151832460733e-05, "loss": 4.763, "step": 9716 }, { "epoch": 0.087714388878859, "grad_norm": 3.2045934200286865, "learning_rate": 4.561473190106518e-05, "loss": 4.8994, "step": 9717 }, { "epoch": 0.08772341577902149, "grad_norm": 3.1762773990631104, "learning_rate": 4.561428055605705e-05, "loss": 4.1581, "step": 9718 }, { "epoch": 0.08773244267918397, "grad_norm": 3.5758793354034424, "learning_rate": 4.561382921104893e-05, "loss": 4.3231, "step": 9719 }, { "epoch": 0.08774146957934645, "grad_norm": 5.335306644439697, "learning_rate": 4.5613377866040803e-05, "loss": 3.5468, "step": 9720 }, { "epoch": 0.08775049647950893, "grad_norm": 3.1267950534820557, "learning_rate": 4.5612926521032676e-05, "loss": 4.5122, "step": 9721 }, { "epoch": 0.08775952337967143, "grad_norm": 3.4330008029937744, "learning_rate": 4.5612475176024555e-05, "loss": 4.3608, "step": 9722 }, { "epoch": 0.0877685502798339, "grad_norm": 2.8795151710510254, "learning_rate": 4.5612023831016434e-05, "loss": 4.5622, "step": 9723 }, { "epoch": 0.08777757717999639, "grad_norm": 3.457340955734253, "learning_rate": 4.561157248600831e-05, "loss": 4.5895, "step": 9724 }, { "epoch": 0.08778660408015887, "grad_norm": 4.004331111907959, "learning_rate": 4.5611121141000186e-05, "loss": 4.9138, "step": 9725 }, { "epoch": 0.08779563098032135, "grad_norm": 2.871480941772461, "learning_rate": 4.561066979599206e-05, "loss": 4.4431, "step": 9726 }, { "epoch": 0.08780465788048385, "grad_norm": 3.8251848220825195, "learning_rate": 4.561021845098393e-05, "loss": 4.813, "step": 9727 }, { "epoch": 0.08781368478064633, "grad_norm": 3.5021817684173584, "learning_rate": 4.560976710597581e-05, "loss": 4.8695, "step": 9728 }, { "epoch": 0.08782271168080881, "grad_norm": 3.8338804244995117, "learning_rate": 4.560931576096768e-05, "loss": 4.8833, "step": 9729 }, { "epoch": 0.08783173858097129, "grad_norm": 3.194902181625366, "learning_rate": 4.560886441595956e-05, "loss": 4.4638, "step": 9730 }, { "epoch": 0.08784076548113377, "grad_norm": 5.486560821533203, "learning_rate": 4.560841307095144e-05, "loss": 5.0241, "step": 9731 }, { "epoch": 0.08784979238129627, "grad_norm": 3.2056095600128174, "learning_rate": 4.5607961725943314e-05, "loss": 4.6168, "step": 9732 }, { "epoch": 0.08785881928145875, "grad_norm": 3.6805830001831055, "learning_rate": 4.5607510380935186e-05, "loss": 4.739, "step": 9733 }, { "epoch": 0.08786784618162123, "grad_norm": 2.417473077774048, "learning_rate": 4.5607059035927065e-05, "loss": 4.8616, "step": 9734 }, { "epoch": 0.08787687308178371, "grad_norm": 3.3051819801330566, "learning_rate": 4.560660769091894e-05, "loss": 4.9578, "step": 9735 }, { "epoch": 0.0878858999819462, "grad_norm": 3.9560019969940186, "learning_rate": 4.560615634591082e-05, "loss": 4.7758, "step": 9736 }, { "epoch": 0.08789492688210869, "grad_norm": 3.395805597305298, "learning_rate": 4.5605705000902696e-05, "loss": 4.2572, "step": 9737 }, { "epoch": 0.08790395378227117, "grad_norm": 3.585061550140381, "learning_rate": 4.560525365589457e-05, "loss": 5.2174, "step": 9738 }, { "epoch": 0.08791298068243365, "grad_norm": 5.748512268066406, "learning_rate": 4.560480231088645e-05, "loss": 4.7714, "step": 9739 }, { "epoch": 0.08792200758259613, "grad_norm": 3.693852424621582, "learning_rate": 4.560435096587832e-05, "loss": 4.6048, "step": 9740 }, { "epoch": 0.08793103448275862, "grad_norm": 3.8787097930908203, "learning_rate": 4.560389962087019e-05, "loss": 5.2099, "step": 9741 }, { "epoch": 0.08794006138292111, "grad_norm": 3.602992057800293, "learning_rate": 4.560344827586207e-05, "loss": 4.6772, "step": 9742 }, { "epoch": 0.08794908828308359, "grad_norm": 4.122610569000244, "learning_rate": 4.5602996930853945e-05, "loss": 4.4829, "step": 9743 }, { "epoch": 0.08795811518324607, "grad_norm": 3.7254951000213623, "learning_rate": 4.5602545585845824e-05, "loss": 4.2685, "step": 9744 }, { "epoch": 0.08796714208340856, "grad_norm": 2.8887763023376465, "learning_rate": 4.56020942408377e-05, "loss": 4.702, "step": 9745 }, { "epoch": 0.08797616898357104, "grad_norm": 6.075344562530518, "learning_rate": 4.560164289582957e-05, "loss": 4.2016, "step": 9746 }, { "epoch": 0.08798519588373352, "grad_norm": 5.3554511070251465, "learning_rate": 4.560119155082145e-05, "loss": 4.2486, "step": 9747 }, { "epoch": 0.08799422278389601, "grad_norm": 3.3032257556915283, "learning_rate": 4.560074020581333e-05, "loss": 4.835, "step": 9748 }, { "epoch": 0.0880032496840585, "grad_norm": 4.2399163246154785, "learning_rate": 4.56002888608052e-05, "loss": 4.4438, "step": 9749 }, { "epoch": 0.08801227658422098, "grad_norm": 3.1548562049865723, "learning_rate": 4.559983751579708e-05, "loss": 4.6088, "step": 9750 }, { "epoch": 0.08802130348438346, "grad_norm": 3.5889317989349365, "learning_rate": 4.559938617078895e-05, "loss": 3.5092, "step": 9751 }, { "epoch": 0.08803033038454594, "grad_norm": 3.22702693939209, "learning_rate": 4.559893482578083e-05, "loss": 5.1578, "step": 9752 }, { "epoch": 0.08803935728470844, "grad_norm": 3.839284896850586, "learning_rate": 4.55984834807727e-05, "loss": 4.4251, "step": 9753 }, { "epoch": 0.08804838418487092, "grad_norm": 4.026360511779785, "learning_rate": 4.5598032135764576e-05, "loss": 4.5773, "step": 9754 }, { "epoch": 0.0880574110850334, "grad_norm": 4.315938949584961, "learning_rate": 4.5597580790756455e-05, "loss": 4.6404, "step": 9755 }, { "epoch": 0.08806643798519588, "grad_norm": 3.251406192779541, "learning_rate": 4.5597129445748334e-05, "loss": 4.8156, "step": 9756 }, { "epoch": 0.08807546488535836, "grad_norm": 3.350646495819092, "learning_rate": 4.559667810074021e-05, "loss": 4.3101, "step": 9757 }, { "epoch": 0.08808449178552086, "grad_norm": 4.237053871154785, "learning_rate": 4.5596226755732086e-05, "loss": 3.982, "step": 9758 }, { "epoch": 0.08809351868568334, "grad_norm": 3.829314708709717, "learning_rate": 4.5595775410723965e-05, "loss": 4.2404, "step": 9759 }, { "epoch": 0.08810254558584582, "grad_norm": 5.551459312438965, "learning_rate": 4.559532406571583e-05, "loss": 2.9175, "step": 9760 }, { "epoch": 0.0881115724860083, "grad_norm": 3.7015066146850586, "learning_rate": 4.559487272070771e-05, "loss": 4.7955, "step": 9761 }, { "epoch": 0.08812059938617078, "grad_norm": 4.023312568664551, "learning_rate": 4.559442137569959e-05, "loss": 4.0887, "step": 9762 }, { "epoch": 0.08812962628633328, "grad_norm": 3.5475716590881348, "learning_rate": 4.559397003069146e-05, "loss": 4.8322, "step": 9763 }, { "epoch": 0.08813865318649576, "grad_norm": 3.8049397468566895, "learning_rate": 4.559351868568334e-05, "loss": 4.9565, "step": 9764 }, { "epoch": 0.08814768008665824, "grad_norm": 3.815842390060425, "learning_rate": 4.5593067340675213e-05, "loss": 4.9469, "step": 9765 }, { "epoch": 0.08815670698682072, "grad_norm": 3.968148708343506, "learning_rate": 4.559261599566709e-05, "loss": 4.1641, "step": 9766 }, { "epoch": 0.0881657338869832, "grad_norm": 3.582704782485962, "learning_rate": 4.5592164650658965e-05, "loss": 4.7826, "step": 9767 }, { "epoch": 0.0881747607871457, "grad_norm": 5.832900524139404, "learning_rate": 4.559171330565084e-05, "loss": 4.7139, "step": 9768 }, { "epoch": 0.08818378768730818, "grad_norm": 4.590658187866211, "learning_rate": 4.559126196064272e-05, "loss": 4.5641, "step": 9769 }, { "epoch": 0.08819281458747066, "grad_norm": 8.05362606048584, "learning_rate": 4.5590810615634596e-05, "loss": 5.1485, "step": 9770 }, { "epoch": 0.08820184148763315, "grad_norm": 3.191896677017212, "learning_rate": 4.559035927062647e-05, "loss": 5.291, "step": 9771 }, { "epoch": 0.08821086838779563, "grad_norm": 3.4364213943481445, "learning_rate": 4.558990792561835e-05, "loss": 4.4367, "step": 9772 }, { "epoch": 0.08821989528795811, "grad_norm": 3.9107398986816406, "learning_rate": 4.558945658061022e-05, "loss": 4.6535, "step": 9773 }, { "epoch": 0.0882289221881206, "grad_norm": 4.57194709777832, "learning_rate": 4.558900523560209e-05, "loss": 3.2583, "step": 9774 }, { "epoch": 0.08823794908828309, "grad_norm": 4.33120059967041, "learning_rate": 4.558855389059397e-05, "loss": 4.0395, "step": 9775 }, { "epoch": 0.08824697598844557, "grad_norm": 3.1741902828216553, "learning_rate": 4.558810254558585e-05, "loss": 4.8077, "step": 9776 }, { "epoch": 0.08825600288860805, "grad_norm": 3.563587188720703, "learning_rate": 4.5587651200577724e-05, "loss": 4.7986, "step": 9777 }, { "epoch": 0.08826502978877053, "grad_norm": 4.275787830352783, "learning_rate": 4.55871998555696e-05, "loss": 4.225, "step": 9778 }, { "epoch": 0.08827405668893303, "grad_norm": 3.908266305923462, "learning_rate": 4.5586748510561475e-05, "loss": 4.5403, "step": 9779 }, { "epoch": 0.0882830835890955, "grad_norm": 3.967529296875, "learning_rate": 4.558629716555335e-05, "loss": 4.6659, "step": 9780 }, { "epoch": 0.08829211048925799, "grad_norm": 3.400639295578003, "learning_rate": 4.558584582054523e-05, "loss": 4.7007, "step": 9781 }, { "epoch": 0.08830113738942047, "grad_norm": 4.095448017120361, "learning_rate": 4.55853944755371e-05, "loss": 4.2721, "step": 9782 }, { "epoch": 0.08831016428958295, "grad_norm": 4.60933780670166, "learning_rate": 4.558494313052898e-05, "loss": 3.1782, "step": 9783 }, { "epoch": 0.08831919118974545, "grad_norm": 3.416090488433838, "learning_rate": 4.558449178552086e-05, "loss": 3.7753, "step": 9784 }, { "epoch": 0.08832821808990793, "grad_norm": 3.844916582107544, "learning_rate": 4.558404044051273e-05, "loss": 4.6962, "step": 9785 }, { "epoch": 0.08833724499007041, "grad_norm": 3.9383952617645264, "learning_rate": 4.558358909550461e-05, "loss": 4.9146, "step": 9786 }, { "epoch": 0.08834627189023289, "grad_norm": 2.9208242893218994, "learning_rate": 4.558313775049648e-05, "loss": 4.6435, "step": 9787 }, { "epoch": 0.08835529879039537, "grad_norm": 3.992751359939575, "learning_rate": 4.5582686405488355e-05, "loss": 4.4503, "step": 9788 }, { "epoch": 0.08836432569055787, "grad_norm": 3.3416852951049805, "learning_rate": 4.5582235060480234e-05, "loss": 4.624, "step": 9789 }, { "epoch": 0.08837335259072035, "grad_norm": 3.6499950885772705, "learning_rate": 4.5581783715472106e-05, "loss": 4.526, "step": 9790 }, { "epoch": 0.08838237949088283, "grad_norm": 2.050729274749756, "learning_rate": 4.5581332370463986e-05, "loss": 4.9594, "step": 9791 }, { "epoch": 0.08839140639104531, "grad_norm": 4.028285503387451, "learning_rate": 4.5580881025455865e-05, "loss": 4.7187, "step": 9792 }, { "epoch": 0.0884004332912078, "grad_norm": 3.7868566513061523, "learning_rate": 4.558042968044773e-05, "loss": 4.5152, "step": 9793 }, { "epoch": 0.08840946019137029, "grad_norm": 2.949226140975952, "learning_rate": 4.557997833543961e-05, "loss": 4.7337, "step": 9794 }, { "epoch": 0.08841848709153277, "grad_norm": 5.654137134552002, "learning_rate": 4.557952699043149e-05, "loss": 4.4072, "step": 9795 }, { "epoch": 0.08842751399169525, "grad_norm": 3.275542736053467, "learning_rate": 4.557907564542336e-05, "loss": 4.3886, "step": 9796 }, { "epoch": 0.08843654089185773, "grad_norm": 3.6244382858276367, "learning_rate": 4.557862430041524e-05, "loss": 4.4172, "step": 9797 }, { "epoch": 0.08844556779202022, "grad_norm": 3.37087082862854, "learning_rate": 4.557817295540712e-05, "loss": 4.3444, "step": 9798 }, { "epoch": 0.0884545946921827, "grad_norm": 3.2123489379882812, "learning_rate": 4.557772161039899e-05, "loss": 4.4822, "step": 9799 }, { "epoch": 0.08846362159234519, "grad_norm": 5.382800579071045, "learning_rate": 4.5577270265390865e-05, "loss": 4.8868, "step": 9800 }, { "epoch": 0.08847264849250767, "grad_norm": 3.4855639934539795, "learning_rate": 4.5576818920382744e-05, "loss": 4.6567, "step": 9801 }, { "epoch": 0.08848167539267016, "grad_norm": 3.535403251647949, "learning_rate": 4.5576367575374617e-05, "loss": 4.683, "step": 9802 }, { "epoch": 0.08849070229283264, "grad_norm": 4.061763763427734, "learning_rate": 4.5575916230366496e-05, "loss": 4.6358, "step": 9803 }, { "epoch": 0.08849972919299512, "grad_norm": 3.7846155166625977, "learning_rate": 4.557546488535837e-05, "loss": 5.0547, "step": 9804 }, { "epoch": 0.08850875609315761, "grad_norm": 4.539268970489502, "learning_rate": 4.557501354035025e-05, "loss": 4.5457, "step": 9805 }, { "epoch": 0.0885177829933201, "grad_norm": 3.0413737297058105, "learning_rate": 4.557456219534213e-05, "loss": 3.9553, "step": 9806 }, { "epoch": 0.08852680989348258, "grad_norm": 3.7545547485351562, "learning_rate": 4.557411085033399e-05, "loss": 4.5868, "step": 9807 }, { "epoch": 0.08853583679364506, "grad_norm": 4.541503429412842, "learning_rate": 4.557365950532587e-05, "loss": 4.6373, "step": 9808 }, { "epoch": 0.08854486369380754, "grad_norm": 3.1987998485565186, "learning_rate": 4.557320816031775e-05, "loss": 4.244, "step": 9809 }, { "epoch": 0.08855389059397004, "grad_norm": 3.7936017513275146, "learning_rate": 4.5572756815309623e-05, "loss": 4.8961, "step": 9810 }, { "epoch": 0.08856291749413252, "grad_norm": 4.686932563781738, "learning_rate": 4.55723054703015e-05, "loss": 4.659, "step": 9811 }, { "epoch": 0.088571944394295, "grad_norm": 5.704789161682129, "learning_rate": 4.5571854125293375e-05, "loss": 4.7447, "step": 9812 }, { "epoch": 0.08858097129445748, "grad_norm": 4.243939399719238, "learning_rate": 4.5571402780285254e-05, "loss": 4.0707, "step": 9813 }, { "epoch": 0.08858999819461996, "grad_norm": 3.1735377311706543, "learning_rate": 4.557095143527713e-05, "loss": 4.6253, "step": 9814 }, { "epoch": 0.08859902509478246, "grad_norm": 3.178849697113037, "learning_rate": 4.5570500090269e-05, "loss": 4.7225, "step": 9815 }, { "epoch": 0.08860805199494494, "grad_norm": 3.9334235191345215, "learning_rate": 4.557004874526088e-05, "loss": 4.7205, "step": 9816 }, { "epoch": 0.08861707889510742, "grad_norm": 4.528887748718262, "learning_rate": 4.556959740025276e-05, "loss": 4.8964, "step": 9817 }, { "epoch": 0.0886261057952699, "grad_norm": 4.70770788192749, "learning_rate": 4.556914605524463e-05, "loss": 4.8881, "step": 9818 }, { "epoch": 0.08863513269543238, "grad_norm": 3.225026845932007, "learning_rate": 4.556869471023651e-05, "loss": 4.941, "step": 9819 }, { "epoch": 0.08864415959559488, "grad_norm": 4.0644450187683105, "learning_rate": 4.556824336522838e-05, "loss": 4.6728, "step": 9820 }, { "epoch": 0.08865318649575736, "grad_norm": 3.750070810317993, "learning_rate": 4.5567792020220254e-05, "loss": 4.3432, "step": 9821 }, { "epoch": 0.08866221339591984, "grad_norm": 4.210346698760986, "learning_rate": 4.5567340675212134e-05, "loss": 4.217, "step": 9822 }, { "epoch": 0.08867124029608232, "grad_norm": 3.545724391937256, "learning_rate": 4.556688933020401e-05, "loss": 4.4219, "step": 9823 }, { "epoch": 0.0886802671962448, "grad_norm": 3.267117738723755, "learning_rate": 4.5566437985195885e-05, "loss": 4.5333, "step": 9824 }, { "epoch": 0.0886892940964073, "grad_norm": 3.3638954162597656, "learning_rate": 4.5565986640187765e-05, "loss": 4.5632, "step": 9825 }, { "epoch": 0.08869832099656978, "grad_norm": 2.938793897628784, "learning_rate": 4.556553529517964e-05, "loss": 5.2106, "step": 9826 }, { "epoch": 0.08870734789673226, "grad_norm": 3.189063549041748, "learning_rate": 4.556508395017151e-05, "loss": 4.4631, "step": 9827 }, { "epoch": 0.08871637479689475, "grad_norm": 3.421377658843994, "learning_rate": 4.556463260516339e-05, "loss": 4.5256, "step": 9828 }, { "epoch": 0.08872540169705723, "grad_norm": 3.6616644859313965, "learning_rate": 4.556418126015526e-05, "loss": 4.5938, "step": 9829 }, { "epoch": 0.08873442859721971, "grad_norm": 2.896247148513794, "learning_rate": 4.556372991514714e-05, "loss": 5.075, "step": 9830 }, { "epoch": 0.0887434554973822, "grad_norm": 3.947572946548462, "learning_rate": 4.556327857013902e-05, "loss": 3.8935, "step": 9831 }, { "epoch": 0.08875248239754469, "grad_norm": 3.518533706665039, "learning_rate": 4.556282722513089e-05, "loss": 4.5573, "step": 9832 }, { "epoch": 0.08876150929770717, "grad_norm": 2.6463239192962646, "learning_rate": 4.556237588012277e-05, "loss": 4.9156, "step": 9833 }, { "epoch": 0.08877053619786965, "grad_norm": 2.7059686183929443, "learning_rate": 4.5561924535114644e-05, "loss": 4.4067, "step": 9834 }, { "epoch": 0.08877956309803213, "grad_norm": 3.296967029571533, "learning_rate": 4.5561473190106516e-05, "loss": 4.4998, "step": 9835 }, { "epoch": 0.08878858999819463, "grad_norm": 3.21854567527771, "learning_rate": 4.5561021845098396e-05, "loss": 5.1779, "step": 9836 }, { "epoch": 0.0887976168983571, "grad_norm": 3.584465503692627, "learning_rate": 4.5560570500090275e-05, "loss": 5.1766, "step": 9837 }, { "epoch": 0.08880664379851959, "grad_norm": 4.308779239654541, "learning_rate": 4.556011915508215e-05, "loss": 4.6328, "step": 9838 }, { "epoch": 0.08881567069868207, "grad_norm": 4.204985618591309, "learning_rate": 4.5559667810074027e-05, "loss": 4.4798, "step": 9839 }, { "epoch": 0.08882469759884455, "grad_norm": 4.270348072052002, "learning_rate": 4.55592164650659e-05, "loss": 4.2232, "step": 9840 }, { "epoch": 0.08883372449900705, "grad_norm": 4.211034774780273, "learning_rate": 4.555876512005777e-05, "loss": 4.9314, "step": 9841 }, { "epoch": 0.08884275139916953, "grad_norm": 4.593605995178223, "learning_rate": 4.555831377504965e-05, "loss": 4.8382, "step": 9842 }, { "epoch": 0.08885177829933201, "grad_norm": 3.4479739665985107, "learning_rate": 4.555786243004152e-05, "loss": 5.0625, "step": 9843 }, { "epoch": 0.08886080519949449, "grad_norm": 3.136714458465576, "learning_rate": 4.55574110850334e-05, "loss": 4.6625, "step": 9844 }, { "epoch": 0.08886983209965697, "grad_norm": 3.4726412296295166, "learning_rate": 4.555695974002528e-05, "loss": 4.2759, "step": 9845 }, { "epoch": 0.08887885899981947, "grad_norm": 3.1731791496276855, "learning_rate": 4.5556508395017154e-05, "loss": 5.1353, "step": 9846 }, { "epoch": 0.08888788589998195, "grad_norm": 3.341811418533325, "learning_rate": 4.5556057050009027e-05, "loss": 4.9459, "step": 9847 }, { "epoch": 0.08889691280014443, "grad_norm": 3.3652842044830322, "learning_rate": 4.5555605705000906e-05, "loss": 4.5779, "step": 9848 }, { "epoch": 0.08890593970030691, "grad_norm": 4.348387718200684, "learning_rate": 4.555515435999278e-05, "loss": 4.4268, "step": 9849 }, { "epoch": 0.0889149666004694, "grad_norm": 3.255431890487671, "learning_rate": 4.555470301498466e-05, "loss": 4.6285, "step": 9850 }, { "epoch": 0.08892399350063189, "grad_norm": 4.131855010986328, "learning_rate": 4.555425166997653e-05, "loss": 4.6228, "step": 9851 }, { "epoch": 0.08893302040079437, "grad_norm": 3.169558048248291, "learning_rate": 4.555380032496841e-05, "loss": 4.9545, "step": 9852 }, { "epoch": 0.08894204730095685, "grad_norm": 3.7924246788024902, "learning_rate": 4.555334897996029e-05, "loss": 4.5314, "step": 9853 }, { "epoch": 0.08895107420111933, "grad_norm": 3.281797409057617, "learning_rate": 4.5552897634952154e-05, "loss": 4.4486, "step": 9854 }, { "epoch": 0.08896010110128182, "grad_norm": 3.234856128692627, "learning_rate": 4.5552446289944033e-05, "loss": 4.6382, "step": 9855 }, { "epoch": 0.0889691280014443, "grad_norm": 3.597783088684082, "learning_rate": 4.555199494493591e-05, "loss": 5.0521, "step": 9856 }, { "epoch": 0.08897815490160679, "grad_norm": 3.905380964279175, "learning_rate": 4.5551543599927785e-05, "loss": 4.8909, "step": 9857 }, { "epoch": 0.08898718180176927, "grad_norm": 3.431971788406372, "learning_rate": 4.5551092254919664e-05, "loss": 4.4617, "step": 9858 }, { "epoch": 0.08899620870193176, "grad_norm": 2.8901877403259277, "learning_rate": 4.5550640909911544e-05, "loss": 5.1813, "step": 9859 }, { "epoch": 0.08900523560209424, "grad_norm": 2.8807079792022705, "learning_rate": 4.555018956490341e-05, "loss": 5.0381, "step": 9860 }, { "epoch": 0.08901426250225672, "grad_norm": 4.035806179046631, "learning_rate": 4.554973821989529e-05, "loss": 3.9923, "step": 9861 }, { "epoch": 0.08902328940241921, "grad_norm": 2.8229641914367676, "learning_rate": 4.554928687488717e-05, "loss": 4.868, "step": 9862 }, { "epoch": 0.0890323163025817, "grad_norm": 3.696002721786499, "learning_rate": 4.554883552987904e-05, "loss": 4.4837, "step": 9863 }, { "epoch": 0.08904134320274418, "grad_norm": 2.9586081504821777, "learning_rate": 4.554838418487092e-05, "loss": 5.0196, "step": 9864 }, { "epoch": 0.08905037010290666, "grad_norm": 3.9685769081115723, "learning_rate": 4.554793283986279e-05, "loss": 5.2324, "step": 9865 }, { "epoch": 0.08905939700306914, "grad_norm": 3.469034433364868, "learning_rate": 4.554748149485467e-05, "loss": 4.5054, "step": 9866 }, { "epoch": 0.08906842390323164, "grad_norm": 3.6309170722961426, "learning_rate": 4.5547030149846544e-05, "loss": 4.7545, "step": 9867 }, { "epoch": 0.08907745080339412, "grad_norm": 3.633612871170044, "learning_rate": 4.5546578804838416e-05, "loss": 4.0226, "step": 9868 }, { "epoch": 0.0890864777035566, "grad_norm": 2.9460055828094482, "learning_rate": 4.5546127459830295e-05, "loss": 4.5214, "step": 9869 }, { "epoch": 0.08909550460371908, "grad_norm": 4.697933673858643, "learning_rate": 4.5545676114822175e-05, "loss": 4.8592, "step": 9870 }, { "epoch": 0.08910453150388156, "grad_norm": 3.5223071575164795, "learning_rate": 4.554522476981405e-05, "loss": 4.5169, "step": 9871 }, { "epoch": 0.08911355840404406, "grad_norm": 7.075228691101074, "learning_rate": 4.5544773424805926e-05, "loss": 3.9456, "step": 9872 }, { "epoch": 0.08912258530420654, "grad_norm": 2.7530126571655273, "learning_rate": 4.55443220797978e-05, "loss": 4.1745, "step": 9873 }, { "epoch": 0.08913161220436902, "grad_norm": 5.07460880279541, "learning_rate": 4.554387073478967e-05, "loss": 4.415, "step": 9874 }, { "epoch": 0.0891406391045315, "grad_norm": 3.85965633392334, "learning_rate": 4.554341938978155e-05, "loss": 4.6277, "step": 9875 }, { "epoch": 0.08914966600469398, "grad_norm": 3.3681821823120117, "learning_rate": 4.554296804477343e-05, "loss": 4.7431, "step": 9876 }, { "epoch": 0.08915869290485648, "grad_norm": 3.646289587020874, "learning_rate": 4.55425166997653e-05, "loss": 4.5127, "step": 9877 }, { "epoch": 0.08916771980501896, "grad_norm": 3.4303042888641357, "learning_rate": 4.554206535475718e-05, "loss": 4.1924, "step": 9878 }, { "epoch": 0.08917674670518144, "grad_norm": 3.497807741165161, "learning_rate": 4.5541614009749054e-05, "loss": 4.601, "step": 9879 }, { "epoch": 0.08918577360534392, "grad_norm": 3.198748826980591, "learning_rate": 4.554116266474093e-05, "loss": 4.8141, "step": 9880 }, { "epoch": 0.0891948005055064, "grad_norm": 2.8766047954559326, "learning_rate": 4.5540711319732806e-05, "loss": 4.2709, "step": 9881 }, { "epoch": 0.08920382740566889, "grad_norm": 3.3085591793060303, "learning_rate": 4.554025997472468e-05, "loss": 4.4528, "step": 9882 }, { "epoch": 0.08921285430583138, "grad_norm": 3.96844744682312, "learning_rate": 4.553980862971656e-05, "loss": 4.8586, "step": 9883 }, { "epoch": 0.08922188120599386, "grad_norm": 4.76986026763916, "learning_rate": 4.5539357284708437e-05, "loss": 5.1182, "step": 9884 }, { "epoch": 0.08923090810615635, "grad_norm": 2.6232359409332275, "learning_rate": 4.553890593970031e-05, "loss": 4.8975, "step": 9885 }, { "epoch": 0.08923993500631883, "grad_norm": 2.9854421615600586, "learning_rate": 4.553845459469219e-05, "loss": 4.7984, "step": 9886 }, { "epoch": 0.08924896190648131, "grad_norm": 4.498281478881836, "learning_rate": 4.553800324968406e-05, "loss": 3.3931, "step": 9887 }, { "epoch": 0.0892579888066438, "grad_norm": 4.603984355926514, "learning_rate": 4.553755190467593e-05, "loss": 4.479, "step": 9888 }, { "epoch": 0.08926701570680629, "grad_norm": 3.764112710952759, "learning_rate": 4.553710055966781e-05, "loss": 5.0423, "step": 9889 }, { "epoch": 0.08927604260696877, "grad_norm": 3.3715903759002686, "learning_rate": 4.5536649214659685e-05, "loss": 4.5894, "step": 9890 }, { "epoch": 0.08928506950713125, "grad_norm": 3.571355104446411, "learning_rate": 4.5536197869651564e-05, "loss": 4.8152, "step": 9891 }, { "epoch": 0.08929409640729373, "grad_norm": 3.133854389190674, "learning_rate": 4.553574652464344e-05, "loss": 4.3256, "step": 9892 }, { "epoch": 0.08930312330745623, "grad_norm": 4.524274826049805, "learning_rate": 4.5535295179635316e-05, "loss": 4.9721, "step": 9893 }, { "epoch": 0.0893121502076187, "grad_norm": 3.7531611919403076, "learning_rate": 4.553484383462719e-05, "loss": 5.0303, "step": 9894 }, { "epoch": 0.08932117710778119, "grad_norm": 3.3766837120056152, "learning_rate": 4.553439248961907e-05, "loss": 4.905, "step": 9895 }, { "epoch": 0.08933020400794367, "grad_norm": 3.67600679397583, "learning_rate": 4.553394114461094e-05, "loss": 4.9571, "step": 9896 }, { "epoch": 0.08933923090810615, "grad_norm": 3.0827128887176514, "learning_rate": 4.553348979960282e-05, "loss": 4.3767, "step": 9897 }, { "epoch": 0.08934825780826865, "grad_norm": 3.410565137863159, "learning_rate": 4.55330384545947e-05, "loss": 4.7289, "step": 9898 }, { "epoch": 0.08935728470843113, "grad_norm": 3.902078151702881, "learning_rate": 4.553258710958657e-05, "loss": 4.2056, "step": 9899 }, { "epoch": 0.08936631160859361, "grad_norm": 3.772935628890991, "learning_rate": 4.553213576457845e-05, "loss": 4.515, "step": 9900 }, { "epoch": 0.08937533850875609, "grad_norm": 4.181002616882324, "learning_rate": 4.553168441957032e-05, "loss": 4.6112, "step": 9901 }, { "epoch": 0.08938436540891857, "grad_norm": 4.339587688446045, "learning_rate": 4.5531233074562195e-05, "loss": 4.4436, "step": 9902 }, { "epoch": 0.08939339230908107, "grad_norm": 3.112352132797241, "learning_rate": 4.5530781729554074e-05, "loss": 4.1827, "step": 9903 }, { "epoch": 0.08940241920924355, "grad_norm": 2.991697311401367, "learning_rate": 4.553033038454595e-05, "loss": 5.0809, "step": 9904 }, { "epoch": 0.08941144610940603, "grad_norm": 3.0994739532470703, "learning_rate": 4.5529879039537826e-05, "loss": 4.798, "step": 9905 }, { "epoch": 0.08942047300956851, "grad_norm": 4.169796466827393, "learning_rate": 4.5529427694529705e-05, "loss": 5.6164, "step": 9906 }, { "epoch": 0.089429499909731, "grad_norm": 3.4587621688842773, "learning_rate": 4.552897634952157e-05, "loss": 5.1899, "step": 9907 }, { "epoch": 0.08943852680989348, "grad_norm": 3.5743463039398193, "learning_rate": 4.552852500451345e-05, "loss": 5.0872, "step": 9908 }, { "epoch": 0.08944755371005597, "grad_norm": 4.6673102378845215, "learning_rate": 4.552807365950533e-05, "loss": 4.3272, "step": 9909 }, { "epoch": 0.08945658061021845, "grad_norm": 4.258767604827881, "learning_rate": 4.55276223144972e-05, "loss": 3.649, "step": 9910 }, { "epoch": 0.08946560751038093, "grad_norm": 3.5397002696990967, "learning_rate": 4.552717096948908e-05, "loss": 4.9284, "step": 9911 }, { "epoch": 0.08947463441054342, "grad_norm": 3.9379208087921143, "learning_rate": 4.5526719624480954e-05, "loss": 4.3873, "step": 9912 }, { "epoch": 0.0894836613107059, "grad_norm": 3.3592419624328613, "learning_rate": 4.552626827947283e-05, "loss": 3.9302, "step": 9913 }, { "epoch": 0.08949268821086839, "grad_norm": 3.29681658744812, "learning_rate": 4.5525816934464705e-05, "loss": 4.7612, "step": 9914 }, { "epoch": 0.08950171511103087, "grad_norm": 3.582529306411743, "learning_rate": 4.552536558945658e-05, "loss": 4.6533, "step": 9915 }, { "epoch": 0.08951074201119336, "grad_norm": 3.9886322021484375, "learning_rate": 4.552491424444846e-05, "loss": 4.9103, "step": 9916 }, { "epoch": 0.08951976891135584, "grad_norm": 3.876162528991699, "learning_rate": 4.5524462899440336e-05, "loss": 4.801, "step": 9917 }, { "epoch": 0.08952879581151832, "grad_norm": 2.824371814727783, "learning_rate": 4.552401155443221e-05, "loss": 5.3964, "step": 9918 }, { "epoch": 0.08953782271168081, "grad_norm": 5.937251567840576, "learning_rate": 4.552356020942409e-05, "loss": 5.437, "step": 9919 }, { "epoch": 0.0895468496118433, "grad_norm": 3.517746686935425, "learning_rate": 4.552310886441597e-05, "loss": 5.3674, "step": 9920 }, { "epoch": 0.08955587651200578, "grad_norm": 3.1086161136627197, "learning_rate": 4.552265751940783e-05, "loss": 4.9258, "step": 9921 }, { "epoch": 0.08956490341216826, "grad_norm": 2.8122520446777344, "learning_rate": 4.552220617439971e-05, "loss": 4.8478, "step": 9922 }, { "epoch": 0.08957393031233074, "grad_norm": 4.46949577331543, "learning_rate": 4.552175482939159e-05, "loss": 5.0054, "step": 9923 }, { "epoch": 0.08958295721249324, "grad_norm": 3.9578614234924316, "learning_rate": 4.5521303484383464e-05, "loss": 4.6217, "step": 9924 }, { "epoch": 0.08959198411265572, "grad_norm": 3.8082425594329834, "learning_rate": 4.552085213937534e-05, "loss": 4.4608, "step": 9925 }, { "epoch": 0.0896010110128182, "grad_norm": 3.0647552013397217, "learning_rate": 4.5520400794367216e-05, "loss": 4.914, "step": 9926 }, { "epoch": 0.08961003791298068, "grad_norm": 3.6027767658233643, "learning_rate": 4.5519949449359095e-05, "loss": 5.3023, "step": 9927 }, { "epoch": 0.08961906481314316, "grad_norm": 3.7150068283081055, "learning_rate": 4.551949810435097e-05, "loss": 4.1533, "step": 9928 }, { "epoch": 0.08962809171330566, "grad_norm": 3.3117780685424805, "learning_rate": 4.551904675934284e-05, "loss": 4.568, "step": 9929 }, { "epoch": 0.08963711861346814, "grad_norm": 3.3281772136688232, "learning_rate": 4.551859541433472e-05, "loss": 4.2379, "step": 9930 }, { "epoch": 0.08964614551363062, "grad_norm": 4.02193021774292, "learning_rate": 4.55181440693266e-05, "loss": 4.9735, "step": 9931 }, { "epoch": 0.0896551724137931, "grad_norm": 3.32987380027771, "learning_rate": 4.551769272431847e-05, "loss": 4.5448, "step": 9932 }, { "epoch": 0.08966419931395558, "grad_norm": 3.330213785171509, "learning_rate": 4.551724137931035e-05, "loss": 4.5844, "step": 9933 }, { "epoch": 0.08967322621411807, "grad_norm": 3.9560065269470215, "learning_rate": 4.551679003430222e-05, "loss": 4.4749, "step": 9934 }, { "epoch": 0.08968225311428056, "grad_norm": 2.940516233444214, "learning_rate": 4.5516338689294095e-05, "loss": 5.2629, "step": 9935 }, { "epoch": 0.08969128001444304, "grad_norm": 4.46031379699707, "learning_rate": 4.5515887344285974e-05, "loss": 4.7042, "step": 9936 }, { "epoch": 0.08970030691460552, "grad_norm": 4.286287307739258, "learning_rate": 4.551543599927785e-05, "loss": 4.0477, "step": 9937 }, { "epoch": 0.089709333814768, "grad_norm": 3.712453842163086, "learning_rate": 4.5514984654269726e-05, "loss": 4.3471, "step": 9938 }, { "epoch": 0.08971836071493049, "grad_norm": 3.1127963066101074, "learning_rate": 4.5514533309261605e-05, "loss": 4.4276, "step": 9939 }, { "epoch": 0.08972738761509298, "grad_norm": 3.463566303253174, "learning_rate": 4.551408196425348e-05, "loss": 4.8822, "step": 9940 }, { "epoch": 0.08973641451525546, "grad_norm": 2.768632650375366, "learning_rate": 4.551363061924535e-05, "loss": 3.8463, "step": 9941 }, { "epoch": 0.08974544141541795, "grad_norm": 3.903707504272461, "learning_rate": 4.551317927423723e-05, "loss": 4.2448, "step": 9942 }, { "epoch": 0.08975446831558043, "grad_norm": 3.7239158153533936, "learning_rate": 4.55127279292291e-05, "loss": 4.1344, "step": 9943 }, { "epoch": 0.08976349521574291, "grad_norm": 3.624849319458008, "learning_rate": 4.551227658422098e-05, "loss": 4.9564, "step": 9944 }, { "epoch": 0.0897725221159054, "grad_norm": 3.4039690494537354, "learning_rate": 4.551182523921286e-05, "loss": 4.2905, "step": 9945 }, { "epoch": 0.08978154901606789, "grad_norm": 2.718151330947876, "learning_rate": 4.551137389420473e-05, "loss": 4.7892, "step": 9946 }, { "epoch": 0.08979057591623037, "grad_norm": 3.6937007904052734, "learning_rate": 4.551092254919661e-05, "loss": 4.388, "step": 9947 }, { "epoch": 0.08979960281639285, "grad_norm": 3.5831902027130127, "learning_rate": 4.5510471204188484e-05, "loss": 4.5261, "step": 9948 }, { "epoch": 0.08980862971655533, "grad_norm": 3.9126946926116943, "learning_rate": 4.551001985918036e-05, "loss": 4.1398, "step": 9949 }, { "epoch": 0.08981765661671783, "grad_norm": 3.5580976009368896, "learning_rate": 4.5509568514172236e-05, "loss": 4.7192, "step": 9950 }, { "epoch": 0.0898266835168803, "grad_norm": 3.4928536415100098, "learning_rate": 4.550911716916411e-05, "loss": 4.2695, "step": 9951 }, { "epoch": 0.08983571041704279, "grad_norm": 3.911123037338257, "learning_rate": 4.550866582415599e-05, "loss": 4.0884, "step": 9952 }, { "epoch": 0.08984473731720527, "grad_norm": 3.3820767402648926, "learning_rate": 4.550821447914787e-05, "loss": 4.0924, "step": 9953 }, { "epoch": 0.08985376421736775, "grad_norm": 3.532928705215454, "learning_rate": 4.550776313413973e-05, "loss": 4.5798, "step": 9954 }, { "epoch": 0.08986279111753025, "grad_norm": 3.4319722652435303, "learning_rate": 4.550731178913161e-05, "loss": 4.8709, "step": 9955 }, { "epoch": 0.08987181801769273, "grad_norm": 4.679045677185059, "learning_rate": 4.550686044412349e-05, "loss": 4.9562, "step": 9956 }, { "epoch": 0.08988084491785521, "grad_norm": 4.054172515869141, "learning_rate": 4.5506409099115364e-05, "loss": 4.541, "step": 9957 }, { "epoch": 0.08988987181801769, "grad_norm": 3.6489925384521484, "learning_rate": 4.550595775410724e-05, "loss": 5.0926, "step": 9958 }, { "epoch": 0.08989889871818017, "grad_norm": 3.17264723777771, "learning_rate": 4.550550640909912e-05, "loss": 4.6454, "step": 9959 }, { "epoch": 0.08990792561834265, "grad_norm": 3.531310796737671, "learning_rate": 4.5505055064090995e-05, "loss": 4.0706, "step": 9960 }, { "epoch": 0.08991695251850515, "grad_norm": 3.0764503479003906, "learning_rate": 4.550460371908287e-05, "loss": 4.7405, "step": 9961 }, { "epoch": 0.08992597941866763, "grad_norm": 3.968533754348755, "learning_rate": 4.5504152374074746e-05, "loss": 5.4437, "step": 9962 }, { "epoch": 0.08993500631883011, "grad_norm": 3.2299439907073975, "learning_rate": 4.550370102906662e-05, "loss": 5.133, "step": 9963 }, { "epoch": 0.0899440332189926, "grad_norm": 2.8611745834350586, "learning_rate": 4.55032496840585e-05, "loss": 5.076, "step": 9964 }, { "epoch": 0.08995306011915508, "grad_norm": 7.186960697174072, "learning_rate": 4.550279833905037e-05, "loss": 4.145, "step": 9965 }, { "epoch": 0.08996208701931757, "grad_norm": 4.592215538024902, "learning_rate": 4.550234699404225e-05, "loss": 4.7711, "step": 9966 }, { "epoch": 0.08997111391948005, "grad_norm": 3.0623931884765625, "learning_rate": 4.550189564903413e-05, "loss": 5.0592, "step": 9967 }, { "epoch": 0.08998014081964253, "grad_norm": 4.292324066162109, "learning_rate": 4.5501444304025995e-05, "loss": 4.853, "step": 9968 }, { "epoch": 0.08998916771980502, "grad_norm": 3.258080244064331, "learning_rate": 4.5500992959017874e-05, "loss": 4.3334, "step": 9969 }, { "epoch": 0.0899981946199675, "grad_norm": 3.1979644298553467, "learning_rate": 4.550054161400975e-05, "loss": 4.9059, "step": 9970 }, { "epoch": 0.09000722152012999, "grad_norm": 3.5340981483459473, "learning_rate": 4.5500090269001626e-05, "loss": 4.7429, "step": 9971 }, { "epoch": 0.09001624842029247, "grad_norm": 2.5737621784210205, "learning_rate": 4.5499638923993505e-05, "loss": 4.9628, "step": 9972 }, { "epoch": 0.09002527532045496, "grad_norm": 3.9461748600006104, "learning_rate": 4.549918757898538e-05, "loss": 4.7135, "step": 9973 }, { "epoch": 0.09003430222061744, "grad_norm": 4.298503398895264, "learning_rate": 4.5498736233977257e-05, "loss": 4.5094, "step": 9974 }, { "epoch": 0.09004332912077992, "grad_norm": 4.421361923217773, "learning_rate": 4.549828488896913e-05, "loss": 3.5729, "step": 9975 }, { "epoch": 0.09005235602094241, "grad_norm": 3.6991026401519775, "learning_rate": 4.5497833543961e-05, "loss": 4.8882, "step": 9976 }, { "epoch": 0.0900613829211049, "grad_norm": 3.86830472946167, "learning_rate": 4.549738219895288e-05, "loss": 4.3999, "step": 9977 }, { "epoch": 0.09007040982126738, "grad_norm": 3.370546817779541, "learning_rate": 4.549693085394476e-05, "loss": 4.3965, "step": 9978 }, { "epoch": 0.09007943672142986, "grad_norm": 3.6608238220214844, "learning_rate": 4.549647950893663e-05, "loss": 4.4759, "step": 9979 }, { "epoch": 0.09008846362159234, "grad_norm": 3.0317771434783936, "learning_rate": 4.549602816392851e-05, "loss": 5.414, "step": 9980 }, { "epoch": 0.09009749052175484, "grad_norm": 3.0392463207244873, "learning_rate": 4.5495576818920384e-05, "loss": 5.1146, "step": 9981 }, { "epoch": 0.09010651742191732, "grad_norm": 3.507902145385742, "learning_rate": 4.5495125473912257e-05, "loss": 4.8748, "step": 9982 }, { "epoch": 0.0901155443220798, "grad_norm": 4.084225654602051, "learning_rate": 4.5494674128904136e-05, "loss": 4.3126, "step": 9983 }, { "epoch": 0.09012457122224228, "grad_norm": 2.9743125438690186, "learning_rate": 4.5494222783896015e-05, "loss": 5.8114, "step": 9984 }, { "epoch": 0.09013359812240476, "grad_norm": 6.84206485748291, "learning_rate": 4.549377143888789e-05, "loss": 4.8973, "step": 9985 }, { "epoch": 0.09014262502256724, "grad_norm": 3.552036762237549, "learning_rate": 4.549332009387977e-05, "loss": 4.7387, "step": 9986 }, { "epoch": 0.09015165192272974, "grad_norm": 3.608969211578369, "learning_rate": 4.549286874887164e-05, "loss": 4.445, "step": 9987 }, { "epoch": 0.09016067882289222, "grad_norm": 3.169265031814575, "learning_rate": 4.549241740386351e-05, "loss": 5.1961, "step": 9988 }, { "epoch": 0.0901697057230547, "grad_norm": 4.411298751831055, "learning_rate": 4.549196605885539e-05, "loss": 4.3119, "step": 9989 }, { "epoch": 0.09017873262321718, "grad_norm": 3.779693841934204, "learning_rate": 4.5491514713847263e-05, "loss": 4.0767, "step": 9990 }, { "epoch": 0.09018775952337967, "grad_norm": 3.511521816253662, "learning_rate": 4.549106336883914e-05, "loss": 4.3213, "step": 9991 }, { "epoch": 0.09019678642354216, "grad_norm": 2.987658977508545, "learning_rate": 4.549061202383102e-05, "loss": 4.7487, "step": 9992 }, { "epoch": 0.09020581332370464, "grad_norm": 4.15426778793335, "learning_rate": 4.5490160678822894e-05, "loss": 4.8955, "step": 9993 }, { "epoch": 0.09021484022386712, "grad_norm": 3.5005245208740234, "learning_rate": 4.5489709333814774e-05, "loss": 4.6077, "step": 9994 }, { "epoch": 0.0902238671240296, "grad_norm": 3.8327560424804688, "learning_rate": 4.5489257988806646e-05, "loss": 4.3799, "step": 9995 }, { "epoch": 0.09023289402419209, "grad_norm": 3.7540180683135986, "learning_rate": 4.548880664379852e-05, "loss": 5.1801, "step": 9996 }, { "epoch": 0.09024192092435458, "grad_norm": 5.443745136260986, "learning_rate": 4.54883552987904e-05, "loss": 3.8033, "step": 9997 }, { "epoch": 0.09025094782451706, "grad_norm": 2.578425407409668, "learning_rate": 4.548790395378228e-05, "loss": 5.097, "step": 9998 }, { "epoch": 0.09025997472467955, "grad_norm": 3.1845853328704834, "learning_rate": 4.548745260877415e-05, "loss": 4.5462, "step": 9999 }, { "epoch": 0.09026900162484203, "grad_norm": 2.6787045001983643, "learning_rate": 4.548700126376603e-05, "loss": 4.3626, "step": 10000 } ], "logging_steps": 1, "max_steps": 110780, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 5000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 6.471642033234739e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }