{ "best_metric": 0.540250366102693, "best_model_checkpoint": "/data0/checkpoints/Qwen2.5-Math-7B-ScalePRM-v3.0/checkpoint-600", "epoch": 0.757934628138323, "eval_steps": 100, "global_step": 600, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0012632243802305385, "grad_norm": 2.429877281188965, "learning_rate": 7e-06, "loss": 1.1465, "step": 1 }, { "epoch": 0.002526448760461077, "grad_norm": 2.2121567726135254, "learning_rate": 7e-06, "loss": 1.0451, "step": 2 }, { "epoch": 0.0037896731406916154, "grad_norm": 1.3855836391448975, "learning_rate": 7e-06, "loss": 0.735, "step": 3 }, { "epoch": 0.005052897520922154, "grad_norm": 0.3500981330871582, "learning_rate": 7e-06, "loss": 0.5603, "step": 4 }, { "epoch": 0.0063161219011526925, "grad_norm": 0.40845438838005066, "learning_rate": 7e-06, "loss": 1.0627, "step": 5 }, { "epoch": 0.007579346281383231, "grad_norm": 0.9035907983779907, "learning_rate": 7e-06, "loss": 5.0569, "step": 6 }, { "epoch": 0.00884257066161377, "grad_norm": 0.6223624348640442, "learning_rate": 7e-06, "loss": 9.4706, "step": 7 }, { "epoch": 0.010105795041844308, "grad_norm": 0.5229220390319824, "learning_rate": 7e-06, "loss": 5.6685, "step": 8 }, { "epoch": 0.011369019422074847, "grad_norm": 0.5141741633415222, "learning_rate": 7e-06, "loss": 5.8831, "step": 9 }, { "epoch": 0.012632243802305385, "grad_norm": 0.6420879364013672, "learning_rate": 7e-06, "loss": 4.6915, "step": 10 }, { "epoch": 0.013895468182535923, "grad_norm": 0.47964179515838623, "learning_rate": 7e-06, "loss": 3.9531, "step": 11 }, { "epoch": 0.015158692562766462, "grad_norm": 0.40398040413856506, "learning_rate": 7e-06, "loss": 1.9501, "step": 12 }, { "epoch": 0.016421916942996998, "grad_norm": 0.5141711235046387, "learning_rate": 7e-06, "loss": 1.9378, "step": 13 }, { "epoch": 0.01768514132322754, "grad_norm": 0.24602794647216797, "learning_rate": 7e-06, "loss": 1.535, "step": 14 }, { "epoch": 0.018948365703458078, "grad_norm": 0.21208855509757996, "learning_rate": 7e-06, "loss": 1.5177, "step": 15 }, { "epoch": 0.020211590083688617, "grad_norm": 0.6067216992378235, "learning_rate": 7e-06, "loss": 0.9104, "step": 16 }, { "epoch": 0.021474814463919155, "grad_norm": 0.422442227602005, "learning_rate": 7e-06, "loss": 0.9266, "step": 17 }, { "epoch": 0.022738038844149693, "grad_norm": 0.659572958946228, "learning_rate": 7e-06, "loss": 0.7499, "step": 18 }, { "epoch": 0.02400126322438023, "grad_norm": 0.4817348122596741, "learning_rate": 7e-06, "loss": 0.7436, "step": 19 }, { "epoch": 0.02526448760461077, "grad_norm": 0.20682591199874878, "learning_rate": 7e-06, "loss": 0.4786, "step": 20 }, { "epoch": 0.026527711984841308, "grad_norm": 0.2613360583782196, "learning_rate": 7e-06, "loss": 0.8464, "step": 21 }, { "epoch": 0.027790936365071846, "grad_norm": 0.2720305621623993, "learning_rate": 7e-06, "loss": 0.7255, "step": 22 }, { "epoch": 0.029054160745302385, "grad_norm": 0.25043392181396484, "learning_rate": 7e-06, "loss": 0.9661, "step": 23 }, { "epoch": 0.030317385125532923, "grad_norm": 0.2801963686943054, "learning_rate": 7e-06, "loss": 0.9602, "step": 24 }, { "epoch": 0.03158060950576346, "grad_norm": 0.2137051522731781, "learning_rate": 7e-06, "loss": 0.8132, "step": 25 }, { "epoch": 0.032843833885993996, "grad_norm": 0.13553065061569214, "learning_rate": 7e-06, "loss": 0.4873, "step": 26 }, { "epoch": 0.03410705826622454, "grad_norm": 0.1350618302822113, "learning_rate": 7e-06, "loss": 0.5186, "step": 27 }, { "epoch": 0.03537028264645508, "grad_norm": 0.1236298605799675, "learning_rate": 7e-06, "loss": 0.3401, "step": 28 }, { "epoch": 0.036633507026685615, "grad_norm": 0.15515856444835663, "learning_rate": 7e-06, "loss": 1.2493, "step": 29 }, { "epoch": 0.037896731406916156, "grad_norm": 0.09012973308563232, "learning_rate": 7e-06, "loss": 0.7651, "step": 30 }, { "epoch": 0.03915995578714669, "grad_norm": 0.14378102123737335, "learning_rate": 7e-06, "loss": 0.4005, "step": 31 }, { "epoch": 0.04042318016737723, "grad_norm": 0.2546883523464203, "learning_rate": 7e-06, "loss": 0.5304, "step": 32 }, { "epoch": 0.04168640454760777, "grad_norm": 0.1023496687412262, "learning_rate": 7e-06, "loss": 0.493, "step": 33 }, { "epoch": 0.04294962892783831, "grad_norm": 0.1719491183757782, "learning_rate": 7e-06, "loss": 0.3707, "step": 34 }, { "epoch": 0.044212853308068845, "grad_norm": 0.08337250351905823, "learning_rate": 7e-06, "loss": 0.3143, "step": 35 }, { "epoch": 0.045476077688299386, "grad_norm": 0.09040359407663345, "learning_rate": 7e-06, "loss": 0.3942, "step": 36 }, { "epoch": 0.04673930206852992, "grad_norm": 0.10850965231657028, "learning_rate": 7e-06, "loss": 0.4453, "step": 37 }, { "epoch": 0.04800252644876046, "grad_norm": 0.08887636661529541, "learning_rate": 7e-06, "loss": 1.2015, "step": 38 }, { "epoch": 0.049265750828991, "grad_norm": 0.1864442229270935, "learning_rate": 7e-06, "loss": 0.3895, "step": 39 }, { "epoch": 0.05052897520922154, "grad_norm": 0.175123393535614, "learning_rate": 7e-06, "loss": 0.3626, "step": 40 }, { "epoch": 0.051792199589452075, "grad_norm": 0.10572918504476547, "learning_rate": 7e-06, "loss": 0.3335, "step": 41 }, { "epoch": 0.053055423969682616, "grad_norm": 0.09624486416578293, "learning_rate": 7e-06, "loss": 0.3065, "step": 42 }, { "epoch": 0.05431864834991315, "grad_norm": 0.13604743778705597, "learning_rate": 7e-06, "loss": 0.317, "step": 43 }, { "epoch": 0.05558187273014369, "grad_norm": 0.15408551692962646, "learning_rate": 7e-06, "loss": 0.7709, "step": 44 }, { "epoch": 0.05684509711037423, "grad_norm": 0.09676961600780487, "learning_rate": 7e-06, "loss": 0.5114, "step": 45 }, { "epoch": 0.05810832149060477, "grad_norm": 0.11936207115650177, "learning_rate": 7e-06, "loss": 0.2785, "step": 46 }, { "epoch": 0.059371545870835304, "grad_norm": 0.1744876503944397, "learning_rate": 7e-06, "loss": 0.2689, "step": 47 }, { "epoch": 0.060634770251065846, "grad_norm": 0.17397810518741608, "learning_rate": 7e-06, "loss": 0.3316, "step": 48 }, { "epoch": 0.06189799463129638, "grad_norm": 0.1329212635755539, "learning_rate": 7e-06, "loss": 0.2853, "step": 49 }, { "epoch": 0.06316121901152692, "grad_norm": 0.09555013477802277, "learning_rate": 7e-06, "loss": 0.3182, "step": 50 }, { "epoch": 0.06442444339175746, "grad_norm": 0.15529152750968933, "learning_rate": 7e-06, "loss": 0.5497, "step": 51 }, { "epoch": 0.06568766777198799, "grad_norm": 0.09599810838699341, "learning_rate": 7e-06, "loss": 0.3102, "step": 52 }, { "epoch": 0.06695089215221854, "grad_norm": 0.12325876951217651, "learning_rate": 7e-06, "loss": 0.2788, "step": 53 }, { "epoch": 0.06821411653244908, "grad_norm": 0.2820286154747009, "learning_rate": 7e-06, "loss": 0.3934, "step": 54 }, { "epoch": 0.06947734091267961, "grad_norm": 0.17912541329860687, "learning_rate": 7e-06, "loss": 0.3709, "step": 55 }, { "epoch": 0.07074056529291016, "grad_norm": 0.14083553850650787, "learning_rate": 7e-06, "loss": 0.4105, "step": 56 }, { "epoch": 0.0720037896731407, "grad_norm": 0.09743569046258926, "learning_rate": 7e-06, "loss": 0.2376, "step": 57 }, { "epoch": 0.07326701405337123, "grad_norm": 0.10704771429300308, "learning_rate": 7e-06, "loss": 0.2714, "step": 58 }, { "epoch": 0.07453023843360176, "grad_norm": 0.11463718861341476, "learning_rate": 7e-06, "loss": 0.456, "step": 59 }, { "epoch": 0.07579346281383231, "grad_norm": 0.12085901200771332, "learning_rate": 7e-06, "loss": 0.3099, "step": 60 }, { "epoch": 0.07705668719406285, "grad_norm": 0.10744248330593109, "learning_rate": 7e-06, "loss": 0.2312, "step": 61 }, { "epoch": 0.07831991157429338, "grad_norm": 0.08374691009521484, "learning_rate": 7e-06, "loss": 0.2685, "step": 62 }, { "epoch": 0.07958313595452392, "grad_norm": 0.10826320946216583, "learning_rate": 7e-06, "loss": 0.3069, "step": 63 }, { "epoch": 0.08084636033475447, "grad_norm": 0.17864489555358887, "learning_rate": 7e-06, "loss": 0.5024, "step": 64 }, { "epoch": 0.082109584714985, "grad_norm": 0.11988472938537598, "learning_rate": 7e-06, "loss": 0.3164, "step": 65 }, { "epoch": 0.08337280909521554, "grad_norm": 0.1612488180398941, "learning_rate": 7e-06, "loss": 0.3295, "step": 66 }, { "epoch": 0.08463603347544607, "grad_norm": 0.13754408061504364, "learning_rate": 7e-06, "loss": 0.2406, "step": 67 }, { "epoch": 0.08589925785567662, "grad_norm": 0.11351214349269867, "learning_rate": 7e-06, "loss": 0.3149, "step": 68 }, { "epoch": 0.08716248223590715, "grad_norm": 0.07585523277521133, "learning_rate": 7e-06, "loss": 0.2706, "step": 69 }, { "epoch": 0.08842570661613769, "grad_norm": 0.0744984969496727, "learning_rate": 7e-06, "loss": 0.214, "step": 70 }, { "epoch": 0.08968893099636822, "grad_norm": 0.2244742512702942, "learning_rate": 7e-06, "loss": 0.4242, "step": 71 }, { "epoch": 0.09095215537659877, "grad_norm": 0.08662209659814835, "learning_rate": 7e-06, "loss": 0.2691, "step": 72 }, { "epoch": 0.09221537975682931, "grad_norm": 0.10564761608839035, "learning_rate": 7e-06, "loss": 0.3228, "step": 73 }, { "epoch": 0.09347860413705984, "grad_norm": 0.18067984282970428, "learning_rate": 7e-06, "loss": 0.3384, "step": 74 }, { "epoch": 0.09474182851729038, "grad_norm": 0.07762212306261063, "learning_rate": 7e-06, "loss": 0.2377, "step": 75 }, { "epoch": 0.09600505289752093, "grad_norm": 0.07793518900871277, "learning_rate": 7e-06, "loss": 0.248, "step": 76 }, { "epoch": 0.09726827727775146, "grad_norm": 0.1307854801416397, "learning_rate": 7e-06, "loss": 0.2667, "step": 77 }, { "epoch": 0.098531501657982, "grad_norm": 0.09771443158388138, "learning_rate": 7e-06, "loss": 0.318, "step": 78 }, { "epoch": 0.09979472603821253, "grad_norm": 0.10437527298927307, "learning_rate": 7e-06, "loss": 0.3303, "step": 79 }, { "epoch": 0.10105795041844308, "grad_norm": 0.11160580813884735, "learning_rate": 7e-06, "loss": 0.2845, "step": 80 }, { "epoch": 0.10232117479867361, "grad_norm": 0.0809980109333992, "learning_rate": 7e-06, "loss": 0.2736, "step": 81 }, { "epoch": 0.10358439917890415, "grad_norm": 0.10574865341186523, "learning_rate": 7e-06, "loss": 0.3012, "step": 82 }, { "epoch": 0.1048476235591347, "grad_norm": 0.07807318866252899, "learning_rate": 7e-06, "loss": 0.2283, "step": 83 }, { "epoch": 0.10611084793936523, "grad_norm": 0.10281991213560104, "learning_rate": 7e-06, "loss": 0.2544, "step": 84 }, { "epoch": 0.10737407231959577, "grad_norm": 0.12749870121479034, "learning_rate": 7e-06, "loss": 0.2973, "step": 85 }, { "epoch": 0.1086372966998263, "grad_norm": 0.13138003647327423, "learning_rate": 7e-06, "loss": 0.3399, "step": 86 }, { "epoch": 0.10990052108005685, "grad_norm": 0.10815514624118805, "learning_rate": 7e-06, "loss": 0.3221, "step": 87 }, { "epoch": 0.11116374546028739, "grad_norm": 0.13537508249282837, "learning_rate": 7e-06, "loss": 0.308, "step": 88 }, { "epoch": 0.11242696984051792, "grad_norm": 0.09689060598611832, "learning_rate": 7e-06, "loss": 0.2511, "step": 89 }, { "epoch": 0.11369019422074846, "grad_norm": 0.08782925456762314, "learning_rate": 7e-06, "loss": 0.2936, "step": 90 }, { "epoch": 0.114953418600979, "grad_norm": 0.12655287981033325, "learning_rate": 7e-06, "loss": 0.4158, "step": 91 }, { "epoch": 0.11621664298120954, "grad_norm": 0.11866717785596848, "learning_rate": 7e-06, "loss": 0.4059, "step": 92 }, { "epoch": 0.11747986736144007, "grad_norm": 0.12691305577754974, "learning_rate": 7e-06, "loss": 0.2453, "step": 93 }, { "epoch": 0.11874309174167061, "grad_norm": 0.11844722181558609, "learning_rate": 7e-06, "loss": 0.2249, "step": 94 }, { "epoch": 0.12000631612190116, "grad_norm": 0.07606595754623413, "learning_rate": 7e-06, "loss": 0.2789, "step": 95 }, { "epoch": 0.12126954050213169, "grad_norm": 0.11529266834259033, "learning_rate": 7e-06, "loss": 0.2654, "step": 96 }, { "epoch": 0.12253276488236223, "grad_norm": 0.12648285925388336, "learning_rate": 7e-06, "loss": 0.2279, "step": 97 }, { "epoch": 0.12379598926259276, "grad_norm": 0.1504458636045456, "learning_rate": 7e-06, "loss": 0.4048, "step": 98 }, { "epoch": 0.1250592136428233, "grad_norm": 0.09578829258680344, "learning_rate": 7e-06, "loss": 0.2912, "step": 99 }, { "epoch": 0.12632243802305385, "grad_norm": 0.10936733335256577, "learning_rate": 7e-06, "loss": 0.2644, "step": 100 }, { "epoch": 0.12632243802305385, "eval_correct_accuracy": 0.5708227311280747, "eval_error_accuracy": 0.45610085547050877, "eval_f1": 0.5070537660000148, "eval_loss": 0.43133699893951416, "eval_runtime": 35.3366, "eval_samples_per_second": 96.217, "eval_steps_per_second": 6.028, "step": 100 }, { "epoch": 0.1275856624032844, "grad_norm": 0.1165054589509964, "learning_rate": 7e-06, "loss": 0.2693, "step": 101 }, { "epoch": 0.12884888678351492, "grad_norm": 0.08343573659658432, "learning_rate": 7e-06, "loss": 0.2388, "step": 102 }, { "epoch": 0.13011211116374546, "grad_norm": 0.10629656910896301, "learning_rate": 7e-06, "loss": 0.2603, "step": 103 }, { "epoch": 0.13137533554397599, "grad_norm": 0.07509850710630417, "learning_rate": 7e-06, "loss": 0.253, "step": 104 }, { "epoch": 0.13263855992420653, "grad_norm": 0.08039335906505585, "learning_rate": 7e-06, "loss": 0.224, "step": 105 }, { "epoch": 0.13390178430443708, "grad_norm": 0.10666981339454651, "learning_rate": 7e-06, "loss": 0.3945, "step": 106 }, { "epoch": 0.1351650086846676, "grad_norm": 0.16490086913108826, "learning_rate": 7e-06, "loss": 0.3087, "step": 107 }, { "epoch": 0.13642823306489815, "grad_norm": 0.09013114124536514, "learning_rate": 7e-06, "loss": 0.3355, "step": 108 }, { "epoch": 0.1376914574451287, "grad_norm": 0.1580226719379425, "learning_rate": 7e-06, "loss": 0.2433, "step": 109 }, { "epoch": 0.13895468182535922, "grad_norm": 0.09130299836397171, "learning_rate": 7e-06, "loss": 0.1928, "step": 110 }, { "epoch": 0.14021790620558977, "grad_norm": 0.07702811807394028, "learning_rate": 7e-06, "loss": 0.2319, "step": 111 }, { "epoch": 0.14148113058582032, "grad_norm": 0.14257381856441498, "learning_rate": 7e-06, "loss": 0.2496, "step": 112 }, { "epoch": 0.14274435496605084, "grad_norm": 0.11546823382377625, "learning_rate": 7e-06, "loss": 0.2592, "step": 113 }, { "epoch": 0.1440075793462814, "grad_norm": 0.12595829367637634, "learning_rate": 7e-06, "loss": 0.2539, "step": 114 }, { "epoch": 0.1452708037265119, "grad_norm": 0.10172153264284134, "learning_rate": 7e-06, "loss": 0.2728, "step": 115 }, { "epoch": 0.14653402810674246, "grad_norm": 0.10145121812820435, "learning_rate": 7e-06, "loss": 0.163, "step": 116 }, { "epoch": 0.147797252486973, "grad_norm": 0.15631917119026184, "learning_rate": 7e-06, "loss": 0.2526, "step": 117 }, { "epoch": 0.14906047686720353, "grad_norm": 0.13442394137382507, "learning_rate": 7e-06, "loss": 0.2591, "step": 118 }, { "epoch": 0.15032370124743408, "grad_norm": 0.08642445504665375, "learning_rate": 7e-06, "loss": 0.2505, "step": 119 }, { "epoch": 0.15158692562766463, "grad_norm": 0.13054709136486053, "learning_rate": 7e-06, "loss": 0.2704, "step": 120 }, { "epoch": 0.15285015000789515, "grad_norm": 0.19653519988059998, "learning_rate": 7e-06, "loss": 0.2384, "step": 121 }, { "epoch": 0.1541133743881257, "grad_norm": 0.20973946154117584, "learning_rate": 7e-06, "loss": 0.2385, "step": 122 }, { "epoch": 0.15537659876835622, "grad_norm": 0.096860371530056, "learning_rate": 7e-06, "loss": 0.241, "step": 123 }, { "epoch": 0.15663982314858677, "grad_norm": 0.10356521606445312, "learning_rate": 7e-06, "loss": 0.4737, "step": 124 }, { "epoch": 0.15790304752881731, "grad_norm": 0.17340725660324097, "learning_rate": 7e-06, "loss": 0.2673, "step": 125 }, { "epoch": 0.15916627190904784, "grad_norm": 0.3000679612159729, "learning_rate": 7e-06, "loss": 0.431, "step": 126 }, { "epoch": 0.16042949628927838, "grad_norm": 0.11215244233608246, "learning_rate": 7e-06, "loss": 0.206, "step": 127 }, { "epoch": 0.16169272066950893, "grad_norm": 0.07078877836465836, "learning_rate": 7e-06, "loss": 0.201, "step": 128 }, { "epoch": 0.16295594504973945, "grad_norm": 0.16037459671497345, "learning_rate": 7e-06, "loss": 0.268, "step": 129 }, { "epoch": 0.16421916942997, "grad_norm": 0.20243118703365326, "learning_rate": 7e-06, "loss": 0.2503, "step": 130 }, { "epoch": 0.16548239381020052, "grad_norm": 0.1389663964509964, "learning_rate": 7e-06, "loss": 0.2517, "step": 131 }, { "epoch": 0.16674561819043107, "grad_norm": 0.12263572961091995, "learning_rate": 7e-06, "loss": 0.2359, "step": 132 }, { "epoch": 0.16800884257066162, "grad_norm": 0.14491412043571472, "learning_rate": 7e-06, "loss": 0.3347, "step": 133 }, { "epoch": 0.16927206695089214, "grad_norm": 0.1378932148218155, "learning_rate": 7e-06, "loss": 0.2638, "step": 134 }, { "epoch": 0.1705352913311227, "grad_norm": 0.07053989171981812, "learning_rate": 7e-06, "loss": 0.2299, "step": 135 }, { "epoch": 0.17179851571135324, "grad_norm": 0.19610151648521423, "learning_rate": 7e-06, "loss": 0.2789, "step": 136 }, { "epoch": 0.17306174009158376, "grad_norm": 0.1290581375360489, "learning_rate": 7e-06, "loss": 0.209, "step": 137 }, { "epoch": 0.1743249644718143, "grad_norm": 0.1481819599866867, "learning_rate": 7e-06, "loss": 0.2723, "step": 138 }, { "epoch": 0.17558818885204486, "grad_norm": 0.1427401453256607, "learning_rate": 7e-06, "loss": 0.2778, "step": 139 }, { "epoch": 0.17685141323227538, "grad_norm": 0.0666273981332779, "learning_rate": 7e-06, "loss": 0.2008, "step": 140 }, { "epoch": 0.17811463761250593, "grad_norm": 0.13182522356510162, "learning_rate": 7e-06, "loss": 0.2441, "step": 141 }, { "epoch": 0.17937786199273645, "grad_norm": 0.08374546468257904, "learning_rate": 7e-06, "loss": 0.2603, "step": 142 }, { "epoch": 0.180641086372967, "grad_norm": 0.10638394951820374, "learning_rate": 7e-06, "loss": 0.2354, "step": 143 }, { "epoch": 0.18190431075319755, "grad_norm": 0.10801179707050323, "learning_rate": 7e-06, "loss": 0.2875, "step": 144 }, { "epoch": 0.18316753513342807, "grad_norm": 0.13121351599693298, "learning_rate": 7e-06, "loss": 0.2304, "step": 145 }, { "epoch": 0.18443075951365862, "grad_norm": 0.10176476836204529, "learning_rate": 7e-06, "loss": 0.2311, "step": 146 }, { "epoch": 0.18569398389388916, "grad_norm": 0.10199464112520218, "learning_rate": 7e-06, "loss": 0.2522, "step": 147 }, { "epoch": 0.18695720827411969, "grad_norm": 0.09650130569934845, "learning_rate": 7e-06, "loss": 0.2351, "step": 148 }, { "epoch": 0.18822043265435023, "grad_norm": 0.12842021882534027, "learning_rate": 7e-06, "loss": 0.2244, "step": 149 }, { "epoch": 0.18948365703458075, "grad_norm": 0.1237226277589798, "learning_rate": 7e-06, "loss": 0.2706, "step": 150 }, { "epoch": 0.1907468814148113, "grad_norm": 0.12939125299453735, "learning_rate": 7e-06, "loss": 0.2445, "step": 151 }, { "epoch": 0.19201010579504185, "grad_norm": 0.11460690945386887, "learning_rate": 7e-06, "loss": 0.2601, "step": 152 }, { "epoch": 0.19327333017527237, "grad_norm": 0.18108275532722473, "learning_rate": 7e-06, "loss": 0.3465, "step": 153 }, { "epoch": 0.19453655455550292, "grad_norm": 0.0727877989411354, "learning_rate": 7e-06, "loss": 0.1878, "step": 154 }, { "epoch": 0.19579977893573347, "grad_norm": 0.12313497066497803, "learning_rate": 7e-06, "loss": 0.2311, "step": 155 }, { "epoch": 0.197063003315964, "grad_norm": 0.1377153992652893, "learning_rate": 7e-06, "loss": 0.2573, "step": 156 }, { "epoch": 0.19832622769619454, "grad_norm": 0.08758647739887238, "learning_rate": 7e-06, "loss": 0.2156, "step": 157 }, { "epoch": 0.19958945207642506, "grad_norm": 0.11441980302333832, "learning_rate": 7e-06, "loss": 0.2801, "step": 158 }, { "epoch": 0.2008526764566556, "grad_norm": 0.12151770293712616, "learning_rate": 7e-06, "loss": 0.242, "step": 159 }, { "epoch": 0.20211590083688616, "grad_norm": 0.159256711602211, "learning_rate": 7e-06, "loss": 0.2612, "step": 160 }, { "epoch": 0.20337912521711668, "grad_norm": 0.08577941358089447, "learning_rate": 7e-06, "loss": 0.2115, "step": 161 }, { "epoch": 0.20464234959734723, "grad_norm": 0.1190810427069664, "learning_rate": 7e-06, "loss": 0.2434, "step": 162 }, { "epoch": 0.20590557397757778, "grad_norm": 0.09624910354614258, "learning_rate": 7e-06, "loss": 0.2438, "step": 163 }, { "epoch": 0.2071687983578083, "grad_norm": 0.16024184226989746, "learning_rate": 7e-06, "loss": 0.2088, "step": 164 }, { "epoch": 0.20843202273803885, "grad_norm": 0.1891951858997345, "learning_rate": 7e-06, "loss": 0.2751, "step": 165 }, { "epoch": 0.2096952471182694, "grad_norm": 0.08837898820638657, "learning_rate": 7e-06, "loss": 0.212, "step": 166 }, { "epoch": 0.21095847149849992, "grad_norm": 0.0905027762055397, "learning_rate": 7e-06, "loss": 0.2189, "step": 167 }, { "epoch": 0.21222169587873047, "grad_norm": 0.07917249947786331, "learning_rate": 7e-06, "loss": 0.2324, "step": 168 }, { "epoch": 0.213484920258961, "grad_norm": 0.13524577021598816, "learning_rate": 7e-06, "loss": 0.2143, "step": 169 }, { "epoch": 0.21474814463919153, "grad_norm": 0.13222923874855042, "learning_rate": 7e-06, "loss": 0.2983, "step": 170 }, { "epoch": 0.21601136901942208, "grad_norm": 0.1525893360376358, "learning_rate": 7e-06, "loss": 0.2408, "step": 171 }, { "epoch": 0.2172745933996526, "grad_norm": 0.08309401571750641, "learning_rate": 7e-06, "loss": 0.1722, "step": 172 }, { "epoch": 0.21853781777988315, "grad_norm": 0.08370368182659149, "learning_rate": 7e-06, "loss": 0.1981, "step": 173 }, { "epoch": 0.2198010421601137, "grad_norm": 0.11228370666503906, "learning_rate": 7e-06, "loss": 0.2336, "step": 174 }, { "epoch": 0.22106426654034422, "grad_norm": 0.19010692834854126, "learning_rate": 7e-06, "loss": 0.3069, "step": 175 }, { "epoch": 0.22232749092057477, "grad_norm": 0.08182361721992493, "learning_rate": 7e-06, "loss": 0.2549, "step": 176 }, { "epoch": 0.2235907153008053, "grad_norm": 0.1046992763876915, "learning_rate": 7e-06, "loss": 0.2458, "step": 177 }, { "epoch": 0.22485393968103584, "grad_norm": 0.11583778262138367, "learning_rate": 7e-06, "loss": 0.269, "step": 178 }, { "epoch": 0.2261171640612664, "grad_norm": 0.07805290818214417, "learning_rate": 7e-06, "loss": 0.1784, "step": 179 }, { "epoch": 0.2273803884414969, "grad_norm": 0.11022092401981354, "learning_rate": 7e-06, "loss": 0.232, "step": 180 }, { "epoch": 0.22864361282172746, "grad_norm": 0.1311209499835968, "learning_rate": 7e-06, "loss": 0.2603, "step": 181 }, { "epoch": 0.229906837201958, "grad_norm": 0.08558022975921631, "learning_rate": 7e-06, "loss": 0.2524, "step": 182 }, { "epoch": 0.23117006158218853, "grad_norm": 0.0957944467663765, "learning_rate": 7e-06, "loss": 0.281, "step": 183 }, { "epoch": 0.23243328596241908, "grad_norm": 0.086683489382267, "learning_rate": 7e-06, "loss": 0.2112, "step": 184 }, { "epoch": 0.2336965103426496, "grad_norm": 0.09485982358455658, "learning_rate": 7e-06, "loss": 0.2146, "step": 185 }, { "epoch": 0.23495973472288015, "grad_norm": 0.14843790233135223, "learning_rate": 7e-06, "loss": 0.2036, "step": 186 }, { "epoch": 0.2362229591031107, "grad_norm": 0.09375383704900742, "learning_rate": 7e-06, "loss": 0.2386, "step": 187 }, { "epoch": 0.23748618348334122, "grad_norm": 0.10639740526676178, "learning_rate": 7e-06, "loss": 0.2202, "step": 188 }, { "epoch": 0.23874940786357177, "grad_norm": 0.10205169022083282, "learning_rate": 7e-06, "loss": 0.2297, "step": 189 }, { "epoch": 0.24001263224380232, "grad_norm": 0.1138874888420105, "learning_rate": 7e-06, "loss": 0.2511, "step": 190 }, { "epoch": 0.24127585662403284, "grad_norm": 0.12742598354816437, "learning_rate": 7e-06, "loss": 0.2247, "step": 191 }, { "epoch": 0.24253908100426338, "grad_norm": 0.14605408906936646, "learning_rate": 7e-06, "loss": 0.2366, "step": 192 }, { "epoch": 0.24380230538449393, "grad_norm": 0.10053393989801407, "learning_rate": 7e-06, "loss": 0.4711, "step": 193 }, { "epoch": 0.24506552976472445, "grad_norm": 0.08829181641340256, "learning_rate": 7e-06, "loss": 0.1501, "step": 194 }, { "epoch": 0.246328754144955, "grad_norm": 0.1484231799840927, "learning_rate": 7e-06, "loss": 0.2063, "step": 195 }, { "epoch": 0.24759197852518552, "grad_norm": 0.17242765426635742, "learning_rate": 7e-06, "loss": 0.2317, "step": 196 }, { "epoch": 0.24885520290541607, "grad_norm": 0.12016981095075607, "learning_rate": 7e-06, "loss": 0.2272, "step": 197 }, { "epoch": 0.2501184272856466, "grad_norm": 0.1021333634853363, "learning_rate": 7e-06, "loss": 0.2402, "step": 198 }, { "epoch": 0.25138165166587717, "grad_norm": 0.11179149895906448, "learning_rate": 7e-06, "loss": 0.246, "step": 199 }, { "epoch": 0.2526448760461077, "grad_norm": 0.10811345279216766, "learning_rate": 7e-06, "loss": 0.2125, "step": 200 }, { "epoch": 0.2526448760461077, "eval_correct_accuracy": 0.5988125530110263, "eval_error_accuracy": 0.4524988743809095, "eval_f1": 0.5154742907624302, "eval_loss": 0.4177984297275543, "eval_runtime": 35.0506, "eval_samples_per_second": 97.003, "eval_steps_per_second": 6.077, "step": 200 }, { "epoch": 0.2539081004263382, "grad_norm": 0.12190552800893784, "learning_rate": 7e-06, "loss": 0.2372, "step": 201 }, { "epoch": 0.2551713248065688, "grad_norm": 0.07629604637622833, "learning_rate": 7e-06, "loss": 0.1976, "step": 202 }, { "epoch": 0.2564345491867993, "grad_norm": 0.10825781524181366, "learning_rate": 7e-06, "loss": 0.2169, "step": 203 }, { "epoch": 0.25769777356702983, "grad_norm": 0.09181591868400574, "learning_rate": 7e-06, "loss": 0.2225, "step": 204 }, { "epoch": 0.2589609979472604, "grad_norm": 0.1266108900308609, "learning_rate": 7e-06, "loss": 0.1858, "step": 205 }, { "epoch": 0.26022422232749093, "grad_norm": 0.11106186360120773, "learning_rate": 7e-06, "loss": 0.2443, "step": 206 }, { "epoch": 0.26148744670772145, "grad_norm": 0.11874532699584961, "learning_rate": 7e-06, "loss": 0.2224, "step": 207 }, { "epoch": 0.26275067108795197, "grad_norm": 0.06901393085718155, "learning_rate": 7e-06, "loss": 0.1683, "step": 208 }, { "epoch": 0.26401389546818255, "grad_norm": 0.1774539351463318, "learning_rate": 7e-06, "loss": 0.2588, "step": 209 }, { "epoch": 0.26527711984841307, "grad_norm": 0.06564710289239883, "learning_rate": 7e-06, "loss": 0.1966, "step": 210 }, { "epoch": 0.2665403442286436, "grad_norm": 0.1348266899585724, "learning_rate": 7e-06, "loss": 0.2094, "step": 211 }, { "epoch": 0.26780356860887417, "grad_norm": 0.10280844569206238, "learning_rate": 7e-06, "loss": 0.2208, "step": 212 }, { "epoch": 0.2690667929891047, "grad_norm": 0.09777519851922989, "learning_rate": 7e-06, "loss": 0.2259, "step": 213 }, { "epoch": 0.2703300173693352, "grad_norm": 0.11480893194675446, "learning_rate": 7e-06, "loss": 0.2402, "step": 214 }, { "epoch": 0.2715932417495658, "grad_norm": 0.17719541490077972, "learning_rate": 7e-06, "loss": 0.2692, "step": 215 }, { "epoch": 0.2728564661297963, "grad_norm": 0.07069459557533264, "learning_rate": 7e-06, "loss": 0.1781, "step": 216 }, { "epoch": 0.2741196905100268, "grad_norm": 0.06251855194568634, "learning_rate": 7e-06, "loss": 0.1819, "step": 217 }, { "epoch": 0.2753829148902574, "grad_norm": 0.1753867119550705, "learning_rate": 7e-06, "loss": 0.2362, "step": 218 }, { "epoch": 0.2766461392704879, "grad_norm": 0.1843274086713791, "learning_rate": 7e-06, "loss": 0.2638, "step": 219 }, { "epoch": 0.27790936365071844, "grad_norm": 0.18026292324066162, "learning_rate": 7e-06, "loss": 0.2274, "step": 220 }, { "epoch": 0.279172588030949, "grad_norm": 0.0640600174665451, "learning_rate": 7e-06, "loss": 0.3739, "step": 221 }, { "epoch": 0.28043581241117954, "grad_norm": 0.091743104159832, "learning_rate": 7e-06, "loss": 0.2274, "step": 222 }, { "epoch": 0.28169903679141006, "grad_norm": 0.10185891389846802, "learning_rate": 7e-06, "loss": 0.471, "step": 223 }, { "epoch": 0.28296226117164064, "grad_norm": 0.08672218769788742, "learning_rate": 7e-06, "loss": 0.2171, "step": 224 }, { "epoch": 0.28422548555187116, "grad_norm": 0.11758771538734436, "learning_rate": 7e-06, "loss": 0.2211, "step": 225 }, { "epoch": 0.2854887099321017, "grad_norm": 0.07176447659730911, "learning_rate": 7e-06, "loss": 0.1967, "step": 226 }, { "epoch": 0.2867519343123322, "grad_norm": 0.1037454828619957, "learning_rate": 7e-06, "loss": 0.2457, "step": 227 }, { "epoch": 0.2880151586925628, "grad_norm": 0.07262658327817917, "learning_rate": 7e-06, "loss": 0.2026, "step": 228 }, { "epoch": 0.2892783830727933, "grad_norm": 0.13171784579753876, "learning_rate": 7e-06, "loss": 0.209, "step": 229 }, { "epoch": 0.2905416074530238, "grad_norm": 0.08208411931991577, "learning_rate": 7e-06, "loss": 0.1964, "step": 230 }, { "epoch": 0.2918048318332544, "grad_norm": 0.10370495170354843, "learning_rate": 7e-06, "loss": 0.2202, "step": 231 }, { "epoch": 0.2930680562134849, "grad_norm": 0.26831239461898804, "learning_rate": 7e-06, "loss": 0.2651, "step": 232 }, { "epoch": 0.29433128059371544, "grad_norm": 0.12230344116687775, "learning_rate": 7e-06, "loss": 0.2265, "step": 233 }, { "epoch": 0.295594504973946, "grad_norm": 0.08064734190702438, "learning_rate": 7e-06, "loss": 0.1711, "step": 234 }, { "epoch": 0.29685772935417654, "grad_norm": 0.10691053420305252, "learning_rate": 7e-06, "loss": 0.1753, "step": 235 }, { "epoch": 0.29812095373440706, "grad_norm": 0.08961788564920425, "learning_rate": 7e-06, "loss": 0.2682, "step": 236 }, { "epoch": 0.29938417811463763, "grad_norm": 0.2417578548192978, "learning_rate": 7e-06, "loss": 0.2622, "step": 237 }, { "epoch": 0.30064740249486815, "grad_norm": 0.09739197045564651, "learning_rate": 7e-06, "loss": 0.1747, "step": 238 }, { "epoch": 0.3019106268750987, "grad_norm": 0.15415729582309723, "learning_rate": 7e-06, "loss": 0.2289, "step": 239 }, { "epoch": 0.30317385125532925, "grad_norm": 0.08798956125974655, "learning_rate": 7e-06, "loss": 0.2076, "step": 240 }, { "epoch": 0.3044370756355598, "grad_norm": 0.09532306343317032, "learning_rate": 7e-06, "loss": 0.3761, "step": 241 }, { "epoch": 0.3057003000157903, "grad_norm": 0.06419141590595245, "learning_rate": 7e-06, "loss": 0.2308, "step": 242 }, { "epoch": 0.30696352439602087, "grad_norm": 0.13766047358512878, "learning_rate": 7e-06, "loss": 0.2203, "step": 243 }, { "epoch": 0.3082267487762514, "grad_norm": 0.09225375950336456, "learning_rate": 7e-06, "loss": 0.2023, "step": 244 }, { "epoch": 0.3094899731564819, "grad_norm": 0.1266135275363922, "learning_rate": 7e-06, "loss": 0.2823, "step": 245 }, { "epoch": 0.31075319753671243, "grad_norm": 0.17997467517852783, "learning_rate": 7e-06, "loss": 0.225, "step": 246 }, { "epoch": 0.312016421916943, "grad_norm": 0.12776713073253632, "learning_rate": 7e-06, "loss": 0.1906, "step": 247 }, { "epoch": 0.31327964629717353, "grad_norm": 0.14866380393505096, "learning_rate": 7e-06, "loss": 0.2119, "step": 248 }, { "epoch": 0.31454287067740405, "grad_norm": 0.11824511736631393, "learning_rate": 7e-06, "loss": 0.2219, "step": 249 }, { "epoch": 0.31580609505763463, "grad_norm": 0.14409460127353668, "learning_rate": 7e-06, "loss": 0.2116, "step": 250 }, { "epoch": 0.31706931943786515, "grad_norm": 0.10304541140794754, "learning_rate": 7e-06, "loss": 0.2073, "step": 251 }, { "epoch": 0.31833254381809567, "grad_norm": 0.09163326770067215, "learning_rate": 7e-06, "loss": 0.1882, "step": 252 }, { "epoch": 0.31959576819832625, "grad_norm": 0.12692378461360931, "learning_rate": 7e-06, "loss": 0.2386, "step": 253 }, { "epoch": 0.32085899257855677, "grad_norm": 0.1747879534959793, "learning_rate": 7e-06, "loss": 0.2054, "step": 254 }, { "epoch": 0.3221222169587873, "grad_norm": 0.12346009910106659, "learning_rate": 7e-06, "loss": 0.2397, "step": 255 }, { "epoch": 0.32338544133901787, "grad_norm": 0.1731298863887787, "learning_rate": 7e-06, "loss": 0.2575, "step": 256 }, { "epoch": 0.3246486657192484, "grad_norm": 0.08011125028133392, "learning_rate": 7e-06, "loss": 0.215, "step": 257 }, { "epoch": 0.3259118900994789, "grad_norm": 0.13160613179206848, "learning_rate": 7e-06, "loss": 0.222, "step": 258 }, { "epoch": 0.3271751144797095, "grad_norm": 0.18522977828979492, "learning_rate": 7e-06, "loss": 0.2548, "step": 259 }, { "epoch": 0.32843833885994, "grad_norm": 0.14212659001350403, "learning_rate": 7e-06, "loss": 0.3002, "step": 260 }, { "epoch": 0.3297015632401705, "grad_norm": 0.13445697724819183, "learning_rate": 7e-06, "loss": 0.2351, "step": 261 }, { "epoch": 0.33096478762040105, "grad_norm": 0.11636935919523239, "learning_rate": 7e-06, "loss": 0.2106, "step": 262 }, { "epoch": 0.3322280120006316, "grad_norm": 0.14159604907035828, "learning_rate": 7e-06, "loss": 0.2531, "step": 263 }, { "epoch": 0.33349123638086214, "grad_norm": 0.10319356620311737, "learning_rate": 7e-06, "loss": 0.2346, "step": 264 }, { "epoch": 0.33475446076109266, "grad_norm": 0.09796885401010513, "learning_rate": 7e-06, "loss": 0.2059, "step": 265 }, { "epoch": 0.33601768514132324, "grad_norm": 0.1082499697804451, "learning_rate": 7e-06, "loss": 0.1863, "step": 266 }, { "epoch": 0.33728090952155376, "grad_norm": 0.12492396682500839, "learning_rate": 7e-06, "loss": 0.215, "step": 267 }, { "epoch": 0.3385441339017843, "grad_norm": 0.06617411971092224, "learning_rate": 7e-06, "loss": 0.1782, "step": 268 }, { "epoch": 0.33980735828201486, "grad_norm": 0.15060101449489594, "learning_rate": 7e-06, "loss": 0.3509, "step": 269 }, { "epoch": 0.3410705826622454, "grad_norm": 0.11944282054901123, "learning_rate": 7e-06, "loss": 0.1862, "step": 270 }, { "epoch": 0.3423338070424759, "grad_norm": 0.12389136850833893, "learning_rate": 7e-06, "loss": 0.2409, "step": 271 }, { "epoch": 0.3435970314227065, "grad_norm": 0.11000983417034149, "learning_rate": 7e-06, "loss": 0.223, "step": 272 }, { "epoch": 0.344860255802937, "grad_norm": 0.09012436866760254, "learning_rate": 7e-06, "loss": 0.2068, "step": 273 }, { "epoch": 0.3461234801831675, "grad_norm": 0.27014490962028503, "learning_rate": 7e-06, "loss": 0.2087, "step": 274 }, { "epoch": 0.3473867045633981, "grad_norm": 0.08035814762115479, "learning_rate": 7e-06, "loss": 0.262, "step": 275 }, { "epoch": 0.3486499289436286, "grad_norm": 0.09129905700683594, "learning_rate": 7e-06, "loss": 0.2082, "step": 276 }, { "epoch": 0.34991315332385914, "grad_norm": 0.11665099114179611, "learning_rate": 7e-06, "loss": 0.1841, "step": 277 }, { "epoch": 0.3511763777040897, "grad_norm": 0.08812276273965836, "learning_rate": 7e-06, "loss": 0.2164, "step": 278 }, { "epoch": 0.35243960208432024, "grad_norm": 0.1272403746843338, "learning_rate": 7e-06, "loss": 0.1927, "step": 279 }, { "epoch": 0.35370282646455076, "grad_norm": 0.11256379634141922, "learning_rate": 7e-06, "loss": 0.2991, "step": 280 }, { "epoch": 0.3549660508447813, "grad_norm": 0.15795424580574036, "learning_rate": 7e-06, "loss": 0.2363, "step": 281 }, { "epoch": 0.35622927522501185, "grad_norm": 0.22632326185703278, "learning_rate": 7e-06, "loss": 0.2088, "step": 282 }, { "epoch": 0.3574924996052424, "grad_norm": 0.18535131216049194, "learning_rate": 7e-06, "loss": 0.2746, "step": 283 }, { "epoch": 0.3587557239854729, "grad_norm": 0.08579732477664948, "learning_rate": 7e-06, "loss": 0.1899, "step": 284 }, { "epoch": 0.3600189483657035, "grad_norm": 0.10859379917383194, "learning_rate": 7e-06, "loss": 0.2067, "step": 285 }, { "epoch": 0.361282172745934, "grad_norm": 0.07765299826860428, "learning_rate": 7e-06, "loss": 0.1761, "step": 286 }, { "epoch": 0.3625453971261645, "grad_norm": 0.17053595185279846, "learning_rate": 7e-06, "loss": 0.2373, "step": 287 }, { "epoch": 0.3638086215063951, "grad_norm": 0.09873699396848679, "learning_rate": 7e-06, "loss": 0.2176, "step": 288 }, { "epoch": 0.3650718458866256, "grad_norm": 0.07418286055326462, "learning_rate": 7e-06, "loss": 0.1797, "step": 289 }, { "epoch": 0.36633507026685613, "grad_norm": 0.11981359124183655, "learning_rate": 7e-06, "loss": 0.1988, "step": 290 }, { "epoch": 0.3675982946470867, "grad_norm": 0.06424502283334732, "learning_rate": 7e-06, "loss": 0.2121, "step": 291 }, { "epoch": 0.36886151902731723, "grad_norm": 0.09006607532501221, "learning_rate": 7e-06, "loss": 0.1945, "step": 292 }, { "epoch": 0.37012474340754775, "grad_norm": 0.10973497480154037, "learning_rate": 7e-06, "loss": 0.2046, "step": 293 }, { "epoch": 0.37138796778777833, "grad_norm": 0.09228470921516418, "learning_rate": 7e-06, "loss": 0.207, "step": 294 }, { "epoch": 0.37265119216800885, "grad_norm": 0.10961271822452545, "learning_rate": 7e-06, "loss": 0.2128, "step": 295 }, { "epoch": 0.37391441654823937, "grad_norm": 0.09072300046682358, "learning_rate": 7e-06, "loss": 0.4585, "step": 296 }, { "epoch": 0.37517764092846995, "grad_norm": 0.08374742418527603, "learning_rate": 7e-06, "loss": 0.2178, "step": 297 }, { "epoch": 0.37644086530870047, "grad_norm": 0.05344458296895027, "learning_rate": 7e-06, "loss": 0.1595, "step": 298 }, { "epoch": 0.377704089688931, "grad_norm": 0.07841549813747406, "learning_rate": 7e-06, "loss": 0.2306, "step": 299 }, { "epoch": 0.3789673140691615, "grad_norm": 0.09865035116672516, "learning_rate": 7e-06, "loss": 0.2274, "step": 300 }, { "epoch": 0.3789673140691615, "eval_correct_accuracy": 0.5818490245971162, "eval_error_accuracy": 0.494822152183701, "eval_f1": 0.5348184158843582, "eval_loss": 0.41273096203804016, "eval_runtime": 35.0595, "eval_samples_per_second": 96.978, "eval_steps_per_second": 6.075, "step": 300 }, { "epoch": 0.3802305384493921, "grad_norm": 0.11520479619503021, "learning_rate": 7e-06, "loss": 0.194, "step": 301 }, { "epoch": 0.3814937628296226, "grad_norm": 0.17391149699687958, "learning_rate": 7e-06, "loss": 0.2053, "step": 302 }, { "epoch": 0.38275698720985313, "grad_norm": 0.08927040547132492, "learning_rate": 7e-06, "loss": 0.1885, "step": 303 }, { "epoch": 0.3840202115900837, "grad_norm": 0.10747874528169632, "learning_rate": 7e-06, "loss": 0.2357, "step": 304 }, { "epoch": 0.3852834359703142, "grad_norm": 0.0821816474199295, "learning_rate": 7e-06, "loss": 0.2017, "step": 305 }, { "epoch": 0.38654666035054475, "grad_norm": 0.08718965202569962, "learning_rate": 7e-06, "loss": 0.2333, "step": 306 }, { "epoch": 0.3878098847307753, "grad_norm": 0.14753767848014832, "learning_rate": 7e-06, "loss": 0.2501, "step": 307 }, { "epoch": 0.38907310911100584, "grad_norm": 0.12474358081817627, "learning_rate": 7e-06, "loss": 0.1895, "step": 308 }, { "epoch": 0.39033633349123636, "grad_norm": 0.14409278333187103, "learning_rate": 7e-06, "loss": 0.208, "step": 309 }, { "epoch": 0.39159955787146694, "grad_norm": 0.06918184459209442, "learning_rate": 7e-06, "loss": 0.1817, "step": 310 }, { "epoch": 0.39286278225169746, "grad_norm": 0.08502199500799179, "learning_rate": 7e-06, "loss": 0.1832, "step": 311 }, { "epoch": 0.394126006631928, "grad_norm": 0.06989938765764236, "learning_rate": 7e-06, "loss": 0.1741, "step": 312 }, { "epoch": 0.39538923101215856, "grad_norm": 0.08131398260593414, "learning_rate": 7e-06, "loss": 0.226, "step": 313 }, { "epoch": 0.3966524553923891, "grad_norm": 0.16150841116905212, "learning_rate": 7e-06, "loss": 0.2081, "step": 314 }, { "epoch": 0.3979156797726196, "grad_norm": 0.10033854097127914, "learning_rate": 7e-06, "loss": 0.1757, "step": 315 }, { "epoch": 0.3991789041528501, "grad_norm": 0.2944275438785553, "learning_rate": 7e-06, "loss": 0.2039, "step": 316 }, { "epoch": 0.4004421285330807, "grad_norm": 0.09300543367862701, "learning_rate": 7e-06, "loss": 0.2403, "step": 317 }, { "epoch": 0.4017053529133112, "grad_norm": 0.089630626142025, "learning_rate": 7e-06, "loss": 0.2457, "step": 318 }, { "epoch": 0.40296857729354174, "grad_norm": 0.06648046523332596, "learning_rate": 7e-06, "loss": 0.2155, "step": 319 }, { "epoch": 0.4042318016737723, "grad_norm": 0.18262338638305664, "learning_rate": 7e-06, "loss": 0.2087, "step": 320 }, { "epoch": 0.40549502605400284, "grad_norm": 0.0919061154127121, "learning_rate": 7e-06, "loss": 0.2062, "step": 321 }, { "epoch": 0.40675825043423336, "grad_norm": 0.113703154027462, "learning_rate": 7e-06, "loss": 0.1859, "step": 322 }, { "epoch": 0.40802147481446394, "grad_norm": 0.20705194771289825, "learning_rate": 7e-06, "loss": 0.1769, "step": 323 }, { "epoch": 0.40928469919469446, "grad_norm": 0.11209185421466827, "learning_rate": 7e-06, "loss": 0.1819, "step": 324 }, { "epoch": 0.410547923574925, "grad_norm": 0.05803574621677399, "learning_rate": 7e-06, "loss": 0.1852, "step": 325 }, { "epoch": 0.41181114795515555, "grad_norm": 0.16077323257923126, "learning_rate": 7e-06, "loss": 0.2117, "step": 326 }, { "epoch": 0.4130743723353861, "grad_norm": 0.10078177601099014, "learning_rate": 7e-06, "loss": 0.193, "step": 327 }, { "epoch": 0.4143375967156166, "grad_norm": 0.09989168494939804, "learning_rate": 7e-06, "loss": 0.2053, "step": 328 }, { "epoch": 0.4156008210958472, "grad_norm": 0.13987579941749573, "learning_rate": 7e-06, "loss": 0.2678, "step": 329 }, { "epoch": 0.4168640454760777, "grad_norm": 0.13039669394493103, "learning_rate": 7e-06, "loss": 0.1998, "step": 330 }, { "epoch": 0.4181272698563082, "grad_norm": 0.1029522716999054, "learning_rate": 7e-06, "loss": 0.2337, "step": 331 }, { "epoch": 0.4193904942365388, "grad_norm": 0.08752740174531937, "learning_rate": 7e-06, "loss": 0.1854, "step": 332 }, { "epoch": 0.4206537186167693, "grad_norm": 0.07876112312078476, "learning_rate": 7e-06, "loss": 0.1909, "step": 333 }, { "epoch": 0.42191694299699983, "grad_norm": 0.2126246988773346, "learning_rate": 7e-06, "loss": 0.2535, "step": 334 }, { "epoch": 0.42318016737723035, "grad_norm": 0.11913909763097763, "learning_rate": 7e-06, "loss": 0.2184, "step": 335 }, { "epoch": 0.42444339175746093, "grad_norm": 0.1513642817735672, "learning_rate": 7e-06, "loss": 0.1994, "step": 336 }, { "epoch": 0.42570661613769145, "grad_norm": 0.1306588500738144, "learning_rate": 7e-06, "loss": 0.2126, "step": 337 }, { "epoch": 0.426969840517922, "grad_norm": 0.1171175092458725, "learning_rate": 7e-06, "loss": 0.1994, "step": 338 }, { "epoch": 0.42823306489815255, "grad_norm": 0.05895727127790451, "learning_rate": 7e-06, "loss": 0.155, "step": 339 }, { "epoch": 0.42949628927838307, "grad_norm": 0.08570288121700287, "learning_rate": 7e-06, "loss": 0.1986, "step": 340 }, { "epoch": 0.4307595136586136, "grad_norm": 0.0765470489859581, "learning_rate": 7e-06, "loss": 0.1696, "step": 341 }, { "epoch": 0.43202273803884417, "grad_norm": 0.08286664634943008, "learning_rate": 7e-06, "loss": 0.1626, "step": 342 }, { "epoch": 0.4332859624190747, "grad_norm": 0.2282284051179886, "learning_rate": 7e-06, "loss": 0.1877, "step": 343 }, { "epoch": 0.4345491867993052, "grad_norm": 0.11943413317203522, "learning_rate": 7e-06, "loss": 0.1897, "step": 344 }, { "epoch": 0.4358124111795358, "grad_norm": 0.10935524851083755, "learning_rate": 7e-06, "loss": 0.1828, "step": 345 }, { "epoch": 0.4370756355597663, "grad_norm": 0.07996437698602676, "learning_rate": 7e-06, "loss": 0.1991, "step": 346 }, { "epoch": 0.4383388599399968, "grad_norm": 0.13088780641555786, "learning_rate": 7e-06, "loss": 0.3873, "step": 347 }, { "epoch": 0.4396020843202274, "grad_norm": 0.15082432329654694, "learning_rate": 7e-06, "loss": 0.2091, "step": 348 }, { "epoch": 0.4408653087004579, "grad_norm": 0.132376566529274, "learning_rate": 7e-06, "loss": 0.2471, "step": 349 }, { "epoch": 0.44212853308068845, "grad_norm": 0.07796452194452286, "learning_rate": 7e-06, "loss": 0.1751, "step": 350 }, { "epoch": 0.443391757460919, "grad_norm": 0.12849055230617523, "learning_rate": 7e-06, "loss": 0.3155, "step": 351 }, { "epoch": 0.44465498184114954, "grad_norm": 0.06422396749258041, "learning_rate": 7e-06, "loss": 0.1486, "step": 352 }, { "epoch": 0.44591820622138006, "grad_norm": 0.1800646334886551, "learning_rate": 7e-06, "loss": 0.1993, "step": 353 }, { "epoch": 0.4471814306016106, "grad_norm": 0.15747664868831635, "learning_rate": 7e-06, "loss": 0.209, "step": 354 }, { "epoch": 0.44844465498184116, "grad_norm": 0.11023043096065521, "learning_rate": 7e-06, "loss": 0.21, "step": 355 }, { "epoch": 0.4497078793620717, "grad_norm": 0.0927424430847168, "learning_rate": 7e-06, "loss": 0.1745, "step": 356 }, { "epoch": 0.4509711037423022, "grad_norm": 0.08278126269578934, "learning_rate": 7e-06, "loss": 0.3105, "step": 357 }, { "epoch": 0.4522343281225328, "grad_norm": 0.08794251829385757, "learning_rate": 7e-06, "loss": 0.1979, "step": 358 }, { "epoch": 0.4534975525027633, "grad_norm": 0.11653570830821991, "learning_rate": 7e-06, "loss": 0.1828, "step": 359 }, { "epoch": 0.4547607768829938, "grad_norm": 0.11114069074392319, "learning_rate": 7e-06, "loss": 0.1826, "step": 360 }, { "epoch": 0.4560240012632244, "grad_norm": 0.2608173191547394, "learning_rate": 7e-06, "loss": 0.2304, "step": 361 }, { "epoch": 0.4572872256434549, "grad_norm": 0.08441725373268127, "learning_rate": 7e-06, "loss": 0.1757, "step": 362 }, { "epoch": 0.45855045002368544, "grad_norm": 0.10891429334878922, "learning_rate": 7e-06, "loss": 0.2122, "step": 363 }, { "epoch": 0.459813674403916, "grad_norm": 0.07106776535511017, "learning_rate": 7e-06, "loss": 0.1721, "step": 364 }, { "epoch": 0.46107689878414654, "grad_norm": 0.08842181414365768, "learning_rate": 7e-06, "loss": 0.2226, "step": 365 }, { "epoch": 0.46234012316437706, "grad_norm": 0.0870131179690361, "learning_rate": 7e-06, "loss": 0.2474, "step": 366 }, { "epoch": 0.46360334754460764, "grad_norm": 0.14521507918834686, "learning_rate": 7e-06, "loss": 0.2615, "step": 367 }, { "epoch": 0.46486657192483816, "grad_norm": 0.09553767740726471, "learning_rate": 7e-06, "loss": 0.1791, "step": 368 }, { "epoch": 0.4661297963050687, "grad_norm": 0.11010967195034027, "learning_rate": 7e-06, "loss": 0.1874, "step": 369 }, { "epoch": 0.4673930206852992, "grad_norm": 0.09533923864364624, "learning_rate": 7e-06, "loss": 0.228, "step": 370 }, { "epoch": 0.4686562450655298, "grad_norm": 0.0890774354338646, "learning_rate": 7e-06, "loss": 0.2345, "step": 371 }, { "epoch": 0.4699194694457603, "grad_norm": 0.12173017859458923, "learning_rate": 7e-06, "loss": 0.2082, "step": 372 }, { "epoch": 0.4711826938259908, "grad_norm": 0.0602993369102478, "learning_rate": 7e-06, "loss": 0.1893, "step": 373 }, { "epoch": 0.4724459182062214, "grad_norm": 0.13122287392616272, "learning_rate": 7e-06, "loss": 0.2178, "step": 374 }, { "epoch": 0.4737091425864519, "grad_norm": 0.07299527525901794, "learning_rate": 7e-06, "loss": 0.1888, "step": 375 }, { "epoch": 0.47497236696668244, "grad_norm": 0.08244926482439041, "learning_rate": 7e-06, "loss": 0.174, "step": 376 }, { "epoch": 0.476235591346913, "grad_norm": 0.08397851884365082, "learning_rate": 7e-06, "loss": 0.2108, "step": 377 }, { "epoch": 0.47749881572714353, "grad_norm": 0.07320383936166763, "learning_rate": 7e-06, "loss": 0.1804, "step": 378 }, { "epoch": 0.47876204010737405, "grad_norm": 0.0849589854478836, "learning_rate": 7e-06, "loss": 0.1829, "step": 379 }, { "epoch": 0.48002526448760463, "grad_norm": 0.10207744687795639, "learning_rate": 7e-06, "loss": 0.2174, "step": 380 }, { "epoch": 0.48128848886783515, "grad_norm": 0.07175120711326599, "learning_rate": 7e-06, "loss": 0.1847, "step": 381 }, { "epoch": 0.4825517132480657, "grad_norm": 0.10446271300315857, "learning_rate": 7e-06, "loss": 0.216, "step": 382 }, { "epoch": 0.48381493762829625, "grad_norm": 0.23799718916416168, "learning_rate": 7e-06, "loss": 0.2136, "step": 383 }, { "epoch": 0.48507816200852677, "grad_norm": 0.11531874537467957, "learning_rate": 7e-06, "loss": 0.2357, "step": 384 }, { "epoch": 0.4863413863887573, "grad_norm": 0.10034700483083725, "learning_rate": 7e-06, "loss": 0.2258, "step": 385 }, { "epoch": 0.48760461076898787, "grad_norm": 0.0934348776936531, "learning_rate": 7e-06, "loss": 0.284, "step": 386 }, { "epoch": 0.4888678351492184, "grad_norm": 0.165315181016922, "learning_rate": 7e-06, "loss": 0.2264, "step": 387 }, { "epoch": 0.4901310595294489, "grad_norm": 0.1086471751332283, "learning_rate": 7e-06, "loss": 0.2028, "step": 388 }, { "epoch": 0.49139428390967943, "grad_norm": 0.14764176309108734, "learning_rate": 7e-06, "loss": 0.1775, "step": 389 }, { "epoch": 0.49265750828991, "grad_norm": 0.21734580397605896, "learning_rate": 7e-06, "loss": 0.1924, "step": 390 }, { "epoch": 0.4939207326701405, "grad_norm": 0.0923137441277504, "learning_rate": 7e-06, "loss": 0.2031, "step": 391 }, { "epoch": 0.49518395705037105, "grad_norm": 0.06933951377868652, "learning_rate": 7e-06, "loss": 0.1739, "step": 392 }, { "epoch": 0.4964471814306016, "grad_norm": 0.0930216833949089, "learning_rate": 7e-06, "loss": 0.2083, "step": 393 }, { "epoch": 0.49771040581083215, "grad_norm": 0.08797884732484818, "learning_rate": 7e-06, "loss": 0.2636, "step": 394 }, { "epoch": 0.49897363019106267, "grad_norm": 0.0919070690870285, "learning_rate": 7e-06, "loss": 0.2154, "step": 395 }, { "epoch": 0.5002368545712932, "grad_norm": 0.07787168025970459, "learning_rate": 7e-06, "loss": 0.2207, "step": 396 }, { "epoch": 0.5015000789515237, "grad_norm": 0.11572758853435516, "learning_rate": 7e-06, "loss": 0.2002, "step": 397 }, { "epoch": 0.5027633033317543, "grad_norm": 0.08295108377933502, "learning_rate": 7e-06, "loss": 0.2614, "step": 398 }, { "epoch": 0.5040265277119849, "grad_norm": 0.0625801831483841, "learning_rate": 7e-06, "loss": 0.1644, "step": 399 }, { "epoch": 0.5052897520922154, "grad_norm": 0.07405094802379608, "learning_rate": 7e-06, "loss": 0.2234, "step": 400 }, { "epoch": 0.5052897520922154, "eval_correct_accuracy": 0.5852417302798982, "eval_error_accuracy": 0.4709590274651058, "eval_f1": 0.5219175883059916, "eval_loss": 0.4229665994644165, "eval_runtime": 35.2003, "eval_samples_per_second": 96.59, "eval_steps_per_second": 6.051, "step": 400 }, { "epoch": 0.5065529764724459, "grad_norm": 0.09175197780132294, "learning_rate": 7e-06, "loss": 0.2027, "step": 401 }, { "epoch": 0.5078162008526764, "grad_norm": 0.1550239622592926, "learning_rate": 7e-06, "loss": 0.2559, "step": 402 }, { "epoch": 0.509079425232907, "grad_norm": 0.139438658952713, "learning_rate": 7e-06, "loss": 0.1762, "step": 403 }, { "epoch": 0.5103426496131376, "grad_norm": 0.11481575667858124, "learning_rate": 7e-06, "loss": 0.1623, "step": 404 }, { "epoch": 0.5116058739933681, "grad_norm": 0.05404340475797653, "learning_rate": 7e-06, "loss": 0.1961, "step": 405 }, { "epoch": 0.5128690983735986, "grad_norm": 0.14743672311306, "learning_rate": 7e-06, "loss": 0.2279, "step": 406 }, { "epoch": 0.5141323227538291, "grad_norm": 0.11647465080022812, "learning_rate": 7e-06, "loss": 0.2001, "step": 407 }, { "epoch": 0.5153955471340597, "grad_norm": 0.08203577995300293, "learning_rate": 7e-06, "loss": 0.1752, "step": 408 }, { "epoch": 0.5166587715142902, "grad_norm": 0.11073414981365204, "learning_rate": 7e-06, "loss": 0.3686, "step": 409 }, { "epoch": 0.5179219958945208, "grad_norm": 0.11331301182508469, "learning_rate": 7e-06, "loss": 0.3378, "step": 410 }, { "epoch": 0.5191852202747513, "grad_norm": 0.09435959905385971, "learning_rate": 7e-06, "loss": 0.1562, "step": 411 }, { "epoch": 0.5204484446549819, "grad_norm": 0.08365237712860107, "learning_rate": 7e-06, "loss": 0.1926, "step": 412 }, { "epoch": 0.5217116690352124, "grad_norm": 0.08092326670885086, "learning_rate": 7e-06, "loss": 0.1931, "step": 413 }, { "epoch": 0.5229748934154429, "grad_norm": 0.07763849943876266, "learning_rate": 7e-06, "loss": 0.2038, "step": 414 }, { "epoch": 0.5242381177956734, "grad_norm": 0.1350603550672531, "learning_rate": 7e-06, "loss": 0.2392, "step": 415 }, { "epoch": 0.5255013421759039, "grad_norm": 0.10287491232156754, "learning_rate": 7e-06, "loss": 0.196, "step": 416 }, { "epoch": 0.5267645665561346, "grad_norm": 0.0719987079501152, "learning_rate": 7e-06, "loss": 0.2016, "step": 417 }, { "epoch": 0.5280277909363651, "grad_norm": 0.22227227687835693, "learning_rate": 7e-06, "loss": 0.2131, "step": 418 }, { "epoch": 0.5292910153165956, "grad_norm": 0.06136275455355644, "learning_rate": 7e-06, "loss": 0.2044, "step": 419 }, { "epoch": 0.5305542396968261, "grad_norm": 0.0627446323633194, "learning_rate": 7e-06, "loss": 0.1793, "step": 420 }, { "epoch": 0.5318174640770567, "grad_norm": 0.20960237085819244, "learning_rate": 7e-06, "loss": 0.1982, "step": 421 }, { "epoch": 0.5330806884572872, "grad_norm": 0.11971580237150192, "learning_rate": 7e-06, "loss": 0.2587, "step": 422 }, { "epoch": 0.5343439128375178, "grad_norm": 0.0932474359869957, "learning_rate": 7e-06, "loss": 0.1947, "step": 423 }, { "epoch": 0.5356071372177483, "grad_norm": 0.09686949849128723, "learning_rate": 7e-06, "loss": 0.1907, "step": 424 }, { "epoch": 0.5368703615979789, "grad_norm": 0.07940957695245743, "learning_rate": 7e-06, "loss": 0.2535, "step": 425 }, { "epoch": 0.5381335859782094, "grad_norm": 0.09676375240087509, "learning_rate": 7e-06, "loss": 0.2038, "step": 426 }, { "epoch": 0.5393968103584399, "grad_norm": 0.09989267587661743, "learning_rate": 7e-06, "loss": 0.1976, "step": 427 }, { "epoch": 0.5406600347386704, "grad_norm": 0.0823327898979187, "learning_rate": 7e-06, "loss": 0.1708, "step": 428 }, { "epoch": 0.541923259118901, "grad_norm": 0.10633084177970886, "learning_rate": 7e-06, "loss": 0.1619, "step": 429 }, { "epoch": 0.5431864834991316, "grad_norm": 0.08448205143213272, "learning_rate": 7e-06, "loss": 0.1854, "step": 430 }, { "epoch": 0.5444497078793621, "grad_norm": 0.07697522640228271, "learning_rate": 7e-06, "loss": 0.188, "step": 431 }, { "epoch": 0.5457129322595926, "grad_norm": 0.1970750391483307, "learning_rate": 7e-06, "loss": 0.2479, "step": 432 }, { "epoch": 0.5469761566398231, "grad_norm": 0.08660274744033813, "learning_rate": 7e-06, "loss": 0.1536, "step": 433 }, { "epoch": 0.5482393810200537, "grad_norm": 0.0931171253323555, "learning_rate": 7e-06, "loss": 0.2169, "step": 434 }, { "epoch": 0.5495026054002842, "grad_norm": 0.07317376136779785, "learning_rate": 7e-06, "loss": 0.1941, "step": 435 }, { "epoch": 0.5507658297805148, "grad_norm": 0.07506151497364044, "learning_rate": 7e-06, "loss": 0.1761, "step": 436 }, { "epoch": 0.5520290541607453, "grad_norm": 0.059854380786418915, "learning_rate": 7e-06, "loss": 0.3068, "step": 437 }, { "epoch": 0.5532922785409758, "grad_norm": 0.1609865128993988, "learning_rate": 7e-06, "loss": 0.1857, "step": 438 }, { "epoch": 0.5545555029212064, "grad_norm": 0.08996118605136871, "learning_rate": 7e-06, "loss": 0.1805, "step": 439 }, { "epoch": 0.5558187273014369, "grad_norm": 0.12089324742555618, "learning_rate": 7e-06, "loss": 0.1826, "step": 440 }, { "epoch": 0.5570819516816674, "grad_norm": 0.08772964775562286, "learning_rate": 7e-06, "loss": 0.1791, "step": 441 }, { "epoch": 0.558345176061898, "grad_norm": 0.10977458208799362, "learning_rate": 7e-06, "loss": 0.2084, "step": 442 }, { "epoch": 0.5596084004421286, "grad_norm": 0.09188458323478699, "learning_rate": 7e-06, "loss": 0.3045, "step": 443 }, { "epoch": 0.5608716248223591, "grad_norm": 0.07033522427082062, "learning_rate": 7e-06, "loss": 0.1926, "step": 444 }, { "epoch": 0.5621348492025896, "grad_norm": 0.0652671530842781, "learning_rate": 7e-06, "loss": 0.1998, "step": 445 }, { "epoch": 0.5633980735828201, "grad_norm": 0.07860173285007477, "learning_rate": 7e-06, "loss": 0.2172, "step": 446 }, { "epoch": 0.5646612979630506, "grad_norm": 0.0679745227098465, "learning_rate": 7e-06, "loss": 0.179, "step": 447 }, { "epoch": 0.5659245223432813, "grad_norm": 0.10545714199542999, "learning_rate": 7e-06, "loss": 0.1547, "step": 448 }, { "epoch": 0.5671877467235118, "grad_norm": 0.07516340911388397, "learning_rate": 7e-06, "loss": 0.1643, "step": 449 }, { "epoch": 0.5684509711037423, "grad_norm": 0.06046690791845322, "learning_rate": 7e-06, "loss": 0.1624, "step": 450 }, { "epoch": 0.5697141954839728, "grad_norm": 0.07889428734779358, "learning_rate": 7e-06, "loss": 0.1808, "step": 451 }, { "epoch": 0.5709774198642034, "grad_norm": 0.08698045462369919, "learning_rate": 7e-06, "loss": 0.2257, "step": 452 }, { "epoch": 0.5722406442444339, "grad_norm": 0.08498376607894897, "learning_rate": 7e-06, "loss": 0.1773, "step": 453 }, { "epoch": 0.5735038686246644, "grad_norm": 0.12781842052936554, "learning_rate": 7e-06, "loss": 0.258, "step": 454 }, { "epoch": 0.574767093004895, "grad_norm": 0.11314232647418976, "learning_rate": 7e-06, "loss": 0.1608, "step": 455 }, { "epoch": 0.5760303173851256, "grad_norm": 0.3507859408855438, "learning_rate": 7e-06, "loss": 0.1656, "step": 456 }, { "epoch": 0.5772935417653561, "grad_norm": 0.08430968970060349, "learning_rate": 7e-06, "loss": 0.201, "step": 457 }, { "epoch": 0.5785567661455866, "grad_norm": 0.09361864626407623, "learning_rate": 7e-06, "loss": 0.2034, "step": 458 }, { "epoch": 0.5798199905258171, "grad_norm": 0.09698746353387833, "learning_rate": 7e-06, "loss": 0.2444, "step": 459 }, { "epoch": 0.5810832149060476, "grad_norm": 0.16023226082324982, "learning_rate": 7e-06, "loss": 0.19, "step": 460 }, { "epoch": 0.5823464392862783, "grad_norm": 0.08157742023468018, "learning_rate": 7e-06, "loss": 0.2149, "step": 461 }, { "epoch": 0.5836096636665088, "grad_norm": 0.09342104941606522, "learning_rate": 7e-06, "loss": 0.2022, "step": 462 }, { "epoch": 0.5848728880467393, "grad_norm": 0.07538167387247086, "learning_rate": 7e-06, "loss": 0.2516, "step": 463 }, { "epoch": 0.5861361124269698, "grad_norm": 0.12720584869384766, "learning_rate": 7e-06, "loss": 0.2208, "step": 464 }, { "epoch": 0.5873993368072004, "grad_norm": 0.08613109588623047, "learning_rate": 7e-06, "loss": 0.2193, "step": 465 }, { "epoch": 0.5886625611874309, "grad_norm": 0.08249358087778091, "learning_rate": 7e-06, "loss": 0.1676, "step": 466 }, { "epoch": 0.5899257855676615, "grad_norm": 0.1288759857416153, "learning_rate": 7e-06, "loss": 0.3238, "step": 467 }, { "epoch": 0.591189009947892, "grad_norm": 0.05164247751235962, "learning_rate": 7e-06, "loss": 0.1418, "step": 468 }, { "epoch": 0.5924522343281226, "grad_norm": 0.0994580090045929, "learning_rate": 7e-06, "loss": 0.2238, "step": 469 }, { "epoch": 0.5937154587083531, "grad_norm": 0.09253129363059998, "learning_rate": 7e-06, "loss": 0.2006, "step": 470 }, { "epoch": 0.5949786830885836, "grad_norm": 0.12927457690238953, "learning_rate": 7e-06, "loss": 0.2027, "step": 471 }, { "epoch": 0.5962419074688141, "grad_norm": 0.11283280700445175, "learning_rate": 7e-06, "loss": 0.2169, "step": 472 }, { "epoch": 0.5975051318490446, "grad_norm": 0.174880713224411, "learning_rate": 7e-06, "loss": 0.1746, "step": 473 }, { "epoch": 0.5987683562292753, "grad_norm": 0.11614017933607101, "learning_rate": 7e-06, "loss": 0.2968, "step": 474 }, { "epoch": 0.6000315806095058, "grad_norm": 0.07818127423524857, "learning_rate": 7e-06, "loss": 0.1643, "step": 475 }, { "epoch": 0.6012948049897363, "grad_norm": 0.08300397545099258, "learning_rate": 7e-06, "loss": 0.1712, "step": 476 }, { "epoch": 0.6025580293699668, "grad_norm": 0.1014489009976387, "learning_rate": 7e-06, "loss": 0.2089, "step": 477 }, { "epoch": 0.6038212537501974, "grad_norm": 0.11591055244207382, "learning_rate": 7e-06, "loss": 0.2231, "step": 478 }, { "epoch": 0.6050844781304279, "grad_norm": 0.13137224316596985, "learning_rate": 7e-06, "loss": 0.1844, "step": 479 }, { "epoch": 0.6063477025106585, "grad_norm": 0.09693000465631485, "learning_rate": 7e-06, "loss": 0.2033, "step": 480 }, { "epoch": 0.607610926890889, "grad_norm": 0.1250012218952179, "learning_rate": 7e-06, "loss": 0.2219, "step": 481 }, { "epoch": 0.6088741512711195, "grad_norm": 0.18828216195106506, "learning_rate": 7e-06, "loss": 0.2427, "step": 482 }, { "epoch": 0.6101373756513501, "grad_norm": 0.11057613044977188, "learning_rate": 7e-06, "loss": 0.2193, "step": 483 }, { "epoch": 0.6114006000315806, "grad_norm": 0.15523040294647217, "learning_rate": 7e-06, "loss": 0.1883, "step": 484 }, { "epoch": 0.6126638244118111, "grad_norm": 0.16174635291099548, "learning_rate": 7e-06, "loss": 0.2067, "step": 485 }, { "epoch": 0.6139270487920417, "grad_norm": 0.2738276422023773, "learning_rate": 7e-06, "loss": 0.2292, "step": 486 }, { "epoch": 0.6151902731722723, "grad_norm": 0.05995164066553116, "learning_rate": 7e-06, "loss": 0.1956, "step": 487 }, { "epoch": 0.6164534975525028, "grad_norm": 0.05519471690058708, "learning_rate": 7e-06, "loss": 0.1501, "step": 488 }, { "epoch": 0.6177167219327333, "grad_norm": 0.08133929967880249, "learning_rate": 7e-06, "loss": 0.2224, "step": 489 }, { "epoch": 0.6189799463129638, "grad_norm": 0.12239203602075577, "learning_rate": 7e-06, "loss": 0.2503, "step": 490 }, { "epoch": 0.6202431706931943, "grad_norm": 0.11004896461963654, "learning_rate": 7e-06, "loss": 0.2796, "step": 491 }, { "epoch": 0.6215063950734249, "grad_norm": 0.1722228229045868, "learning_rate": 7e-06, "loss": 0.1807, "step": 492 }, { "epoch": 0.6227696194536555, "grad_norm": 0.10695190727710724, "learning_rate": 7e-06, "loss": 0.1782, "step": 493 }, { "epoch": 0.624032843833886, "grad_norm": 0.08578750491142273, "learning_rate": 7e-06, "loss": 0.2389, "step": 494 }, { "epoch": 0.6252960682141165, "grad_norm": 0.10670057684183121, "learning_rate": 7e-06, "loss": 0.1801, "step": 495 }, { "epoch": 0.6265592925943471, "grad_norm": 0.04314388707280159, "learning_rate": 7e-06, "loss": 0.1366, "step": 496 }, { "epoch": 0.6278225169745776, "grad_norm": 0.11937737464904785, "learning_rate": 7e-06, "loss": 0.2017, "step": 497 }, { "epoch": 0.6290857413548081, "grad_norm": 0.08274619281291962, "learning_rate": 7e-06, "loss": 0.1721, "step": 498 }, { "epoch": 0.6303489657350387, "grad_norm": 0.07380262762308121, "learning_rate": 7e-06, "loss": 0.1962, "step": 499 }, { "epoch": 0.6316121901152693, "grad_norm": 0.04727354645729065, "learning_rate": 7e-06, "loss": 0.1509, "step": 500 }, { "epoch": 0.6316121901152693, "eval_correct_accuracy": 0.7056827820186599, "eval_error_accuracy": 0.4376407023863125, "eval_f1": 0.5402416946684214, "eval_loss": 0.362473726272583, "eval_runtime": 35.262, "eval_samples_per_second": 96.421, "eval_steps_per_second": 6.04, "step": 500 }, { "epoch": 0.6328754144954998, "grad_norm": 0.08614058047533035, "learning_rate": 7e-06, "loss": 0.1706, "step": 501 }, { "epoch": 0.6341386388757303, "grad_norm": 0.06968270987272263, "learning_rate": 7e-06, "loss": 0.2028, "step": 502 }, { "epoch": 0.6354018632559608, "grad_norm": 0.13758571445941925, "learning_rate": 7e-06, "loss": 0.2296, "step": 503 }, { "epoch": 0.6366650876361913, "grad_norm": 0.07083171606063843, "learning_rate": 7e-06, "loss": 0.1807, "step": 504 }, { "epoch": 0.637928312016422, "grad_norm": 0.06689167022705078, "learning_rate": 7e-06, "loss": 0.1997, "step": 505 }, { "epoch": 0.6391915363966525, "grad_norm": 0.07969733327627182, "learning_rate": 7e-06, "loss": 0.1803, "step": 506 }, { "epoch": 0.640454760776883, "grad_norm": 0.095677949488163, "learning_rate": 7e-06, "loss": 0.1874, "step": 507 }, { "epoch": 0.6417179851571135, "grad_norm": 0.10759231448173523, "learning_rate": 7e-06, "loss": 0.1656, "step": 508 }, { "epoch": 0.6429812095373441, "grad_norm": 0.13282425701618195, "learning_rate": 7e-06, "loss": 0.2538, "step": 509 }, { "epoch": 0.6442444339175746, "grad_norm": 0.09664168953895569, "learning_rate": 7e-06, "loss": 0.1768, "step": 510 }, { "epoch": 0.6455076582978051, "grad_norm": 0.11897934973239899, "learning_rate": 7e-06, "loss": 0.2023, "step": 511 }, { "epoch": 0.6467708826780357, "grad_norm": 0.05450622737407684, "learning_rate": 7e-06, "loss": 0.1277, "step": 512 }, { "epoch": 0.6480341070582663, "grad_norm": 0.07977665215730667, "learning_rate": 7e-06, "loss": 0.2231, "step": 513 }, { "epoch": 0.6492973314384968, "grad_norm": 0.19492259621620178, "learning_rate": 7e-06, "loss": 0.2253, "step": 514 }, { "epoch": 0.6505605558187273, "grad_norm": 0.09466379135847092, "learning_rate": 7e-06, "loss": 0.3611, "step": 515 }, { "epoch": 0.6518237801989578, "grad_norm": 0.17244236171245575, "learning_rate": 7e-06, "loss": 0.2149, "step": 516 }, { "epoch": 0.6530870045791883, "grad_norm": 0.08291974663734436, "learning_rate": 7e-06, "loss": 0.1848, "step": 517 }, { "epoch": 0.654350228959419, "grad_norm": 0.06109621748328209, "learning_rate": 7e-06, "loss": 0.1435, "step": 518 }, { "epoch": 0.6556134533396495, "grad_norm": 0.06171726807951927, "learning_rate": 7e-06, "loss": 0.1749, "step": 519 }, { "epoch": 0.65687667771988, "grad_norm": 0.09645943343639374, "learning_rate": 7e-06, "loss": 0.197, "step": 520 }, { "epoch": 0.6581399021001105, "grad_norm": 0.09050124883651733, "learning_rate": 7e-06, "loss": 0.1609, "step": 521 }, { "epoch": 0.659403126480341, "grad_norm": 0.09600576758384705, "learning_rate": 7e-06, "loss": 0.1826, "step": 522 }, { "epoch": 0.6606663508605716, "grad_norm": 0.1261880248785019, "learning_rate": 7e-06, "loss": 0.1875, "step": 523 }, { "epoch": 0.6619295752408021, "grad_norm": 0.13587896525859833, "learning_rate": 7e-06, "loss": 0.1905, "step": 524 }, { "epoch": 0.6631927996210327, "grad_norm": 0.12359704077243805, "learning_rate": 7e-06, "loss": 0.2087, "step": 525 }, { "epoch": 0.6644560240012632, "grad_norm": 0.10092345625162125, "learning_rate": 7e-06, "loss": 0.2041, "step": 526 }, { "epoch": 0.6657192483814938, "grad_norm": 0.12595926225185394, "learning_rate": 7e-06, "loss": 0.1928, "step": 527 }, { "epoch": 0.6669824727617243, "grad_norm": 0.08753985911607742, "learning_rate": 7e-06, "loss": 0.1444, "step": 528 }, { "epoch": 0.6682456971419548, "grad_norm": 0.08193645626306534, "learning_rate": 7e-06, "loss": 0.1545, "step": 529 }, { "epoch": 0.6695089215221853, "grad_norm": 0.07170840352773666, "learning_rate": 7e-06, "loss": 0.1652, "step": 530 }, { "epoch": 0.670772145902416, "grad_norm": 0.18759992718696594, "learning_rate": 7e-06, "loss": 0.232, "step": 531 }, { "epoch": 0.6720353702826465, "grad_norm": 0.13691110908985138, "learning_rate": 7e-06, "loss": 0.1905, "step": 532 }, { "epoch": 0.673298594662877, "grad_norm": 0.06453829258680344, "learning_rate": 7e-06, "loss": 0.2283, "step": 533 }, { "epoch": 0.6745618190431075, "grad_norm": 0.12694236636161804, "learning_rate": 7e-06, "loss": 0.2013, "step": 534 }, { "epoch": 0.675825043423338, "grad_norm": 0.06403839588165283, "learning_rate": 7e-06, "loss": 0.1585, "step": 535 }, { "epoch": 0.6770882678035686, "grad_norm": 0.13636727631092072, "learning_rate": 7e-06, "loss": 0.2156, "step": 536 }, { "epoch": 0.6783514921837992, "grad_norm": 0.12285730242729187, "learning_rate": 7e-06, "loss": 0.1967, "step": 537 }, { "epoch": 0.6796147165640297, "grad_norm": 0.0780211091041565, "learning_rate": 7e-06, "loss": 0.1751, "step": 538 }, { "epoch": 0.6808779409442602, "grad_norm": 0.09688100218772888, "learning_rate": 7e-06, "loss": 0.2141, "step": 539 }, { "epoch": 0.6821411653244908, "grad_norm": 0.07864505052566528, "learning_rate": 7e-06, "loss": 0.2138, "step": 540 }, { "epoch": 0.6834043897047213, "grad_norm": 0.060981281101703644, "learning_rate": 7e-06, "loss": 0.187, "step": 541 }, { "epoch": 0.6846676140849518, "grad_norm": 0.06510937958955765, "learning_rate": 7e-06, "loss": 0.1771, "step": 542 }, { "epoch": 0.6859308384651823, "grad_norm": 0.07638704031705856, "learning_rate": 7e-06, "loss": 0.2129, "step": 543 }, { "epoch": 0.687194062845413, "grad_norm": 0.11518476903438568, "learning_rate": 7e-06, "loss": 0.1854, "step": 544 }, { "epoch": 0.6884572872256435, "grad_norm": 0.06868738681077957, "learning_rate": 7e-06, "loss": 0.1581, "step": 545 }, { "epoch": 0.689720511605874, "grad_norm": 0.09059899300336838, "learning_rate": 7e-06, "loss": 0.2245, "step": 546 }, { "epoch": 0.6909837359861045, "grad_norm": 0.06422233581542969, "learning_rate": 7e-06, "loss": 0.1559, "step": 547 }, { "epoch": 0.692246960366335, "grad_norm": 0.10189103335142136, "learning_rate": 7e-06, "loss": 0.193, "step": 548 }, { "epoch": 0.6935101847465656, "grad_norm": 0.08199501782655716, "learning_rate": 7e-06, "loss": 0.1908, "step": 549 }, { "epoch": 0.6947734091267962, "grad_norm": 0.20546898245811462, "learning_rate": 7e-06, "loss": 0.2011, "step": 550 }, { "epoch": 0.6960366335070267, "grad_norm": 0.14664340019226074, "learning_rate": 7e-06, "loss": 0.1892, "step": 551 }, { "epoch": 0.6972998578872572, "grad_norm": 0.08695843815803528, "learning_rate": 7e-06, "loss": 0.1871, "step": 552 }, { "epoch": 0.6985630822674878, "grad_norm": 0.08112246543169022, "learning_rate": 7e-06, "loss": 0.1616, "step": 553 }, { "epoch": 0.6998263066477183, "grad_norm": 0.08381661772727966, "learning_rate": 7e-06, "loss": 0.2231, "step": 554 }, { "epoch": 0.7010895310279488, "grad_norm": 0.09177428483963013, "learning_rate": 7e-06, "loss": 0.1956, "step": 555 }, { "epoch": 0.7023527554081794, "grad_norm": 0.08766631782054901, "learning_rate": 7e-06, "loss": 0.1871, "step": 556 }, { "epoch": 0.70361597978841, "grad_norm": 0.07755694538354874, "learning_rate": 7e-06, "loss": 0.2718, "step": 557 }, { "epoch": 0.7048792041686405, "grad_norm": 0.08710070699453354, "learning_rate": 7e-06, "loss": 0.1909, "step": 558 }, { "epoch": 0.706142428548871, "grad_norm": 0.07648595422506332, "learning_rate": 7e-06, "loss": 0.1948, "step": 559 }, { "epoch": 0.7074056529291015, "grad_norm": 0.10871299356222153, "learning_rate": 7e-06, "loss": 0.2093, "step": 560 }, { "epoch": 0.708668877309332, "grad_norm": 0.07032714784145355, "learning_rate": 7e-06, "loss": 0.1699, "step": 561 }, { "epoch": 0.7099321016895626, "grad_norm": 0.0873897522687912, "learning_rate": 7e-06, "loss": 0.1372, "step": 562 }, { "epoch": 0.7111953260697932, "grad_norm": 0.07188841700553894, "learning_rate": 7e-06, "loss": 0.1794, "step": 563 }, { "epoch": 0.7124585504500237, "grad_norm": 0.07733464986085892, "learning_rate": 7e-06, "loss": 0.2043, "step": 564 }, { "epoch": 0.7137217748302542, "grad_norm": 0.07270821928977966, "learning_rate": 7e-06, "loss": 0.177, "step": 565 }, { "epoch": 0.7149849992104848, "grad_norm": 0.1570441722869873, "learning_rate": 7e-06, "loss": 0.2411, "step": 566 }, { "epoch": 0.7162482235907153, "grad_norm": 0.2707260549068451, "learning_rate": 7e-06, "loss": 0.2307, "step": 567 }, { "epoch": 0.7175114479709458, "grad_norm": 0.07656281441450119, "learning_rate": 7e-06, "loss": 0.1759, "step": 568 }, { "epoch": 0.7187746723511764, "grad_norm": 0.09973770380020142, "learning_rate": 7e-06, "loss": 0.1965, "step": 569 }, { "epoch": 0.720037896731407, "grad_norm": 0.06791306287050247, "learning_rate": 7e-06, "loss": 0.1749, "step": 570 }, { "epoch": 0.7213011211116375, "grad_norm": 0.17801041901111603, "learning_rate": 7e-06, "loss": 0.1941, "step": 571 }, { "epoch": 0.722564345491868, "grad_norm": 0.18452543020248413, "learning_rate": 7e-06, "loss": 0.2446, "step": 572 }, { "epoch": 0.7238275698720985, "grad_norm": 0.12178942561149597, "learning_rate": 7e-06, "loss": 0.1583, "step": 573 }, { "epoch": 0.725090794252329, "grad_norm": 0.13167473673820496, "learning_rate": 7e-06, "loss": 0.2324, "step": 574 }, { "epoch": 0.7263540186325597, "grad_norm": 0.05255408585071564, "learning_rate": 7e-06, "loss": 0.195, "step": 575 }, { "epoch": 0.7276172430127902, "grad_norm": 0.09154222905635834, "learning_rate": 7e-06, "loss": 0.1878, "step": 576 }, { "epoch": 0.7288804673930207, "grad_norm": 0.0887879729270935, "learning_rate": 7e-06, "loss": 0.2052, "step": 577 }, { "epoch": 0.7301436917732512, "grad_norm": 0.1336040198802948, "learning_rate": 7e-06, "loss": 0.1952, "step": 578 }, { "epoch": 0.7314069161534817, "grad_norm": 0.08207479119300842, "learning_rate": 7e-06, "loss": 0.185, "step": 579 }, { "epoch": 0.7326701405337123, "grad_norm": 0.05941140279173851, "learning_rate": 7e-06, "loss": 0.204, "step": 580 }, { "epoch": 0.7339333649139428, "grad_norm": 0.06899949908256531, "learning_rate": 7e-06, "loss": 0.1408, "step": 581 }, { "epoch": 0.7351965892941734, "grad_norm": 0.09259360283613205, "learning_rate": 7e-06, "loss": 0.1823, "step": 582 }, { "epoch": 0.7364598136744039, "grad_norm": 0.1346062421798706, "learning_rate": 7e-06, "loss": 0.2174, "step": 583 }, { "epoch": 0.7377230380546345, "grad_norm": 0.1547420769929886, "learning_rate": 7e-06, "loss": 0.1864, "step": 584 }, { "epoch": 0.738986262434865, "grad_norm": 0.10551164299249649, "learning_rate": 7e-06, "loss": 0.1554, "step": 585 }, { "epoch": 0.7402494868150955, "grad_norm": 0.08826129138469696, "learning_rate": 7e-06, "loss": 0.2044, "step": 586 }, { "epoch": 0.741512711195326, "grad_norm": 0.07170785963535309, "learning_rate": 7e-06, "loss": 0.1777, "step": 587 }, { "epoch": 0.7427759355755567, "grad_norm": 0.1085812896490097, "learning_rate": 7e-06, "loss": 0.2822, "step": 588 }, { "epoch": 0.7440391599557872, "grad_norm": 0.08545360714197159, "learning_rate": 7e-06, "loss": 0.19, "step": 589 }, { "epoch": 0.7453023843360177, "grad_norm": 0.05576294660568237, "learning_rate": 7e-06, "loss": 0.1826, "step": 590 }, { "epoch": 0.7465656087162482, "grad_norm": 0.056626636534929276, "learning_rate": 7e-06, "loss": 0.2191, "step": 591 }, { "epoch": 0.7478288330964787, "grad_norm": 0.06961087882518768, "learning_rate": 7e-06, "loss": 0.2004, "step": 592 }, { "epoch": 0.7490920574767093, "grad_norm": 0.09317582845687866, "learning_rate": 7e-06, "loss": 0.1465, "step": 593 }, { "epoch": 0.7503552818569399, "grad_norm": 0.13993658125400543, "learning_rate": 7e-06, "loss": 0.1876, "step": 594 }, { "epoch": 0.7516185062371704, "grad_norm": 0.06080286204814911, "learning_rate": 7e-06, "loss": 0.2012, "step": 595 }, { "epoch": 0.7528817306174009, "grad_norm": 0.060514189302921295, "learning_rate": 7e-06, "loss": 0.1658, "step": 596 }, { "epoch": 0.7541449549976315, "grad_norm": 0.09004813432693481, "learning_rate": 7e-06, "loss": 0.3195, "step": 597 }, { "epoch": 0.755408179377862, "grad_norm": 0.07283802330493927, "learning_rate": 7e-06, "loss": 0.2021, "step": 598 }, { "epoch": 0.7566714037580925, "grad_norm": 0.08824078738689423, "learning_rate": 7e-06, "loss": 0.1941, "step": 599 }, { "epoch": 0.757934628138323, "grad_norm": 0.09339006245136261, "learning_rate": 7e-06, "loss": 0.4697, "step": 600 }, { "epoch": 0.757934628138323, "eval_correct_accuracy": 0.6098388464800678, "eval_error_accuracy": 0.48491670418730304, "eval_f1": 0.540250366102693, "eval_loss": 0.42047378420829773, "eval_runtime": 35.1541, "eval_samples_per_second": 96.717, "eval_steps_per_second": 6.059, "step": 600 } ], "logging_steps": 1.0, "max_steps": 791, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.7774891537268736e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }